Next Article in Journal
Mapping Properties of Associate Laguerre Polynomial in Symmetric Domains
Previous Article in Journal
Application of a Fractional Order PI Controller for a Speed Servo Drive Control
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Symmetrical Convergence Rates and Asymptotic Properties of Estimators in a Semi-Parametric Errors-in-Variables Model with Strong Mixing Errors and Missing Responses

1
College of Science, University of Shanghai for Science and Technology, Shanghai 200093, China
2
United Automotive Electronics Co., Ltd., Shanghai 201206, China
*
Author to whom correspondence should be addressed.
Symmetry 2024, 16(11), 1544; https://doi.org/10.3390/sym16111544
Submission received: 9 October 2024 / Revised: 13 November 2024 / Accepted: 15 November 2024 / Published: 18 November 2024
(This article belongs to the Section Mathematics)

Abstract

:
This paper considers a semi-parametric errors-in-variables (EV) model, η i = x i β + g ( τ i ) + ϵ i , ξ i = x i + δ i , 1 i n . The properties of estimators are investigated under conditions of missing data and strong mixing errors. Three approaches are used to handle missing data: direct deletion, imputation, and the regression surrogate. Furthermore, estimators for the coefficient β and the nonparametric function g ( · ) are obtained. Notably, both estimators achieve strong consistency at a rate of o ( n 1 / 4 ) , exhibiting a symmetry in their convergence rates, and they also demonstrate asymptotic normality. Additionally, the validity of our theoretical results is supported by simulations demonstrating the finite sample behaviour of these estimators.

1. Introduction

We examine the semi-parametric EV model given by
η i = x i β + g ( τ i ) + ϵ i , ξ i = x i + δ i .
where the response variables η i are determined by a linear combination of the known design point ( x i , τ i ) , unobservable latent variables x i , an unknown function g ( τ i ) , and random errors ϵ i with a mean of zero. Additionally, observations of x i , denoted as ξ i , are subject to measurement errors δ i that also have a mean of zero. The estimation of the unknown real parameter β is required, and the unknown function g ( τ i ) is defined on the closed interval [ 0 , 1 ] . Furthermore, a known function f ( · ) is defined on the same interval and relates to the latent variables x i through the following equation:
x i = f ( τ i ) + u i .
Here, u i represents design points related to x i .
The practicality of Model (1) has garnered increasing scholarly attention. In the case where x i is observed without a measurement error, the model simplifies to the standard semi-parametric regression model, as initially proposed by Engle and his colleagues [1] for examining the relationship between climatic conditions and electricity usage. In recent years, numerous other scholars have conducted research on this topic. For instance, Roozbeh et al. [2] proposed a robust estimator for semi-parametric models, proven to outperform traditional estimators in terms of asymptotic normality and consistency under certain conditions through simulations and real data analysis. Ding and co-authors [3] analyzed the long-run behavior of wavelet-based estimators applied to semi-parametric models with varying error variances and randomness. Wei et al. [4] explored variance estimation in partial linear variable-coefficient semi-parametric models, introducing a ridge estimator to enhance the least squares method. Fu et al. [5] examined a generalized semi-parametric model, developing a difference-based M-estimator for its parametric component. They proved the asymptotic normality of this estimator and evaluated the convergence behavior of a wavelet-based estimator for the model’s nonparametric function. These outcomes were subsequently applied to models exhibiting dependent errors.
Measurement errors in covariates present a significant challenge to statistical analyses, underscoring the importance of model (1). In cases where g ( · ) is identically zero, model (1) reduces to the basic linear errors-in-variables (EV) regression model, which Deaton [6] effectively used to mitigate the impacts of sampling error. Chen et al. [7] established pivotal results on convergence and the strong law of large numbers for sums of i.i.d. random variables, providing valuable insights into the strong consistency and convergence rates of nonparametric regression and least squares estimators in linear EV models. Research has since expanded into semi-parametric EV models with independent errors and complete data. For instance, Zhou et al. [8] investigated a semi-parametric partially linear model for panel data, introducing heteroscedasticity detection methods and an efficient weighted least squares estimator. Zhang et al. [9] addressed heteroscedastic partially linear EV models, defining Berry–Esseen bounds for estimators and evaluating their performance under independent errors. Emami [10] examined ridge and restricted ridge estimation in semi-parametric EV models in the presence of covariate measurement errors and unstable covariance matrices, establishing conditions and regions of optimality through simulations based on quadratic risks. Further relevant discussions can be found in the works of Miao and colleagues [11], Hu et al. [12], Zhang et al. [13], and others.
While semi-parametric regression and errors-in-variables models have traditionally relied on complete data, real-world applications frequently face the challenge of incomplete data due to missing observations or data losses. Recognizing the critical need for advanced methodologies, recent studies have made noteworthy contributions. Zou et al. [14] developed methods for estimation and inference in partial functional linear models with measurement error and incomplete data, demonstrating their asymptotic validity and practical utility through simulations and real data analysis. Similarly, Xiao et al. [15] proposed two estimation techniques for varying-coefficient models, showing the superior accuracy and convergence of the imputation-based method through simulations and real-world applications. Additionally, Zou et al. [16] introduced novel approaches for analyzing heteroscedastic partially linear varying-coefficient models with right-censored and randomly missing data. Through the integration of regression calibration, imputation techniques, and inverse probability weighting, they improved estimators for parameters and coefficient functions, demonstrating their asymptotic properties. These efforts underscore the evolving strategies to address data incompleteness, enriching the toolset for statistical analysis in the presence of incomplete information.
It is crucial to recognize that error independence is not consistently a valid assumption in practical applications, especially when dealing with economic data collected sequentially. Such data frequently exhibit significant error dependence. A variety of correlation structures have been identified, including autoregressive series, negatively associated sequences, ρ -mixing sequences, α -mixing sequences, and moving average series. There has been substantial interest among researchers in exploring models characterized by α -mixing errors. For example, Xi et al. [17] demonstrated the asymptotic properties of β , g, and f under conditions of α -mixing errors. Zhang et al. [18] examined the asymptotic performance of estimators in a semi-parametric EV model where errors are negatively associated (NA) and data are incomplete. Similarly, Zhang and Yang [19] analyzed the asymptotic behavior of estimators in a semi-parametric EV model with linear process sequence errors under incomplete data.
In conclusion, there exists a significant gap in the research: no theoretical foundation has been established for semi-parametric EV models with strong mixing errors under incomplete data. Our study fills this gap, focusing on such a model and addressing both strong mixing errors and the challenge of missing data. This effort is driven by curiosity and a desire to shed light on the unknown.
This paper unfolds in several key parts. Section 2 lays out our assumptions. In Section 3, we explore three methodologies for estimating β and g ( · ) while addressing incomplete responses. In this section, we also unveil strong consistency rates and the asymptotic normality of our estimators. In Section 4, we dedicate our focus to a simulation study, delving into the behavior of our suggested estimators under limited sample sizes. In Section 5, we present some propositions. The justifications for the outcomes and propositions are detailed in Section 6 and Section 7. Finally, Section 8 concludes this paper, summarizing the key findings.

2. Assumptions

In this study, we analyze incomplete datasets denoted by { ( η i , ρ i , ξ i , τ i ) } that are derived from model (1). Here, ρ i indicates the presence of data, with ρ i = 0 representing instances where the values of η i are randomly missing, and ρ i = 1 indicating observed values. Additionally, ϵ i are strong mixing errors. It is crucial to distinguish between different types of missing data mechanisms, as these influence the validity of statistical methods applied to incomplete data:
  • Missing Completely at Random (MCAR): The probability of missingness is independent of both the observed and unobserved data values.
  • Missing at Random (MAR): The probability of missingness depends on the observed values, but not on the unobserved (missing) values themselves.
  • Not Missing at Random (NMAR): Missingness depends on the unobserved (missing) values themselves.
In this work, we assume a Missing at Random (MAR) mechanism, as it aligns with the methods employed to handle incomplete data, including direct deletion, imputation, and regression surrogate techniques. Next, we will present several assumptions, which will serve to substantiate the principal results outlined below.
Definition 1.
In a probability space Γ , B , Q , consider a sequence of real-valued random variables Y i : i X . Let G p n denote the σ-algebra generated by the collection of random variables Y i : p i n . Set
α n = sup Q C D Q C Q D : C G n , D G p + n
The sequence Y i is termed strongly mixing when α n approaches zero as n tends to infinity.
(H0)
Consider the sequence { ϵ i } i = 1 n of random variables which are both stationary and exhibit strong mixing properties with mixing coefficients α n . Additionally, consider the sequence { δ i } i = 1 n of independent random variables that satisfy the following:
(i)
E ϵ i = 0 , E ϵ i 2 = 1 , E δ i = 0 , E δ i 2 = Ξ δ 2 > 0 ;
(ii)
sup i E ϵ i 4 + 2 η < , sup i E δ i 4 + 2 η < for some η 0 ;
(iii)
The sequences { ϵ i } i = 1 n and { δ i } i = 1 n are mutually independent.
(H1)
Let { μ i } i = 1 n be a sequence as defined in (2) satisfying the following:
(i)
As n , 1 n i = 1 n μ i 2 Λ 0 , 1 n i = 1 n ρ i μ i 2 Λ 1 where 0 < Λ 0 , Λ 1 < .
(ii)
As n , sup n max 1 m n | i = 1 m μ k i | n log n < , { k 1 , k 2 , , k n } denotes any order of 1 to n.
(H2)
On [ 0 , 1 ] , both g ( · ) and f ( · ) are continuous and Lipschitz continuous of order one.
(H3)
For each integer k with 1 k n , there exists a weight function V n k c τ that is confined to the domain [0, 1] and satisfies the following:
(i)
Let S n k τ i = i = 1 n ρ k V n k c ( τ i ) . Then, max 1 k n S n k τ i = O ( 1 ) .
(ii)
Assume for all a > 0 : sup τ j = 1 n ρ j V n k c τ I | τ τ k | > a n 1 / 4 = O n 1 4 .
(iii)
sup τ V n k c τ = o 1 n log n .
(H4)
Let V n k τ , for 1 k n , denote the probability weight functions defined over the interval [ 0 , 1 ] , which are subject to the following conditions:
(i)
Let T n k τ i = i = 1 n ρ k V n k ( τ i ) . Then, max 1 k n T n k τ i = O ( 1 ) .
(ii)
Assume for all a > 0 : sup τ k = 1 n V n k τ I | τ τ k | > a · n 1 4 = o n 1 4 .
(iii)
sup τ max k V n k ( τ ) = o 1 n log n .
(H5)
(i)
Given b > 0 , it holds that sup τ k = 1 n ρ k V n k c τ I | τ τ k | > b · n 1 2 = o n 1 2 .
(ii)
Given b > 0 , it holds that sup τ k = 1 n V n k τ I | τ τ k | > b · n 1 2 = o n 1 2 .
(H6)
(i)
k = 1 n V n k c ( τ ) 2 = O ( n 3 / 5 ) for any 0 τ 1 .
(ii)
k = 1 n V n k ( τ ) 2 = O ( n 3 / 5 ) for any 0 τ 1 .
(H7)
Positive integers r : = r n and s : = s n exist, satisfying the following:
(i)
r + s n , r , s , s r 0 , r 2 = O n s .
(ii)
Let μ max = max 1 i n μ i ; it follows that s r μ max 0 as n .
Remark 1.
Assumptions (H0) to (H5) are commonly accepted conditions that are frequently cited in scholarly articles, including works by Zhang et al. [9] and Liang et al. [20].
Remark 2.
Meeting assumption (H7) is relatively effortless. For instance, one can choose r = n 1 / 3 and s = log n , and ensure that max 1 i n μ i = o n 1 / 6 .

3. Main Results

3.1. Building Estimators: Methods and Techniques

3.1.1. Direct Deletion Method

In model (1), the direct deletion method simplifies the model by excluding missing data and using only complete observations. This results in the following simplified model: ρ i η i = ρ i x i β + ρ i g ( τ i ) + ρ i ϵ i When x i is observable (i.e., ρ i = 1 ), the least squares estimation (LSE) method can be used to estimate the parameter β . Given a known β , the estimator for g ( · ) is constructed using the complete data, which consist of the observed triplets { ( ρ i η i , ρ i ξ i , ρ i τ i ) } i = 1 n . Specifically, the estimator g n * ( τ , β ) is given by
g n * ( τ , β ) = k = 1 n V n k c ( τ ) ρ k η k ρ k ξ k β ,
where V n k c ( t ) are weight functions that satisfy assumption (H3).
Computational Complexity: The computational complexity of the direct deletion method is O ( n ) , as it only requires performing least squares regression on the complete data. This method is suitable when the proportion of missing data is small, but it may introduce bias if the missing data are not missing at random.

3.1.2. Modified Least Squares Estimation (LSE)

Liang et al. [21] proposed a standard LSE method for parameter estimation in partial linear models. However, their method does not fully account for missing data or incomplete observations. To address this, we modify their LSE approach by incorporating an additional term that utilizes weighted information from the available data, thereby reducing bias caused by missing values. The modified estimator β ^ c is obtained by minimizing the following loss function:
S S ( β ) = i = 1 n ρ i η i ξ i β g n * ( τ i , β ) 2 Ξ δ 2 β 2 = min .
This modification incorporates the function g n * ( τ i , β ) , which accounts for missing data, by using weights derived from available data. The resulting modified estimator for β is
β ^ c = i = 1 n ( ρ i ξ i c ˜ 2 ρ i Ξ δ 2 ) 1 i = 1 n ρ i ξ i c ˜ η i c ˜ ,
Substituting β ^ c into g n * ( τ , β ) gives the estimator for g ( · ) :
g ^ n c τ = k = 1 n ρ k V n k c τ ( η k ξ k β ^ c ) .
Computational Complexity: The computational complexity of the modified LSE is O ( n 2 ) due to the need to perform weighted least squares regression across all available data points. This method takes advantage of all available data, making it more robust in the presence of missing values, though it requires more computational resources.

3.1.3. Imputation Method

To fully utilize the incomplete data, the imputation method fills in the missing values. Following the approach proposed by Wang and Sun [22], the imputed value U i [ I ] is given by
U i [ I ] = ρ i η i + ( 1 ρ i ) [ ξ i β ^ c + g ^ n c ( τ i ) ] .
After imputation, the complete dataset { ( U i [ I ] , ξ i , τ i ) } i = 1 n is used to construct the estimators β ^ I and g ^ n [ I ] ( τ ) :
β ^ I = i = 1 n ξ ˜ i 2 ρ i Ξ δ 2 1 i = 1 n ξ ˜ i U ˜ i I ,
g ^ n I τ = k = 1 n V n k τ ( U k I ξ k β ^ I ) .
Computational Complexity: The computational complexity of the imputation method is O ( n 2 ) , as it involves the iterative imputation of missing values and the performance of regression analysis for each iteration. While the computational cost is higher, this method allows for a better utilization of the incomplete data and can reduce bias compared to methods that discard missing observations.

3.1.4. Regression Substitution Method

The regression substitution method replaces missing values with predicted values using a regression model. The imputed value U i [ R ] is given by
U i [ R ] = ξ i β ^ c + g ^ n c ( τ i ) .
After imputation, the complete dataset { ( U i [ R ] , ξ i , τ i ) } i = 1 n is used to estimate β ^ R and g ^ n [ R ] ( τ ) :
β ^ R = i = 1 n ξ ˜ i 2 1 i = 1 n ξ ˜ i U ˜ i [ R ] ,
g ^ n R τ = k = 1 n V n k τ ( U k R ξ k β ^ R ) .
Computational Complexity: The computational complexity of the regression substitution method is also O ( n 2 ) due to the regression imputation and subsequent updates. Similar to the imputation method, this method benefits from the complete use of available data but at the cost of increased computational time.

3.1.5. Summary of Computational Complexities

The direct deletion method, with its computational complexity of O ( n ) , is efficient and suitable for cases with minimal missing data. However, it may introduce bias if data are not Missing at Random. In contrast, the imputation and regression substitution methods, which operate at O ( n 2 ) , effectively utilize incomplete data and reduce bias but require significantly more computational resources due to their iterative nature.
In practical decision-making scenarios, such as policy formulation or resource allocation, data constraints and resource limitations are common. In these contexts, selecting the appropriate estimation method is crucial. Practitioners should consider the extent of missing data and the available computational resources when making this choice. For example, when the proportion of missing data is low, the direct deletion method provides a quick and low-cost solution. However, when missing data is more substantial, imputation and regression substitution methods, though computationally more demanding, offer more accurate estimates and help ensure that decisions are based on more reliable data.
In cases of large datasets or limited computational resources, it is advisable to set an upper limit on the number of iterations or to use the direct deletion method for initial estimates. Techniques like early stopping or gradually reducing the sample size can help manage computational load when convergence is slow. These strategies allow for efficient decision-making even under data and resource constraints, ensuring that robust decisions can still be made without overburdening computational resources.

3.2. Strong Consistency in Estimator Behavior

Our focus in this part of the paper is on studying the strong consistency of the two estimators, β and g · , assuming that the missingness of η i is random. We emphasize that strong consistency is a critical criterion for evaluating estimator quality, allowing us to distinguish effective estimators from ineffective ones based on their consistency rates. Below, we present three theorems that exhibit symmetrical structures.
Theorem 1.
Assume that (H0)–(H3), (H6) and sup i E ϵ i 10 < are fulfilled. Given τ in [ 0 , 1 ] , it holds that
(a) 
β ^ c β = o ( n 1 4 ) a . s .
(b) 
g ^ n c ( τ ) g ( τ ) = o ( n 1 4 ) a . s .
Theorem 2.
Assume that (H0)–(H4), (H6) and sup i E ϵ i 10 < are fulfilled. Given τ in [ 0 , 1 ] , it holds that
(a) 
β ^ I β = o ( n 1 4 ) a . s .
(b) 
g ^ n [ I ] ( τ ) g ( τ ) = o ( n 1 4 ) a . s .
Theorem 3.
Assume that (H0)–(H4), (H6) and sup i E ϵ i 10 < are fulfilled. Given τ in [ 0 , 1 ] , it holds that
(a) 
β ^ R β = o ( n 1 4 ) a . s .
(b) 
g ^ n [ R ] ( τ ) g ( τ ) = o ( n 1 4 ) a . s .

3.3. Asymptotic Normality in Estimator Behavior

Normal distribution occupies a central position not only in mathematics but also in physics, engineering, and numerous other fields of study. Its significance extends to the realm of asymptotic normality of estimators, which is essential for constructing confidence bands for these estimators. Our further analysis delves into the asymptotic normality behavior of the β and g · estimators.
To begin, we shall review key terms that will hold significant importance in the subsequent conclusions and demonstrations:
x ˜ i c = x i k = 1 n ρ k V n k c τ i x k , δ ˜ i c = δ i k = 1 n δ k V n k c τ i δ k , g ˜ i c = g τ i k = 1 n ρ k V n k c τ i g τ k , ϵ ˜ i c = ϵ i k = 1 n ρ k V n k c τ i ϵ k , x ˜ i = x i k = 1 n V n k τ i x k , δ ˜ i = δ i k = 1 n V n k τ i δ k , χ i = ϵ i δ i β , g ˜ i = g τ i k = 1 n V n k τ i g τ k , ϵ ˜ i = ϵ i k = 1 n V n k τ i ϵ k , R n 2 = i = 1 n x ˜ i 2 , R 1 n 2 = i = 1 n ρ i ξ ˜ i 2 ρ i Ξ δ 2 , R 2 n 2 = i = 1 n ξ ˜ i 2 Ξ δ 2 , R 3 n 2 = i = 1 n ξ ˜ i 2 , Θ 1 n 2 = V a r i = 1 n ρ i ( x ˜ i c + δ i ) ϵ i δ i β , Λ 1 n 2 τ = V a r [ k = 1 n V n k c τ ρ k + k = 1 n V n k ( τ ) ( 1 ρ k ) i = 1 n V n i c ( τ k ) ρ i ) ϵ k δ k β , Ω = i = 1 n 1 ρ i Ξ δ 2 R 1 n 2 , Φ = i = 1 n x ˜ i 2 1 ρ i R 1 n 2 , Θ 2 n 2 = V a r i = 1 n ρ i ( x ˜ i + Ω + Φ x ˜ i c + 1 + Ω + Φ δ i ) ϵ i δ i β , M = R 3 n 2 R 1 n 2 i = 1 n x i 2 ˜ , N = R 3 n 2 R 1 n 2 E δ i 2 , Θ 3 n 2 = V a r M + N i = 1 n ρ i ( x ˜ i c + δ i ) ϵ i δ i β , Λ 2 n 2 τ = V a r k = 1 n V n k τ ρ k ϵ k δ k β , Λ 3 n 2 τ = V a r k = 1 n V n k τ i = 1 n V n i c τ k ρ i ϵ i δ i β .
Theorem 4.
Assume that (H0)–(H3), (H5) and (H7) are fulfilled. Given τ in [ 0 , 1 ] ,
(a) 
Given Θ 1 n 2 C n , it follows that R 1 n 2 ( β ^ c β ) / Θ 1 n D N ( 0 , 1 ) .
(b) 
Under the condition n Λ 1 n 2 ( τ ) as n , it holds that g ^ n c ( τ ) g ( τ ) / Λ 1 n ( τ ) D N ( 0 , 1 ) .
Theorem 5.
Assume that (H0)–(H5) and (H7) are fulfilled. Given τ in [ 0 , 1 ] ,
(a) 
Given Θ 2 n 2 C n , it follows that R 2 n 2 ( β ^ I β ) / Θ 2 n D N ( 0 , 1 ) .
(b) 
Under the condition n Λ 2 n 2 ( τ ) as n , it holds that g ^ n [ I ] ( τ ) g ( τ ) / Λ 2 n ( t ) D N ( 0 , 1 ) .
Theorem 6.
Assume that (H0)–(H5) and (H7) are fulfilled. Given τ in [ 0 , 1 ] ,
(a) 
Given Θ 1 n 2 C n , it follows that R 3 n 2 ( β ^ R β ) / Θ 3 n D N ( 0 , 1 ) .
(b) 
Under the condition n Λ 3 n 2 ( τ ) as n , it holds that g ^ n [ R ] ( τ ) g ( τ ) / Λ 3 n ( τ ) D N ( 0 , 1 ) .

4. Simulation Study

The main focus of this investigation is to assess how well the proposed estimators perform when dealing with finite samples using simulated data. Specifically, our investigation focuses on the following two aspects:
(i)
To evaluate the efficacy of three estimators for β   β ^ c , β ^ I , β ^ R and three estimators for g ( · )   g ^ n c ( τ ) , g ^ n [ I ] ( τ ) , g ^ n [ R ] ( τ ) . For the β estimators, the performance metric utilized in this study is the MSE. For the estimators of g ( · ) , the assessment is carried out using the GMSE as the evaluation criterion.
(ii)
Q-Q plots are utilized to graphically represent and compare the distributional properties of the three estimators for β and g ( · ) are presented.
To conduct the simulation study, we designed the following data generation process for the observation i = 1 , 2 , , n . The variables η i and ξ i are generated as follows:
η i = x i + sin 2 π τ i + ε i , ξ i = x i + δ i
Simply put, we have β = 1 , τ i = i 0.5 n , x i = τ i 2 + μ i with μ i = sin i n 1 / 3 for 1 i n . The sequence { μ i } i = 1 n consists of independent and identically distributed random variables, all adhering to a normal distribution N ( 0 , 0 . 2 2 ) . Additionally, { ϵ i } i = 1 n represents a strong mixing sequence. Assume that the sequence { ϵ i } i = 1 n is governed by a multivariate normal distribution with expected values E ϵ 1 , , ϵ n = 0 , , 0 , covariance C o v ϵ i , ϵ j = 0 . 3 2 exp ( α j i ) for i j and variance V a r ϵ i = 0 . 3 2 for 1 i n . The decay function in the covariance decreases exponentially with the absolute difference in indices j i , and is controlled by the parameter α . The following are the specified weight functions for the estimators we propose:
V n k c τ = Q τ τ i a n k = 1 n ρ k Q τ τ k a n , V n k τ = T τ τ i d n k = 1 n T τ τ k d n .
Here, Q ( · ) and T ( · ) refer to Gaussian kernels, a n and d n are defined as the corresponding bandwidths.

4.1. MSE/GMSE Evaluation for β and g ( · ) Estimators

In this section, datasets were generated with the sample size sets n = 100 , n = 300 , and n = 500 . We assumed that the response variables had missing probabilities of p = 0.1 , 0.25, and 0.5 based on M = 500 replications. For the estimation of β and the function g · , the corresponding MSE and GMSE of the estimators are
MSE β ^ = 1 M l = 1 M [ β ^ l β ] 2 , GMSE g ^ · = 1 M n l = 1 M k = 1 n g ^ τ k , l g τ k 2
For each estimator, we compute the Mean Squared Error (MSE) or Generalized Mean Squared Error (GMSE) using M = 500 replications. The bandwidth grids, denoted as a n and d n , range from 0 to 1 and are discretized into 20 points with a step size of 0.05. We then proceed to select the bandwidths that yield the smallest MSE or GMSE. The results for the smallest MSE and GMSE obtained with α = 0.1 and α = 0.3 are presented below, respectively.
By referring to Table 1 and Table 2, one can observe the following:
(i)
All estimators demonstrated notably strong consistency.
(ii)
For every fixed n, a higher missing probability led to an increased MSE or GMSE for every estimator considered.
(iii)
For every fixed p, a larger sample size n brought about a reduction in both the MSE and GMSE for every estimator considered.
(iv)
Compared to α = 0.1 , almost all estimators showed a rise in both MSE and GMSE when α = 0.3 .
(v)
In comparison to β ^ c , the estimated values of β ^ I and β ^ R exhibited a greater proximity to the real value, suggesting that addressing incomplete data is indeed beneficial. A similar result was observed for the estimators of g · .

4.2. Simulation Evidence for Asymptotic Normality

Below, we focus on this subsection, where we have developed quantile–quantile plots to illustrate the behavior of β and g · , using the parameter α = 0.3. We investigate the performance of all estimators for β and g ( · ) under the conditions p = 0.25 with sample sizes of n = 100 , 500 . In Figure 1, Figure 2 and Figure 3, Q-Q plots are showcased for the estimators β ^ c , β ^ I , and β ^ R with sample sizes of 100 and 500. Additionally, in Figure 4, Figure 5 and Figure 6, Q-Q plots are presented for the estimators g ^ n c 0.5 , g ^ n [ I ] ( 0.5 ) , and g ^ n R 0.5 with sample sizes of 100 and 500, respectively. In all Q-Q plots, the blue points represent the sample quantiles, and the red dashed line indicates the theoretical normal distribution for comparison. This visualization helps to evaluate how closely the estimated values follow a normal distribution.
By examining Figure 1 through Figure 6, the following become evident:
(i)
Derived from a sample size of 500, the estimated values for β and g 0.5 closely resemble a normal distribution, indicating asymptotic normality.
(ii)
As sample size increases from 100 to 500, all estimators’ sample variances decrease, reflecting improved precision and stability in larger datasets.
(iii)
The simulation outcomes corroborated the findings predicted by our theory.

5. Preliminary Lemmas

In the following section, we use the notation C , C 1 , to represent various finite positive constants, each assigned a distinct value based on the specific context. We then present several essential lemmas vital for proving our main results. Consider a sequence of random variables, denoted as { Y i } i = 1 n , which exhibit strong mixing properties with associated mixing coefficients given by α n .
Lemma 1
(Härdle et al. [23], Lemma A.3). Consider a sequence of n independent random variables X 1 , , X n , each having zero mean, finite variance and s-th moments uniformly bounded for some s > 2 . Given a sequence { b k j , k , j = 1 , , n } satisfying sup 1 j , k n | b k j | = O n q 1 with 0 < q 1 < 1 and i = 1 n b i j = O n q 2 for q 2 max 0 , 2 / s q 1 . Then,
max 1 j n | k = 1 n b k j X k | = O n q 1 q 2 / 2 log n a . s .
Lemma 2
(Xu Bing [24]). Consider a sequence of strongly mixing random variables Y i , such that E Y i = 0 and s u p i 1 E Y i q < for q > 2 . Suppose that the mixing coefficients α n and the real-valued sequences { b n i } i = 1 n satisfy the conditions:
n = 1 i = 1 n b n i 2 log n q / 2 <
n = 1 α n q 2 / q < .
Under these assumptions, it follows that
i = 1 b n i Y i = o 1 a . s .
Lemma 3
(Volkonskii and Rozanov [25]). Consider a collection of strongly mixing random variables { Y 1 , , Y m } , each associated with a distinct σ-algebra F r j s j that satisfies the conditions 1 r 1 < s 1 < < s m N and r j + 1 s j γ 1 . Assume that each Y j satisfies the inequality 1 Y j 1 . Then,
E j = 1 m Y j j = 1 m E Y j 16 ( m 1 ) ρ ( γ ) ,
where ρ ( γ ) represents the mixing coefficient associated with the gap γ, and F r j s j denotes the σ-algebra generated by the random variables { Y i : r j i s j } .
Lemma 4
(Hall and Heyde [26]). Suppose we have two random variables, denoted as U and V, which fulfill the conditions that E | U | r < , E | V | s < , where r , s > 1 are such that r 1 + s 1 < 1 . Thus,
| E U V E U E V | 8 U r V s sup C σ ( U ) , D σ ( V ) | P ( C D ) P ( C ) P ( D ) | 1 r 1 s 1 .
Lemma 5
(Yang [27], Therorem 2.2).
(a) 
Given v > 2 , γ > 0 , the assumptions E [ Y i ] = 0 and E [ | Y i | v + γ ] < as well as λ > v ( v + γ ) 2 γ and α ( n ) = O ( n λ ) , it follows that for any arbitrarily small positive number η, a constant ( C ) exists, which is dependent on ( η , v , γ , λ ) , such that the following inequality holds:
E max 1 m n i = 1 m Y i s C n η i = 1 n E [ | Y i | r ] + i = 1 n Y i v + γ 2 s 2 .
(b) 
Assume that E [ Y i ] = 0 and E [ | Y i | 2 + γ ] < when γ > 0 , we have
E i = 1 n Y i 2 1 + 16 l = 1 n α γ 2 + γ ( l ) i = 1 n Y i 2 + γ 2 .
In this case, by setting r = v = 2 + γ , the proof for (b) is directly obtainable from the Lemma 4.
Lemma 6.
Let Y i be stationary with E Y i = 0 and E | Y 1 | p < for a certain p > 2 . Suppose further that α ( n ) decreases at a rate of O n λ for some positive λ. Consider an array of real numbers b n i , 1 i n , n 1 which satisfies the conditions that max 1 i n b n i = O n u for some u > 0 and i = 1 n b n i = O 1 . Let p = 4 + 2 ρ for some ρ > 0 . If λ > v v + ρ / 2 ρ for some v > p . max 1 i n b n i = O n u L 1 ( n ) and i = 1 m b n i 2 = O n τ for a positive τ, then
i = 1 n b n i Y i = o p n ψ
for ψ < m i n { 1 1 / p , u , u 1 / p + τ 2 u + 1 / v , τ / 2 1 / p ρ / [ v v + ρ ] + 1 v } .
Here, L 1 x is a positive slowly varying function as n .
Lemma 7
(Zhang et al. [28]).
(a) 
Let L ˜ i = L ( τ i ) k = 1 n V n k ( τ i ) L ( τ k ) , where L ( · ) can be eithor g ( · ) or f ( · ) . Let L ˜ i c = L ( τ i ) k = 1 n ρ k V n k c ( τ i ) L ( τ k ) , where L ( · ) = g ( · ) or f ( · ) . According to conditions (H0)–(H4), it follows that max 1 i n | L ˜ i | = o n 1 / 4 and max 1 i n | L ˜ i c | = o n 1 / 4 .
(b) 
Based on (H0)–(H4), we conclude that n 1 i = 1 n x ˜ i 2 Δ 0 , n 1 i = 1 n ρ i ( x ˜ i c ) 2 Δ 1 , i = 1 n | x ˜ i | C 1 n , and i = 1 n | ρ i x ˜ i c | C 2 n .

6. Proof of Main Results

6.1. Proof of Strong Consistency

Proof of Theorem 1(a).
Taking into account (3), we may conclude that
β ^ c β = R 1 n 2 { i = 1 n ρ i ξ i ˜ c ( η i ˜ c ξ i ˜ c β ) + i = 1 n ρ i Ξ δ 2 β } = R 1 n 2 i = 1 n ρ i ( x ˜ i c + δ ˜ i c ) ( ϵ ˜ i c δ ˜ i c β ) + ρ i Ξ δ 2 β + i = 1 n ρ i x ˜ i c g ˜ i c + i = 1 n ρ i δ ˜ i c g ˜ i c = R 1 n 2 { i = 1 n ρ i ( x ˜ i c + δ i ) ( ϵ i δ i β ) + ρ i Ξ δ 2 β + i = 1 n ρ i x ˜ i c g ˜ i c + i = 1 n ρ i δ ˜ i c g ˜ i c i = 1 n k = 1 n ρ i ρ k V n k c ( τ i ) x ˜ i c ϵ k + i = 1 n k = 1 n ρ i ρ k V n k c ( τ i ) x ˜ i c δ k β i = 1 n k = 1 n ρ i ρ k V n k c ( τ i ) ϵ i δ k i = 1 n k = 1 n ρ i ρ k V n k c ( τ i ) δ i ϵ k + i = 1 n k = 1 n j = 1 n ρ i ρ k ρ j V n k c ( τ i ) V n j c ( τ i ) δ k ϵ j + 2 i = 1 n k = 1 n ρ i ρ k V n k c ( τ i ) δ i δ k β i = 1 n k = 1 n j = 1 n ρ i ρ k ρ j V n k c ( τ i ) V n j c ( τ i ) δ k δ j β } : = R 1 n 2 k = 1 10 A n .
Therefore, in order to verify Theorem 1(a), we firstly deduce, almost surely, based on (11), R 1 n 2 R n 2 0 . Furthermore, for each l in the set { 1 , 2 , , 10 } , we show that n 1 A l n = o ( n 1 4 ) , almost surely.
Stage 1. We demonstrate that R 1 n 2 R n 2 0 a . s . Also, keep in mind that
R 1 n 2 = i = 1 n ρ i ( x ˜ i c + δ ˜ i c ) 2 ρ i Ξ δ 2 = i = 1 n ρ i ( x ˜ i c ) 2 + i = 1 n ρ i δ ˜ i c 2 Ξ δ 2 + 2 i = 1 n ρ i x ˜ i c δ ˜ i c = i = 1 n ρ i ( x ˜ i c ) 2 + i = 1 n ρ i ( δ i 2 Ξ δ 2 ) + 2 i = 1 n ρ i x ˜ i c δ i + i = 1 n ρ i k = 1 n ρ k V n k c τ i δ k 2 2 i = 1 n ρ i δ i k = 1 n ρ k V n k c ( τ i ) δ k 2 i = 1 n ρ i x ˜ i c k = 1 n ρ k V n k c ( τ i ) δ k : = R n 2 + R 11 n + 2 R 12 n + R 13 n 2 R 14 n 2 R 15 n .
Using Lemma 7(b), one can determine that n 1 R n 2 Δ 1 . Therefore, it is enough to confirm that R 1 k n = o ( n ) a.s. over the range k = 1 to 5. Let s = 2 + δ , q 1 = 0 , and q 2 = 1 in Lemma 1. By applying assumption (H0), we have that R 11 n = o ( n ) a.s. Next, by choosing s = 4 + 2 δ , q 1 = 1 / 4 , and q 2 = 1 / 4 in Lemma 1 and applying assumption (H0), (H1), and (H3), it can be concluded that
| R 12 n | = n 3 4 · | i = 1 n n 3 4 ρ i x ˜ i c δ i | = o ( n ) a . s .
similarly to (12). With the help of Lemma 1 and the assumption (H0)(i) and (H3)(i)(iii), it is obvious that
max 1 i n | k = 1 n ρ k V n k c τ i δ k | = O n 1 4 log n a . s . , i = 1 n ( | δ i | E | δ i | ) = o ( n ) a . s .
By referring to Lemma 7(b) and considering assumption (H0)(i), (H3)(i)(ii)(iii), as well as (13). One can determine that
| R 13 n | i = 1 n | ρ i | · max 1 i n | k = 1 n ρ k V n k c ( τ i ) δ k | 2 = O ( n 1 / 2 log 2 n ) = o ( n ) a . s . | R 14 n | 2 i = 1 n E | ρ i δ i | + i = 1 n ( | ρ i δ i | E | ρ i δ i | ) · max 1 i n | k = 1 n ρ k V n k c ( τ i ) δ k | = o ( n ) a . s . | R 15 n | 2 i = 1 n | ρ i x ˜ i c | · max 1 i n | k = 1 n ρ k V n k c ( τ i ) δ k | = o ( n ) a . s .
Therefore, one can deduce that R 1 n 2 = R n 2 + o ( n ) = R n 2 + o ( R n 2 ) a . s . , which yields R 1 n 2 C n 1 a . s .
Stage 2. Note the following:
A 1 n = i = 1 n ρ i ( x ˜ i c + δ i ) ( ϵ i δ i β ) + i = 1 n ρ i Ξ δ 2 β = i = 1 n ρ i x ˜ i c ( ϵ i δ i β ) + i = 1 n ρ i δ i ϵ i i = 1 n ρ i ( δ i 2 Ξ δ 2 ) β : = A 11 n + A 12 n + A 13 n
Based on (H0), the sequence { χ i = ϵ i δ i β } i = 1 n exhibits strong mixing and has a mean of zero, where p = 4 + 2 ρ . We can take ρ > 0 , u = 1 / 2 , τ = 1 , and v = 4 + 3 ρ in Lemma 6. Consequently,
n 1 A 11 n = n 1 i = 1 n ρ i x ˜ i c ( ϵ i δ i β ) : = i = 1 n ρ i Q i n χ i , max 1 i n | Q i n | n 1 max 1 i n | x ˜ i c | = o ( n 1 / 2 ) i = 1 n Q i n 2 = n 2 i = 1 n x ˜ i c 2 = O n 1
Then,
n 1 A 11 n = n 1 i = 1 n ρ i x ˜ i c ( ϵ i δ i β ) = o ( n 1 / 4 ) a . s .
The sequence { δ i ϵ i } i = 1 n consists of strongly mixing random variables. Let q = 4 + 2 η > 2 and b n i = n 3 / 4 in Lemma 2. We then obtain that n = 1 ( i = 1 ( b n i 2 log n ) q / 2 < and n = 1 α ( n ) ( q 2 ) / q < ; therefore,
n 1 A 12 n = 1 n i = 1 n δ i ϵ i = n 1 4 · n 3 / 4 i = 1 n δ i ϵ i = o ( n 1 4 ) a . s .
From (H0) to (H3) and let q 1 = 1 and q 2 = 0 in Lemma 1, we obtain
n 1 A 13 n C n | i = 1 n ρ i ( δ i 2 Ξ δ 2 ) β | = O ( n 1 4 ) a . s .
Hence, from (14) to (16), one can deduce that n 1 A 1 n = o ( n 1 / 4 ) a . s .
Stage 3. The results indicate that n 1 A k n = o ( n 1 / 4 ) a . s . over the range k = 2 to 10. Using the conditions of Theorem 1, and let u = 1 / 2 , τ = 3 / 5 , ρ = 2 , and v = 10 in Lemma 6 and (H6), this results in
| k = 1 n V n k c ( τ ) ρ k ϵ k | = o ( n 1 / 4 ) a . s . , | k = 1 n V n k ( τ ) ρ k ϵ k | = o ( n 1 / 4 ) a . s .
Thus, by Lemma 7, (13), (17), (H0)–(H3), and Abel’s inequality [23], we prove that
n 1 A 2 n C n | i = 1 n ρ i x ˜ i c g ˜ i c | C n i = 1 n | ρ i x ˜ i c | · max 1 i n | g ˜ i c | = o ( n 1 / 4 ) n 1 A 3 n C n | i = 1 n ρ i δ ˜ i c g ˜ i c | C n i = 1 n | δ i | · | ρ i g ˜ i c | + i = 1 n | ρ i g ˜ i c | · max 1 i n | k = 1 n ρ k V n k c τ i δ k | C n · i = 1 n ( | δ i | E | δ i | ) + i = 1 n E | δ i | · max 1 i n | ρ i g ˜ i c | + o ( n 1 2 log n ) = o ( n 1 / 4 ) a . s . n 1 A 4 n C n i = 1 n ρ i ( f i ˜ + μ i j = 1 n V n j τ i μ j ) k = 1 n ρ k V n k c τ i ϵ k C n i = 1 n δ i | f i ˜ | · max 1 i n | j = 1 n ρ k V n k c τ i ϵ k | + C n · max 1 i n | j = 1 m μ j i | · max 1 i n | k = 1 n ρ k V n k c τ i ϵ k | + o ( n 1 / 2 log n ) = ( n 1 / 4 ) a . s . n 1 A 5 n C n | i = 1 n ρ i x ˜ i c β | · max 1 i n | k = 1 n ρ k V n k c τ i δ k | = O ( n 1 / 4 log n ) n 1 A 6 n C n i = 1 n | ρ i ϵ i | · max 1 i n | k = 1 n V n k c τ i ρ k δ k | = o ( n 1 / 4 ) a . s .
Similarly, we can demonstrate that n 1 A k n = o ( n 1 / 4 ) almost surely when k = 7 , 8 , 9 , 10 . As a result, we have proven Theorem 1(a). □
Proof of Theorem 2(a).
Based on Equations (5) and (6), one can write the following:
β ^ I β = R 2 n 2 { i = 1 n x ˜ i ( x i + δ i ) ( 1 ρ i ) ( β ^ c β ) + i = 1 n ( 1 ρ i ) δ ˜ i ( x i + δ i ) ( β ^ c β ) i = 1 n ( 1 ρ i ) x ˜ i [ g ( τ i ) g ^ n c ( τ i ) ] i = 1 n ( 1 ρ i ) δ ˜ i [ g ( τ i ) g ^ n c ( τ i ) ] + i = 1 n ρ i x ˜ i ( ϵ i δ i β ) i = 1 n ρ i ( δ i 2 Ξ δ 2 ) β + i = 1 n δ ˜ i g ˜ i c + i = 1 n x ˜ i g ˜ i c . + i = 1 n ρ i δ i ϵ i i = 1 n k = 1 n V n k ( τ i ) ( 1 ρ k ) x ˜ i ( x k + δ k ) ( β ^ c β ) . i = 1 n k = 1 n V n k ( τ i ) ( 1 ρ k ) δ ˜ i ( x k + δ k ) ( β ^ c β ) + i = 1 n k = 1 n V n k ( τ i ) ( 1 ρ k ) x ˜ i [ g ( τ k ) g ^ n c ( τ k ) ] + i = 1 n k = 1 n V n k ( τ i ) ( 1 ρ k ) δ ˜ i [ g ( τ k ) g ^ n c ( τ k ) ] i = 1 n k = 1 n V n k ( τ i ) ρ k x ˜ i ϵ k + i = 1 n k = 1 n V n k ( τ i ) ρ k x ˜ i δ k β + i = 1 n k = 1 n V n k ( τ i ) ρ i δ i δ k β + i = 1 n k = 1 n V n k ( τ i ) ρ k δ i δ k β i = 1 n k = 1 n j = 1 n V n k ( τ i ) V n j ( τ i ) ρ k δ k δ j β i = 1 n k = 1 n V n k ( τ i ) ρ i δ k ϵ i i = 1 n k = 1 n V n k ( τ i ) ρ k δ i ϵ k + j = 1 n k = 1 n V n k ( τ i ) V n j ( τ i ) ρ k δ j ϵ k } : = R 2 n 2 k = 1 21 D k n
Similar to S 1 n 2 C n 1 a . s . in the proof of Theorem 1(a), we can establish that S 2 n 2 C n 1 a . s . Hence, it is sufficient to confirm that n 1 D k n = o ( n 1 / 4 ) a.s. over the range k = 1 to 21. Utilizing (H0)–(H4), Lemmas 1, 2, and 7, as well as Theorem 1(a) and Equations (12)–(17), we derive that
n 1 D 1 n = n 1 i = 1 n x ˜ i 2 1 ρ i ( β ^ c β ) + n 1 i = 1 n | x ˜ i 1 ρ i | · max 1 i n | k = 1 n V n k τ i x k | ( β ^ c β ) + n 1 | i = 1 n x ˜ i δ i ( 1 ρ i ) | ( β ^ c β ) = o ( n 1 / 4 ) a . s . n 1 D 2 n = n 1 i = 1 n δ ˜ i ( 1 ρ i ) ( x i + δ i ) ( β ^ c β ) n 1 | i = 1 n ( 1 ρ i ) x i δ i | · | β ^ c β | + n 1 | i = 1 n ( 1 ρ i ) ( δ i 2 Ξ δ 2 ) | · | β ^ c β | + C n i = 1 n ( 1 ρ i ) · | β ^ c β | + n 1 | i = 1 n x i | · max 1 i n | k = 1 n V n k ( τ i ) δ k | · | β ^ c β | + n 1 | i = 1 n δ i | · max 1 i n | k = 1 n V n k ( τ i ) δ k | · | β ^ c β | = o ( n 1 / 4 ) a . s . n 1 D 3 n n 1 i = 1 n | ( 1 ρ i ) x ˜ i | · max 1 i n | g ( τ i ) g ^ n c ( τ i ) | = o ( n 1 / 4 ) a . s . n 1 D 5 n = n 1 i = 1 n ρ i x ˜ i ( ϵ i δ i β ) = o ( n 1 / 4 ) a . s . n 1 D 10 n = n 1 i = 1 n | x ˜ i | · max 1 i n | k = 1 n V n k ( τ i ) | · max 1 k n | f ( τ k ) + μ k | · | β ^ c β | + n 1 i = 1 n | x ˜ i | · max 1 i n | k = 1 n V n k ( τ i ) ( 1 ρ k ) δ k | · | β ^ c β | = o ( n 1 / 4 ) a . s . n 1 D 11 n = n 1 i = 1 n | δ ˜ i | · max 1 i n | k = 1 n V n k ( τ i ) | · max 1 k n | f ( τ k ) + μ k | · | β ^ c β | + n 1 i = 1 n | δ ˜ i | · max 1 i n | k = 1 n V n k ( τ i ) ( 1 ρ k ) δ k | · | β ^ c β | n 1 i = 1 n | δ i | E | δ i | + i = 1 n E | δ i | i = 1 n | k = 1 n V n k τ i δ k | · o ( n 1 4 ) + o ( n 1 4 ) = o ( n 1 / 4 ) a . s . n 1 D 13 n n 1 i = 1 n δ ˜ i · max 1 i n | k = 1 n 1 ρ k V n k τ i | · max 1 i n g τ i g ^ n c τ i n 1 i = 1 n | δ i | E | δ i | + i = 1 n E | δ i | i = 1 n | k = 1 n V n k τ i δ k | · o ( n 1 / 4 ) = o ( n 1 / 4 ) a . s . n 1 D 14 n n 1 i = 1 n ( f i ˜ + μ i j = 1 n V n j τ i μ j ) · max 1 i n | k = 1 n ρ k V n k ( τ i ) ϵ k | = o ( n 1 / 4 ) a . s .
The almost sure convergence of n 1 D k n = o n 1 / 4 for n = 4 , 6 , , 9 , 12 , 15 , , 21 is completely similar. Hence, we have concluded Theorem 2(a). □
Proof of Theorem 2(b).
Based on (5)–(7), it is possible to state that
g ^ n I ( τ ) g ( τ ) = k = 1 n V n k ( τ ) ( 1 ρ k ) x k ( β ^ c β ) + k = 1 n V n k ( τ ) x k ( β β ^ I ) k = 1 n V n k ( τ ) δ k ( β β ^ c ) + k = 1 n V n k ( τ ) δ k ( β β ^ I ) k = 1 n V n k ( τ ) ρ k δ k β ^ c + k = 1 n V n k ( τ ) ρ k ϵ k + k = 1 n V n k ( τ ) ( 1 ρ k ) [ g ^ n c ( τ k ) g ( τ k ) ] + k = 1 n V n k ( τ ) [ g ( τ k ) g ( τ ) ] : = k = 1 8 I k n
Thus, it is sufficient to demonstrate that I k n ( τ ) = o ( n 1 4 ) a.s. over the range k = 1 to 8. Using (H0)–(H4), Lemma 7, Equation (13), and Theorem 1 and Theorem 2(a), we find that
I 1 n ( τ ) | β β ^ c | · | k = 1 n V n k ( τ ) ( 1 ρ k ) x k | | β β ^ c | · | k = 1 n V n k ( τ ) ( 1 ρ k ) [ f τ k + μ k ] | o ( n 1 / 4 ) · | k = 1 n V n k ( τ ) | + max 1 k n | V n k ( τ ) ( 1 ρ k ) | · max 1 m n | j = 1 m μ k j | = o ( n 1 / 4 ) a . s . I 7 n ( τ ) max 1 k n | g ( τ k ) g ^ n c ( τ k ) | · | k = 1 n V n k ( τ ) ( 1 ρ k ) | = o ( n 1 / 4 ) a . s . I 8 n ( τ ) | k = 1 n V n k ( τ ) ( g ( τ k ) g ( τ ) ) | | k = 1 n V n k ( τ ) [ g ( τ k ) g ( τ ) ] · I | τ k τ | > a · n 1 / 4 | + | k = 1 n V n k ( τ ) [ g ( τ k ) g ( τ ) ] · I | τ k τ | < a · n 1 / 4 | = o ( n 1 / 4 ) a . s .
By utilizing (17), we demonstrate that n 1 I k n τ = o n 1 / 4 , which holds almost surely for k values ranging from 2 to 6, is analogous. Therefore, the demonstration of Theorem 2(b) has been finalized. □
Proof of Theorem 1(b).
To establish Theorem 1(b), we employ a proof strategy analogous to that used in our derivation of Theorem 2(b). □
Proof of Theorem 3.
To establish Theorem 3, we employ a proof strategy analogous to that used in our derivation of Theorem 2. □

6.2. Proof of Asymptotic Normality

Proof of Theorem 5(a).
The demonstration of Theorem 4(a) follows a comparable pattern. Next, we proceed to prove Theorem 5(a). Based on (6), it is feasible to express that
β ^ I β = R 2 n 2 { i = 1 n ρ i ( x ˜ i + Ω + Φ x ˜ i c + 1 + Ω + Φ δ i ) ϵ i δ i β + 1 + Ω + Φ Ξ δ 2 β . + Ω + Φ k = 4 10 A k n + i = 1 n k = 1 n x ˜ i 1 ρ i V n k τ i x k . + i = 1 n x ˜ i 1 ρ i δ i ( β ^ c β ) + [ i = 1 n 1 ρ i δ i 2 E δ i 2 + i = 1 n 1 ρ i x i δ i i = 1 n 1 ρ i x i k = 1 n V n k τ i δ k . . i = 1 n 1 ρ i δ i k = 1 n V n k τ i δ k ] ( β ^ c β ) i = 1 n 1 ρ i x ˜ i g τ i g ^ n c τ i + i = 1 n δ ˜ i g ˜ i c + i = 1 n x ˜ i g ˜ i c i = 1 n 1 ρ i δ ˜ i g τ i g ^ n c τ i i = 1 n k = 1 n V n k ( τ i ) ( 1 ρ k ) x ˜ i ( x k + δ k ) ( β ^ c β ) . i = 1 n k = 1 n V n k ( τ i ) ( 1 ρ k ) δ ˜ i ( x k + δ k ) ( β ^ c β ) + i = 1 n k = 1 n V n k ( τ i ) ( 1 ρ k ) x ˜ i [ g ( τ k ) g ^ n c ( τ k ) ] + i = 1 n k = 1 n V n k ( τ i ) ( 1 ρ k ) δ ˜ i [ g ( τ k ) g ^ n c ( τ k ) ] i = 1 n k = 1 n V n k ( τ i ) ρ k x ˜ i ϵ k + i = 1 n k = 1 n V n k ( τ i ) ρ k x ˜ i δ k β + i = 1 n k = 1 n V n k ( τ i ) ρ i δ i δ k β + i = 1 n k = 1 n V n k ( τ i ) ρ k δ i δ k β i = 1 n k = 1 n j = 1 n V n k ( τ i ) V n j ( τ i ) ρ k δ k δ j β i = 1 n k = 1 n V n k ( τ i ) ρ i δ k ϵ i i = 1 n k = 1 n V n k ( τ i ) ρ k δ i ϵ k + i = 1 n k = 1 n V n k ( τ i ) V n j ( τ i ) ρ k δ j ϵ k } : = R 2 n 2 k = 1 20 D k n
Set ω i = ρ i ( x ˜ i + Ω + Φ x ˜ i c + 1 + Ω + Φ δ i ) ϵ i δ i β + 1 + Ω + Φ Ξ δ 2 β , Z n i = ω i / Θ 2 n . With E Z n i = 0 , V a r i = 1 n Z n i = 1 , E Z n i 2 + δ < , based on (H0) and Lemma 7 along with the condition Θ 2 n 2 C n , we utilize Bernstein’s big-block and small-block technique. Setting k = n / s + r , we proceed as follows:
y n b = i = k b k b + s 1 Z n i , y n b = i = l b l b + r 1 Z n i , y n k + 1 = i = k s + r + 1 n Z n i ,
where k b = b + 1 s + r + 1 , l b = b + 1 s + r + s + 1 , b ranges from 1 to k. Consequently,
D 1 n Θ 2 n = i = 1 n Z n i = b = 1 k y n b + b = 1 k y n b + y n k + 1 : = D n + D n + D n
To establish that R 2 n 2 ( β ^ I β ) / Θ 2 n D N ( 0 , 1 ) , it is enough to demonstrate that
D n D N 0 , 1 , E ( D n ) 2 0 , E ( D n ) 2 0 , D k n Θ 2 n P 0 f o r k = 2 , , 20 .
Stage 1. We demonstrate that E ( D n ) 2 0 , E ( D n ) 2 0 and D n D N 0 , 1 . Applying Lemma 5(b), from (H0) to (H3) and (H7), accordingly, we verify that
E ( D n ) 2 C n · E [ b = 1 k i = l b l b + r 1 ρ i ( x ˜ i + δ i ) ( ϵ i δ i β ) + Ξ δ 2 β + Ω + Φ i = 1 n i = l b l b + r 1 ρ i ( x ˜ i c + δ i ) ( ϵ i δ i β ) + Ξ δ 2 β ] 2 C 1 n · E b = 1 k i = l b l b + r 1 ρ i x ˜ i ϵ i ρ i x ˜ i δ i β + ρ i δ i ϵ i ρ i ( δ i 2 Ξ δ 2 ) β 2 + C 2 n · E b = 1 k i = l b l b + r 1 ρ i x ˜ i c ϵ i ρ i x ˜ i c δ i β + ρ i δ i ϵ i ρ i ( δ i 2 Ξ δ 2 ) β 2 C 1 q p · max 1 i n | x ˜ i | 2 + C 2 s r · max 1 i n | x ˜ i c | 2 + C 2 k r n = o ( 1 )
E ( D n ) 2 C 1 n · E [ i = k ( s + r ) + 1 n ρ i ( x ˜ i + δ i ) ( ϵ i δ i β ) + Ξ δ 2 β 2 + C 2 n · E [ i = k ( s + r ) + 1 n ρ i ( x ˜ i c + δ i ) ( ϵ i δ i β ) + Ξ δ 2 β 2 C 1 s n · max 1 i n | x ˜ i | 2 + C 2 s n · max 1 i n | x ˜ i c | 2 + C 2 s n = o ( 1 )
Next, we verify that D n D N ( 0 , 1 ) . It is enough to demonstrate that
V a r ( D n ) 1
| E exp ( i t b = 1 k y n b ) b = 1 k E exp ( i t y n b ) | 0
U n ( ω ) = b = 1 k E y n b 2 I y n b > ω 0 , ω > 0
It is apparent that E D n + D n + D n 2 = 1 and
E D n + D n + D n 2 = E D n 2 + E D n 2 + E D n 2 + 2 E D n D n 2 + 2 E D n D n 2 + E D n D n 2
From the given definition of D n , one can hence straightforwardly confirm that E D n 2 < by referring to Equations (18) and (19).
It is straightforward to verify
V a r D n 1 E ( D n + D n ) 2 + 2 E ( D n D n ) + 2 E ( D n D n ) 2 E ( D n ) 2 + E ( D n ) 2 + 2 E ( D n ) 2 1 2 + 2 E ( D n ) 2 1 2 = o ( 1 )
This results in (20). For (21), by applying Lemma 3 and (H7), we obtain
| E exp ( i t b = 1 k y n b ) b = 1 k E exp ( i t y n b ) | 0
We thus need to verify (22). Suppose that λ > 2 + γ / γ . Set v = 2 1 + t and η = γ 2 t , where 0 < t < γ 2 . We then select t in such a way that
2 + γ γ < 2 + γ 1 + t γ 2 t = v v + η 2 η < λ
Obviously, v + η = 2 + γ , by applying Lemmas 5 and 7 from (H0) to (H7) and Θ 2 n 2 C n . For small values of ϵ where 0 < ε < t , we discover that
U n ( ω ) C b = 1 k E y n b v = C b = 1 k E | i = k b k b + s 1 Z n i | v C n v 2 b = 1 k s ε i = k b k b + s 1 E ω i v + i = k b k b + s 1 ω i 2 + γ 2 v 2 C n 1 + t { s ε b = 1 k i = k b k b + s 1 | x ˜ i c | 2 ( 1 + t ) + | x ˜ i | 2 ( 1 + t ) + O ( 1 ) + b = 1 k [ i = k b k b + s 1 | x ˜ i c | 2 + | x ˜ i | 2 + O ( 1 ) ) ] 1 + t } C n 1 + t [ n s ε · ( max 1 i n | x ˜ i c | ) 2 t + n s ε · ( max 1 i n | x ˜ i | ) 2 t + n s t · ( max 1 i n | x ˜ i c | ) 2 t + n p t · ( max 1 i n | x ˜ i | ) 2 t + C 1 n s t ] = o ( 1 )
Thus, we know that D 1 n / Θ 2 n D N ( 0 , 1 ) .
Stage 2. We demonstrate that D k n / Θ 2 n P 0 , k = 2 , , 20 .
D 3 n = P 1 n + P 2 n ( β ^ c β ) P 1 n Θ 2 n C n · i = 1 n | f ˜ i + μ i j = 1 n V n j τ i μ j | · max 1 i n 1 ρ i · max 1 i n | k = 1 n V n k τ i f τ k + μ k | = o ( n 1 / 2 ) P 2 n Θ 2 n C n | i = 1 n x ˜ i 1 ρ i δ i | = O ( n 1 / 4 log n ) D 5 n Θ 2 n C n { max 1 i n | 1 ρ i | [ n · max 1 i n | f ˜ i | · max 1 i n | g ^ n c τ i g τ i | + max 1 m n | i = 1 m μ j i | · max 1 i n | g ^ n c τ i g τ i | + max 1 i n | j = 1 n V n k ( τ i ) | · max 1 i n | g ^ n c τ i g τ i | · max 1 m n | i = 1 m μ j k | ] } = o ( 1 ) a . s . E D 6 n Θ 2 n 2 C n E i = 1 n δ i g ˜ i c 2 + E i = 1 n g ˜ i c k = 1 n V n k τ i δ k 2 = o ( n 1 / 2 ) D 7 n Θ 2 n C n | i = 1 n x ˜ i g ˜ i c | C n | i = 1 n f ˜ i g ˜ i c | + | i = 1 n μ i g ˜ i c | + | i = 1 n k = 1 n V n k τ i μ k g ˜ i c | C n [ n · max 1 i n | f ˜ i | · max 1 i n | g ˜ i c | + max 1 m n | i = 1 m μ j i | · max 1 i n | g ˜ i c | + max 1 k n | k = 1 n V n k τ i | · max 1 i n | g ˜ i c | · max 1 m n | k = 1 m μ j k | ] = o ( 1 ) D 9 n = i = 1 n x i ˜ k = 1 n V n k τ i 1 ρ j x k + i = 1 n x i ˜ k = 1 n V n k τ i 1 ρ k δ k β ^ c β : = T 1 n + T 2 n ( β ^ c β ) T 1 n Θ 2 n C n i = 1 n | f ˜ i + μ i j = 1 n V n j τ i μ j | · max 1 i n | k = 1 n V n k τ i 1 ρ i x k | = O ( n 1 / 2 ) E T 2 n Θ 1 n 2 C n i = 1 n k = 1 n x i ˜ V n k τ i ( 1 ρ k ) 2 Ξ δ 2 = O ( 1 ) D 14 n = i = 1 n f i ˜ k = 1 n ρ k V n k τ i δ k β + i = 1 n μ i k = 1 n ρ k V n k τ i δ k β i = 1 n s = 1 n V n s τ i μ s k = 1 n δ k V n k τ i δ k β : = Q 1 n + Q 2 n + Q 3 n E Q 1 n Θ 2 n 2 C n k = 1 n i = 1 n V n k τ i f i ˜ 2 C · max 1 i n | f i ˜ | 2 · max 1 k n | i = 1 n V n k ( τ i ) | 2 = o ( n 1 / 2 ) E Q 2 n Θ 2 n 2 C n k = 1 n i = 1 n V n k τ i μ i 2 C · max 1 m n | i = 1 m μ j i | 2 · max 1 i , k n V n k 2 τ i = o ( 1 ) Q 3 n Θ 3 n C n · max 1 m n | s = 1 m v j s | · max 1 s n i = 1 n V n s τ i · max 1 i n | k = 1 n V n k τ i δ k | = o p ( 1 ) E D 16 n Θ 2 n 2 C n i 1 = 1 n i 2 = 1 n k 1 = 1 n k 2 = 1 n V n k 1 τ i 1 V n k 2 τ i 2 E δ i 1 δ i 2 δ k 1 δ k 2 C n i = 1 n k = 1 n V n k 1 2 τ i E δ i 2 δ k 2 + i 1 = 1 n i 2 = 1 n V n i 1 τ i 1 V n i 2 τ i 2 E δ i 1 2 δ i 2 2 = O ( n 1 ) E D 18 n Θ 2 n 2 C n · E i = 1 n k = 1 n V n k τ i ϵ i δ k 2 C n i = 1 n k = 1 n V n k τ i δ k 2 + η 2 · ϵ i 2 + η 2 C n i = 1 n k = 1 n V n k τ i 2 + η + k = 1 n V n k 2 τ i ( 2 + η ) / 2 2 / ( 2 + η ) = o ( n 1 / 2 log 1 n ) .
The demonstration that D k n = o p 1 for k = 2 , 4 , 8 , 10 , 11 , 12 , 13 , 15 , 17 , 19 , 20 proceeds in a similar manner. Thus, Theorem 5(a) is thereby proven. □
Proof of Theorem 5(b).
Based on (7), we have that
g ^ n [ I ] τ g τ = k = 1 n V n k ( τ ) ρ k + k = 1 n V n k ( τ ) ( 1 ρ k ) i = 1 n V n i c ( τ k ) ρ i ϵ k δ k β + [ k = 1 n V n k ( τ ) ( 1 ρ k ) x k + k = 1 n V n k ( τ ) ( 1 ρ k ) δ k k = 1 n V n k ( τ ) ( 1 ρ k ) i = 1 n V n i c ( τ k ) ρ i δ k k = 1 n V n k ( τ ) ( 1 ρ k ) i = 1 n V n i c ( τ k ) ρ i x k ] β ^ c β + k = 1 n V n k ( τ ) x k + k = 1 n V n k ( τ ) δ k β β ^ I + k = 1 n V n k ( τ ) ( 1 ρ k ) i = 1 n V n i c ( τ k ) ρ i + k = 1 n V n k ( τ ) δ k ( g τ k g τ ) : = U 1 n ( τ ) + [ U 2 n ( τ ) + U 3 n ( τ ) + U 4 n ( τ ) + U 5 n ( τ ) ] β ^ c β + [ U 6 n ( τ ) + U 7 n ( τ ) ] β β ^ I + ( U 8 n ( τ ) + U 9 n ( τ ) )
Analogous to the proof of 4(a) and 5(a). One can verify that U 1 n τ / Λ 2 n τ D N 0 , 1 . Observe that n Λ 2 n 2 τ as n . According to Theorems 4 and 5, we can obtain β β ^ c = O p ( n 1 / 2 ) and β β ^ I = O p ( n 1 / 2 ) . Thus, it is adequate to demonstrate that U k n τ = O ( n 1 / 2 ) for k = 8 , 9 . U s n τ = O p ( 1 ) for s = 2 , , 7 . For b > 0 ,
U 2 n τ | k = 1 n V n k τ 1 ρ k x k | | k = 1 n V n k τ 1 ρ k f τ k + μ k | max 1 k n | f τ k | · | k = 1 n V n k τ 1 ρ k | + max 1 k n | V n k τ 1 ρ k | · max 1 m n | k = 1 m μ j k | = O 1 E U 3 n τ 2 k = 1 n V n k τ 1 ρ k 2 Ξ δ 2 = O ( n 1 / 2 log 1 n ) E U 4 n τ 2 k = 1 n V n k τ 1 ρ k i = 1 n V n i c τ k ρ i 2 Ξ δ 2 = O ( n 1 / 2 log 1 n ) U 5 n τ | k = 1 n V n k τ 1 ρ k i = 1 n V n i c τ ρ i x k | | k = 1 n V n k τ 1 ρ k i = 1 n V n i c τ i ρ k f τ k + μ k | max 1 j n | f τ k | · | k = 1 n V n k τ 1 ρ k i = 1 n V n i c τ k ρ i | + max 1 k n | V n k τ 1 ρ k i = 1 n V n k c τ ρ i | · max 1 m n | k = 1 m μ j k | = O 1 U 8 n τ k = 1 n V n k τ 1 ρ k i = 1 n V n i c τ k ρ i g τ k g τ · I | τ k τ | > b · n 1 / 2 + k = 1 n V n k τ 1 ρ k i = 1 n V n i c τ k ρ i g τ k g τ · I | τ k τ | b · n 1 / 2 = O n 1 / 2
The argument for showing U k n = O p 1 for k = 6 , 7 , 9 follows in a similar manner. Therefore, Theorem 5(b) is proven. □
Proof of Theorem 6.
The proof of Theorem 6 bears a resemblance to the previous one. □

7. Appendix Proofs of Lemmas

Proof of Lemma 6. Set υ n i = Y i I ( | Y i | n 1 / p ) , υ n i = Y i I ( | Y i | > n 1 / p ) . Then,
i = 1 n b n i Y i i = 1 n b n i υ n i + i = 1 n b n i υ n i : = G 1 n + G 2 n
Simply put, if E | Y 1 | p < , it follows that i = 1 P ( | Y i | > i 1 / p ) < . This implies that i = 1 n | Y i | I ( | Y i | > i 1 / p ) < a.s. Consequently,
G 2 n = O ( n u ) i = 1 n | Y i | I ( | Y i | > n 1 / p ) = o p ( n ψ L 1 ( n ) ) b y ψ < u .
From the conditions E Y i = 0 , E | Y i | p < and ψ 1 1 / p , we can deduce that
| i = 1 n b n i E υ n i | = | i = 1 n b n i E Y i I ( | Y i | > n 1 / p ) | C E | Y 1 | I ( | Y 1 | > n 1 / p ) = O ( n 1 p p ) = o ( n ψ log n ) .
Next, we prove that G 1 n = o p ( n ψ ) . Based on (23), we establish that
P | i = 1 n b n i υ n i E υ n i | > τ n ψ 0 f o r a n y τ > 0 .
Drawing upon Lemma 5(a), by selecting a sufficiently large value for v > p and a sufficiently small value for η > 0 , we obtain
P | i = 1 n b n i ( υ n i E υ n i ) | > τ n ψ ( τ n ψ ) v E | i = 1 n b n i ( υ n i E υ n i ) | v C n ψ v · n η = 1 n E | b n i ( υ n i E υ n i ) | v + i = 1 n b n i ( υ n i E υ n i ) v + ρ 2 v 2 C n ψ v · n η i = 1 n | b n i | v · E | Y 1 | r I ( | Y 1 | n 1 / p ) + i = 1 n b n i 2 v 2 E | Y 1 | r + ρ I ( | Y 1 | n 1 / p ) v v + ρ C n ψ v · n η · n u ( v 2 ) · n τ · n v p p · L 1 ( n ) v 2 + C n ψ v · n τ v 2 · n v + ρ p p · v v + ρ C n η + ψ v u v + 2 u τ + v / p 1 L 1 ( n ) v 2 + n ψ v τ v / 2 + v / p + ρ / ( v + ρ ) 1
Let I 1 = ψ v u v + 2 u τ + v / p 1 and I 2 = ψ v τ v / 2 + v / p + ρ / ( r + ρ ) 1 . According to the premises in Lemma 6, it is shown that I 1 < 0 and I 2 < 0 , resulting in (25). Hence, the proof of Lemma 6 is finished.

8. Conclusions

This study presents robust methods for handling incomplete data with strong mixing errors, focusing on estimating β and g ( · ) under the Missing at Random (MAR) assumption. We establish the strong consistency and asymptotic normality of the proposed estimators, demonstrating their effectiveness in addressing data uncertainty. Among these approaches, the imputation estimator g ^ n [ I ] ( τ ) and β ^ I provide the most accurate estimates by leveraging both observed and imputed data. These results suggest that the imputation method is the most effective for managing incomplete data in the context of the semi-parametric EV model. In practical applications, such as policy formulation and resource allocation, where incomplete data are common, the imputed estimator is a reliable and efficient tool. By utilizing available information, it ensures more accurate parameter estimates, which are crucial for making informed decisions under data uncertainty.

Author Contributions

Conceptualization, J.Z.; methodology, J.Z.; software, H.Y.; validation, J.Z., H.Y. and T.H.; formal analysis, J.Z., H.Y. and T.H.; investigation, J.Z., H.Y. and T.H.; resources, J.Z.; data curation, J.Z., H.Y. and T.H.; writing—original draft preparation, J.Z. and H.Y.; writing—review and editing, J.Z., H.Y. and T.H.; visualization, H.Y.; supervision, J.Z.; project administration, J.Z.; funding acquisition, J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China (11701368).

Data Availability Statement

This paper used simulated data generated by MATLAB R2021a, not actual data.

Acknowledgments

The authors are deeply thankful for the insightful comments and recommendations provided by the Editor and referee.

Conflicts of Interest

Author Tingting Hu was employed by the company United Automotive Electronics Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Engle, R.F.; Granger, C.W.J.; Rice, J.; Weiss, A. Semiparametric estimates of the relation between weather and electricity sales. J. Am. Stat. Assoc. 1986, 81, 310–320. [Google Scholar] [CrossRef]
  2. Roozbeh, M.; Arashi, M. Least-trimmed squares: Asymptotic normality of robust estimator in semiparametric regression models. J. Stat. Comput. Simul. 2017, 87, 1130–1147. [Google Scholar] [CrossRef]
  3. Ding, L.W.; Chen, P.; Zhang, Q.; Li, Y.M. Asymptotic normality for wavelet estimators in heteroscedastic semi-parametric model with random errors. J. Syst. Sci. Complex. 2020, 33, 1212–1243. [Google Scholar] [CrossRef]
  4. Wei, C.; Wu, X. Error variance estimation in partially linear varying coefficient models. Math. Appl. 2008, 21, 378–383. [Google Scholar]
  5. Fu, F.; Zeng, Z.; Liu, X. Difference-based M-estimator of generalized semiparametric model with NSD errors. J. Inequal. Appl. 2019, 2019, 61. [Google Scholar] [CrossRef]
  6. Deaton, A. Panel data from time series of cross-sections. J. Econom. 1985, 30, 109–126. [Google Scholar] [CrossRef]
  7. Chen, P.; Kong, N.; Sung, S.H. Complete convergence for weighted sums of i.i.d. random variables with applications in regression estimation and EV model. Commun. Stat. 2016, 3599–3613. [Google Scholar] [CrossRef]
  8. You, J.; Zhou, X.; Zhou, Y. Statistical inference for panel data semiparametric partially linear regression models with heteroscedastic errors. J. Multiv. Anal. 2010, 101, 1079–1101. [Google Scholar] [CrossRef]
  9. Zhang, J.J.; Liang, H.Y.; Amei, A. Asymptotic normality of estimators in heteroscedastic errors-in-variables model. AStA Adv. Stat. Anal. 2014, 98, 165–195. [Google Scholar] [CrossRef]
  10. Emami, H. Ridge estimation in semiparametric linear measurement error models. Lin. Algebra Appl. 2018, 552, 127–146. [Google Scholar] [CrossRef]
  11. Miao, Y.W. Moderate deviations for LS estimator in simple linear EV regression model. J. Stat. Plan. Inference 2009, 139, 2263–2272. [Google Scholar] [CrossRef]
  12. Hu, D.; Chen, P.Y.; Sung, S.H. Strong laws for weighted sums of -mixing random variables and applications in errors-in-variables regression models. Stat. Methods Appl. 2017, 26, 600–617. [Google Scholar] [CrossRef]
  13. Zhang, Y.; Liu, X.; Yu, Y.; Hu, H. Asymptotic normality and mean consistency of LS estimators in the errors-in-variables model with dependent errors. Open Math. 2020, 18, 930–947. [Google Scholar] [CrossRef]
  14. Zou, Y.Y.; Wu, C.X.; Fan, G.L.; Zhang, R.Q. Estimation for a hybrid model of functional and linear measurement errors regression with missing response. Statistics 2022, 56, 271–296. [Google Scholar] [CrossRef]
  15. Xiao, Y.T.; Li, F.X. Estimation in partially linear varying-coefficient errors-in-variables models with missing response variables. Comput. Stat. 2020, 35, 1637–1658. [Google Scholar] [CrossRef]
  16. Zou, Y.Y.; Wu, C.X. Statistical inference for the heteroscedastic partially linear varying-coefficient errors-in-variables model with missing censoring indicators. Discret. Dyn. Nat. Soc. 2021, 2021, 1141022. [Google Scholar] [CrossRef]
  17. Xi, M.M.; Wang, R.; Yu, W.; Shen, Y.; Wang, X.Y. Asymptotic properties for the estimators in heteroscedastic semiparametric EV models with α-mixing errors. Statistics 2021, 54, 1232–1254. [Google Scholar] [CrossRef]
  18. Zhang, J.J.; Liu, C.L. Asymptotic properties for estimators in a semiparametric EV model with NA errors and missing responses. Discret. Dyn. Nat. Soc. 2022, 2022, 4862820. [Google Scholar] [CrossRef]
  19. Zhang, J.J.; Yang, X. Statistical inference for estimators in a semiparametric EV model with linear process errors and missing responses. Math. Probl. Eng. 2023, 2023, 2547329. [Google Scholar] [CrossRef]
  20. Liang, H.Y.; Fan, G.L. Berry-Esseen type bounds of estimators in a semiparametric model with linear process errors. J. Multiv. Anal. 2009, 100, 1–15. [Google Scholar] [CrossRef]
  21. Liang, H.Y.; Baek, J.I. Convergence for weighted sums of negatively associated random variables. Stat. Probab. Lett. 2004, 41, 883–894. [Google Scholar] [CrossRef]
  22. Wang, Q.H.; Sun, Z.H. Estimation in partially linear models with missing responses at random. J. Multiv. Anal. 2007, 98, 1470–1493. [Google Scholar] [CrossRef]
  23. Härdle, W.; Liang, H.; Gao, J.T. Partial Linear Models; Physica-Verlag: Heidelberg, Germany, 2000. [Google Scholar]
  24. Xu, B. The convergence of the weighted sum for strong mixing dependent variable and its application. J. Math. 2002, 45, 1025–1034. [Google Scholar]
  25. Volkonskii, V.A.; Rozanov, Y.A. Some limit theorems for random functions. Theory Probab. 1959, 4, 178–197. [Google Scholar] [CrossRef]
  26. Hall, P.; Heyde, C.C. Martingale Limit Theory and Its Applications; Academic Press: Cambridge, MA, USA, 1980. [Google Scholar]
  27. Miao, Y.; Yang, G.; Shen, L. The central limit theorem for LS estimator in simple linear EV regression models. Theory Methods 2007, 36, 2263–2272. [Google Scholar] [CrossRef]
  28. Zhang, J.J.; Liang, H.Y. Asymptotic normality of estimators in heteroscedastic semi-parametric model with strong mixing errors. Commun. Stat. Theory Methods 2012, 41, 2172–2201. [Google Scholar] [CrossRef]
Figure 1. Q-Q plots of β ^ c for M = 500 and sample sizes n = 100 and n = 500 .
Figure 1. Q-Q plots of β ^ c for M = 500 and sample sizes n = 100 and n = 500 .
Symmetry 16 01544 g001
Figure 2. Q-Q plots of β ^ I for M = 500 and sample sizes n = 100 and n = 500 .
Figure 2. Q-Q plots of β ^ I for M = 500 and sample sizes n = 100 and n = 500 .
Symmetry 16 01544 g002
Figure 3. Q-Q plots of β ^ R for M = 500 and sample sizes n = 100 and n = 500 .
Figure 3. Q-Q plots of β ^ R for M = 500 and sample sizes n = 100 and n = 500 .
Symmetry 16 01544 g003
Figure 4. Q-Q plots of g ^ n c 0.5 for M = 500 and sample sizes n = 100 and n = 500 .
Figure 4. Q-Q plots of g ^ n c 0.5 for M = 500 and sample sizes n = 100 and n = 500 .
Symmetry 16 01544 g004
Figure 5. Q-Q plots of g ^ n I 0.5 for M = 500 and sample sizes n = 100 and n = 500 .
Figure 5. Q-Q plots of g ^ n I 0.5 for M = 500 and sample sizes n = 100 and n = 500 .
Symmetry 16 01544 g005
Figure 6. Q-Q plots of g ^ n R 0.5 for M = 500 and sample sizes n = 100 and n = 500 .
Figure 6. Q-Q plots of g ^ n R 0.5 for M = 500 and sample sizes n = 100 and n = 500 .
Symmetry 16 01544 g006
Table 1. The MSE comparison for three estimators of β with α = 0.1 and α = 0.3 .
Table 1. The MSE comparison for three estimators of β with α = 0.1 and α = 0.3 .
np β ^ c ( α = 0.1 ) β ^ c ( α = 0.3 ) β ^ I ( α = 0.1 ) β ^ I ( α = 0.3 ) β ^ R ( α = 0.1 ) β ^ R ( α = 0.3 )
1000.1 5.2900 × 10 4 7.8500 × 10 4 4.2470 × 10 4 5.9300 × 10 4 3.7200 × 10 4 5.0100 × 10 4
3000.1 3.5000 × 10 4 4.3800 × 10 4 2.9600 × 10 4 3.7300 × 10 4 2.6800 × 10 4 3.5500 × 10 4
5000.1 1.4100 × 10 4 2.4100 × 10 4 1.3690 × 10 4 2.0900 × 10 4 1.1590 × 10 4 1.9900 × 10 4
1000.25 8.4900 × 10 4 1.6000 × 10 3 7.4200 × 10 4 1.4200 × 10 3 7.0100 × 10 4 1.3000 × 10 3
3000.25 5.9000 × 10 4 6.5200 × 10 4 3.1800 × 10 4 4.0790 × 10 4 4.2300 × 10 4 3.9400 × 10 4
5000.25 2.2800 × 10 4 3.3790 × 10 4 1.8000 × 10 4 2.2760 × 10 4 2.6700 × 10 4 2.9500 × 10 4
1000.5 1.7300 × 10 3 1.8300 × 10 3 1.2000 × 10 3 1.500 × 10 3 1.2400 × 10 3 1.6000 × 10 3
3000.5 3.8500 × 10 4 8.2600 × 10 4 3.3200 × 10 4 6.1200 × 10 4 3.3900 × 10 4 6.4500 × 10 4
5000.5 3.8300 × 10 4 4.2000 × 10 4 2.5000 × 10 4 3.3700 × 10 4 3.1000 × 10 4 3.3500 × 10 4
Table 2. The MSE comparison for three estimators of g ( τ ) with α = 0.1 and α = 0.3 .
Table 2. The MSE comparison for three estimators of g ( τ ) with α = 0.1 and α = 0.3 .
np g ^ n c ( α = 0.1 ) g ^ n c ( α = 0.3 ) g ^ n [ I ] ( α = 0.1 ) g ^ n [ I ] ( α = 0.3 ) g ^ n [ R ] ( α = 0.1 ) g ^ n [ R ] ( α = 0.3 )
1000.1 2.6900 × 10 2 3.5800 × 10 2 2.4400 × 10 2 3.4700 × 10 2 2.6200 × 10 2 3.4500 × 10 2
3000.1 3.4700 × 10 3 1.9800 × 10 2 3.3200 × 10 3 1.6100 × 10 2 3.2700 × 10 3 1.6100 × 10 2
5000.1 2.6100 × 10 3 1.5100 × 10 2 2.3700 × 10 3 1.1700 × 10 2 2.3500 × 10 3 1.1500 × 10 2
1000.25 2.9000 × 10 2 3.9500 × 10 2 2.6600 × 10 2 3.5300 × 10 2 2.8300 × 10 2 3.6200 × 10 2
3000.25 3.6500 × 10 3 2.1100 × 10 2 3.3400 × 10 3 1.7700 × 10 2 3.3300 × 10 3 1.8200 × 10 2
5000.25 2.7200 × 10 3 1.7700 × 10 2 2.4400 × 10 3 1.3100 × 10 2 2.4000 × 10 3 1.2500 × 10 2
1000.5 3.1300 × 10 2 4.3400 × 10 2 2.7500 × 10 2 4.0000 × 10 2 2.9400 × 10 2 4.1000 × 10 2
3000.5 3.8500 × 10 3 2.3800 × 10 2 3.3500 × 10 3 2.0800 × 10 2 3.3900 × 10 3 2.0500 × 10 2
5000.5 2.8300 × 10 3 1.8700 × 10 2 2.5000 × 10 3 1.7400 × 10 2 2.6500 × 10 3 1.8200 × 10 2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, J.; Yan, H.; Hu, T. Symmetrical Convergence Rates and Asymptotic Properties of Estimators in a Semi-Parametric Errors-in-Variables Model with Strong Mixing Errors and Missing Responses. Symmetry 2024, 16, 1544. https://doi.org/10.3390/sym16111544

AMA Style

Zhang J, Yan H, Hu T. Symmetrical Convergence Rates and Asymptotic Properties of Estimators in a Semi-Parametric Errors-in-Variables Model with Strong Mixing Errors and Missing Responses. Symmetry. 2024; 16(11):1544. https://doi.org/10.3390/sym16111544

Chicago/Turabian Style

Zhang, Jingjing, Haiqin Yan, and Tingting Hu. 2024. "Symmetrical Convergence Rates and Asymptotic Properties of Estimators in a Semi-Parametric Errors-in-Variables Model with Strong Mixing Errors and Missing Responses" Symmetry 16, no. 11: 1544. https://doi.org/10.3390/sym16111544

APA Style

Zhang, J., Yan, H., & Hu, T. (2024). Symmetrical Convergence Rates and Asymptotic Properties of Estimators in a Semi-Parametric Errors-in-Variables Model with Strong Mixing Errors and Missing Responses. Symmetry, 16(11), 1544. https://doi.org/10.3390/sym16111544

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop