Next Article in Journal
A Comparative Analysis of Compression and Transfer Learning Techniques in DeepFake Detection Models
Previous Article in Journal
Center of Trapezoid Graph: Application in Selecting Center Location to Set up a Private Hospital
Previous Article in Special Issue
Assessing Scientific Text Similarity: A Novel Approach Utilizing Non-Negative Matrix Factorization and Bidirectional Encoder Representations from Transformer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Limit Theorems for Kernel Regression Estimator for Quasi-Associated Functional Censored Time Series Within Single Index Structure

1
Department of Mathematics, University of Sciences and Technology Mohamed Boudiaf, Oran BP 1505, El Mnaouar-Oran 31000, Algeria
2
LMAC (Laboratory of Applied Mathematics of Compiègne), Université de Technologie de Compiègne, CS 60 319-60 203 Compiègne Cedex, 60203 Compiègne, France
3
Department of Mathematics, College of Science, King Khalid University, P.O. Box 9004, Abha 62529, Saudi Arabia
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2025, 13(5), 886; https://doi.org/10.3390/math13050886
Submission received: 24 January 2025 / Revised: 19 February 2025 / Accepted: 5 March 2025 / Published: 6 March 2025
(This article belongs to the Special Issue Probability, Stochastic Processes and Machine Learning)

Abstract

:
In this paper, we develop kernel-based estimators for regression functions under a functional single-index model, applied to censored time series data. By capitalizing on the single-index structure, we reduce the dimensionality of the covariate-response relationship, thereby preserving the ability to capture intricate dependencies while maintaining a relatively parsimonious form. Specifically, our framework utilizes nonparametric kernel estimation within a quasi-association setting to characterize the underlying relationships. Under mild regularity conditions, we demonstrate that these estimators attain both strong uniform consistency and asymptotic normality. Through extensive simulation experiments, we confirm their robust finite-sample performance. Moreover, an empirical examination using intraday Nikkei stock index returns illustrates that the proposed method significantly outperforms traditional nonparametric regression approaches.
MSC:
62G20; 62G05; 62G32; 62G08; 62G35; 62G07; 62E20

1. Introduction

The statistical analysis of single-index models has been thoroughly developed and examined from both theoretical and practical perspectives. Initial studies concentrated on models with vector explanatory variables, as emphasized by [1,2]. A significant advancement to the functional framework was introduced by [3], where explanatory variables are presented in an infinite-dimensional space, such as a Hilbert space. This advancement, known as the single functional index model, has attracted considerable attention due to its capacity to address two pivotal challenges: dimensionality reduction and interpretability. Specifically, the model alleviates the curse of dimensionality and offers meaningful insights through the interpretable index parameter, θ .
The model investigated in this study lies between traditional single-index models and functional regression models, functioning as a generalization of the functional linear model. For instance, ref. [4] introduced a nonparametric estimation method for the link function using adaptive linear projections, demonstrating polynomial convergence rates. Building on this work, ref. [5] further extended functional linear regression models by proposing functional partially linear single-index models, utilizing least squares estimation to achieve similar convergence rates.
The single functional index model integrates the benefits of single-index models with the application potential of functional linear models [6] and nonparametric functional regression models [7,8,9,10,11,12,13,14]. Functional data modeling has become increasingly prominent, driven by technological advancements that allow for precise measurements and the generation of extensive datasets over continuous domains. However, most theoretical progress assumes independence among observations, a condition seldom met in practice. Instead, functional time series often display dependence, highlighting the necessity to explore models that relax the independence assumption.
This paper tackles the nonparametric estimation problem for regression functions in single functional index models under conditions of weak dependence. While strong mixing conditions are typically used to model dependence, this study adopts the more general quasi-association framework, initially introduced for real-valued random fields by [15]. Quasi-association extends the concept of positive association, as defined by [16], and can be considered a specific instance within the weak dependence framework outlined by [17].
From a practical perspective, quasi-associated data are crucial in areas such as reliability theory, mathematical physics, and percolation theory [18]. Theoretically, this correlation concept provides a unified approach to addressing both mixing and association dependencies in stochastic processes.
The single-index model, as a semi-parametric framework, bridges the strengths of parametric and nonparametric regression. Its application in multivariate statistics has been well documented by [2,19,20], with further developments made by [21,22,23]. In the realm of functional statistics, single functional index models have mainly been investigated for functional data, beginning with the foundational work of [3] on independent and identically distributed (i.i.d.) cases. Extensions to dependent data have been explored by [24,25]. Additionally, ref. [26] proposed a cross-validation-based estimation method for both nonparametric regression functions and linear coefficients, demonstrating its asymptotic optimality.
Subsequent research has introduced innovative estimation techniques. For example, ref. [27] utilized Bayesian methods to estimate bandwidths in kernel density and regression functions under autoregressive structures, showing enhanced predictive performance. ref. [28] developed a location-adaptive k-nearest neighbors approach for functional single-index models. Furthermore, ref. [29] created functional varying-coefficient single-index models for analyzing imaging data, while ref. [30] investigated Hilbertian regressors for the nonparametric estimation of conditional cumulative distributions. ref. [31] integrated multivariate random variables and random curves within a functional partial linear single-index framework. For recent works, refer to [32,33].
The concept of quasi-association for Hilbert space-valued random variables was introduced by [34], followed by limit theorems and regression-related advances, such as conditional density estimation [35] and asymptotic normality under dependence [36]. Other developments include local quadratic approximations for second-derivative link functions [37] and asymptotic normality for quasi-associated data in Hilbert spaces [38].
In this paper, we establish the almost complete convergence rates and asymptotic normality for the kernel single-index estimator in the context of censored data. This investigation represents a fundamentally unresolved open problem in the literature. Our objective is to bridge this gap by integrating the results of [34,39,40] with techniques designed to handle associated functional data and censored data. However, as will be demonstrated, the problem requires more than merely combining existing ideas. Instead, it requires intricate mathematical derivations to manage the associated functional data within our specific context, necessitating the effective application of large-sample theory techniques. Additionally, we establish the asymptotic distribution along with the corresponding confidence interval. This paper also includes a comprehensive section on the computational aspects of the proposed method, applied to both simulated and real data.
The remainder of this work is organized as follows. In Section 2, we describe the nonparametric model and the kernel estimator. Section 3 presents our primary results on the almost complete convergence of the Nadaraya–Watson kernel estimator under fairly general conditions. Section 4 discusses the asymptotic distribution. Section 5 provides the simulation results to illustrate the performance of the proposed method. In Section 6, we present a real data example. We conclude this paper with some remarks and suggestions for future work in Section 7. To maintain the flow of the presentation, all mathematical developments are deferred to Section 8.

2. Model and Estimator

2.1. Quasi-Associated Dependence

Stochastic models often involve families of random variables characterized by positive or negative dependence properties, or their variations. Foundational studies on this topic include [16,41,42,43]. To illustrate the concept, consider two finite samples separated by a gap r, with indices in the past denoted by { Pa } , and in the future by { Fu } . For certain classes of measurable functions, the independence of { Pa } and { Fu } can be expressed as the covariance cov ( f ( Fu ) , g ( Pa ) ) = 0 . A weaker form of this condition involves controlling these covariances as r increases, with the rate of decay defining the degree of weak dependence. For practical applications, this condition is often restricted to Lipschitz functions, making it easier to verify across a wide range of models. To define the quasi-association dependence in a Hilbert space H we define, in a first, and according to [15], this dependence in the real-valued random fields is as follows. Define a function F : R n R where R n is equipped with its 1 -norm (i.e., ( x 1 , , x n ) 1 = | x 1 | + + | x n | ) and define the Lipschitz modulus of F ( · ) as
Lip ( F ) = sup x y | F ( x ) F ( y ) | x y 1 .
A sequence ( Z n ) n N of random variables is said to be quasi-associated if for any disjoint subsets I and J of N , and all bounded Lipschitz functions F 1 : R | I | R and F 2 : R | J | R satisfying, this covariance exists:
Cov ( F 1 ( Z i , i I ) , F 2 ( Z j , j J ) ) Lip F 1 Lip F 2 i I j J Cov ( Z i , Z j ) ,
where | I | denotes the cardinality of a finite set I, and Cov ( Z i , Z j ) the covariance of the pair ( Z i , Z j ) , respectively. Equation (1) is satisfied for positively associated or negatively associated random variables (see [44]). For example, ref. [45] demonstrated that any Gaussian random sequence with a covariance function taking both positive and negative values is quasi-associated.
Covariance inequalities are pivotal in establishing moment inequalities and limit theorems for dependent random variables [15,46] (and references therein). Independent real-valued random variables are inherently associated [47] and negatively associated [43]. A Gaussian random field is positively associated if its covariance function is non-negative [48], and negatively associated if Cov ( X s , X t ) 0 for all s t [43]. The computational simplicity of covariance functions, as highlighted by [49] and others, often makes them preferable to mixing coefficients, which can be challenging to estimate.
The interplay between association and mixing is non-trivial, with overlapping yet distinct characteristics [17]. For instance, in linear time series, sequences defined by X k = j = 0 a j v k j , where ( v i ) i Z are i.i.d. centered random variables with variance σ 2 , are associated if a j 0 . However, under suitable conditions on a j , ref. [50] established α -mixing for such sequences. Notably, ref. [51] showed that when ( v i ) are i.i.d. Bernoulli variables with a j = ς and 0 < ς < 1 / 2 , the sequence is associated but not α -mixing. ref.  [34] extended the definition to functional random variables. Let us denote by the symbols · , · , and · the inner product and the associated norm on a Hilbert space H , respectively, and we associate with this space a complete orthonormal system ( e p ) p 1 of H . A sequence ( Z i ) i N of random variables taking values in a Hilbert space H is called quasi-associated relative to an orthonormal basis { e p : p 1 } of H , if for any p 1 , Z i , e 1 , , Z i , e p i N is a sequence of quasi-associated random vectors.
Remark 1. 
Association and mixing are two distinct yet interrelated categories of stochastic processes, as elaborated on in [17]. The exploration of their respective dependence structures has drawn considerable academic attention. A fundamental difference lies in their interpretation of relationships among random variables: within the framework of association, independence can be inferred from the absence of correlation. This property, however, holds exclusively in the Gaussian setting, as noted in [47]. Thus, in Gaussian contexts, the covariance structure provides a complete description of dependence. This makes association an attractive property to study, since covariance calculations are typically more straightforward than those involving mixing coefficients.
In contrast, a major drawback of mixing processes is the limited number of models for which mixing coefficients can be explicitly determined. For example, consider the results in [52]: Let ε = { ε i , i Z } be a stationary sequence of independent random variables with E ε 0 = 0 and σ 0 2 = E ε 1 2 < . Let a = { a n } n Z be a sequence of real numbers such that
r Z a r 2 < .
Define the linear process { X k , k Z } as
X k : = j = a j ε k j .
This process is well-defined, strictly stationary, and satisfies E X 1 2 < . This is denoted by L ˜ 2 ( a , ε ) , the class of such linear sequences, and by L 2 ( a , ε ) , the subset of L ˜ 2 ( a , ε ) , where a i = 0 for i < 0 . If  ε 0 has an absolutely continuous density in L 1 , then any linear process { X k , k Z } within L 2 ( a , ε ) is absolutely regular (β-mixing) with coefficients satisfying the following:
β n K j n k j | a k | 2 / 3 , for some constant K ,
provided that
j 0 k j | a k | 2 / 3 < ,
as established in [50].
1. 
Associated but not mixing sequences: Consider the sequence ε = { ε i , i Z } , where ε i are i.i.d. random variables with P ( ε i = 1 ) = P ( ε i = 0 ) = 1 / 2 . Let a = { 2 k , k 1 } . The linear process { X k } L 2 ( a , ε ) is associated (see ( P 2 ) and ( P 4 ) in [16]) but not mixing, as  β n 1 / 4 (refer to [53]). Furthermore, X 1 follows a uniform distribution on [ 0 , 1 ] , and its covariance function satisfies
r ( n ) : = Cov ( X 1 , X n ) = c 2 σ 0 2 2 n ,
exhibiting exponential decay. The associated empirical process { G n ( x ) : x R } converges, as demonstrated in [54].
2. 
Associated and mixing sequences: When the conditions of Theorem 2.1 in [50] hold, and  a k 0 for all k N , the linear process { X k } L 2 ( a , ε ) exhibits both association and β-mixing properties.
3. 
Mixing but not associated sequences: Consider a sequence a = { a k , k N } satisfying a 0 a 1 < 0 and a i = 0 for i 2 . Then, X n = a 0 ε n + a 1 ε n 1 is not associated, as 
Cov ( X n , X n 1 ) = a 0 a 1 σ 0 2 < 0 ,
violating the association condition. However, the sequence { X n } is mixing due to its m-dependence.

2.2. The Single Functional Index Model

Let { ( X i , C i , Y i 0 ) , 1 i n } be n copies i.i.d. of a triplet of random variables ( X , C , Y 0 ) , where X takes values in a separable real Hilbert space H endowed with the inner product · , · and its norm · = · , · 1 / 2 , and C is a censoring variable with real values and Y 0 is the variable of interest, typically called a lifetime variable. We consider the semi-metric d θ , associated with the single-index θ , which is defined by d θ ( χ 1 , χ 2 ) : = | χ 1 χ 2 , θ | . In the given topological structure, and for any θ , our focus is to estimate nonparametrically the conditional mean (regression) function of Y 0 given X , θ , denoted by
r θ ( χ ) : = I E ( Y 0 | θ , X = θ , χ ) .
Since the single functional index θ plays a crucial role in such a model, the identifiability of this model will be important. Clearly, the relation in (2) can be reformulated as Y 0 = r ( X , θ ) + ϵ with I E ( ϵ | X ) = 0 . To ensure the identifiability of the model θ (i.e., θ 1 = θ 2 ), we require that r j ( χ , θ j ) is differentiable with respect to (wrt) the variable χ , θ j , j = 1 , 2 , respectively, and that θ , e 1 = 1 , where e 1 is the first element of an orthonormal basis of the space H . Then, for all χ H , r 1 ( χ , θ 1 ) = r 2 ( χ , θ 2 ) implies r 1 = r 2 and θ 1 = θ 2 , respectively. For more details on the proof of the identifiability of the functional single model, see [3].
It is worth highlighting that when censored data are considered, the smoothing nonparametric regression estimator r ^ ( θ , χ ) of r ( θ , χ ) can be viewed as a local least squares estimator that is
r ^ ( θ , χ ) = arg min t I R i = 1 n W i , n ( θ , χ ) ( Y i t ) 2 .
where W i , n ( θ , χ ) are the Nadaraya–Watson weights [55,56], given as
W i , n ( θ , χ ) = K ( h n 1 | χ X i , θ | ) i = 1 n K ( h n 1 | χ X i , θ | ) .
A natural estimator of r ( θ , χ ) is defined as
r ^ ( θ , χ ) = i = 1 n Y i 0 W i , n ( θ , χ ) = i = 1 n Y i 0 K ( h n 1 | χ X i , θ | ) i = 1 n K ( h n 1 | χ X i , θ | ) .
where K ( · ) is a kernel function, and  h n : = h n , K is a sequence of bandwidths (positive real numbers) decreasing to zero as n goes to infinity.
As is known, when dealing with censoring data, the first problem arising is understanding and managing the censorship itself, due to incomplete or partial collected observations. One of the suggested approaches to tackling censored data is to apply the method of synthetic data, which allows us to take into account the censoring effect on the lifetime distribution, for instance, [57,58,59]. Theoretically, synthetic data transformation provides equal expected values for both variables. In this case, the synthetic response variable (transformed variable) is given by δ i Y i G ¯ ( Y i ) , 1 i n , where G ¯ = 1 G , and if we assume ( Y i 0 , X i ) and ( C i ) are independent (that is usually supposed for censored data), we obtain
I E δ i Y i G ¯ ( Y i ) | X i = I E I E δ i Y i G ¯ ( Y i ) | Y i 0 , X i | X i = I E I E 1 { Y i 0 C i } Y i G ¯ ( Y i ) | Y i 0 , X i | X i = I E Y i 0 G ¯ ( Y i 0 ) I E 1 { Y i 0 C i } | Y i 0 | X i = I E Y i 0 | X i .
By the same manner, we get by conditioning wrt θ , X i
I E δ i Y i G ¯ ( Y i ) | θ , X i = I E Y i 0 | θ , X i .
Thus, for the infinite dimension case, one can define the pseudo-estimator of r θ ( χ ) as
r ˜ ( θ , χ ) = i = 1 n δ i Y i G ¯ ( Y i ) K ( h n 1 | χ X i , θ | ) i = 1 n K ( h n 1 | χ X i , θ | ) = r ˜ N ( θ , χ ) r ^ D ( θ , χ ) .
where
r ˜ N ( θ , χ ) = 1 n I E [ K 1 ( θ , χ ) ] i = 1 n δ i Y i G ¯ ( Y i ) K i ( θ , χ ) r ^ D ( θ , χ ) = 1 n I E [ K 1 ( θ , χ ) ] i = 1 n K i ( θ , χ ) K i ( θ , χ ) = K ( h n 1 | χ X i , θ | ) = K ( h n 1 d θ ( χ , X i ) ) .
In practice, since the values Y 0 are censored observations, the survival function G ¯ ( · ) = 1 G ( · ) is unknown (because G ( · ) is too), so it is impossible to use the estimator (6). To complete building a good estimator of r ( θ , χ ) , we may replace the function G ¯ ( · ) by its corresponding [60] estimator G ¯ n ( · ) : = 1 G n ( · ) , defined by
G ¯ n ( t ) = i = 1 n 1 1 δ ( i ) n i + 1 1 { Y ( i ) t } , if t < Y ( n ) , 0 , if t Y ( n ) .
where Y ( 1 ) Y ( 2 ) Y ( n ) are the order statistics of Y i and δ ( i ) is the indicator of no censoring associated with Y ( i ) . We will set log 2 u = log + log + u , log + u = log ( u e ) . ref. [61] obtained the law of iterated logarithms for the censored case. More precisely, for a continuous G, we have as n ;
sup 0 t Y ( n ) 1 G n ( t ) 1 G ( t ) = O log 2 n n 1 2 ,
almost surely. Such a result will be useful to establish the main results of this paper. Now, plugging the estimator (7) in (6), one can define an estimator of r ( θ , χ ) as
r ^ n ( θ , χ ) = 1 n I E [ K 1 ( θ , χ ) ] i = 1 n δ i Y i G ¯ n ( Y i ) K ( h n 1 | χ X i , θ | ) 1 n I E [ K 1 ( θ , χ ) ] i = 1 n K ( h n 1 | χ X i , θ | ) = : r ^ N ( θ , χ ) r ^ D ( θ , χ ) .
The covariance coefficient λ r defined as
λ r : = sup s r | i j | s λ i , j ,
where
λ i , j = k 1 l 1 C o v ( X i k , X j l ) + k 1 C o v ( X i k , Y j ) + l 1 C o v ( Y i , X j l ) + C o v ( Y i , Y j ) ,
with X i p : = X i , e p .

3. Strong Uniform Consistency

3.1. Results

Throughout this manuscript, provided that no ambiguity arises, we will denote by C, C , and/or C θ , x generic positive constants, which may vary with each occurrence. For any sequences ( u n ) and ( v n ) , we write u n = O ( v n ) , provided that | u n | C | v n | for all n. If this condition holds almost completely, we use the notation O a . c o . . Let N χ represent a fixed neighborhood of a point χ in the space H , and define the ball centred at χ with radius r by
B θ ( χ , r ) : = { χ 1 H | χ χ 1 , θ | < r } .
Define the upper endpoints of the distributions F and G by
τ F = sup { y F ¯ ( y ) > 0 } and τ G = sup { y G ¯ ( y ) > 0 } .
We assume that τ F is finite and that G ¯ ( τ F ) > 0 , which implies that τ F < τ G .
Consider the following coverings of the compact sets S H and Θ H :
S H k = 1 d n S H B ( χ k , l n ) and Θ H j = 1 d n Θ H B ( t j , l n ) ,
where χ k , t j H . For each χ H and θ Θ H , define
k ( χ ) = arg min k { 1 , , d n S H } χ χ k and j ( θ ) = arg min j { 1 , , d n Θ H } θ t j .
It is important to note that covering the compact subsets in this manner is essential for establishing our uniform consistency. Additionally, this approach is crucial to maintaining the geometric relationship between the numbers d n S H , d n Θ H of balls and the sequence of the radius l n . In the context of abstract semi-metric spaces, it is generally assumed that d n S H l n and d n Θ H l n remain bounded (see [26], for further discussion).
We now present the assumptions necessary to derive our initial result:
(A0)
The random variables ( X , Y 0 ) and C are independent.
(A1)
(i)
The probability I P ( X B θ ( χ , ϵ ) ) , denoted by ϕ θ , χ ( ϵ ) , is positive for all ϵ > 0 , and satisfies
lim ϵ 0 ϕ θ , χ ( ϵ ) = 0 .
(ii)
There exists a differentiable function ϕ ( · ) such that for all χ S H and θ Θ H ,
0 < C ϕ ( h n ) ϕ θ , χ ( h n ) C ϕ ( h n ) < ,
and there exists a constant ξ 0 > 0 for which
ϕ ( ξ ) < C for all ξ < ξ 0 .
(A2)
The function r ( · , · ) satisfies a Hölder condition. Specifically, for all θ H and for all ( χ 1 , χ 2 ) N χ × N χ ,
| r ( θ , χ 1 ) r ( θ , χ 2 ) | C θ , χ d θ ( χ 1 , χ 2 ) γ ,
where γ > 0 .
(A3)
The function K ( · ) is bounded and continuous, and satisfies the following conditions:
(i)
There exist constants C , C > 0 such that
0 < C 1 [ 0 , 1 ] ( t ) < K ( t ) < C 1 [ 0 , 1 ] ( t ) < for all t .
(ii)
For all t 1 , t 2 R ,
| K ( t 1 ) K ( t 2 ) | C | t 1 t 2 | .
(A4)
Let l n = O log n n . The sequences d n S H and d n Θ H satisfy
(i)
( log n ) 2 n ϕ ( h n ) < log d n S H + log d n Θ H < n ϕ ( h n ) log n .
(ii)
n = 1 d n S H d n Θ H 1 β < for some β > 2 .
(iii)
n ϕ ( h n ) = O ( ( log n ) 2 ) and lim n n ϕ ( h n ) = .
(A5)
The process ( X i , Y i ) i N is quasi-associated with a covariance coefficient λ k that satisfies
a > 0 , C > 0 such that k N , λ k C e a k .
(A6)
0 < sup i j I P ( X i , X j ) B θ ( χ , h n ) × B θ ( χ , h n ) = O ϕ θ , χ ( h n ) 2 .
(A7)
(i)
There exists a constant 0 < ρ < 1 2 such that
ϕ θ , χ ( h n ) h n 2 ρ 1 2 ρ .
(ii)
lim n log 5 ( n ) n ϕ θ , χ ( h n ) = 0 .
Remark 2. 
Assumption (A0) typically serves as the independence condition necessary to ensure identifiability within the random censoring framework. In the absence of (A0), additional information regarding the structure of censoring must be provided. Furthermore, the condition τ F < τ G (which implies G ¯ ( Y 0 ) > G ¯ ( τ F ) ) is a standard assumption for establishing asymptotic results in censoring contexts. Assumption (A1) characterizes the concentration of the explanatory variable within small neighborhoods, and it is fundamentally important for deriving various asymptotic properties, especially when dealing with functional data. Assumption (A2) is necessary to control the bias term in convergence rates and to ensure the regularity of the model. Assumptions (A3) and (A4) are technical in nature and facilitate the evaluation of the asymptotic variance component. Assumption (A5) provides an upper bound for the polynomial decay of the association coefficient, while Assumption (A6) describes the asymptotic behavior of the joint distribution of the pair ( X i , X j ) . Finally, Assumption (A7) is included for the sake of brevity in the proofs, as it satisfies technical requirements, particularly when applying the exponential inequality introduced by [39].
The proof of Theorem 1 relies on the subsequent lemmas, whose proofs are provided in Section 8.
Lemma 1. 
Under assumptions (A0), (A1), (A3), and (A4), we have
sup χ S H sup θ Θ H r ˜ 1 ( θ , χ ) E r ˜ 1 ( θ , χ ) = O a . c o . log d n S H + log d n Θ H n ϕ ( h n ) .
Lemma 2. 
Under assumptions (A1) and (A3) through (A7), the following holds:
sup χ S H sup θ Θ H r ^ 2 ( θ , χ ) 1 = O a . c o . log d n S H + log d n Θ H n ϕ ( h n ) .
Corollary 1. 
From Lemma 2, it follows that
n = 1 P inf θ Θ H inf χ H r ^ 2 ( θ , χ ) < 1 2 < .
Lemma 3. 
Assuming (A0), (A1), (A3), and (A4), we have
sup χ S H sup θ Θ H r ^ 1 ( θ , χ ) r ˜ 1 ( θ , χ ) = O a . c o . log d n S H + log d n Θ H n ϕ ( h n ) .
Lemma 4. 
Under assumptions (A0), (A1)-(i), and (A2), it holds that
sup χ S H sup θ Θ H E r ˜ 1 ( θ , χ ) r ( θ , χ ) = O ( h n γ ) .
We present the following main theorem on this section.
Theorem 1. 
Assume that conditions (A0) through (A7) are satisfied. Then, the following holds:
sup χ S H sup θ Θ H r ^ n ( θ , χ ) r ( θ , χ ) = O ( h n γ ) + O a . c o . log d n S H + log d n Θ H n ϕ ( h n ) .
Proof of Theorem 1. 
We begin by considering the following decomposition:
r ^ n ( θ , χ ) r ( θ , χ ) = r ^ 1 ( θ , χ ) r ( θ , χ ) r ^ 2 ( θ , χ ) r ^ 2 ( θ , χ ) = r ^ 1 ( θ , χ ) r ˜ 1 ( θ , χ ) r ^ 2 ( θ , χ ) + r ˜ 1 ( θ , χ ) E r ˜ 1 ( θ , χ ) r ^ 2 ( θ , χ ) + E r ˜ 1 ( θ , χ ) r ( θ , χ ) r ^ 2 ( θ , χ ) + r ( θ , χ ) 1 r ^ 2 ( θ , χ ) r ^ 2 ( θ , χ ) .
The result of the theorem is obtained by synthesizing the findings of the preceding lemmas.    □
In the specific case where the functional single index is held fixed, the following corollary applies.
Corollary 2. 
Under assumptions (A0) through (A7), it holds that
sup χ S H r ^ n ( θ , χ ) r ( θ , χ ) = O ( h n γ ) + O a . c o . log d n S H n ϕ ( h n ) .

3.2. Methodology for Estimating a Single Functional Index

In practice, the functional index is unknown and must be estimated. Two main approaches are available for this purpose. The first is the Least Squares Cross-Validation (LSCV) method, defined by
θ ^ = arg min θ Θ 1 n i = 1 n Y i r ^ n i ( θ , X i ) 2 ,
where r ^ n i ( θ , χ ) denotes the leave-one-out estimator of r ^ n ( θ , χ ) . The second is the Maximum Likelihood Cross-Validation (MLCV) method, given by
θ ^ = arg min θ Θ 1 n i = 1 n log f ^ Y i r ^ n i ( θ , X i ) ,
where f ^ ( · · ) is the estimator of the conditional density of Y given θ , X . In practice, both of these methods require optimizing over a finite subset Θ .
Following [26], we propose selecting the optimal index from the set
Θ = Θ n = θ H : θ = i = 1 k c i e i , θ = 1 , j { 1 , , k } such that θ , e j > 0 ,
where { e i } i = 1 , , k is a finite basis of the Hilbert subspace spanned by the covariates { X i } i , and  { c i } i are real-valued constants calibrated to ensure the identifiability of the model (commonly, c i { 1 , 0 , 1 } ). It is well known that cross-validation methods can be used to estimate any parameter within this framework.
An alternative approach—which will be used in this section—is to choose θ ( t ) from the eigenfunctions of the covariance operator
E X E [ X ] X , · H ,
where X ( t ) is, for instance, a diffusion-type process defined by [ a , b ] , and  X ( t ) denotes its first derivative (see [30]). Given a training sample L , the covariance operator can be estimated by its empirical version
1 | L | i L X i E [ X ] t X i E [ X ] .
A discretized version of the eigenfunctions { θ i ( t ) } can then be obtained using principal component analysis, as described in [62]. In the simulation steps for computing the regression estimator, we let θ * be the first eigenfunction corresponding to the largest eigenvalue of the empirical covariance operator, thus taking the place of θ .

4. Asymptotic Distribution

To establish our second main result, we introduce a revised set of assumptions, replacing (A1), (A3), (A4), and (A7) with the following:
(A1’)
The probability measure satisfies
P ( X B θ ( χ , h n ) ) = : ϕ θ , χ ( h n ) > 0 ,
and there exists a function β θ , χ ( · ) such that
s [ 0 , 1 ] , lim n ϕ θ , χ ( s h n ) ϕ θ , χ ( h n ) = β θ , χ ( s ) .
(A3’)
The kernel function K ( · ) is a continuous and bounded function supported on [ 0 , 1 ] . Moreover, K ( · ) is differentiable, and its derivative K ( · ) exists, satisfying the condition that there exist constants C and C such that
< C < K ( t ) < C < 0 , for all t [ 0 , 1 ] .
(A4’)
The bandwidth sequence h n and the function ϕ θ , χ ( · ) satisfy the following conditions:
(i)
lim n h n = 0 and lim n log n n ϕ θ , χ ( h n ) = 0 .
(ii)
lim n n h n 2 γ ϕ θ , χ ( h n ) = 0 .
(A7’)
There exists a sequence p n that diverges to infinity while satisfying p n = o ( n ) , such that
lim n p n n ϕ θ , χ ( h n ) = 0 .
Remark 3. 
Assumption (A1’) characterizes the concentration behaviour of the explanatory variable within small neighborhoods. This is a crucial condition in deriving asymptotic properties, particularly in functional data analysis. The technical conditions in (A3’) and (A4’) facilitate explicit computation of the asymptotic variance. The function β θ , χ ( · ) plays a central role in the results, as it determines the precise constant terms appearing in asymptotic expansions (for additional choices of this function, see [63]). Additionally, the condition (A4’- i i ) ensures the elimination of the bias term in the derivation of asymptotic normality. The sequence p n introduced in (A7’) is instrumental for applying Doob’s decomposition theorem, which is key in establishing the limiting distribution.
To establish Theorem 2, we rely on the following auxiliary results, proved in Section 8.
Lemma 5. 
Under assumptions (A0), (A1’), (A3’)–(A4’), (A5)–(A6), and (A7’), we have
n ϕ θ , χ ( h n ) r ^ 1 ( θ , χ ) E [ r ^ 1 ( θ , χ ) ] D N ( 0 , σ r 2 ( θ , χ ) ) .
Lemma 6. 
Under assumptions (A1’), (A3’), (A4’-i), (A5)–(A6), and (A7’), we have
r ^ 2 ( θ , χ ) 1 P 0 .
Lemma 7. 
Under assumptions (A0), (A1’), and (A2), and if (A4’- i i ) holds, then we obtain
n ϕ θ , χ ( h n ) E [ r ^ 1 ( θ , χ ) ] r ( θ , χ ) r ^ 2 ( θ , χ ) 0 .
Our main result in this section is summarized in the following theorem.
Theorem 2. 
Suppose that the assumptions (A0), (A1’), (A2), (A3’)–(A4’), (A5)–(A6), and (A7’) hold. Then, we have the following asymptotic normality result:
n ϕ θ , χ ( h n ) r ^ n ( θ , χ ) r ( θ , χ ) D N ( 0 , σ r 2 ( θ , χ ) ) ,
where the asymptotic variance is given by
σ r 2 ( θ , χ ) = c 2 r ( θ , χ ) 2 c 1 2 ,
with
c l = K l ( 1 ) 0 1 ( K l ) ( s ) β θ , χ ( s ) d s , for l = 1 , 2 .
Proof of Theorem 2. 
We begin by considering the decomposition:
r ^ n ( θ , χ ) r ( θ , χ ) = r ^ 1 ( θ , χ ) r ( θ , χ ) r ^ 2 ( θ , χ ) r ^ 2 ( θ , χ ) .
Rewriting the numerator,
r ^ 1 ( θ , χ ) r ( θ , χ ) r ^ 2 ( θ , χ ) = r ^ 1 ( θ , χ ) E [ r ^ 1 ( θ , χ ) ] + E [ r ^ 1 ( θ , χ ) ] r ( θ , χ ) r ^ 2 ( θ , χ ) .
By applying Slutsky’s theorem, one can establish the asymptotic normality of r ( θ , χ ) by showing that its numerator converges in distribution to N ( 0 , σ r 2 ( θ , χ ) ) while its denominator converges in probability to 1.    □

Confidence Interval

Observe in Theorem 2 that the asymptotic variance depends on the unknown function r ( · ) , as well as on the normalization involving ϕ θ , χ ( h n ) , which cannot be specified directly. Hence, Corollary 3, which provides a minor modification of Theorem 2, yields a formulation that is more suitable for practical use.
As is customary, the conditional variance σ r 2 ( θ , χ ) is estimated by
σ r , n 2 ( θ , χ ) = c 2 , n ( r ^ n ( θ , χ ) ) 2 c 1 , n 2 ,
where
c l , n = K l ( 1 ) 0 1 ( K l ) ( s ) β θ , χ , n ( s ) d s , for l = 1 , 2 ,
and
β θ , χ , n ( s ) = ϕ θ , χ , n ( s h n ) ϕ θ , χ , n ( h n ) ,
with
ϕ θ , χ , n ( u ) = 1 n i = 1 n 1 { d θ ( χ , X i ) u } .
The following corollary is more or less a straightforward consequence of Theorem 2 and Slutsky’s theorem.
Corollary 3. 
Suppose that the assumptions (A0), (A1’), (A2), (A3’)–(A4’), (A5)–(A6), and (A7’) hold. Then, we have the following asymptotic normality result:
n ϕ θ , χ , n ( h n ) σ r , n 1 ( θ , χ ) r ^ n ( θ , χ ) r ( θ , χ ) D N ( 0 , 1 ) .
Remark 4. 
Using Corollary 3, the asymptotic 100 ( 1 α ) % confidence band for the regression function r is given by
r ^ n ( θ , χ ) ± c α σ r , n ( θ , χ ) n ϕ θ , χ , n ( h n ) ,
where c α is the upper α / 2 quantile of the distribution of N ( 0 , 1 ) .

5. Numerical Results

This study examines the performance of an estimator under a variety of theoretical configurations. Four key frameworks form the basis of the analysis: single-index modeling, dependence structures, and the censored functional setting. To conduct this analysis, we construct a functional random variable by generating diverse functional time series data.
The first dataset is generated from a functional autoregressive process of order 2 (FAR(2)), specified as
X i = j = 1 2 0 1 ψ ( t , s ) X i j ( s ) d s + ε i ,
where
ψ ( t , s ) = exp t 2 + s 2 2 .
The second dataset utilizes a functional autoregressive conditional heteroscedastic (FARCH(1,1)) model, defined as
X i ( t ) = σ i ( t ) ϵ i ( t ) ,
where    
σ i 2 ( t ) = w ( t ) + 0 1 α ( t , s ) X i 1 2 ( s ) d s ,
with w ( t ) = 0.1 t ( 1 t ) and α ( t , s ) = t ( 1 t ) s ( 1 s ) .
The third dataset is derived from a functional general autoregressive conditional heteroscedastic (FGARCH(1,1)) model:
X i ( t ) = σ i ( t ) ϵ i ( t ) ,
where    
σ i 2 ( t ) = w ( t ) + 0 1 α ( t , s ) X i 1 2 ( s ) d s + 0 1 α ( t , s ) σ i 1 2 ( s ) d s .
To evaluate the impact of correlation structures on estimation quality, the outcomes from these three types of functional time series are compared with those from an independent (i.i.d.) setting. The different functional covariates are generated using the R package FTSgof; for instance, dgp.far is employed to produce FAR(1)-based regressors, dgp.fgarch creates FGARCH(1,1) processes, and dgp.ou provides i.i.d. data. A visual summary of these generated functional time series appears in Figure 1.
In the theoretical segment of this work, we generate the response variable { Y i } via a functional single-index model of the form
Y i = r θ , X i + ϵ i , where r ( z ) = 2 2 + z 2 .
Here, ϵ i is an independent white noise term (assumed independent of X i ) drawn from a normal distribution N ( 0 , 1 ) . The true functional index θ is selected as the first Karhunen–Loève (KL) [64,65,66] basis function of { X i } . Under this construction, the conditional distribution of Y i given X i follows N r θ , X i , 1 , ensuring that r θ , X i is the true conditional mean.
To introduce censoring, a random variable C is drawn from an exponential distribution Exp ( 1 / λ ) , where the parameter λ controls the censoring rate. High ( λ = 2 ), moderate ( λ = 0.5 ), and low ( λ = 0.05 ) rates correspond to approximate censoring rates of 70–80%, 50–56%, and 4–8%, respectively. The primary objective of this numerical study is to assess the accuracy of the asymptotic normality of the statistic.
N T ( χ 0 , θ ) = n ϕ θ ^ ( χ 0 , h n ) σ ^ r 2 θ ^ , χ 0 r ^ n θ ^ , χ 0 r θ , χ 0 ,
where χ 0 is a fixed conditioning curve (e.g., X i 0 ), and σ ^ r 2 θ ^ , χ 0 is a plug-in estimator of the asymptotic variance,
σ ^ r 2 ( θ , χ ) = C 2 ^ ( θ , χ ) r ^ ( θ , χ ) 2 C 1 ^ ( θ , χ ) 2 ,
with
C ^ ( θ , χ ) = 1 n ϕ θ ^ ( χ , h n ) i = 1 n K i ( θ , χ ) , = 1 , 2 ,
and
ϕ θ ^ ( χ , h n ) = 1 n i = 1 n 1 χ X i , θ < h n .
Empirically, these estimators are calculated by first selecting both the optimal functional index θ and bandwidth h via a one-leave-out cross-validation approach:
( θ C V , h C V ) = arg min θ Θ n , h n H n 1 n i = 1 n Y i r ^ n i θ , X i 2 .
The set Θ n consists of vectors θ expressible as finite combinations of basis functions { e i } that span the Hilbert subspace of { X i } , subject to θ = 1 and certain nontriviality constraints on the coefficients. Meanwhile, H n is chosen so that the ball of radius h n around each point χ contains a fixed number k { 5 , 15 , 25 , , 55 } of neighboring observations. Finally, we generate 100 independent replications of the statistic N T X i 0 , θ C V and quantify its departure from normality through its empirical bias and variance:
Bias = 1 100 b = 1 100 N T b X i 0 , θ C V , Var = 1 100 b = 1 100 N T b X i 0 , θ C V Bias 2 .
These measures provide a comprehensive evaluation of how closely N T ( χ 0 , θ ) adheres to its theoretical asymptotic distribution under different levels of censoring and model configurations. The results of bias and the variance are reported in the following Table 1.
This study demonstrates that the asymptotic normality of the estimator is shaped by several key elements, including the structure of the functional time series data, the choice of smoothing parameter, censoring, and the functional index. Notably, high levels of dependence and substantial censoring rates can adversely affect the estimator’s performance in finite samples. Evidence supporting this conclusion emerges from a comparison between the results obtained in an i.i.d. setting (with λ = 0.05 ) and those from an FGARCH(1,1) process (also with λ = 0.05 ). In the former case, the bias is nearly zero and the variance is close to one, whereas under the FGARCH(1,1) specification, both bias and variance increase substantially. In conclusion, the estimator’s finite-sample performance is notably influenced by the sample size, n. Our simulation studies considered several values of n (namely, 25, 50, 100, 150, 350, and 500). For the sake of brevity, we present results only for n = 50 , 100, and 350, since the other cases yielded similar outcomes. Overall, while the estimator performs adequately even with smaller samples, its effectiveness improves markedly as n increases.

6. Real Data Example

In the second phase of our empirical investigation, we demonstrate how the proposed methodology can be employed in real-world contexts. The study of incomplete functional time series data has garnered significant attention across various applied disciplines, particularly finance, where high-frequency market data frequently exhibit interruptions or extraordinary events. Such interruptions often yield censored observations. To illustrate, we consider the intraday returns of the Nikkei stock index from 1 January 1984, to 31 December 2021. Single-index modeling is extensively used in econometrics because it offers a middle ground between overly rigid parametric models and highly flexible nonparametric alternatives that are challenging to estimate. This approach has been popularized by early studies, such as those in [2,67]. Moreover, financial time series naturally embody the characteristics of quasi-associated functional data. Financial datasets typically display features like high variability, non-normal distributions, excess kurtosis, and irregular volatility. As a result, these series are often modeled using generalized autoregressive conditional heteroskedasticity (GARCH) processes, which provide a clear example of the mathematical formulation of quasi-associated processes (see [17]). Therefore, applying the present model to such complex functional structures enhances its practical applicability. In addition, since trading on this index is suspended whenever prices fall below a specified threshold (an occurrence observed during crisis periods such as the COVID-19 pandemic), the resulting dataset is censored. Throughout the observation period, numerous trading halts led to a substantial censoring rate of approximately 20%. The data, publicly available at https://fred.stlouisfed.org/series/NIKKEI225 (accessed on 5 January 2025), are treated as functional observations by dividing the entire time series into 30-day segments, with each segment reflecting a functional curve for one month. This procedure yields 444 functional observations.
Before conducting our main analysis, we address heteroscedasticity by measuring logarithmic differences. For each month i and day d, we define
X i ( d ) = 100 log s ( d ) s ( d 1 ) ,
where s ( t ) denotes the daily return. We then set Y i = max d X i + 1 ( d ) , reflecting the maximum price observed in the subsequent month. The forecasting objective is thus to predict this maximum one month ahead using the functional trajectory of the previous month. We compare two estimation strategies: a single-index estimator r ^ and a nonparametric estimator r ˜ . The latter employs a kernel-smoothing approach
r ˜ ( x ) = i = 1 n K h 1 d ( x , X i ) Y i j = 1 n K h 1 d ( x , X j ) ,
where the bandwidth h and single-index parameter θ are chosen via cross-validation. The distance d ( · , · ) is defined through a PCA-based metric, which is suitable for handling discontinuities in the curves.
To evaluate both methods, we estimate confidence intervals by exploiting the asymptotic normality derived in this work. Specifically, we construct a ( 1 ζ ) confidence interval for r ( θ , χ ) as
r ^ n ( θ , χ ) ± t ζ / 2 σ ^ ( θ , χ ) n ϕ θ * ^ ( χ , h n ) ,
where t ζ / 2 is the ζ / 2 upper quantile of the standard normal distribution and σ ^ ( θ , χ ) is an estimate of the asymptotic variance. The performance of these intervals is assessed by their average length and empirical coverage probability. To facilitate comparisons between the two models, we partition the dataset into a training set of 300 observations and a test set of 140 observations, then repeat this split 100 times with different random partitions. The numerical results, including coverage rates and interval lengths, are presented in the subsequent figures.
The results in Figure 2 and Figure 3 indicate that the single-index model substantially outperforms its kernel-based counterpart. In particular, the semi-parametric approach produces narrower confidence intervals while maintaining accurate coverage probabilities. Notably, the average confidence interval length for the estimator r ^ n is 3.56, compared with 5.09 for the kernel estimator r ˜ ( x ) . Moreover, the coverage probabilities associated with r ^ n align more closely with the nominal level than those derived from r ˜ ( x ) .

7. Concluding Remarks

This study proposes kernel estimators for regression functions in settings where the data comprise censored time series governed by a functional single-index model. By leveraging the single-index structure, the relationship between the response and covariates is dimensionally reduced, thereby simplifying the model while retaining its ability to capture complex relationships. The proposed methodology employs nonparametric kernel estimation within a quasi-association framework to effectively characterize these dependencies.
We establish that, under mild regularity conditions, the estimators converge uniformly almost surely, and we further demonstrate their asymptotic normality. In addition, we provide practical insights into bandwidth selection and discuss the implications of our theoretical findings for real-world applications. Finally, the proposed methods are evaluated in a comprehensive simulation study, which highlights their robustness and favorable finite-sample performance across various scenarios.
A key issue that has not yet been addressed in this study is how to optimally determine the smoothing parameters so as to minimize the mean square errors (MSEs) of the proposed kernel estimators. This problem is highly relevant within the broader sphere of kernel-based nonparametric estimation, and it merits dedicated investigation. Although this topic is crucial for further enhancing the proposed methodology, we have decided to postpone a comprehensive examination to future work. In that subsequent research, we plan to systematically analyze the selection of smoothing parameters, potentially employing cross-validation or other established methods for optimal parameter choice.
Several other directions also warrant further exploration. The principal drawback of conventional kernel estimators lies in their tendency to incur bias, a challenge that remains central to kernel smoothing. While local linear fitting is theoretically capable of mitigating bias, this issue still requires substantial attention in nonparametric functional data analysis (NFDA). This need becomes especially salient when dealing with complex functional data, where simpler smoothing techniques may produce suboptimal results. Our computational findings illustrate that local linear bias correction significantly enhances estimation performance by reducing overall error. However, incorporating any form of bias correction (CB) into a nonparametric setting demands strong mathematical underpinnings to guarantee control over its asymptotic behavior, a key issue that remains insufficiently studied in the current literature. To our knowledge, no comprehensive investigation of the asymptotic nature of bias correction methods has been conducted in NFDA. We therefore believe this represents an exciting and largely unexplored area for future work. In particular, it would be highly beneficial to assess how the proposed kernel-based approach might be paired with established CB algorithms, such as those introduced by [68,69], which have been shown to improve estimation accuracy in related contexts. Integrating our local linear smoothing framework with these CB algorithms could substantially enhance estimation quality, especially for more complex functional data, by balancing bias reduction against variance control—an essential consideration in nonparametric functional data analysis.
Another natural question is whether our results might be adapted to alternative nonparametric estimation methods, such as wavelet-based estimators, delta sequence estimators [70], K ( · ) -nearest neighbor ( K ( · ) NN) estimators [12], and other local linear estimators [71]. These methods, widely used across various fields of functional data analysis, could benefit from the bias correction techniques introduced in this paper. Investigating how these estimators might be extended or combined with our proposed framework could thus reveal valuable synergies, further advancing the capabilities of nonparametric functional data analysis.
Finally, we propose relaxing the stationarity assumption that underpins the current study. Although stationarity is commonly assumed in many functional data models, it may not be appropriate for data exhibiting dynamic or time-varying behavior; see [72]. Extending our approach to accommodate locally stationary functional ergodic processes would be a natural and informative progression; see [73]. Under these more general conditions, the conditional quantile estimator would smoothly vary over time, thereby capturing nonstationary or evolving data structures. Exploring uniform limit theorems for such processes would be an especially promising avenue for future investigation, as it would shed additional light on the asymptotic properties of kernel estimators in more flexible, nonstationary contexts.

8. Proofs

This section is dedicated to presenting the proofs of our findings. The notation previously introduced will continue to be utilized in the ensuing discussions.
The proof Lemma 1 is based on the application of the Bernstein-type inequality for weakly dependent random variables, as proved by [39] in Section 2. Theorem 2.1 stated in the following lemma.
Lemma 8 ([39]). 
Suppose that X 1 , , X n are real-valued random variables with I E ( X i ) = 0 and I P ( | X i | < M ) = 1 , for all i = 1 , , n and some M < . Let σ n 2 = Var ( X 1 + + X n ) . Assume, furthermore, that there exist constants K 1 and β > 0 such that, for all u-tuples ( s 1 , , s u ) and all v-tuples ( t 1 , , t v ) with 1 s 1 s u t 1 t v n , the following inequality is fulfilled:
| Cov ( X s 1 X s u , X t 1 X t v ) | K 2 M u + v v exp β ( t 1 s u ) .
Then
I P i = 1 n X i t exp t 2 / 2 A n + B 1 / 3 t 5 / 3 ,
where A n can be chosen as any value greater than or equal to σ n 2 , and
B = 16 n K 2 9 A n ( 1 e β ) 1 2 ( K M ) 1 e β .
Proof of Lemma 1. 
Next, let us consider the following decomposition:
sup χ S H sup θ Θ H r ˜ 1 ( θ , χ ) I E r ˜ 1 ( θ , χ ) sup χ S H sup θ Θ H r ˜ 1 ( θ , χ ) r ˜ 1 ( θ , χ k ( χ ) ) F 1 + sup χ S H sup θ Θ H r ˜ 1 ( θ , χ k ( χ ) ) r ˜ 1 ( t j ( θ ) , χ k ( χ ) ) F 2 + sup χ S H sup θ Θ H r ˜ 1 ( t j ( θ ) , χ k ( χ ) ) I E r ˜ 1 ( t j ( θ ) , χ k ( χ ) ) F 3 + sup χ S H sup θ Θ H I E r ˜ 1 ( t j ( θ ) , χ k ( χ ) ) I E r ˜ 1 ( θ , χ k ( χ ) ) F 4 + sup χ S H sup θ Θ H I E r ˜ 1 ( θ , χ k ( χ ) ) I E r ˜ 1 ( θ , χ ) F 5 .
Since both F 1 and F 5 can be treated in the same way, we only deal with F 1 . Indeed, by (A3- ( i i ) ) and Cauchy Schwartz’s inequality, we have
sup χ S H sup θ Θ H r ˜ 1 ( θ , χ ) r ˜ 1 ( θ , χ k ( χ ) ) 1 n sup χ S H sup θ Θ H i = 1 n δ i | Y i | G ¯ ( Y i ) K i ( θ , χ ) I E K 1 ( θ , χ ) K i ( θ , χ k ( χ ) ) I E K 1 ( θ , χ k ( χ ) ) 1 n ϕ ( h n ) sup χ S H sup θ Θ H i = 1 n 1 { Y i 0 C i } | Y i 0 | G ¯ ( Y i 0 ) K i ( θ , χ ) K i ( θ , χ k ( χ ) ) C τ F G ¯ ( τ F ) sup χ S H sup θ Θ H θ · χ χ k ( χ ) h n ϕ ( h n ) C τ F G ¯ ( τ F ) l n ϕ ( h n ) .
Thus, by assumption (A4-( i i )), since l n = O ( log n n ) , by Jensen’s inequality, we see that n goes to infinity
F 5 F 1 = O log d n S H + log d n Θ H n ϕ ( h n ) .
Analogously, dealing with F 2 and F 4 in the same way, one may easily obtain
sup χ S H sup θ Θ H r ˜ 1 ( θ , χ ) r ˜ 1 ( θ , χ k ( χ ) ) 1 n ϕ ( h n ) sup χ S H sup θ Θ H i = 1 n 1 { Y i 0 C i } | Y i 0 | G ¯ ( Y i 0 ) K i ( θ , χ ) K i ( t j ( θ ) , χ ) C τ F G ¯ ( τ F ) sup χ S H sup θ Θ H χ · θ t j ( θ ) h n ϕ ( h n ) C τ F G ¯ ( τ F ) l n ϕ ( h n ) = O log n n ϕ ( h n ) .
Hence, we conclude that
F 4 F 2 = O log d n S H + log d n Θ H n ϕ ( h n ) .
Now, our focus turns to the upper bound of the term F 3 . Firstly, we express
I P F 3 > ϵ d n S H d n Θ H max 1 k d n S H max 1 j d n Θ H I P r ˜ N ( t j ( θ ) , χ k ( χ ) ) I E r ˜ N ( t j ( θ ) , χ k ( χ ) ) > ϵ d n S H d n Θ H max 1 k d n S H max 1 j d n Θ H I P i = 1 n Ψ i > ϵ ,
where
Ψ i = 1 n I E [ K 1 ( t j ( θ ) , χ k ( χ ) ) ] φ ( X i , Y i ) ,
with
φ ( X i , Y i ) = δ i Y i G ¯ ( Y i ) K i ( t j ( θ ) , χ k ( χ ) ) I E δ 1 Y 1 G ¯ ( Y 1 ) K 1 ( t j ( θ ) , χ k ( χ ) ) , 1 i n .
Clearly, we have
I E ( Ψ i ) = 0 φ 2 C K a n d Lip ( φ ) 2 C h n 1 Lip ( K ) .
It is important to highlight that the Kallabis and Neumann inequality in [39] relies on the asymptotic evaluation of both quantities:
Var i = 1 n Ψ i and Cov Ψ s 1 Ψ s u , Ψ t 1 Ψ t v , ( s 1 s u ) I N u ; ( t 1 t v ) I N v .
We begin by studying the variance term:
Var i = 1 n Ψ i = n Var Ψ 1 + i = 1 n l = 1 i l n Cov Ψ i , Ψ l = : J 1 , n + J 2 , n .
For J 1 , n , we have
J 1 , n = n ( n I E [ K 1 ( t j ( θ ) , χ k ( χ ) ) ] ) 2 Var φ ( X i , Y i ) O 1 n ( I E [ K 1 ( t j ( θ ) , χ k ( χ ) ) ] ) 2 I E K 1 2 ( t j ( θ ) , χ k ( χ ) ) .
Since K ( · ) is bounded with support [ 0 , 1 ] , Assumption (A1) implies
I E K 1 2 ( θ , χ ) C ϕ θ , χ ( h n ) , ( θ , χ ) H × H .
Hence,
J 1 , n = O 1 n ϕ θ , χ ( h n ) .
We will now focus on the quantity J 2 , n , and thus we have
J 2 , n = i = 1 n l = 1 i l n Cov Ψ i , Ψ l .
To deal with this last quantity, we merely need to split it into
J 2 , n = i = 1 n l = 1 0 < | l i | m n n Cov Ψ i , Ψ l + i = 1 n l = 1 m n + 1 | l i | n 1 n Cov Ψ i , Ψ l = : J 2 , n + J 2 , n ,
where m n is a sequence of natural numbers going to infinity as n . First, we deal with J 2 , n , so
J 2 , n C n ( n I E [ K 1 ( t j ( θ ) , χ k ( χ ) ) ] ) 2 l = 1 0 < | l i | m n n Cov K i ( t j ( θ ) , χ k ( χ ) ) , K l ( t j ( θ ) , χ k ( χ ) ) C n ( I E [ K 1 ( t j ( θ ) , χ k ( χ ) ) ] ) 2 j = 1 0 < | l i | m n n I E K i ( t j ( θ ) χ k ( χ ) ) K l ( t j ( θ ) , χ k ( χ ) ) .
Assumptions (A1), (A3), and (A6) allow us to write
I E K i ( t j ( θ ) χ k ( χ ) ) K l ( t j ( θ ) , χ k ( χ ) ) C I P ( X i , X l ) B θ ( χ , h n ) × B θ ( χ , h n ) = O ϕ θ , χ ( h n ) 2 .
Therefore,
J 2 , n = O m n n .
Second, by applying the quasi-association of the sequence ( X i , Y i ) , we have by assumption (A5):
J 2 , n C ( n I E [ K 1 ( t j ( θ ) , χ k ( χ ) ) ] ) 2 i = 1 n l = 1 | l i | m n + 1 n λ i , l .
Thus, since | l i | m n + 1 > m n
J 2 , n O 1 ( n ϕ θ , χ ( h n ) ) 2 n 2 λ m n = O ϕ θ , χ ( h n ) 2 e a m n .
So,
J 2 , n = J 2 , n + J 2 , n C m n n + ϕ θ , χ ( h n ) 2 e a m n = O m n n + 1 ϕ θ , χ ( h n ) 2 e a m n .
Choosing m n = 1 a log n ϕ θ , χ ( h n ) 2 , a > 0 , we have
J 2 , n = O o ( 1 ) + 1 n = o ( 1 ) .
Combining (19) and (21), we get
Var i = 1 n Ψ i = O 1 n ϕ θ , χ ( h n ) .
Now, we treat the covariance Cov Ψ s 1 Ψ s u , Ψ t 1 Ψ t v , ( s 1 s u ) I N u ; ( t 1 t v ) I N v .
To proceed, we consider the following cases:
  • If t 1 = s u , we obtain
    Cov Ψ s 1 Ψ s u , Ψ t 1 Ψ t v C K n I E [ K 1 ( t j ( θ ) , χ k ( χ ) ) ] u + v I E [ K 1 2 ( t j ( θ ) , χ k ( χ ) ) ] ϕ θ , χ ( h n ) C n ϕ θ , χ ( h n ) u + v .
  • If t 1 > s u , we apply the quasi-association condition to show
    | Cov Ψ s 1 Ψ s u , Ψ t 1 Ψ t v | C h n 1 Lip ( K ) ) n I E [ K 1 ( t j ( θ ) , χ k ( χ ) ) ] 2 2 K n I E [ K 1 ( t j ( θ ) , χ k ( χ ) ) ] u + v 2 i = 1 u j = 1 v λ s i , t j h n 2 C n ϕ θ , χ ( h n ) u + v v λ t 1 s u h n 2 C n ϕ θ , χ ( h n ) u + v v e a ( t 1 s u ) .
Conversely, we have
| Cov Ψ s 1 Ψ s u , Ψ t 1 Ψ t v | 2 C K n I E [ K 1 ( t j ( θ ) , χ k ( χ ) ) ] u + v 2 I E Ψ s u Ψ t 1 2 C K n I E [ K 1 ( t j ( θ ) , χ k ( χ ) ) ] u + v sup i l I P ( X i , X l ) B θ ( χ , h n ) × B θ ( χ , h n ) ϕ θ , χ ( h n ) 2 C n ϕ θ , χ ( h n ) u + v .
Now, taking a ρ -power of (24), ( 1 ρ ) -power of (25), under assumption (A7)- ( i ) , we obtain an upper bound of the three terms as follows
Cov Ψ s 1 Ψ s u , Ψ t 1 Ψ t v ϕ θ , χ ( h n ) C n ϕ θ , χ ( h n ) u + v v e a ρ ( t 1 s u ) .
Therefore, the conditions of Theorem 2.1 of [39] are fulfilled by the variables Ψ i , i = 1 , , n , where
K n = C n ϕ θ , χ ( h n ) , M n = C n ϕ θ , χ ( h n ) and A n = Var i = 1 n Ψ i = O 1 n ϕ θ , χ ( h n ) .
As a result, it follows that
I P i = 1 n Ψ i > η log d n S H + log d n Θ H n ϕ ( h n ) 1 / 2 exp η 2 log ( d n S H d n Θ H ) n ϕ ( h n ) C ( n ϕ ( h n ) ) 1 + ( n ϕ ( h n ) ) 1 η 5 log ( d n S H d n Θ H ) n ϕ ( h n ) 5 / 2 1 / 3 exp η 2 log ( d n S H d n Θ H ) C + n ϕ ( h n ) C ( n ϕ ( h n ) ) 1 η 5 log ( d n S H d n Θ H ) n ϕ ( h n ) 5 / 2 1 / 3 exp η 2 log ( d n S H d n Θ H ) C + η 5 / 3 ( n ϕ ( h n ) ) 4 / 5 log ( d n S H d n Θ H ) n ϕ ( h n ) 5 / 6 exp η 2 log ( d n S H d n Θ H ) C + η 5 / 3 log 5 ( d n S H d n Θ H ) ( n ϕ ( h n ) ) 1 / 6 .
From assumption (A7)- ( i i ) , we get
I P i = 1 n Ψ i > η log d n S H + log d n Θ H n ϕ ( h n ) 1 / 2 exp η 2 log ( d n S H d n Θ H ) C + o ( 1 ) = ( d n S H d n Θ H ) C η 2 .
An appropriate choice of η ensures the existence of some real β = C η 2 with β > 2 :
I P r ^ 1 ( t j ( θ ) , χ k ( χ ) ) I E r ^ 1 ( t j ( θ ) , χ k ( χ ) ) > η log d n S H + log d n Θ H n ϕ ( h n ) C d n S H d n Θ H β .
Therefore,
I P F 3 > ϵ C d n S H d n Θ H 1 β .
Assumption (A5)- ( i i ) implies
F 3 = O a . c o . log d n S H + log d n Θ H n ϕ ( h n ) .
Finally, results (26), (17), and (16) complete the proof of Lemma 3. □
Proof of Lemma 2. 
Since
I E r ^ 2 ( θ , χ ) = I E r ^ 2 ( t j ( θ ) , χ k ( χ ) ) = 1 ,
we can write
sup χ S H sup θ Θ H r ^ 2 ( θ , χ ) I E r ^ 2 ( θ , χ ) sup χ S H sup θ Θ H r ^ 2 ( θ , χ ) r ^ 2 ( θ , χ k ( χ ) ) + sup χ S H sup θ Θ H r ^ 2 ( θ , χ k ( χ ) ) r ^ 2 ( t j ( θ ) , χ k ( χ ) ) + sup χ S H sup θ Θ H r ^ 2 ( t j ( θ ) , χ k ( χ ) ) 1 .
For the first and the second terms on the right-hand side of the previous inequality, we use the Hölder continuity condition of K ( · ) (in view of (A3- ( i i ) )) and the Cauchy–Schwartz inequality to obtain
sup χ S H sup θ Θ H r ^ 2 ( θ , χ ) r ^ 2 ( θ , χ k ( χ ) ) = O log d n S H + log d n Θ H n ϕ ( h n ) ,
and
sup χ S H sup θ Θ H r ^ 2 ( θ , χ k ( χ ) ) r ^ 2 ( t j ( θ ) , χ k ( χ ) ) = O log d n S H + log d n Θ H n ϕ ( h n ) .
Clearly, by the Lipschitz property of K ( · ) imposed in (A3- ( i i ) ) and Cauchy–Schwartz inequality, we can write
sup χ S H sup θ Θ H r ^ 2 ( θ , χ ) r ^ 2 ( θ , χ k ( χ ) ) C ϕ θ , χ ( h n ) sup θ Θ H 1 n i = 1 n K d θ ( χ , X i ) h n K d θ ( χ k ( χ ) , X i ) h n C h n ϕ θ , χ ( h n ) sup θ Θ H 1 n i = 1 n | | θ , χ X i | θ , χ k ( χ ) X i | C sup θ Θ H θ · χ χ k ( χ ) h n ϕ ( h n ) C l n ϕ ( h n ) .
Because l n = O ( log n n ) , we obtain for n an infinitely large value
sup χ S H sup θ Θ H r ^ 2 ( θ , χ ) r ^ 2 ( θ , χ k ( χ ) ) = O log n n ϕ ( h n ) = o log d n S H + log d n Θ H n ϕ ( h n ) .
Similarly, following the same steps leading to (30), we have
sup χ S H sup θ Θ H r ^ 2 ( θ , χ k ( χ ) ) r ^ 2 ( t j ( θ ) , χ k ( χ ) ) = O log n n ϕ ( h n ) = o log d n S H + log d n Θ H n ϕ ( h n ) .
For the last term on the right-hand side of (27), we have
I P sup χ S H sup θ Θ H r ^ 2 ( t j ( θ ) , χ k ( χ ) ) 1 > η log d n S H + log d n Θ H n ϕ ( h n ) 1 / 2 I P max k = 1 , , d n S H max j = 1 , , d n Θ H r ^ 2 ( t j ( θ ) , χ k ( χ ) ) 1 > η log d n S H + log d n Θ H n ϕ ( h n ) 1 / 2 C d n S H d n Θ H I P r ^ 2 ( t j ( θ ) , χ k ( χ ) ) 1 > η log d n S H + log d n Θ H n ϕ ( h n ) 1 / 2 .
It worth nothing that the probability
I P r ^ 2 ( t j ( θ ) , χ k ( χ ) ) 1 > η log d n S H + log d n Θ H n ϕ ( h n ) 1 / 2 ,
represents a special case of the following probability when we take Y 1 :
I P r ˜ 1 ( t j ( θ ) , χ k ( χ ) ) I E r ˜ 1 ( t j ( θ ) , χ k ( χ ) ) > η log d n S H + log d n Θ H n ϕ ( h n ) 1 / 2 .
Thus,
I P sup χ S H sup θ Θ H r ^ 2 ( t j ( θ ) , χ k ( χ ) ) 1 > η log d n S H + log d n Θ H n ϕ ( h n ) C ( d n S H d n Θ H ) 1 β .
Therefore, by assumption (A5)- ( i i ) ,
sup χ S H sup θ Θ H r ^ 2 ( t j ( θ ) , χ k ( χ ) ) 1 = O log d n S H + log d n Θ H n ϕ ( h n ) .
Combining this last result together with results (31) and (30) ends the proof of the lemma. □
Proof of Corollary 1. 
Observe that if
inf χ S H inf θ Θ H r ^ 2 ( θ , χ ) 1 2 ,
then there must exist some χ S H and θ Θ H such that
sup χ S H sup θ Θ H 1 r ^ 2 ( θ , χ ) > 1 2 .
Hence, it follows that
P inf χ S H inf θ Θ H r ^ 2 ( θ , χ ) 1 2 P sup χ S H sup θ Θ H 1 r ^ 2 ( θ , χ ) > 1 2 .
By invoking Lemma 2, the proof of the corollary is thus established. □
Proof of Lemma 3. 
First, observe that the difference r ^ 1 ( θ , χ ) r ˜ 1 ( θ , χ ) can be written as
r ^ 1 ( θ , χ ) r ˜ 1 ( θ , χ ) = 1 n E [ K 1 ( θ , χ ) ] i = 1 n δ i Y i G ¯ n ( Y i ) δ i Y i G ¯ ( Y i ) K i ( θ , χ ) = 1 n E [ K 1 ( θ , χ ) ] i = 1 n 1 { Y i 0 C i } Y i 0 G ¯ n ( Y i 0 ) 1 { Y i 0 C i } Y i 0 G ¯ ( Y i 0 ) K i ( θ , χ ) .
Since δ i = 1 { Y i 0 C i } , we may factor out Y i 0 K i ( θ , χ ) and obtain
r ^ 1 ( θ , χ ) r ˜ 1 ( θ , χ ) 1 n E [ K 1 ( θ , χ ) ] i = 1 n | Y i 0 K i ( θ , χ ) | G ¯ n ( Y i 0 ) G ¯ ( Y i 0 ) G ¯ n ( Y i 0 ) G ¯ ( Y i 0 ) .
Noting that Y i 0 τ F and employing G ¯ n ( t ) G ¯ n ( τ F ) and G ¯ ( t ) G ¯ ( τ F ) for t < τ F , one obtains
r ^ 1 ( θ , χ ) r ˜ 1 ( θ , χ ) τ F G ¯ n ( τ F ) G ¯ ( τ F ) sup t < τ F G ¯ n ( t ) G ¯ ( t ) 1 n E [ K 1 ( θ , χ ) ] i = 1 n K i ( θ , χ ) .
Recognizing that 1 n E [ K 1 ( θ , χ ) ] i = 1 n K i ( θ , χ ) = r ^ 2 ( θ , χ ) , so we have
r ^ 1 ( θ , χ ) r ˜ 1 ( θ , χ ) r ^ 2 ( θ , χ ) τ F G ¯ n ( τ F ) G ¯ ( τ F ) sup t < τ F G ¯ n ( t ) G ¯ ( t ) .
One can further split r ^ 2 ( θ , χ ) as r ^ 2 ( θ , χ ) 1 + 1 , which yields
r ^ 1 ( θ , χ ) r ˜ 1 ( θ , χ ) ( r ^ 2 ( θ , χ ) 1 ) τ F G ¯ n ( τ F ) G ¯ ( τ F ) sup t < τ F G ¯ n ( t ) G ¯ ( t ) + τ F G ¯ n ( τ F ) G ¯ ( τ F ) sup t < τ F G ¯ n ( t ) G ¯ ( t ) .
Consequently,
sup χ S H sup θ Θ H r ^ 1 ( θ , χ ) r ˜ 1 ( θ , χ ) sup χ S H sup θ Θ H r ^ 1 ( θ , χ ) 1 τ F G ¯ n ( τ F ) G ¯ ( τ F ) sup t < τ F G ¯ n ( t ) G ¯ ( t ) + τ F G ¯ n ( τ F ) G ¯ ( τ F ) sup t < τ F G ¯ n ( t ) G ¯ ( t ) .
Next, by combining Lemma 2 with the result of [61], we know
sup t < τ F G ¯ n ( t ) G ¯ ( t ) = O log 2 ( n ) n .
Since G ¯ n ( τ F ) G ¯ ( τ F ) converges asymptotically to ( G ¯ ( τ F ) ) 2 as n , and
log 2 ( n ) n = o log ( n ) n ϕ ( h n ) by ( A 4 ) - ( ii ) ,
it follows that
sup χ S H sup θ Θ H r ^ 1 ( θ , χ ) r ˜ 1 ( θ , χ ) = o log d n S H + log d n Θ H n ϕ ( h n ) .
This completes the argument. □
Proof of Lemma 4. 
Leveraging the stationarity assumption, conditioning on θ , X 1 , and noting that
E 1 { Y 1 0 C 1 } | Y 1 0 = G ¯ Y 1 0 ,
we deduce the following sequence of identities:
E r ˜ 1 ( θ , χ ) r ( θ , χ ) = 1 E [ K 1 ( θ , χ ) ] E δ 1 Y 1 G ¯ ( Y 1 ) K 1 ( θ , χ ) r ( θ , χ ) = E E δ 1 Y 1 G ¯ ( Y 1 ) | θ , X 1 K 1 ( θ , χ ) r ( θ , χ ) E [ K 1 ( θ , χ ) ] E [ K 1 ( θ , χ ) ] = E E Y 1 0 G ¯ ( Y 1 0 ) E { 1 { Y 1 0 C 1 } Y 1 0 } | θ , X 1 K 1 ( θ , χ ) r ( θ , χ ) E [ K 1 ( θ , χ ) ] E [ K 1 ( θ , χ ) ] = E r ( θ , X 1 ) K 1 ( θ , χ ) r ( θ , χ ) E [ K 1 ( θ , χ ) ] E [ K 1 ( θ , χ ) ] = 1 E [ K 1 ( θ , χ ) ] E K 1 ( θ , χ ) 1 { B θ ( χ , h n ) } ( X 1 ) r ( θ , X 1 ) r ( θ , χ ) .
Under the Lipschitz condition (Assumption (A2)), there exists a constant C such that
1 B θ ( χ , h n ) ( X 1 ) r ( θ , X 1 ) r ( θ , χ ) C d θ ( χ , X 1 ) γ C h n γ .
Hence, uniformly in both χ and θ , it follows that
sup χ S H sup θ Θ H E r ˜ 1 ( θ , χ ) r ( θ , χ ) = O h n γ .
This completes the proof. □
Proof of Lemma 5. 
Below, we restate some auxiliary results from [63] that will be instrumental in our subsequent proofs:
1 ϕ θ , χ ( h n ) I E K 1 l ( θ , χ ) c l = K l ( 1 ) 0 1 ( K l ) ( s ) β θ , χ ( s ) d s , as n , l { 1 , 2 } .
For any θ , χ H , define
Z n i ( θ , χ ) = ϕ θ , χ ( h n ) n I E K 1 ( θ , χ ) δ i Y i G ¯ n ( Y i ) K i ( θ , χ ) I E δ 1 Y 1 G ¯ n ( Y 1 ) K 1 ( θ , χ ) ,
and let
S n : = i = 1 n Z n i ( θ , χ ) .
It follows immediately that
S n = n ϕ θ , χ ( h n ) r ^ 1 ( θ , χ ) I E r ^ 1 ( θ , χ ) .
Thus, our goal now is to show
S n D N 0 , σ r ( θ , χ ) ,
where D denotes convergence in distribution and σ r ( θ , χ ) is the limiting variance.
Our strategy for establishing the asymptotic behavior of S n is based on Doob’s methodology (see [74], pp. 228–231), which decomposes S n into a predictable component and a fluctuation component. This decomposition illuminates both the main stochastic fluctuations (controlled via conditional expectations) and the predictable variation (analyzed via limit theorems), providing a structured way to precisely characterize the limiting distribution of S n . We introduce two sequences of natural numbers p = p n and q = q n , both tending to infinity as n , with the additional restriction that q = o ( p ) . Using these, we decompose S n as follows:
S n = T n + T n + ζ k ,
where
T n = j = 1 k η j , T n = j = 1 k ξ j , ζ k = i = k ( p + q ) + 1 n Z n i ( θ , χ ) ,
and the index sets are defined by
I j = { ( j 1 ) ( p + q ) + 1 , , ( j 1 ) ( p + q ) + p } , J j = { ( j 1 ) ( p + q ) + p + 1 , , j ( p + q ) } ,
with
η j = i I j Z n i ( θ , χ ) , ξ j = i J j Z n i ( θ , χ ) .
Let k = n / ( p + q ) . It then follows that
p q n 1 and k q n 0 , as n .
The main asymptotic properties we need to establish can be summarized as verifying the following two statements:
I E ( T n ) 2 + I E ζ k 2 0 ,
and
T n D N 0 , σ r 2 ( θ , χ ) .
These results collectively imply the convergence in (34).
  • Proof of (35):
  • Proof of I E ( T n ) 2 0 : Using stationarity and basic covariance bounds, we have
    I E ( T n ) 2 = k Var ( ξ 1 ) + 2 1 i < j k | Cov ( ξ i , ξ j ) |
  • Bound for k Var ( ξ 1 ) :
    k Var ( ξ 1 ) k q Var ( Z n 1 ) + 2 k 1 i < j q | Cov ( Z n i , Z n j ) | .
From the definition of the Z n 1 , conditioning w.r.t θ , X 1 , and again the fact
I E 1 { Y 1 0 C 1 } | Y 1 0 = G ¯ ( Y 1 0 ) ,
and the fact G n ( t ) G ( t ) ; for instance, see [75]. We obtain
Var ( Z n 1 ) = ϕ θ , χ ( h n ) n I E K 1 ( θ , χ ) 2 Var δ 1 Y 1 G ¯ n ( Y 1 ) K 1 ( θ , χ ) I E δ 1 Y 1 G ¯ n ( Y 1 ) K 1 ( θ , χ ) = ϕ θ , χ ( h n ) n I E K 1 ( θ , χ ) 2 Var δ 1 Y 1 G ¯ n ( Y 1 ) K 1 ( θ , χ ) = ϕ θ , χ ( h n ) n I E K 1 ( θ , χ ) 2 Var I E δ 1 Y 1 G ¯ ( Y 1 ) | θ , X 1 G ¯ n ( Y 1 ) K 1 ( θ , χ ) = ϕ θ , χ ( h n ) n I E K 1 ( θ , χ ) 2 Var I E Y 1 0 G ¯ ( Y 1 0 ) I E 1 { Y 1 0 C 1 } | Y 1 0 | θ , X 1 G ¯ n ( Y 1 ) K 1 ( θ , χ ) = ϕ θ , χ ( h n ) n I E K 1 ( θ , χ ) 2 Var r ( θ , χ ) K 1 ( θ , χ ) .
Thanks to (33), we write
Var ( Z n 1 ) r ( θ , χ ) ϕ θ , χ ( h n ) n I E K 1 ( θ , χ ) 2 I E K 1 2 ( θ , χ ) r ( θ , χ ) ϕ θ , χ ( h n ) n c 1 ϕ θ , χ ( h n ) 2 c 2 ϕ θ , χ ( h n ) r ( θ , χ ) c 1 n ϕ θ , χ ( h n ) 2 c 2 ϕ θ , χ ( h n ) = O 1 n .
Hence, we have
Var ( Z n 1 ) = O k q n 0 , a s n .
Similarly, for the second term, we obtain
k 1 i < j q | Cov ( Z n i , Z n j ) | = k r ( θ , χ ) ϕ θ , χ ( h n ) n I E K 1 ( θ , χ ) 2 1 i < j q | Cov ( K i ( θ , χ ) , K j ( θ , χ ) ) | .
Now, we infer
1 i < j q | Cov ( K i ( θ , χ ) , K j ( θ , χ ) ) | = i = 1 q j = 1 j i q Cov ( K i ( θ , χ ) , K j ( θ , χ ) ) .
To deal with this last quantity, we merely need to split it into
i = 1 q j = 1 j i q Cov ( K i ( θ , χ ) , K j ( θ , χ ) ) = i = 1 q j = 1 0 < | j i | m q q Cov ( K i ( θ , χ ) , K j ( θ , χ ) ) + i = 1 q j = 1 m q + 1 | j i | q 1 q Cov ( K i ( θ , χ ) , K j ( θ , χ ) ) = : q ( Q 1 , q + Q 2 , q ) ,
where m q is a sequence of natural numbers going to infinity as q . First, we deal with Q 1 , q , so
Q 1 , q j = 1 0 < | j i | m q q Cov K i ( θ , χ ) , K j ( θ , χ ) j = 1 0 < | j i | m q q I E K i ( θ , χ ) K j ( θ , χ ) .
Assumptions (A1’), (A3’), and (A6) allow us to write
I E K i ( θ , χ ) K j ( θ , χ ) C I P ( X i , X l ) B θ ( χ , h n ) × B θ ( χ , h n ) = O ϕ θ , χ ( h n ) 2 .
Therefore,
Q 1 , q = O m q ϕ θ , χ ( h n ) 2 .
Second, by applying the quasi-association of the sequence ( X i , Y i ) , we have by assumption (A5):
Q 2 , q j = 1 | j i | m q + 1 q λ i , j .
Thus, since | j i | m q + 1 > m q
Q 2 , n O q λ m q = O q e a m q .
So,
i = 1 q j = 1 j i q Cov ( K i ( θ , χ ) , K j ( θ , χ ) ) = q O m q ϕ θ , χ ( h n ) 2 + O q e a m q .
Choosing m q = 1 a log q ϕ θ , χ ( h n ) 2 , a > 0 , we have
i = 1 q j = 1 j i q Cov ( K i ( θ , χ ) , K j ( θ , χ ) ) = q O o ( 1 ) + O ( ϕ θ , χ ( h n ) 2 ) = o ( q ϕ θ , χ ( h n ) ) .
Thus,
k 1 i < j q | Cov ( Z n i , Z n j ) | = O k n ϕ θ , χ ( h n ) o ( q ϕ θ , χ ( h n ) ) .
Therefore,
k 1 i < j q | Cov ( Z n i , Z n j ) | = o k q n 0 , a s n .
We use stationarity, to evaluate the second term on the right-hand side of (37), as follows
1 i < j k | Cov ( ξ i , ξ j ) | = l = 1 k 1 ( k l ) | Cov ( ξ 1 , ξ l + 1 ) | k l = 1 k 1 | Cov ( ξ 1 , ξ l + 1 ) | k l = 1 k 1 ( i , j ) J 1 × J l + 1 Cov ( Z n i , Z n j ) .
Clearly, for all ( i , j ) J 1 × J j , we have | i j | p + 1 > p , so
1 i < j k | Cov ( ξ i , ξ j ) | O k n ϕ θ , χ ( h n ) i = 1 p j = 2 p + q + 1 , | i j | > p k ( p + q ) λ i , j O k p n ϕ θ , χ ( h n ) λ p O k p n e a p ϕ θ , χ ( h n ) 0 .
Hence, by combining this last result with (38), (41), and (44), we can write
I E ( ζ k ) 2 0 , a s n .
Finally, the result (35) has been successfully established.
  • Proof of I E ( T n ) 2 0 : Since ( n k ( p + q ) ) p , we have
    I E ( ζ k ) 2 ( n k ( p + q ) ) Var ( Z n 1 ) + 2 1 i < j n | Cov ( Z n i , Z n j ) | p Var ( Z n 1 ) + 2 1 i < j n | Cov ( Z n i , Z n j ) | = O p n + o 1 n = O p n .
    Then,
    I E ( ζ k ) 2 0 , a s n .
  • Proof of (36): The proof is based on the following two results
    I E e i t j = 1 k η j j = 1 k I E e i t η j 0 ,
    and
    k Var ( η 1 ) σ r 2 ( θ , χ ) , k I E ( η 1 2 1 { η 1 > ϵ σ r ( θ , χ ) } ) 0 .
  • Proof of (45): First noting that
    | I E ( e i t j = 1 k η j ) j = 1 k I E e i t η j |       I E e i t j = 1 k η j I E e i t j = 1 k 1 η j I E e i t η k                 + I E e i t j = 1 k 1 η j j = 1 k 1 I E e i t η j           = Cov e i t j = 1 k 1 η j , e i t η k + I E e i t j = 1 k 1 η j j = 1 k 1 I E e i t η j ,
    we can repeatedly apply the same reasoning to obtain
    | I E ( e i t j = 1 k η j ) j = 1 k I E e i t η j |         Cov e i t j = 1 k 1 η j , e i t η k + Cov e i t j = 1 k 2 η j , e i t η k 1                   + + Cov e i t η 2 , e i t η 1 .
Next, using the quasi-association property, we deduce that
Cov e i t η 2 , e i t η 1 O t 2 n ϕ θ , χ ( h n ) i I 1 j I 2 λ i , j .
Applying a similar bound to each covariance term on the right-hand side of (47) gives
| I E e i t j = 1 k η j j = 1 k I E e i t η j | O t 2 n ϕ θ , χ ( h n ) i I 1 j I 2 λ i , j + i I 1 I 2 j I 3 λ i , j + + i I 1 I k 1 j I k λ i , j .
Observe that for any 2 l k 1 and ( i , j ) I l × I l + 1 , the condition | i j | q + 1 > q holds, implying
i I 1 I l 1 j I l λ i , j p λ q .
Hence, combining these estimates leads to
I E e i t j = 1 k η j j = 1 k I E e i t η j O t 2 n ϕ θ , χ ( h n ) k p λ q O t 2 e a q ϕ θ , χ ( h n ) 0 ,
which establishes the desired result.
Proof of (46): Building on the result in (39), we derive the following expression for k Var ( η 1 ) :
k Var ( η 1 ) = k p Var Z n 1 = k p r ( θ , χ ) ϕ θ , χ h n n E K 1 ( θ , χ ) 2 E K 1 2 ( θ , χ ) E [ K 1 ( θ , χ ) ] 2 = k p r 2 ( θ , χ ) ϕ θ , χ h n n E [ K 1 2 ( θ , χ ) ] E [ K 1 ( θ , χ ) ] 2 1 = k p n r 2 ( θ , χ ) ϕ θ , χ h n c 2 ϕ θ , χ h n c 1 ϕ θ , χ h n 2 1 = k p n r 2 ( θ , χ ) c 2 c 1 2 ϕ θ , χ h n .
It follows immediately that
k Var ( η 1 ) c 2 c 1 2 r 2 ( θ , χ ) = σ r 2 ( θ , χ ) .
Turning to the second term in (46), we note that
| η 1 | C p Z n 1 C p n ϕ θ , χ h n .
By applying Chebyshev’s inequality, we obtain
k E η 1 2 1 { η 1 > ϵ σ ( θ , χ ) } C k p 2 n ϕ θ , χ h n P η 1 > ϵ σ ( θ , χ ) C p 2 n ϕ θ , χ h n k Var ( η 1 ) ϵ 2 σ 2 ( θ , χ ) = O p 2 n ϕ θ , χ h n 0 as a consequence of ( A 7 ) .
This completes the proof. □
Proof of Lemma 6. 
By invoking Chebychev’s inequality, for any ϵ > 0 , we have
P r ^ 2 ( θ , χ ) 1 > ϵ = P r ^ 2 ( θ , χ ) E r ^ 2 ( θ , χ ) > ϵ E r ^ 2 ( θ , χ ) E r ^ 2 ( θ , χ ) 2 ϵ 2 Var r ^ 2 ( θ , χ ) ϵ 2 .
Apply the estimate in (22); it follows that
P r ^ 2 ( θ , χ ) 1 > ϵ O 1 n ϕ θ , χ ( h n ) 0 as n .
Hence, we conclude that
r ^ 2 ( θ , χ ) P 1 as n .
Proof of Lemma 7. 
Building on the conclusion of (48), one obtains
n ϕ θ , χ ( h n ) E r ^ 1 ( θ , χ ) r ( θ , χ ) r ^ 2 ( θ , χ )    = n ϕ θ , χ ( h n ) E r ^ 1 ( θ , χ ) r ( θ , χ )    = n h n 2 γ ϕ θ , χ ( h n ) 0 .
Hence, the proof is complete. □

Author Contributions

S.A., O.E.B., S.B. and A.L.: Conceptualization, formal analysis, investigation, methodology, project administration, original draft writing, review, and edition. All authors have read and agreed to the published version of the manuscript.

Funding

This project was supported by the PRFU of the Ministry of Higher Education and Scientific Research of Algeria (MESRS) and the University of Sciences and Technology Oran Mohamed Boudiaf (USTO-MB), under the code C00L03UN310220230005. The authors thank and extend their appreciation to the funders of this work. This work was supported by the Deanship of Scientific Research and Graduate Studies at King Khalid University through the Research Groups Program under grant number R.G.P. 1/163/45.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study.

Acknowledgments

The authors extend their sincere gratitude to the Editor-in-Chief, the Associate Editor, and the three referees for their invaluable feedback and for pointing out a number of oversights in the version initially submitted. Their insightful comments have greatly refined and focused the original work, resulting in a markedly improved presentation.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Härdle, W.; Hall, P.; Ichimura, H. Optimal smoothing in single-index models. Ann. Statist. 1993, 21, 157–178. [Google Scholar] [CrossRef]
  2. Hristache, M.; Juditsky, A.; Spokoiny, V. Direct estimation of the index coefficient in a single-index model. Ann. Statist. 2001, 29, 595–623. [Google Scholar] [CrossRef]
  3. Ferraty, F.; Peuch, A.; Vieu, P. Modèle à indice fonctionnel simple. Comptes Rendus Math. 2003, 336, 1025–1028. [Google Scholar] [CrossRef]
  4. Chen, D.; Hall, P.; Müller, H.G. Single and multiple index functional regression models with nonparametric link. Ann. Statist. 2011, 39, 1720–1747. [Google Scholar] [CrossRef]
  5. Ding, H.; Liu, Y.; Xu, W.; Zhang, R. A class of functional partially linear single-index models. J. Multivar. Anal. 2017, 161, 68–82. [Google Scholar] [CrossRef]
  6. Ramsay, J.O.; Silverman, B.W. Functional Data Analysis, 2nd ed.; Springer Series in Statistics; Springer: New York, NY, USA, 2005. [Google Scholar] [CrossRef]
  7. Ferraty, F.; Vieu, P. The functional nonparametric model and application to spectrometric data. Comput. Statist. 2002, 17, 545–564. [Google Scholar] [CrossRef]
  8. Ferraty, F.; Vieu, P. Nonparametric Functional Data Analysis. Theory and Practice; Springer Series in Statistics; Springer: New York, NY, USA, 2006. [Google Scholar] [CrossRef]
  9. Ferraty, F.; Park, J.; Vieu, P. Estimation of a functional single index model. In Recent Advances in Functional Data Analysis and Related Topics; Contributions to Statistics; Physica-Verlag/Springer: Berlin/Heidelberg, Germany, 2011; pp. 111–116. [Google Scholar] [CrossRef]
  10. Bouzebda, S. Limit Theorems in the Nonparametric Conditional Single-Index U-Processes for Locally Stationary Functional Random Fields under Stochastic Sampling Design. Mathematics 2024, 12, 1996. [Google Scholar] [CrossRef]
  11. Bouzebda, S. General tests of conditional independence based on empirical processes indexed by functions. Jpn. J. Stat. Data Sci. 2023, 6, 115–177. [Google Scholar] [CrossRef]
  12. Bouzebda, S. Uniform in Number of Neighbor Consistency and Weak Convergence of k-Nearest Neighbor Single Index Conditional Processes and k-Nearest Neighbor Single Index Conditional U-Processes Involving Functional Mixing Data. Symmetry 2024, 16, 1576. [Google Scholar] [CrossRef]
  13. Bouzebda, S.; Taachouche, N. Oracle inequalities and upper bounds for kernel conditional U-statistics estimators on manifolds and more general metric spaces associated with operators. Stochastics 2024, 96, 2135–2198. [Google Scholar] [CrossRef]
  14. Bouzebda, S.; Taachouche, N. On the variable bandwidth kernel estimation of conditional U-statistics at optimal rates in sup-norm. Phys. A 2023, 625, 129000. [Google Scholar] [CrossRef]
  15. Bulinski, A.; Suquet, C. Normal approximation for quasi-associated random fields. Statist. Probab. Lett. 2001, 54, 215–226. [Google Scholar] [CrossRef]
  16. Esary, J.D.; Proschan, F.; Walkup, D.W. Association of random variables, with applications. Ann. Math. Statist. 1967, 38, 1466–1474. [Google Scholar] [CrossRef]
  17. Doukhan, P.; Louhichi, S. A new weak dependence condition and applications to moment inequalities. Stoch. Process. Appl. 1999, 84, 313–342. [Google Scholar] [CrossRef]
  18. Richard, E.; Barlow, F.P. Statistical Theory of Reliability and Life Testing: Probability Models; International Series in Decision Processes Series in Quantitative Methods for Decision Making; Rinehart and Winston: Holt, MI, USA, 1975; p. 290. [Google Scholar]
  19. Ichimura, H. Semiparametric least squares (SLS) and weighted SLS estimation of single-index models. J. Econ. 1993, 58, 71–120. [Google Scholar] [CrossRef]
  20. Delecroix, M.; Härdle, W.; Hristache, M. Efficient estimation in conditional single-index regression. J. Multivar. Anal. 2003, 86, 213–226. [Google Scholar] [CrossRef]
  21. Huang, Z. Empirical likelihood for single-index varying-coefficient models with right-censored data. J. Korean Statist. Soc. 2010, 39, 533–544. [Google Scholar] [CrossRef]
  22. Li, T.T.; Yang, H.; Wang, J.L.; Xue, L.G.; Zhu, L.X. Estimation for a partial-linear single-index model [correction on MR2589322]. Ann. Statist. 2011, 39, 3441–3443. [Google Scholar] [CrossRef]
  23. Strzalkowska-Kominiak, E.; Cao, R. Maximum likelihood estimation for conditional distribution single-index models under censoring. J. Multivar. Anal. 2013, 114, 74–98. [Google Scholar] [CrossRef]
  24. Saidi, A.A.; Ferraty, F.; Kassa, R. Single functional index model for a time series. Rev. Roum. Math. Pures Appl. 2005, 50, 321–330. [Google Scholar]
  25. Masry, E. Nonparametric regression estimation for dependent functional data: Asymptotic normality. Stoch. Process. Appl. 2005, 115, 155–177. [Google Scholar] [CrossRef]
  26. Ait-Saïdi, A.; Ferraty, F.; Kassa, R.; Vieu, P. Cross-validated estimations in the single-functional index model. Statistics 2008, 42, 475–494. [Google Scholar] [CrossRef]
  27. Shang, H.L. Estimation of a functional single index model with dependent errors and unknown error density. arXiv 2018, arXiv:1810.08714. [Google Scholar] [CrossRef]
  28. Novo, S.; Aneiros, G.; Vieu, P. Automatic and location-adaptive estimation in functional single-index regression. J. Nonparametr. Stat. 2019, 31, 364–392. [Google Scholar] [CrossRef]
  29. Li, J.; Huang, C.; Zhu, H. A functional varying-coefficient single-index model for functional response data. J. Amer. Statist. Assoc. 2017, 112, 1169–1181. [Google Scholar] [CrossRef]
  30. Attaoui, S.; Ling, N. Asymptotic results of a nonparametric conditional cumulative distribution estimator in the single functional index modeling for time series data with applications. Metrika 2016, 79, 485–511. [Google Scholar] [CrossRef]
  31. Wang, G.; Feng, X.N.; Chen, M. Functional partial linear single-index model. Scand. J. Stat. 2016, 43, 261–274. [Google Scholar] [CrossRef]
  32. Attaoui, S.; Bentata, B.; Bouzebda, S.; Laksaci, A. The strong consistency and asymptotic normality of the kernel estimator type in functional single index model in presence of censored data. AIMS Math. 2024, 9, 7340–7371. [Google Scholar] [CrossRef]
  33. Elmezouar, Z.C.; Alshahrani, F.; Almanjahie, I.M.; Bouzebda, S.; Kaid, Z.; Laksaci, A. Strong consistency rate in functional single index expectile model for spatial data. AIMS Math. 2024, 9, 5550–5581. [Google Scholar] [CrossRef]
  34. Douge, L. Théorèmes limites pour des variables quasi-associées hilbertiennes. Ann. ISUP 2010, 54, 51–60. [Google Scholar]
  35. Attaoui, S.; Laksaci, A.; Ould Said, E. A note on the conditional density estimate in the single functional index model. Statist. Probab. Lett. 2011, 81, 45–53. [Google Scholar] [CrossRef]
  36. Ling, N.; Xu, Q. Asymptotic normality of conditional density estimation in the single index model for functional time series data. Statist. Probab. Lett. 2012, 82, 2235–2243. [Google Scholar] [CrossRef]
  37. Ye, Z.; Hooker, G. Local Quadratic Estimation of the Curvature in a Functional Single Index Model. arXiv 2018, arXiv:1803.09321. [Google Scholar] [CrossRef]
  38. Douge, L. Nonparametric regression estimation for quasi-associated Hilbertian processes. arXiv 2018, arXiv:1805.02422. [Google Scholar]
  39. Kallabis, R.S.; Neumann, M.H. An exponential inequality under weak dependence. Bernoulli 2006, 12, 333–350. [Google Scholar] [CrossRef]
  40. Bouzebda, S.; Laksaci, A.; Mohammedi, M. The k-nearest neighbors method in single index regression model for functional quasi-associated time series data. Rev. Mat. Complut. 2023, 36, 361–391. [Google Scholar] [CrossRef]
  41. Harris, T.E. A lower bound for the critical probability in a certain percolation process. Proc. Camb. Philos. Soc. 1960, 56, 13–20. [Google Scholar] [CrossRef]
  42. Lehmann, E.L. Some concepts of dependence. Ann. Math. Statist. 1966, 37, 1137–1153. [Google Scholar] [CrossRef]
  43. Joag-Dev, K.; Proschan, F. Negative association of random variables, with applications. Ann. Statist. 1983, 11, 286–295. [Google Scholar] [CrossRef]
  44. Bulinski, A.; Sabanovitch, E. Asymptotical behaviour of some functionals of positively and negatively random fields. Fundam. Appl. Math. 1998, 4, 479–492. [Google Scholar]
  45. Shashkin, A.P. Quasi-associatedness of a Gaussian system of random vectors. Uspekhi Mat. Nauk 2002, 57, 199–200. [Google Scholar] [CrossRef]
  46. Bulinskii, A.V.; Shashkin, A.P. Rates in the CLT for sums of dependent multiindexed random vectors. J. Math. Sci. 2004, 122, 3343–3358. [Google Scholar] [CrossRef]
  47. Newman, C.M. Asymptotic independence and limit theorems for positively and negatively dependent random variables. In Inequalities in Statistics and Probability (Lincoln, Neb., 1982); IMS Lecture Notes Monograph Series; Institute of Mathematical Statistics: Hayward, CA, USA, 1984; Volume 5, pp. 127–140. [Google Scholar] [CrossRef]
  48. Pitt, L.D. Positively correlated normal variables are associated. Ann. Probab. 1982, 10, 496–499. [Google Scholar] [CrossRef]
  49. Doukhan, P. Mixing; Lecture Notes in Statistics; Springer: New York, NY, USA, 1994; Volume 85, pp. xii+142. [Google Scholar] [CrossRef]
  50. Pham, T.D.; Tran, L.T. Some mixing properties of time series models. Stoch. Process. Appl. 1985, 19, 297–303. [Google Scholar] [CrossRef]
  51. Andrews, D.W.K. Nonstrong mixing autoregressive processes. J. Appl. Probab. 1984, 21, 930–934. [Google Scholar] [CrossRef]
  52. Louhichi, S. Weak convergence for empirical processes of associated sequences. Ann. Inst. H. Poincaré Probab. Statist. 2000, 36, 547–567. [Google Scholar] [CrossRef]
  53. Bradley, R.C. Basic properties of strong mixing conditions. A survey and some open questions. Probab. Surv. 2005, 2, 107–144. [Google Scholar] [CrossRef]
  54. Yu, H. A Glivenko-Cantelli lemma and weak convergence for empirical processes of associated sequences. Probab. Theory Relat. Fields 1993, 95, 357–370. [Google Scholar] [CrossRef]
  55. Nadaraja, E.A. On a regression estimate. Teor. Verojatnost. Primenen. 1964, 9, 157–159. [Google Scholar]
  56. Watson, G.S. Smooth regression analysis. Sankhyā Ser. A 1964, 26, 359–372. [Google Scholar]
  57. Carbonez, A.; Györfi, L.; van der Meulen, E.C. Partitioning-estimates of a regression function under random censoring. Statist. Decis. 1995, 13, 21–37. [Google Scholar] [CrossRef]
  58. Kohler, M.; Máthé, K.; Pintér, M. Prediction from randomly right censored data. J. Multivar. Anal. 2002, 80, 73–100. [Google Scholar] [CrossRef]
  59. Bouzebda, S.; Chaouch, M. Uniform limit theorems for a class of conditional Z-estimators when covariates are functions. J. Multivar. Anal. 2022, 189, 104872. [Google Scholar] [CrossRef]
  60. Kaplan, E.L.; Meier, P. Nonparametric estimation from incomplete observations. J. Am. Stat. Assoc. 1958, 53, 457–481. [Google Scholar] [CrossRef]
  61. Deheuvels, P.; Einmahl, J.H.J. Functional limit laws for the increments of Kaplan-Meier product-limit processes and applications. Ann. Probab. 2000, 28, 1301–1335. [Google Scholar] [CrossRef]
  62. Hall, P.; Hosseini-Nasab, M. On properties of functional principal components analysis. J. R. Stat. Soc. Ser. B Stat. Methodol. 2006, 68, 109–126. [Google Scholar] [CrossRef]
  63. Ferraty, F.; Mas, A.; Vieu, P. Nonparametric regression on functional data: Inference and practical aspects. Aust. N. Z. J. Stat. 2007, 49, 267–286. [Google Scholar] [CrossRef]
  64. Karhunen, K. Über lineare Methoden in der Wahrscheinlichkeitsrechnung. Ann. Acad. Sci. Fenn. Ser. A I Math.-Phys. 1947, 1947, 79. [Google Scholar]
  65. Loève, M. Probability Theory II, 4th ed.; Graduate Texts in Mathematics; Springer: New York, NY, USA; Berlin/Heidelberg, Germany, 1978; Volume 46, pp. xvi+413. [Google Scholar]
  66. Loève, M. Probability Theory. Foundations. Random Sequences; D. Van Nostrand Co., Inc.: Toronto, ON, Canada; New York, NY, USA; London, UK, 1955; pp. xv+515. [Google Scholar]
  67. McCullagh, P.; Nelder, J.A. Generalized Linear Models, 2nd ed.; Monographs on Statistics and Applied Probability; Chapman & Hall: London, UK, 1989; pp. xix+511. [Google Scholar] [CrossRef]
  68. Yao, W. A bias corrected nonparametric regression estimator. Statist. Probab. Lett. 2012, 82, 274–282. [Google Scholar] [CrossRef]
  69. Karlsson, A. Bootstrap methods for bias correction and confidence interval estimation for nonlinear quantile regression of longitudinal data. J. Stat. Comput. Simul. 2009, 79, 1205–1218. [Google Scholar] [CrossRef]
  70. Bouzebda, S.; Nezzal, A.; Zari, T. Uniform Consistency for Functional Conditional U-Statistics Using Delta-Sequences. Mathematics 2023, 11, 161. [Google Scholar] [CrossRef]
  71. Cheng, M.Y.; Wu, H.T. Local linear regression on manifolds and its geometric interpretation. J. Amer. Statist. Assoc. 2013, 108, 1421–1434. [Google Scholar] [CrossRef]
  72. Soukarieh, I.; Bouzebda, S. Weak convergence of the conditional U-statistics for locally stationary functional time series. Stat. Inference Stoch. Process. 2024, 27, 227–304. [Google Scholar] [CrossRef]
  73. Bouzebda, S.; Didi, S. Multivariate wavelet density and regression estimators for stationary and ergodic discrete time processes: Asymptotic results. Comm. Statist. Theory Methods 2017, 46, 1367–1406. [Google Scholar] [CrossRef]
  74. Doob, J.L. Stochastic Processes; John Wiley & Sons, Inc.: New York, NY, USA; Chapman & Hall, Ltd.: London, UK, 1953; pp. viii+654. [Google Scholar]
  75. Cai, Z. Estimating a distribution function for censored time series data. J. Multivar. Anal. 2001, 78, 299–318. [Google Scholar] [CrossRef]
Figure 1. A sample of 100 curves of different functional time series.
Figure 1. A sample of 100 curves of different functional time series.
Mathematics 13 00886 g001
Figure 2. Comparison of coverage probability of 95% confidence interval.
Figure 2. Comparison of coverage probability of 95% confidence interval.
Mathematics 13 00886 g002
Figure 3. Comparison of 95% confidence interval-length.
Figure 3. Comparison of 95% confidence interval-length.
Mathematics 13 00886 g003
Table 1. The bias and the variance for different scenarios.
Table 1. The bias and the variance for different scenarios.
Functional Time Series Casen λ BiasVar
5020.14360.783
500.50.2300.952
500.050.1110.998
10020.1960.874
Independent case1000.50.1120.831
1000.050.0510.967
35020.1210.697
3500.50.1010.715
3500.050.0220.881
5020.2470.861
500.50.3310.753
500.050.1020.977
10020.1660.699
FAR(1)1000.50.1260.702
1000.050.0990.913
35020.1350.578
3500.50.8810.677
3500.050.0850.863
5020.2810.841
500.50.1730.919
500.050.1840.942
10020.1690.768
FARCH(1,1)1000.50.1420.823
1000.050.1060.893
35020.1170.673
3500.50.1100.671
3500.050.0980.855
5020.3571.811
500.50.1751.182
500.050.1941.121
10020.21261.630
FGARCH(1,1)1000.50.1431.071
1000.050.1361.018
35020.1811.541
3500.50.1211.131
3500.050.1191.009
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Attaoui, S.; Benouda, O.E.; Bouzebda, S.; Laksaci, A. Limit Theorems for Kernel Regression Estimator for Quasi-Associated Functional Censored Time Series Within Single Index Structure. Mathematics 2025, 13, 886. https://doi.org/10.3390/math13050886

AMA Style

Attaoui S, Benouda OE, Bouzebda S, Laksaci A. Limit Theorems for Kernel Regression Estimator for Quasi-Associated Functional Censored Time Series Within Single Index Structure. Mathematics. 2025; 13(5):886. https://doi.org/10.3390/math13050886

Chicago/Turabian Style

Attaoui, Said, Oum Elkheir Benouda, Salim Bouzebda, and Ali Laksaci. 2025. "Limit Theorems for Kernel Regression Estimator for Quasi-Associated Functional Censored Time Series Within Single Index Structure" Mathematics 13, no. 5: 886. https://doi.org/10.3390/math13050886

APA Style

Attaoui, S., Benouda, O. E., Bouzebda, S., & Laksaci, A. (2025). Limit Theorems for Kernel Regression Estimator for Quasi-Associated Functional Censored Time Series Within Single Index Structure. Mathematics, 13(5), 886. https://doi.org/10.3390/math13050886

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop