Next Article in Journal
Analysis and Demonstration of Control Scheme for Multiple Operating Modes of Energy Storage Converters to Enhance Power Factor
Next Article in Special Issue
Functional Ergodic Time Series Analysis Using Expectile Regression
Previous Article in Journal
Predicting Subgrade Resistance Value of Hydrated Lime-Activated Rice Husk Ash-Treated Expansive Soil: A Comparison between M5P, Support Vector Machine, and Gaussian Process Regression Algorithms
Previous Article in Special Issue
On Some Improved Class of Estimators by Using Stratified Ranked Set Sampling
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Wavelet Density and Regression Estimators for Functional Stationary and Ergodic Data: Discrete Time

1
Department of Statistics, College of Sciences, Qassim University, P.O. Box 6688, Buraydah 51452, Saudi Arabia
2
Department of Mathematics, College of Sciences, Qassim University, P.O. Box 6688, Buraydah 51452, Saudi Arabia
3
LMAC (Laboratory of Applied Mathematics of Compiègne), Université de Technologie de Compiègne, 60200 Compiègne, France
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2022, 10(19), 3433; https://doi.org/10.3390/math10193433
Submission received: 6 August 2022 / Revised: 14 September 2022 / Accepted: 17 September 2022 / Published: 21 September 2022

Abstract

:
The nonparametric estimation of density and regression function based on functional stationary processes using wavelet bases for Hilbert spaces of functions is investigated in this paper. The mean integrated square error over adapted decomposition spaces is given. To obtain the asymptotic properties of wavelet density and regression estimators, the Martingale method is used. These results are obtained under some mild conditions on the model; aside from ergodicity, no other assumptions are imposed on the data. This paper extends the scope of some previous results for wavelet density and regression estimators by relaxing the independence or the mixing condition to the ergodicity. Potential applications include the conditional distribution, curve discrimination, and time series prediction from a continuous set of past values.

1. Introduction

The statistical literature has recently become more interested in statistical issues concerning studying functional random variables or variables with values in an infinite-dimensional space. The availability of data measured on ever-finer temporal/spatial grids, such as in meteorology, medicine, satellite images, and many other research disciplines, is driving the growth of this research issue, statistically modeling these data as random functions revealed many complex theoretical and numerical research challenges. The reader may consult the monographs for a summary of the theoretical and practical aspects of functional data analysis. The work in Bosq [1] concerns linear models for random variables with values in a Hilbert space, Ramsay and Silverman [2] discussed scalar-on-function and function-on-function linear models, functional principal component analysis, and parametric discriminant analysis. The work in [3], on the other hand, concentrates on nonparametric methods, particularly kernel-type estimation for scalar-on-function nonlinear regression models. Such tools were extended to classification and discrimination analysis. Horváth and Kokoszka [4] discussed the application of several interesting statistical concepts to the functional data framework, including goodness-of-fit tests, portmanteau tests, and change point problems. The work in [5] focuses on analyzing variance for functional data, whereas that in [6] is more concerned with regression analysis for Gaussian processes. Recent studies and surveys on functional data modeling and analysis can be found in [7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22].
Motivated by diverse applications and their helpful role in statistical inference, the problem of estimating conditional models has been subjected to a wide range of statistical literature, employing many types of estimation approaches, the most common of which are the traditional kernel methods. Such methods, however, may have some limitations when estimating compactly supported or discontinuous curves at boundary points. Alternative wavelet methods are prominent due to their adaptability to discontinuities in the curve to be estimated. In practice, the wavelet procedure provides a simple estimation algorithm to implement and compute. For more information on wavelets theory, we refer to [23,24,25,26] and others. The work in [27] discusses wavelet approximation properties in detail, and surveys the use of wavelets in various curve estimation problems. Some wavelet theory applications are discussed in [28] by considering the estimation of the integrated squared derivative density function in the independent unidimensional case. The results in [28] were then extended by [29] to estimate the derivatives of a density for negatively and positively associated sequences, respectively. Rao [30] proposed wavelet estimators for the partial derivatives of a multivariate probability density function, where the rates of almost certain convergence for the independence case are obtained. We cite [31] on estimating partial derivatives of a multivariate probability density function in the presence of additive noise. At this point, we refer to [32]. In the i.i.d. framework, Ref. [33] investigated the density and regression estimation problems for functional data. The authors [33] developed a new adaptive procedure based on the term-by-term selection of wavelet coefficient estimators using wavelet bases for Hilbert spaces of functions. The primary goal of this paper is to extend the previous reference to stationary ergodic processes. To our knowledge, the consideration of the general dependence framework for wavelet analysis is unexplored, which motivates this study. Allow X n , n Z to be a stationary sequence. Consider the backward field A n = σ X k : k n and the forward field B n = σ X k : k n . The sequence is strongly mixing if
sup A A 0 , B B n | P ( A B ) P ( A ) P ( B ) | = α ( n ) 0 as n .
The sequence is ergodic if
lim n 1 n k = 0 n 1 P A τ k B P ( A ) P ( B ) = 0 ,
where τ is the time-evolution or shift transformation. The naming of strong mixing in the above definition is more stringent than what is ordinarily referred to (when using the vocabulary of measure-preserving dynamical systems) as strong mixing, namely to that
lim n P A τ n B = P ( A ) P ( B )
for any two measurable sets A , B , see, for instance, Ref. [34]. As a result, strong mixing implies ergodicity, whereas the converse is not always true (see, for example, Remark 2.6 on page 50 concerning Proposition 2.8 on page 51 in [35]). Some reasons for considering ergodic dependence structure in data rather than a mixing structure are discussed in [36,37,38,39,40,41,42,43,44], where details on the definition of ergodic property of processes are given, as well as illustrative examples of such processes. One of the arguments used in [45] to justify the ergodic setting is that for certain classes of processes, proving ergodic properties rather than the mixing condition can be much easier. As a result, the ergodicity hypothesis appears to be the best fit and offers a better framework for studying data series generated by noisy chaos. The work in [45] provided an example of an ergodic but non-mixing process in their discussion, which can be summarized as follows: Let ( T i , λ i ) : i Z be a strictly stationary process such that T i T i 1 is a Poisson process with parameter λ i , where T i is the s i g m a -field generated by ( T i , λ i , T i 1 , ) . Assume that
λ i = f ( λ i 1 , T i 1 ) ,
and f : [ 0 , ] × N ( 0 , ) is a given function. This process is not mixing in general (see Remark 3 of [46]). It is known that any sequence ( ε i ) i Z of i.i.d. random variables is ergodic. Hence, according to Proposition 2.10 in [35], it is easy to see that ( Y i ) i Z with
Y i = ϑ ( ( , ε i 1 , ε i ) , ( ε i + 1 , ε i + 2 , ) ) ,
for some Borel-measurable function ϑ ( · ) .
The primary goal of this paper is to provide the first complete theoretical justification for wavelet-based functional density and regression function estimation for stationary processes. To our knowledge, the mean integrated square error over adapted decomposition spaces using wavelet estimators in functional ergodic data frameworks has not yet been considered in the literature and thus remains a fundamentally unsolved open problem. By combining several Martingale theory techniques used in the mathematical development of the proofs, we hope to fill this gap in the literature. These tools are not the same as those used in regression estimation under strong mixing or in an independent setting. However, as we will see later, combining existing ideas and results is not enough to solve the problem. Dealing with wavelet estimators in an ergodic setting will involve detailed mathematical derivations.
The following is how the paper is structured. The multiresolution analysis is introduced in Section 2. The main results for density estimation are presented in Section 3. The main results for the regression estimation are presented in Section 4. Some potential applications are listed in Section 5. Section 6 contains some concluding remarks. Section 7 contains a collection of all proofs.

2. Multiresolution Analysis

We will now introduce some basic notation to define wavelet bases for Hilbert spaces of functions following [33,47] with some changes necessary for our setting. In our work, we consider nonlinear, thresholded, wavelet-based estimators. Firstly, we initiate our study by describing elements of the basic theory of wavelet methods and introducing nonlinear wavelet-based estimators; the interested reader may refer to [23,24], see also [48,49] and the references therein, although the wavelet bases on a separable Hilbert space H of real or complex-valued functions on a complete separable metric space was introduced later by [47], which we briefly recall here for the sake of reader’s convenience. Let H be a separable Hilbert space of real-valued functions defined on a complete separable metric space S . Since the space H is separable, it has an orthonormal basis
E = e j : j Δ ,
where Δ is a countable index set. The space H is equipped with an inner product · , · and a norm · .
Consider the sequence of subsets I k ; k 0 an increasing sequence of finite subsets of Δ such that
k 0 I k = Δ .
The subset J k denotes the orthogonal complement of I k in I k + 1 , i.e.,
J k = I k + 1 I k .
Choose, for any k 0 , ζ k , S , I k and η k , S , J k , such that the following matrices
A k = e j ζ k , ( j , ) I k × I k , B k = e j η k , ( j , ) J k × J k ,
satisfy one of the two following conditions, for instance, see [33,47] and the references therein.
(A.1)
A k * A k = diag ( a k , ) I k and B k * B k = diag ( b k , ) J k where a k , and b k , for I k and J k are positive constants.
(A.2)
A k A k * = diag ( c k , ) I k and B k B k * = diag ( d k , ) J k where c k , and d k , for I k and J k are positive constants.
The condition (A.1) implies that
a k , = j I k | e j ζ k , | 2 , I k , and b k , = j J k | e j η k , | 2 , I k ,
which means that all the columns of A k and B k are not the zero vector. As for (A.2), it gives
c k , = I k | e j ζ k , | 2 , j I k , and d k , = J k | e j η k , | 2 , j I k ,
indicating that all the rows of A k and B k are not the zero vector. For any x S , we set
ϕ k ( · ; ζ k , ) = j I k 1 g j , k , e j ( ζ k , ) ¯ e j ( · ) , ψ k ( · ; η k , ) = j J k 1 h j , k , e j ( η k , ) ¯ e j ( · ) ,
where
g j , k , = a k , if ( A . 1 ) , c k , if ( A . 2 ) , h j , k , = b k , if ( A . 1 ) , d k , if ( A . 2 ) ,
The following collection form an orthonormal basis of H (see Theorem 2 of [47]):
B = ϕ 0 x , ζ 0 , , I 0 ; ψ k x , η k , , k 0 , J k .
For more details, see [33,47,50]. Hence, we conclude that for any f H , we have
f ( x ) = I 0 α 0 , ϕ 0 x ; ζ 0 , + k 0 J k β k , ψ k x ; η k ; ,
where
α 0 , = f , ϕ 0 · ; ζ 0 , , β k , = f , ψ k · ; η k ; .
In the following, we add two assumptions on the orthonormal basis E :
(E.1) 
There exists a constant C 1 > 0 such that, for any integer k 0 , one has
(i)
j I k 1 g j , k , e j ( ζ k , ) 2 C 1 ,
(ii)
j I k 1 h j , k , e j ( η k , ) 2 C 1 .
(E.2)
There exists a constant C 2 > 0 such that, for any integer k 0 , one has
sup x S j J k e j ( x ) 2 C 2 | J k | .
Remark 1.
Clearly, one can see that assumption (E.1)is satisfied under assumption (A.1)when taking C 1 = 1 , we may also refer to [47], Section 4, Example 2 and its applications for more details. [47,50] have presented three examples verifying the assumption (E.2)taking
sup x S j J k e j ( x ) 2 1 ,
see also [50] Theorem 3.2. Moreover, Ref. [33] have used both assumptions in the case of i.i.d and functional data.

Besov Space

Over the years, many researchers have addressed, from a statistical point of view, the following question: given an estimation method and a prescribed estimation rate for a given loss function, what is the maximal space over which this rate is achieved, for instance, see [27,51] and the references therein. We are interested in the estimation methods based on wavelet bases’ thresholding procedures in a natural setting. It is well known that wavelet bases provide characterizations of smoothness spaces such as the Hölder spaces C s , Sobolev spaces W s ( L p ) , and Besov spaces B q s ( L p ) for a range of indices s that depend both on the smoothness properties of ψ and its dual function ψ ˜ , for instance, we refer to [51] for more detail and examples, at this point, we may refer to [52]. From a statistical point of view, the following definition is used in approximation theory for the study of nonlinear procedures such as thresholding and greedy algorithms, for instance we refer to [27,49,51,53].
Definition 1
(Besov space). Let s > 0 . We say that the function f H , defined by (7), belongs to the Besov space B s ( H ) if and only if:
sup m 0 | J m | 2 s k m J k | β k , | 2 < .
Definition 2
(Weak Besov space). Let r > 0 . We say that the function f H , defined by (7), belongs to the weak Besov space W r ( H ) if and only if:
sup λ 0 λ r k 0 J k 𝟙 | β k , | λ < .

3. Problem Definition of the Density Estimation

Let { X i , Y i } i 1 be a sequence strictly stationary ergodic pairs of random elements where Y i is a real or complex-valued variable and X i takes values in a complete separable metric space of Hilbert space S associated with the corresponding Borel σ -algebra B . Let P X be the probability measure induced by X 1 on ( S , B ) . Suppose that there exists σ -finite measure ν on the measurable space ( S , B ) such that P X is dominated by ν . The Radon–Nikodym theorem ensures the existence of a non-negative measurable function f ( · ) such that
P X ( B ) = B f ( x ) ν ( d x ) , B B .
In this context, we aim to estimate f ( · ) based on n observed functional data X 1 , , X n – Examples of such random elements X 1 are stochastic processes with continuous sample paths on a finite interval [ a , b ] with S = C [ a , b ] associated with supremum norm and processes with square integrable sample paths on the real line when S = L 2 ( R ) . We suppose that f H , where H is a separable Hilbert space of real or complex-valued functions defined on S and square-integrable with respect to the σ -finite measure ν . In this paper, we are particularly interested in the wavelet estimation procedures developed in the 1990s; see Meyer’s work for the functional data of a Hilbert space, more precisely, the nonlinear estimators. The majority of the approaches carried out in this model consist in introducing kernel estimators techniques to estimate the model’s functional part, refer to [54]. Let f ( · ) be the common density function of the sample X 1 , , X n , which is assumed to be
(F.1) 
C f > 0 a known constant such that
sup x S f ( x ) C f .

3.1. Density Function Estimator

From now, we assume that the density function f ( · ) H , a separable Hilbert space. Then, f ( · ) fulfills the wavelet representation (7). Suppose that we observe a sequence ( X i , Y i ) i = 1 n of copies of ( X , Y ) that is assumed to be functional stationary and ergodic with X admitting the density function f ( · ) . We study density estimation through wavelet bases for Hilbert spaces of functions developed by [47]. We consider the estimates of the coefficients { α k , } and { β k , } given, respectively, by (14) and (15). For any j 0 m . Here, the resolution level m = m ( n ) at a rate specified below, since we assume that ϕ ( · ) and ψ i ( · ) have a compact support so that the summations in (7) are finite for each fixed x (note that in this case the support of ϕ ( · ) and ψ i ( · ) is a monotonically increasing function of their degree of differentiability [24]). We focus our attention on the nonlinear estimators (13) which will be studied in the mean integrated squared error over adapted decomposition spaces, in a similar way as in [33] in the setting of the i.i.d functional processes. The density wavelet hard thresholding estimator f ^ ( · ) is defined, for all x S , by
f ^ n ( x ) = I 0 α ^ 0 , ϕ 0 x ; ζ 0 , + k = 0 m n J k β ^ k , 𝟙 | β ^ k , | κ ln n n ψ k x ; η k ; ,
where
α ^ k , = 1 n i = 1 n ϕ k X i ; ζ k , ,
β ^ k , = 1 n i = 1 n ψ k X i ; η k , .
Here, κ is a large enough constant and m n is the integer satisfying
1 2 n ln n | J m n | n ln n .

3.2. Estimation Procedure Steps

Our estimation method is divided into three steps:
1.
Estimation of the wavelet coefficients α k , and β k , , see (8), by the estimators α ^ k , and β ^ k , defined by Equations (14) and (15);
2.
Applying the hard thresholding to select the greatest β ^ k , ;
3.
Reconstructing the selected elements of the initial wavelet basis.
It is important to note that our choice is the universal threshold κ ln n n 1 / 2 and the definition of m n is based on theoretical considerations. The considered estimator does not depend on the smoothness of f ( · ) ; we may refer the reader to [50] for more details in the case of the linear wavelet estimator of f ( · ) . Furthermore, for more details on the case of H = L ( [ a , b ] ) and more standard nonparametric models, see [27,55]. To state the results, we need some notation. Throughout the paper, we will denote by F i the σ field generated by { X j : 0 j i } and G i the σ field generated by { ( X j , Y j ) : X i + 1 , 0 j i } . Let B B , be an open set of the Borel σ algebra B . For any i = 1 , , n define f X i F i 1 ( · ) as the conditional density of X i given the σ field F i 1 . Define
F X i = P X i B = P X B , see   ( 11 ) ,
and
F X i F i 1 = P X i B | F i 1
as the distribution function and the conditional distribution function, given the σ field F i 1 , respectively. The following assumptions will be needed throughout the paper.
(C.0) 
There is a non-negative measurable function f F i 1 such that
P X F i 1 ( B ) = B f F i 1 ( x ) ν ( d x ) , B B .
(C.1) 
For any x S
lim n 1 n i = 1 n f F i 1 ( x ) = f ( x ) , in   the a . s . and L 2 sense .
At this point, we may refer to [56] for further details.
Theorem 1.
Under the conditions (C.0), (C.1), (F.1), (E.1)and (E.2), and (16) for any θ ( 0 , 1 ) ,
f B θ / 2 ( H ) W 2 ( 1 θ ) ( H ) ,
there exists a constant C 1 > 0 in such a way that
E f ^ ( x ) f ( x ) 2 C 1 ln n n θ ,
for large enough n.
A direct consequence is the following upper bound result: for s > 0 , if
f B s / ( 2 s + 1 ) ( H ) W 2 / ( 2 s + 1 ) ( H ) ,
then there exists a constant C 2 > 0 such that
E f ^ f 2 C 2 ln n n 2 s / ( 2 s + 1 ) .
This rate of convergence corresponds to the near optimal one in the “standard” minimax setting (see, e.g., [27]). Moreover, applying [[49], Theorem 3.2], one can see that B θ / 2 ( H ) W 2 ( 1 θ ) ( H ) is the “maxiset” associated to f ^ ( · ) at the rate of convergence ( ln n / n ) θ , i.e.,
lim n n ln n θ E f ^ f 2 < f B θ / 2 ( H ) W 2 ( 1 θ ) ( H ) .

4. Problem Definition of the Regression Estimation

For a measurable function ρ : R q R , we define the regression function m ( · , ρ ) by
ρ ( Y ) = m ( X , ρ ) + ϵ ,
where ϵ is a random variable independent of X with N ( 0 , 1 ) . We suppose that m ( · , ρ ) H , where H is a separable Hilbert space of real or complex-valued functions defined on S and square integrable with respect to the σ -finite measure ν . We shall suppose that there exist a known constant and C m > 0 such that
sup x S m ( x , ρ ) C m .
In this context, we redefine the probability measure P X in (11) and suppose that f ( · ) is a non-negative measurable known function.
(M.1) 
We shall suppose that there exist two known constant C m > 0 such that
sup x S m ( x ; ρ ) C m .
(M.2) 
We shall suppose that there exist two known constant c f > 0 such that
inf x S f ( x ) c f .

Regression Function Estimator

In this context, we aim to estimate m ( · , ρ ) based on n observed functional data ( X 1 , Y 1 ) , , ( X n , Y n ) . The kernel estimator of the regression function for functional data was proposed by [57]
m ^ n ; h n ( x , ρ ) : = i = 1 n ρ ( Y i ) K ( d ( x , X i ) / h n ) i = 1 n K ( d ( x , X i ) / h n ) .
As combined with the work of [55], we define the wavelet hard thresholding estimator m ^ ( · , ρ ) , for all x S , by
m ^ ( x , ρ ) = I 0 η ^ 0 , ϕ 0 x ; ζ 0 , + k = 0 m n J k θ ^ k , 𝟙 | θ ^ k , | κ ln n n ψ k x ; η k ; ,
where
η ^ k , = 1 n i = 1 n ρ ( Y i ) f ( X i ) ϕ k X i ; ζ k , ,
θ ^ k , = 1 n i = 1 n ρ ( Y i ) f ( X i ) ψ k X i ; η k , .
where κ is large enough constant and m n is the integer satisfying
1 2 n ( ln n ) 2 | J m n | n ( ln n ) 2 .
Theorem 2.
Under the conditions (E.1), (E.2) (M.1)(M.2), (C.0)and (C.1), combined with the assumption (24), for any θ ( 0 , 1 ) , m ( · , ρ ) B θ / 2 ( H ) W 2 ( 1 θ ) ( H ) , there exists a constant C 3 > 0 in such a way that
E m ^ ( · , ρ ) m ( · ; ρ ) 2 C ln n n θ ,
for large enough n.
Suppose that m ( · , ρ ) and f satisfy (M.1) and, for any θ ( 0 , 1 ) , m ( · , ρ ) B θ / 2 ( H ) W 2 ( 1 θ ) ( H ) , where B θ / 2 ( H ) is in (Definition 1) with s = θ / 2 and W 2 ( 1 θ ) ( H ) is in (Definition 2) with r = 2 ( 1 θ ) . Then, there exists a constant C 3 > 0 such that
E m ^ ( · , ρ ) m ( · ; ρ ) 2 C 3 ( ln n ) 2 n θ
for n large enough. Again, note that, for s > 0 , if
m ( · , ρ ) B s / ( 2 s + 1 ) ( H ) W 2 / ( 2 s + 1 ) ( H ) ,
then there exists a constant C 4 > 0 such that
E m ^ ( · , ρ ) m ( · ; ρ ) 2 C 4 ( ln n ) 2 n 2 s / ( 2 s + 1 ) .
Up to an additional logarithmic term, this rate of convergence corresponds to the near-optimal one in the “standard” minimax setting (see, for example, [27]). Theorem 2 is the first to investigate an adaptive wavelet-based estimator for functional data in the context of nonparametric regression for ergodic processes.
Since the coefficients defined by (22) and (23) depend on the unknown function f ( · ) , one can use
η ˜ k , = 1 n i = 1 n ρ ( Y i ) f ^ ( X i ) ϕ k X i ; ζ k , ,
θ ˜ k , = 1 n i = 1 n ρ ( Y i ) f ^ ( X i ) ψ k X i ; η k , .
We define the wavelet hard thresholding estimator m ˜ ( · , ρ ) , for all x S , by
m ˜ ( x , ρ ) = I 0 η ˜ 0 , ϕ 0 x ; ζ 0 , + k = 0 m n J k θ ˜ k , 𝟙 | θ ˜ k , | κ ln n n ψ k x ; η k ; .
Recall the following elementary observation
1 f ^ ( · ) = 1 f ( · ) + ( f ( · ) f ^ ( · ) ) f ( · ) f ^ ( · ) .
From the last equation, we can infer the following:
η ˜ k , = η ^ k , + 1 n i = 1 n ( f ( X i ) f ^ ( X i ) ) f ( X i ) f ^ ( X i ) ρ ( Y i ) ϕ k X i ; ζ k , , θ ˜ k , = θ ^ k , + 1 n i = 1 n ( f ( X i ) f ^ ( X i ) ) f ( X i ) f ^ ( X i ) ρ ( Y i ) ψ k X i ; η k , .
By combining Theorem 1 with Theorem 2, we obtain the following corollary.
Corollary 1.
Under the conditions of Theorems 1 and 2, there exists a constant C 5 > 0 such that
E m ˜ ( x ; ρ ) m ( x ; ρ ) 2 C 5 ln n n θ ,
for large enough n.
Remark 2.
In our previous paper [40], we were concerned with the nonparametric estimation of the density and the regression function in a finite-dimensional setting using orthonormal wavelet bases. Our findings differ significantly from those presented in the present paper. In [40], we provided the strong uniform consistency properties with rates of these estimators, over compact subsets of R d , under a general ergodic condition on the underlying processes. We also establish the asymptotic normality of wavelet-based estimators. We used the Burkholder–Rosenthal inequality as the main ingredient in this paper, which is a more complicated tool than the exponential inequality used in the previous paper. More importantly, in the present paper, we look into the mean integrated square error over compact subsets, which is entirely different from the results of the previous paper.

5. Applications

5.1. The Conditional Distribution

Our result can be used to investigate the conditional distribution F ( y x ) for y R d . To be more precise, let ρ ( y ) = 𝟙 { y t } . We define the wavelet hard thresholding estimator m ˘ ( · , ρ ) , for all x S , by
F ^ ( y x ) = I 0 η ˘ 0 , ϕ 0 x ; ζ 0 , + k = 0 m n J k θ ˘ k , 𝟙 | θ ˘ k , | κ ln n n ψ k x ; η k ; ,
where
η ˘ k , = 1 n i = 1 n 𝟙 { Y i t } f ( X i ) ϕ k X i ; ζ k , ,
θ ˘ k , = 1 n i = 1 n 𝟙 { Y i t } f ( X i ) ψ k X i ; η k , .
A direct consequence of Theorem 2 is
E F ^ ( y x ) F ( y x ) 2 C ln n n θ .

5.2. The Curve Discrimination

We can state the curve discrimination problem in the following way. Let X i i = 1 , , n be a sample of curves, and each of them is known to belong to one among G groups ι = 1 , , G . Let us denote by T i the group of the curve X i . Assume that each pair of variables X i , T i has the same distribution as the pair ( X , T ) . Given a new curve x , the question is to know its class membership, and for that, we will estimate, for any ι = 1 , , G , the conditional probability:
p ι ( x ) = P ( T = ι X = x ) .
Following the idea proposed in [58,59] permitting the estimation of these probabilities by
p ^ ι ( x ) = I 0 η ˘ ˘ 0 , ϕ 0 x ; ζ 0 , + k = 0 m n J k θ ˘ ˘ k , 𝟙 | θ ˘ ˘ k , | κ ln n n ψ k x ; η k ; ,
where
η ˘ ˘ k , = 1 n i = 1 n 𝟙 { T i = ι } f ( X i ) ϕ k X i ; ζ k , ,
θ ˘ ˘ k , = 1 n i = 1 n 𝟙 { T i = ι } f ( X i ) ψ k X i ; η k , .
A direct consequence of Theorem 2 is
E F ^ ( y x ) F ( y x ) 2 C ln n n θ .
As remarked by [58,59], for each ι , we make use of the notation
Y = 1 if T = ι 0 otherwise ,
then, we can write
p ι ( x ) = E ( Y X = x ) .
An application of Theorem 2 is
E p ^ ι ( x ) p ι ( x 2 C ln n n θ .

5.3. Time Series Prediction from Continuous Set of Past Values

A direct consequence of our results is the prediction of future values of some real time series, which we will follow from [59]. One of the purposes of the proposed functional method is to predict the future from a continuous set of past values of the process. Let { Z ( t ) } t R denote a real-valued process and s denote a fixed non-negative real number. We are interested in the prediction’s problem of a future value Z ( τ ) given some past values Z ( t ) for τ T t < τ , at some time τ > 0 . Our goal can be seen as the estimation of the operator r:
Z ( τ + s ) = r ( Z ( t ) for τ T t < τ ) + ε ,
whenever it exists. Let us describe the model. Suppose that the process has been observed from t = 0 until t = t max and without loss of generality, assume that
t max = n T + s < τ .
The methodology consists of splitting the observed process into n pieces of fixed length. Each piece of the process is denoted by
X i = { Z ( t ) , ( i 1 ) T t < i T } .
Let us denote the response value Y i = Z ( i T + s ) . This can be formulated by a regression problem in the following way:
Y i = m X i , I d + ε i for i = 1 , , n ,
where I d denotes the identity function. To fit the theoretical setting of the present paper, we assume that such a function m · , I d does not depend on i; this is the case for the stationary processes. Hence, at time τ , we can use for predicting the value at time τ + s the following predictor, which is directly derived from (21):
m ^ ( x , I d ) = I 0 η ^ 0 , ϕ 0 x ; ζ 0 , + k = 0 m n J k θ ^ k , 𝟙 | θ ^ k , | κ ln n n ψ k x ; η k ; ,
where
η ^ k , = 1 n i = 1 n Z ( i T + s ) f ( X i ) ϕ k X i ; ζ k , ,
θ ^ k , = 1 n i = 1 n Z ( i T + s ) f ( X i ) ψ k X i ; η k , ,
where x = { Z ( t ) , for τ T t < τ } . Our Theorem 2 gives mathematical support to this nonparametric functional predictor, and provides a different way of solving the prediction problem investigated in [3,59].

6. Concluding Remarks

In this work, we have investigated the nonparametric estimation of the density and the regression function based on the functional stationary processes, using wavelet bases for Hilbert spaces of functions. We have characterized the mean integrated square error over compact subsets. The asymptotic properties of these estimators are obtained employing the Martingale approach, which is completely different from the mixing and the independent setting. The assumption on the dependence of the process is ergodicity. To motivate the present paper, we have presented the conditional distribution, the curve discrimination and the time series prediction from a continuous set of past values. Extending the nonparametric functional ideas to the local stationary process is a somewhat underdeveloped field. It would be interesting to extend our work to the case of the functional local stationary process, which requires nontrivial mathematics; this would go well beyond the scope of the present paper.

7. Proofs

7.1. Proof of Theorem 1

In this paper, we need an upper bound inequality for partial sums of unbounded martingale differences that we use to derive the asymptotic results for the density and the regression functions estimates built upon functional strictly stationary and ergodic data. Here and in the sequel, we denote by “C” a positive constant that may be different from line to line. This inequality is given in the following lemmas. This lemma is stated following Notation 1 in [60].
Lemma 1
(Burkholder-Rosenthal inequality). Let ( X i ) i 1 be a stationary Martingale adapted to the filtration ( F i ) i 1 , define ( d i ) i 1 is the sequence of Martingale differences adapted to ( F i ) i 1 and
S n = i = 1 n d i .
Then, for any positive integer n,
max 1 j n | S j | ) p n 1 / p d 1 p + k = 1 n E ( d k 2 / F k 1 ) p / 2 1 / 2 , f o r   a n y p 2 ;
where, as usual, the norm · p = E [ | · | p ] 1 / p .
Lemma 2
([61]). Let { Z i , i 1 } be a sequence of Martingale differences such that
| Z i | B , a . s . ,
then, for all ϵ > 0 and all sufficiently large n, we have
P i = 1 n Z i > ϵ 2 exp ϵ 2 2 n B 2 .
The following lemmas describe the asymptotic behavior of the estimators α ^ k , and β ^ k , .
Lemma 3.
For any k { 0 , , m n } and any I k , under assumptions (C.0), (C.1), (F.1)and (E.1)(i), there exists a constant C > 0 such that
E α ^ k , α k , 2 C ln n n .
Lemma 4.
For any k { 0 , , m n } and any J k , and under assumptions (C.0), (C.1), (F.1), (E.1)and (E.2), and condition (16), there exists a constant C > 0 such that
E β ^ k , β k , 4 = C ln n n 2 , a . s .
Lemma 5.
For any k { 0 , , m n } and any J k , for κ > 0 large enough and under assumptions (C.0), (C.1), (F.1), (E.1)and (E.2), and condition (16), there exists a constant C > 0 such that
P β ^ k , β k , κ 2 ln n n C ln n n 2 .

7.1.1. Proof of Theorem 1

Observe that the proof of Theorem 1 is a direct application of ([49], Theorem 3.1) with c ( n ) = ( ln n / n ) 1 / 2 , σ i = 1 , r = 2 and the Lemmas 3–5. We adapted and extended the method of demonstration of [33], Theorem 3.1 to the stationary ergodic process. □

7.1.2. Proof of Lemmas

Proof of Lemma 3

Consider the following decomposition
α ^ k , α k , = α ^ k , α ˜ k , + α ˜ k , α k , = A k , , 1 + A k , , 2 ,
where
α ˜ k , = 1 n i = 1 n E ϕ k x i ; ζ k , | F i 1 .
Under the assumptions (C.0) and (C.2), we have
α ˜ k , = 1 n i = 1 n S ϕ k x ; ζ k , f F i 1 ( x ) ν ( d x ) = S ϕ k x ; ζ k , 1 n i = 1 n f F i 1 ( x ) ν ( d x ) = S ϕ k x ; ζ k , f ( x ) + o ( 1 ) ν ( d x ) = S ϕ k x ; ζ k , f ( x ) ν ( d x ) + o ( 1 ) = α k , + o ( 1 ) .
We readily obtain
α ˜ k , = α k , , as , n ,
implying that
A k , , 2 = o ( 1 ) , a . s .
Therefore, we infer that
α ^ k , α k , = A k , , 1 + o ( 1 ) , a . s .
Let us now consider the term A k , , 1 . We have
A k , , 1 = α ^ k , α ˜ k , = 1 n i = 1 n ϕ k x i ; ζ k , E ϕ k x i ; ζ k , | F i 1 = 1 n i = 1 n Φ k x i ; ζ k , .
Notice that ( Φ k x i ; ζ k , ) 0 k m n is a sequence of Martingale differences with respect to the sequence of σ fields ( F i ) 0 k m n . It is obvious, by Lemma 1 inequality, to see that
E | A k , , 1 | 2 = 1 n 2 E i = 1 n Φ k x i ; ζ k , 2 .
Applying the Burkholder–Rosenthal inequality (1), for p = 2 , we obtain
E i = 1 n Φ k x i ; ζ k , 2 1 2 n 1 / 2 Φ k x 1 ; ζ k , 2 + i = 1 n E Φ k 2 x i ; ζ k , | F i 1 1 1 / 2 = Φ ( 1 ) + Φ ( 2 )
On one hand, using a very famous decomposition combined with the fact that F 0 is the trivial σ field, we obtain
1 n Φ ( 1 ) 2 = Φ k x 1 ; ζ k , 2 2 = E ϕ k ( x 1 ; ζ k , ) E ϕ k x 1 ; ζ k , | F 0 2 E j = 0 2 ϕ k x 1 ; ζ k , j E ϕ k x 1 ; ζ k , 2 j = j = 0 2 C 2 j E ϕ k x 1 ; ζ k , j . E ϕ k x 1 ; ζ k , 2 j = C 2 2 E ϕ k x 1 ; ζ k , 2 + C 2 1 E ϕ k x 1 ; ζ k , 2 + C 2 0 E ϕ k x 1 ; ζ k , 2 ,
By the Cauchy–Schwarz inequality, together with assumptions (E.1)(i), (E.2) and the condition (16), we obtain
sup x S ϕ k x i ; ζ k , sup x S j J k 1 g j , k , | e j ( ζ k , ) | | e j ( x ) | j J k 1 g j , k , | e j ( ζ k , ) | 2 1 / 2 sup x S j J k | e j ( x ) | 2 1 / 2 C 1 1 / 2 C 2 1 / 2 | J k | C 3 | J m n | C 3 n ln n .
Observe that, under the assumptions (F.1) and (E.1)(i) and the fact that E is an orthonormal basis of H, we have
E ϕ k x 1 ; ζ k , 2 = S ϕ k x ; ζ k , 2 f ( x ) ν ( d x ) C f S ϕ k x ; ζ k , 2 ν ( d x ) = C f S j I k 1 g j , k , e j ( ζ k , ) e j ( x ) 2 ν ( d x ) = C f S j I k 1 g j , k , e j ( ζ k , ) 2 ν ( d x ) C f C 1 .
where C 1 is a positive constant,
E ϕ k x 1 ; ζ k , = O n ln n ,
E ϕ k x 1 ; ζ k , 2 = O ( 1 )
therefore,
Φ ( 1 ) = O ( n 1 / 2 ) .
On the other hand, we consider the second term of decomposition (48), observe that
Φ 2 = E i = 1 n E Φ k 2 x i ; ζ k , | F i 1 1 / 2 = i = 1 n E E Φ k 2 x i ; ζ k , | F i 1 1 / 2 = i = 1 n E Φ k 2 x i ; ζ k , 1 / 2
using the notable identity, we obtain
E Φ k 2 x i ; ζ k , = E ϕ k x i ; ζ k , E ϕ k x i ; ζ k , | F i 1 2 E ϕ k x i ; ζ k , 2 + 2 ϕ k x i ; ζ k , E ϕ k x i ; ζ k , | F i 1 + E ϕ k x i ; ζ k , 2 | F i 1 2 E ϕ k x i ; ζ k , 2 + 2 E E ϕ k x i ; ζ k , 2 | F i 1 4 E ϕ k x i ; ζ k , 2
observe that, using (50) and (51), we obtain
Φ 2 = O ( n 1 / 2 ) .
therefore, we combine (54) and (55) to obtain
E i = 1 n Φ k x i ; ζ k , 2 1 / 2 = O ( n 1 / 2 ) .
Hence,
E | A k , , 1 | 2 = 1 n 2 E i = 1 n Φ k x i ; ζ k , 2 = 1 n 2 O n C ln n n
Therefore, there exists a constant C = C f C 1 > 0 , such that
E α ^ k , α k , 2 4 C n 4 C ln n n .
Hence, the proof is complete. □

Proof of Lemma 4

Consider the following decomposition
β ^ k , β k , = β ^ k , β ˜ k , + β ˜ k , β k , = B k , , 1 + B k , , 2 ,
where
β ˜ k , = 1 n i = 1 n E ψ k x i ; η k , | F i 1 .
Observe that, under the assumptions (F.1) and (E.1)(i) and the fact that E is an orthonormal basis of H, and proceeding in a similar way as in (46), we show that
β ˜ k , = β k , , as , n .
This, in turn, implies that
B k , , 2 = o ( 1 ) , a . s .
Therefore, we obtain
β ^ k , β k , = B k , , 1 + o ( 1 ) , a . s .
Hence, we readily infer
E β ^ k , β k , 4 = 1 n 4 E i = 1 n Ψ i , k , 4 ,
where
Ψ i , k , = ψ k x i ; η k , E ψ k x i ; η k , | F i 1 .
Notice that ( Ψ i , k , ) 0 k n is a sequence of Martingale differences with respect to the sequence of σ fields ( F i ) 0 k n , applying the Burkholder–Rosenthal inequality for p = 4 (see Lemma 1), we obtain
E i = 1 n Ψ i , k , 4 1 / 4 max 1 j n i = 1 j Ψ i , k , 4 n 1 / 4 Ψ 1 , k , 4 + i = 1 n E ( Ψ i , k , 2 | F i 1 ) 4 / 2 1 / 2 = Ψ k , ( 1 ) + Ψ k , ( 2 ) .
Consider the first term of Equation (61). We have
1 n Ψ k , ( 1 ) 4 = Ψ 1 , k , 4 4 = E ψ k x 1 ; η k , E ψ k x 1 ; η k , | F 0 4 E ψ k x 1 ; η k , + E ψ k x 1 ; η k , 4 .
Using the classical identity
( a + b ) n = k = 0 n C n k a k b n k
in connection with the Jensen inequality and taking n = 4 , we obtain
ψ k x 1 ; η k , + E ψ k x 1 ; η k , 4 = k = 0 4 C 4 k ψ k x 1 ; η k , k E ψ k x 1 ; η k , 4 k k = 0 4 C 4 k ψ k x 1 ; η k , k E ψ k x 1 ; η k , 4 k .
This gives that
1 n Ψ k , ( 1 ) 4 E k = 0 4 C 4 k ψ k x 1 ; η k , k E ψ k x 1 ; η k , 4 k = k = 0 4 C 4 k E ψ k x 1 ; η k , k E ψ k x 1 ; η k , 4 k .
By proceeding in a similar way as in (51) and making use of the assumptions (F.1) and (E.1)(i), we infer that
E ψ k x 1 ; η k , 2 C ,
where C is a positive constant. Moreover, by the Cauchy–Schwarz inequality together with assumptions (E.1)(ii), (E.2) and the condition (16), we obtain
sup x S ψ k x ; η k , sup x S j J k 1 h j , k , | e j ( η k , ) | | e j ( x ) | j J k 1 h j , k , | e j ( η k , ) | 2 1 / 2 sup x S j J k | e j ( x ) | 2 1 / 2 C 1 1 / 2 C 2 1 / 2 | J k | C 4 | J m n | C 4 n ln n .
We then obtain
E ψ k x 1 ; η k , = O n ln n ,
E ψ k x 1 ; η k , 2 = O ( 1 ) ,
E ψ k x 1 ; η k , 3 = O n ln n ,
E ψ k x 1 ; η k , 4 = O n ln n .
Observe that the largest term is (68); now, using (68) in Equation (62), we deduce that
1 n Ψ k , ( 1 ) 4 = O n ln n .
This implies that
Ψ k , ( 1 ) = O n 1 / 2 ln n 1 / 4 .
Let us now investigate the upper bound of Ψ k , ( 2 ) in (61). Observe that
Ψ k , ( 2 ) = i = 1 n E ( Ψ i , k , 2 / F i 1 ) 2 1 / 2 = E i = 1 n E Ψ i , k , 2 | F i 1 2 1 / 4 ,
for all i = 1 , , n . Making use of the Jensen inequality with the fact that ( a b ) 2 = a 2 2 a b + b 2 , it follows that
i = 1 n E Ψ i , k , 2 | F i 1 = i = 1 n E ψ k x i ; η k , E ψ k x i ; η k , | F i 1 2 | F i 1 4 i = 1 n E ψ k x i ; η k , 2 | F i 1 .
Observe, under the assumptions (F.1), (E.1)(i), (C.0) and (C.1) and (63):
i = 1 n E ψ k x i ; η k , 2 | F i 1 = n S ψ k x ; η k , 2 1 n i = 1 n f F i 1 ( x ) ν ( d x ) = n S ψ k x ; η k , 2 f ( x ) + o ( 1 ) ν ( d x ) n C f + o ( 1 ) S ψ k x ; ζ k , 2 ν ( d x ) n C f C 1 .
It follows that
Ψ k , ( 2 ) = O n 1 / 2 .
Combining (61), (69) and (71), we obtain
E i = 1 n Ψ i , k , 4 = O n 2 ln n + O n 2 .
We conclude that
E β ^ k , β k , 4 = O 1 n 2 ln n + O 1 n 2 .
This implies that there exists a constant C > 0 , such that
E β ^ k , β k , 4 C ln n n 2 .
The proof is achieved. □

Proof of Lemma 5

Consider the previous decomposition in Lemma 4, to write that
β ^ k , β k , = β ^ k , β ˜ k , + β ˜ k , β k , = B k , , 1 + B k , , 2 ,
where
B k , , 1 = 1 n i = 1 n Ψ i , k , = 1 n i = 1 n ψ k x i ; η k , E ψ k x i ; η k , | F i 1 , B k , , 2 = 1 n i = 1 n E ψ k x i ; η k , | F i 1 β k , .
By using (59), we obtain the desired result for the term B k , , 2
β ^ k , β k , = B k , , 1 + o ( 1 ) .
Now, observe that
P 1 n i = 1 n Ψ i , k , κ 2 ln n n = P i = 1 n Ψ i , k , κ 2 n ln n .
An application of Lemma 2 implies that
Ψ i , k , = ψ k x i ; η k , E ψ k x i ; η k , | F i 1 2 sup x S ψ k x ; η k , C 3 n ln n C 3 n .
Let B = C 3 n . Then, for all ϵ n = κ 2 n ln n where n is sufficiently large, we have
P i = 1 n Ψ i , k , κ 2 n ln n 2 exp ϵ n 2 2 n B 2 = 2 exp κ 2 n ln n 2 2 n C 3 n 2 = 2 exp κ 2 ln n 8 C 3 2 n = 2 exp ln n κ 2 8 C 3 2 n = 2 n w ( κ , n ) ,
where
w ( κ , n ) = κ 2 8 C 3 2 n .
By choosing κ such that w ( n , κ ) = 2 , we have
P β ^ k , β k , κ 2 ln n n C 1 n 2 + o ( 1 ) C ln n n 2 .
The proof of (44) is achieved. □

7.1.3. Proof of Theorem 2

Recall that
m ^ ( x , ρ ) = I 0 η ^ 0 , ϕ 0 x ; ζ 0 , + k = 0 m n J k θ ^ k , 𝟙 | θ ^ k , | κ ln n n ψ k x ; η k ; ,
where
η ^ k , = 1 n i = 1 n ρ ( Y i ) f ( X i ) ϕ k X i ; ζ k , , θ ^ k , = 1 n i = 1 n ρ ( Y i ) f ( X i ) ψ k X i ; η k , .
Lemma 6.
For any k { 0 , , m n } and any I k and under the assumptions (E.1)(i), (M.1)- (M.2), (C.0)and (C.1), there exists a constant C > 0 such that
E η ^ k , η k , 2 C ln n n .
Lemma 7.
For any k { 0 , , m n } and any J k , and under the assumptions (E.1), (E.2)(M.1)-(M.2), (C.0)and (C.1), combined with the condition (24), there exists a constant C > 0 such that
E θ ^ k , θ k , 4 = C ln n n 2 , a . s .
Lemma 8.
For any k { 0 , , m n } and any J k , for κ > 0 large enough, (E.1), (E.2) (M.1)(M.2), (C.0)and (C.1), combined with the condition (24), there exists a constant C > 0 such that
P θ ^ k , θ k , κ 2 ln n n C ln n n 2 .

7.1.4. Proof of Theorem 2

Observe that the proof of Theorem 2 is a direct application of ([49], Theorem 3.1) with c ( n ) = ( ln n / n ) 1 / 2 , σ i = 1 , r = 2 and the following Lemmas 6–8. We extended the method of the proof in [33], Theorem 4.1. □

7.1.5. Proof of Lemmas

Proof of Lemma 6

Consider the following decomposition
η ^ k , η k , = η ^ k , η ˜ k , + η ˜ k , η k , = A k , , 1 + A k , , 2 ,
where
η ˜ k , = 1 n i = 1 n E ρ ( Y i ) f ( X i ) ϕ k X i ; ζ k , | F i 1 = 1 n i = 1 n E m ( X i , ρ ) + ϵ i f ( X i ) ϕ k X i ; ζ k , | F i 1 = 1 n i = 1 n E m ( X i , ρ ) f ( X i ) ϕ k X i ; ζ k , | F i 1 + 1 n i = 1 n E ϵ i f ( X i ) ϕ k X i ; ζ k , | F i 1 .
From the independence between ϵ i and X i , we have
E ϵ i | G i 1 = E ϵ i | X i = E ϵ i = 0 .
Observe that
E ϵ f ( X i ) ϕ k X i ; ζ k , | F i 1 = E E ϵ | G i 1 f ( X i ) ϕ k X i ; ζ k , | F i 1 = E E ϵ | X i f ( X i ) ϕ k X i ; ζ k , | F i 1 = E E ϵ f ( X i ) ϕ k X i ; ζ k , | F i 1 = 0 .
This implies that
η ˜ k , = 1 n i = 1 n E ρ ( Y i ) f ( X i ) ϕ k X i ; ζ k , | F i 1 = 1 n i = 1 n E m ( X i , ρ ) f ( X i ) ϕ k X i ; ζ k , | F i 1 .
Making use of the assumptions (M.1)(M.2), (C.0) and (C.1), we have
η ˜ k , = 1 n i = 1 n S m ( x , ρ ) f ( x ) ϕ k x ; ζ k , f F i 1 ( x ) ν ( d x ) = S m ( x , ρ ) f ( x ) ϕ k x ; ζ k , 1 n i = 1 n f F i 1 ( x ) ν ( d x ) = S m ( x , ρ ) f ( x ) ϕ k x ; ζ k , f ( x ) + o ( 1 ) ν ( d x ) = S m ( x , ρ ) f ( x ) ϕ k x ; ζ k , f ( x ) ν ( d x ) + o ( 1 ) = η k , + o ( 1 ) .
We readily obtain that
η ˜ k , = η k , , as , n .
implying that
A k , , 2 = o ( 1 ) , a . s .
Therefore,
η ^ k , η k , = A k , , 1 + o ( 1 ) , a . s .
Let us now turn our attention to the term A k , , 1 in (80), we have
A k , , 1 = η ^ k , η ˜ k , = 1 n i = 1 n ρ ( Y i ) f ( X i ) ϕ k X i ; ζ k , E ρ ( Y i ) f ( X i ) ϕ k X i ; ζ k , | F i 1 = 1 n i = 1 n Φ k X i ; ζ k , .
Notice that ( Φ k x i ; ζ k , ) 0 k m n is a sequence of Martingale differences with respect to the sequence of σ fields ( F i ) 0 k m n . It is obvious, proceeding as the proof of (56), that
E | A k , , 1 | 2 = 1 n 2 E i = 1 n Φ k X i ; ζ k , 2
where
E i = 1 n Φ k x i ; ζ k , 2 1 2 n 1 / 2 Φ k x 1 ; ζ k , 2 + i = 1 n E Φ k 2 x i ; ζ k , | F i 1 1 1 / 2 = Φ ( 1 ) + Φ ( 2 )
On the one hand, using a very famous decomposition combined with the fact that F 0 is the trivial σ field, we obtain
1 n Φ ( 1 ) 2 = Φ k x 1 ; ζ k , 2 2 = E ρ ( Y i ) f ( X i ) ϕ k ( x 1 ; ζ k , ) E ρ ( Y 1 ) f ( X 1 ) ϕ k x 1 ; ζ k , | F 0 2 E j = 0 2 ρ ( Y 1 ) f ( X 1 ) ϕ k x 1 ; ζ k , j E ρ ( Y 1 ) f ( X 1 ) ϕ k x 1 ; ζ k , 2 j = j = 0 2 C 2 j E ρ ( Y 1 ) f ( X 1 ) ϕ k x 1 ; ζ k , j . E ρ ( Y 1 ) f ( X 1 ) ϕ k x 1 ; ζ k , 2 j = C 2 2 E ρ ( Y 1 ) f ( X 1 ) ϕ k x 1 ; ζ k , 2 + C 2 1 E ρ ( Y 1 ) f ( X 1 ) ϕ k x 1 ; ζ k , 2 + C 2 0 E ρ ( Y 1 ) f ( X 1 ) ϕ k x 1 ; ζ k , 2 ,
It follows, from assumptions (M.1) and (M.2), that
| ρ ( Y i ) | C m + | ϵ i | ,
combined with the independence between X 1 and ϵ 1 , E [ ϵ 1 2 ] = 1 . Observe that, under assumption (E.1)(i) and the fact that E is an orthonormal basis of H, we have
E ρ ( Y 1 ) f ( X 1 ) ϕ k X 1 ; ζ k , 2 C m 2 + 1 c f E 1 f ( X 1 ) ϕ k x 1 ; ζ k , 2 C m 2 + 1 c f S 1 f ( x ) ϕ k x ; ζ k , 2 f ( x ) ν ( d x ) = C m 2 + 1 c f S j I k 1 g j , k , e j ( ζ k , ) e j ( x ) 2 ν ( d x ) = C m 2 + 1 c f S j I k 1 g j , k , e j ( ζ k , ) 2 ν ( d x ) C m 2 + 1 c f C 1 = O ( 1 ) .
Moreover, from Assumptions (M.1) and (M.2) and using (52), (85), combined with the independence between X 1 and ϵ 1 , E [ ϵ 1 ] = 0 . we have
E ρ ( Y 1 ) f ( X 1 ) ϕ k X 1 ; ζ k , = O n ln n .
Therefore,
Φ ( 1 ) = O ( n 1 / 2 ) .
On the other hand, we consider the second term of decomposition (84), and proceeding as in the proof of (55) and considering (86) and (87)
Φ 2 = O ( n 1 / 2 ) .
therefore, combining (88) and (89) to obtain
E i = 1 n Φ k X i ; ζ k , 2 1 / 2 = O ( n 1 / 2 ) .
Hence,
E | A k , , 1 | 2 = 1 n 2 E i = 1 n Φ k x i ; ζ k , 2 = 1 n 2 O n C ln n n
Therefore, combining (83) and (90), there exists a constant C such as
E α ^ k , α k , 2 C n C ln n n .
Hence, the proof is complete. □

Proof of Lemma 7

Consider the following decomposition
θ ^ k , θ k , = θ ^ k , θ ˘ k , + θ ˘ k , θ k , = B k , , 1 + B k , , 2 ,
where
θ ˘ k , = 1 n i = 1 n E ρ ( Y i ) f ( X i ) ψ k x i ; η k , | F i 1 .
Observe that, under the assumptions (M.1)(M.2) and (C.1), the Equation (81) and the fact that E is an orthonormal basis of H, by proceeding as in (82), we show that
θ ˘ k , = θ k , , as , n ,
implying that
B k , , 2 = o ( 1 ) , a . s .
Therefore,
θ ^ k , θ k , = B k , , 1 + o ( 1 ) , a . s .
Hence, we obtain
E θ ^ k , θ k , 4 = 1 n 4 E i = 1 n Ψ i , k , 4 ,
where
Ψ i , k , = Y i f ( x i ) ψ k x i ; η k , E Y i f ( x i ) ψ k x i ; η k , | F i 1 .
Notice that ( Ψ i , k , ) 0 k n is a sequence of Martingale differences with respect to the sequence of σ fields ( F i ) 0 k n , applying the Burkholder–Rosenthal inequality for p = 4 (see Lemma 1), we obtain
E i = 1 n Ψ i , k , 4 1 / 4 max 1 j n i = 1 j Ψ i , k , 4 n 1 / 4 Ψ 1 , k , 4 + i = 1 n E ( Ψ i , k , 2 | F i 1 ) 4 / 2 1 / 2 = Ψ k , ( 1 ) + Ψ k , ( 2 ) .
Consider the first term of Equation (96),
1 n Ψ k , ( 1 ) 4 = Ψ 1 , k , 4 4 = E ρ ( Y 1 ) f ( X 1 ) ψ k X 1 ; η k , E ρ ( Y 1 ) f ( X 1 ) ψ k X 1 ; η k , | F 0 4 E ρ ( Y 1 ) f ( X 1 ) ψ k X 1 ; η k , + E ρ ( Y 1 ) f ( X 1 ) ψ k X 1 ; η k , 4 .
By combining the identity
( a + b ) n = k = 0 n C n k a k b n k
and the Jensen inequality and taking n = 4 , we obtain
ρ ( Y 1 ) f ( x 1 ) ψ k X 1 ; η k , + E ρ ( Y 1 ) f ( X 1 ) ψ k X 1 ; η k , 4 = k = 0 4 C 4 k ρ ( Y 1 ) f ( X 1 ) ψ k X 1 ; η k , k E ρ ( Y 1 ) f ( X 1 ) ψ k X 1 ; η k , 4 k k = 0 4 C 4 k ρ ( Y 1 ) f ( X 1 ) ψ k X 1 ; η k , k E ρ ( Y 1 ) f ( X 1 ) ψ k X 1 ; η k , 4 k .
Then, we have
1 n Ψ k , ( 1 ) 4 E k = 0 4 C 4 k ρ ( Y 1 ) f ( X 1 ) ψ k X 1 ; η k , k E ρ ( Y 1 ) f ( X 1 ) ψ k X 1 ; η k , 4 k = k = 0 4 C 4 k E ρ ( Y 1 ) f ( X 1 ) ψ k X 1 ; η k , k E ρ ( Y 1 ) f ( X 1 ) ψ k X 1 ; η k , 4 k .
By following the same reasoning as in (86) and under the same assumptions (M.1)-(M.2) and (E.1)(i), we have
E ρ ( Y 1 ) f ( X 1 ) ψ k X 1 ; η k , 2 C ,
where C is a positive constant. Moreover, by the Cauchy–Schwarz inequality together with assumptions (E.1)(ii), (E.2) and condition (24), we obtain
sup x S ψ k x ; η k , sup x S j J k 1 h j , k , | e j ( η k , ) | | e j ( x ) | j J k 1 h j , k , | e j ( η k , ) | 2 1 / 2 sup x S j J k | e j ( x ) | 2 1 / 2 C 1 1 / 2 C 2 1 / 2 | J k | C 3 | J m n | C 3 n ln n 2 .
Hence, we infer that
E ρ ( Y 1 ) f ( X 1 ) ψ k X 1 ; η k , = O n ln n 2 ,
E ρ ( Y 1 ) f ( X 1 ) ψ k X 1 ; η k , 2 = O ( 1 ) ,
E ρ ( Y 1 ) f ( X 1 ) ψ k X 1 ; η k , 3 = O n ln n 2 ,
E ρ ( Y 1 ) f ( X 1 ) ψ k X 1 ; η k , 4 = O n ln n 2 .
Observe that the largest term is (103), now, using that same statement (103) in Equation (62), we deduce that
1 n Ψ k , ( 1 ) 4 C 4 E ψ k X 1 ; η k , 4 = O n ln n 2 .
It follows that
Ψ k , ( 1 ) = O n ln n 1 / 2 .
Let us now investigate the upper bound of Ψ k , ( 2 ) of (61). Observe that
Ψ k , ( 2 ) = i = 1 n E ( Ψ i , k , 2 / F i 1 ) 2 1 / 2 = E i = 1 n E Ψ i , k , 2 | F i 1 2 1 / 4 ,
for all i = 1 , , n , using the Jensen inequality and the fact that
( a b ) 2 = a 2 2 a b + b 2 ,
it follows
i = 1 n E Ψ i , k , 2 | F i 1 = i = 1 n E ρ ( Y i ) f ( X i ) ψ k x i ; η k , E ρ ( Y i ) f ( x i ) ψ k X i ; η k , | F i 1 2 | F i 1 4 i = 1 n E ρ ( Y i ) f ( X i ) ψ k X i ; η k , 2 | F i 1 .
From the independence between ϵ i and X i , we have
E ϵ 2 | G i 1 = E ϵ 2 | X i = E ϵ 2 = 1 .
Under assumptions (M.1)(M.2) and (E.1)(i) and (C.1), (63) and (105), we have
i = 1 n E ρ ( Y i ) f ( X i ) ψ k X i ; η k , 2 | F i 1 = i = 1 n E E ρ ( Y i ) f ( X i ) ψ k X i ; η k , 2 | G i 1 | F i 1 n C m 2 + 1 c f S ψ k x ; η k , 2 1 n i = 1 n f F i 1 ( x ) f ( x ) ν ( d x ) = n C m 2 + 1 c f ( 1 + o ( 1 ) ) S ψ k x ; η k , 2 ν ( d x ) n C ,
where
C = C m 2 + 1 c f ( 1 + o ( 1 ) ) .
It follows
Ψ k , ( 2 ) = C n 1 / 2 .
Combining (61), (69) and (71), we obtain
E i = 1 n Ψ i , k , 4 = O n ln n 2 + O n 2 ,
combining this with (95), we conclude
E θ ^ k , θ k , 4 = O 1 n 4 n ln n 2 + O 1 n 4 n 2 .
Hence, there exists a constant C > 0 , such that
E θ ^ k , θ k , 4 C 1 n 2 C ln n n 2 .
The proof is achieved. □

Proof of Lemma 8

Considering the previous decomposition (92) in Lemma 7, we have
θ ^ k , θ k , = θ ^ k , θ ˘ k , + θ ˘ k , θ k , = B k , , 1 + B k , , 2 ,
where
B k , , 1 = 1 n i = 1 n Ψ i , k , = 1 n i = 1 n ρ ( Y i ) f ( X i ) ψ k X i ; η k , E ρ ( Y i ) f ( X i ) ψ k X i ; η k , | F i 1 , B k , , 2 = 1 n i = 1 n E ρ ( Y i ) f ( X i ) ψ k X i ; η k , | F i 1 θ k , .
Statement (94) achieves the desired result for the term B k , , 2
θ ^ k , θ k , = B k , , 1 + o ( 1 ) .
We consider the next decomposition
Ψ i , k , = V i , k , + W i , k , ,
where
V i , k , = ρ ( Y i ) f ( X i ) ψ k x i ; η k , 𝟙 A i E ρ ( Y i ) f ( X i ) ψ k X i ; η k , 𝟙 A i | F i 1 ,
W i , k , = ρ ( Y i ) f ( X i ) ψ k X i ; η k , 𝟙 A i c E ρ ( Y i ) f ( X i ) ψ k X i ; η k , 𝟙 A i c | F i 1 ,
and
𝟙 A i = | ϵ i | c * ln n ,
and c * denotes a constant which will be chosen later. Now, observe that
P θ ^ k , θ ˘ k , κ 2 ln n n P B k , , 1 κ 2 ln n n + o ( 1 ) = P i = 1 n Ψ i , k , κ 2 n ln n + o ( 1 ) = I 1 + I 2 + o ( 1 ) ,
where
I 1 = P i = 1 n V i , k , κ 2 n ln n , I 2 = P i = 1 n W i , k , κ 2 n ln n .
First, we aim to bound the term I 1 of Equation (110). The Markov inequality and the Cauchy–Schwarz inequality yield
I 1 2 κ n ln n E i = 1 n V i , k , 2 κ n ln n i = 1 n E V i , k , .
Observe that
E V i , k , = E ρ ( Y i ) f ( X i ) ψ k X i ; η k , 𝟙 A i E ρ ( Y i ) f ( X i ) ψ k X i ; η k , 𝟙 A i | F i 1
E ρ ( Y i ) f ( X i ) ψ k X i ; η k , 𝟙 A i + E E ρ ( Y i ) f ( X i ) ψ k X i ; η k , 𝟙 A i | F i 1 2 E ρ ( Y i ) f ( X i ) ψ k X i ; η k , 𝟙 A i
2 E ρ ( Y i ) f ( X i ) ψ k X i ; η k , 2 1 / 2 P A i 1 / 2 .
We use (98) combined with an elementary Gaussian inequality and take c * to have
c * 2 4 1 / 2 = 2 .
We obtain
I 1 2 C κ n ln n exp c * 2 ln n 4 2 C κ n c * 2 4 1 / 2 ln n C κ 1 n 2 ,
where C κ = 2 C κ . We now intend to investigate an upper bound for I 2 of decomposition (110). We start by verifying the condition of Lemma 2. Suppose that assumptions (M.1) and (M.2) are satisfied combined with (99), we obtain
| Y i 𝟙 A i c | C m + c * ln n C ln n ,
which implies
W i , k , ρ ( Y i ) f ( X i ) ψ k X i ; η k , 𝟙 A i c E ρ ( Y i ) f ( X i ) ψ k X i ; η k , 𝟙 A i c 𝟙 A i c | F i 1 2 C ln n c f sup x S ψ k x ; η k , 2 C c f ln n n ln n 2 C 3 n ln n C 3 n ,
where C 3 = 2 C C f , let B = C 3 n , then, for all ϵ n = κ 2 n ln n sufficiently large n, we have
I 2 = P i = 1 n W i , k , κ 2 n ln n 2 exp ϵ n 2 2 n B 2 = 2 exp κ 2 n ln n 2 2 n C 3 n 2 = 2 exp κ 2 ln n 4 C 3 2 n = 2 exp ln n κ 2 4 C 3 2 n = 2 n w ( κ , n ) ,
where
w ( κ , n ) = κ 2 4 C 3 2 n .
Taking κ such that w ( n , κ ) = 2 , we have
I 2 C 1 n 2 .
It follows from (110), (114) and (118) that
P θ ^ k , θ ˜ k , κ 2 ln n n C 1 n 2 + o ( 1 )
C ln n n 2 .
The proof of (79) is achieved. □

Author Contributions

S.D., A.A.H. and S.B.: conceptualization, methodology, investigation, writing-original draft, writing—review and editing. All authors contributed equally to the writing of this paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the Special Issue Editor of the Special Issue on “Functional Data Analysis: Theory and Applications to Different Scenarios”, Mustapha Rachdi for the invitation. The authors are indebted to the Editor-in-Chief and the two referees for their very generous comments and suggestions on the first version of our article, which helped us to improve content, presentation and layout of the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bosq, D. Linear Processes in Function Spaces; Volume 149 of Lecture Notes in Statistics; Theory and applications; Springer: New York, NY, USA, 2000. [Google Scholar]
  2. Ramsay, J.; Silverman, B. Functional Data Analysis, 2nd ed.; Springer Series in Statistics; Springer: New York, NY, USA, 2005. [Google Scholar]
  3. Ferraty, F.; Vieu, P. Nonparametric Functional Data Analysis; Springer Series in Statistics; Springer: New York, NY, USA, 2006. [Google Scholar]
  4. Horváth, L.; Kokoszka, P. Inference for Functional Data with Applications; Springer Series in Statistics; Springer: New York, NY, USA, 2012. [Google Scholar]
  5. Zhang, J. Analysis of Variance for Functional Data; Volume 127 of Monographs on Statistics and Applied Probability; CRC Press: Boca Raton, FL, USA, 2014. [Google Scholar]
  6. Qing, S.; Choi, T. Gaussian Process Regression Analysis for Functional Data; CRC Press: Boca Raton, FL, USA, 2011. [Google Scholar]
  7. Geenens, G. Curse of dimensionality and related issues in nonparametric functional regression. Stat. Surv. 2011, 5, 30–43. [Google Scholar] [CrossRef]
  8. Cuevas, A. A partial overview of the theory of statistics with functional data. J. Statist. Plann. Inference 2014, 147, 1–23. [Google Scholar] [CrossRef]
  9. Shang, H. A survey of functional principal component analysis. AStA Adv. Stat. Anal. 2014, 98, 121–142. [Google Scholar] [CrossRef]
  10. Horváth, L.; Rice, G. An introduction to functional data analysis and a principal component approach for testing the equality of mean curves. Rev. Mat. Complut. 2015, 28, 505–548. [Google Scholar] [CrossRef]
  11. Müller, H. Peter hall, functional data analysis and random objects. Ann. Statist. 2016, 44, 1867–1887. [Google Scholar] [CrossRef]
  12. Nagy, S. An overview of consistency results for depth functionals. In Functional Statistics and Related Fields; Contrib. Stat.; Springer: Cham, Switzerland, 2017; pp. 189–196. [Google Scholar]
  13. Vieu, P. On dimension reduction models for functional data. Statist. Probab. Lett. 2018, 136, 134–138. [Google Scholar] [CrossRef]
  14. Aneiros-Pérez, G.; Cao, R.; Fraiman, R.; Genest, C.; Vieu, P. Recent advances in functional data analysis and high-dimensional statistics. J. Multivar. Anal. 2018, 170, 3–9. [Google Scholar] [CrossRef]
  15. Ling, N.; Vieu, P. Nonparametric modelling for functional data: Selected survey and tracks for future. Statistics 2018, 52, 934–949. [Google Scholar] [CrossRef]
  16. Goia, A.; Vieu, P. An introduction to recent advances in high/infinite dimensional statistics. J. Multivar. Anal. 2016, 146, 1–6. [Google Scholar] [CrossRef]
  17. Almanjahie, I.M.; Bouzebda, S.; ChikrElmezouar, Z.; Laksaci, A. The functional kNN estimator of the conditional expectile: Uniform consistency in number of neighbors. Stat. Risk Model. 2022, 38, 47–63. [Google Scholar] [CrossRef]
  18. Almanjahie, I.M.; Bouzebda, S.; Kaid, Z.; Laksaci, A. Nonparametric estimation of expectile regression in functional dependent data. J. Nonparametr. Stat. 2022, 34, 250–281. [Google Scholar] [CrossRef]
  19. Bouzebda, S.; Nezzal, A. Uniform consistency and uniform in number of neighbors consistency for nonparametric regression estimates and conditional U-statistics involving functional data. Jpn. J. Stat. Data Sci. 2022, 1–103. [Google Scholar] [CrossRef]
  20. Bouzebda, S.; Nemouchi, B. Weak-convergence of empirical conditional processes and conditional U-processes involving functional mixing data. Stat. Inference Stoch. Process. 2022, 1–56. [Google Scholar] [CrossRef]
  21. Bouzebda, S.; Nemouchi, B. Uniform consistency and uniform in bandwidth consistency for nonparametric regression estimates and conditional U-statistics involving functional data. J. Nonparametr. Stat. 2020, 32, 452–509. [Google Scholar] [CrossRef]
  22. Almanjahie, I.M.; Kaid, Z.; Laksaci, A.; Rachdi, M. Estimating the conditional density in scalar-on-function regression structure: k-NN local linear approach. Mathematics 2022, 10, 902. [Google Scholar] [CrossRef]
  23. Meyer, Y. Wavelets and operators. In Different Perspectives on Wavelets (San Antonio, TX, 1993); Volume 47 of Proc. Sympos. Appl. Math.; Amer. Math. Soc.: Providence, RI, USA, 1993; pp. 35–58. [Google Scholar]
  24. Daubechies, I. Ten Lectures on Wavelets; Volume 61 of CBMS-NSF Regional Conference Series in Applied Mathematics; Society for Industrial and Applied Mathematics (SIAM): Philadelphia, PA, USA, 1992. [Google Scholar]
  25. Mallat, S. A Wavelet Tour of Signal Processing, 3rd ed.; Elsevier/Academic Press: Amsterdam, The Netherlands, 2009. [Google Scholar]
  26. Vidakovic, B. Statistical Modeling by Wavelets; Wiley Series in Probability and Statistics: Applied Probability and Statistics; John Wiley & Sons Inc.: New York, NY, USA, 1999. [Google Scholar]
  27. Härdle, W.; Kerkyacharian, G.; Picard, D.; Tsybakov, A. Wavelets, Approximation, and Statistical Applications; Volume 129 of Lecture Notes in Statistics; Springer: New York, NY, USA, 1998. [Google Scholar]
  28. Rao, B.L.S.P. Nonparametric estimation of the derivatives of a density by the method of wavelets. Bull. Inform. Cybernet. 1996, 28, 91–100. [Google Scholar] [CrossRef]
  29. Chaubey, Y.P.; Doosti, H.; Rao, B.P. Wavelet based estimation of the derivatives of a density for a negatively associated process. J. Stat. Theory Pract. 2008, 2, 453–463. [Google Scholar] [CrossRef]
  30. Rao, B.L.S.P. Nonparametric Estimation of Partial Derivatives of a Multivariate Probability Density by the Method of Wavelets; De Gruyter: Berlin, Germany, 2018; pp. 321–330. [Google Scholar]
  31. Prakasa Rao, B.L.S. Wavelet estimation for derivative of a density in the presence of additive noise. Braz. J. Probab. Stat. 2018, 32, 834–850. [Google Scholar] [CrossRef]
  32. Allaoui, S.; Bouzebda, S.; Chesneau, C.; Liu, J. Uniform almost sure convergence and asymptotic distribution of the wavelet-based estimators of partial derivatives of multivariate density function under weak dependence. J. Nonparametr. Stat. 2021, 33, 170–196. [Google Scholar] [CrossRef]
  33. Chesneau, C.; Kachour, M.; Maillot, B. Nonparametric estimation for functional data by wavelet thresholding. REVSTAT 2013, 11, 211–230. [Google Scholar]
  34. Rosenblatt, M. Uniform ergodicity and strong mixing. Z. Wahrscheinlichkeitstheorie Verw. Gebiete 1972, 24, 79–84. [Google Scholar] [CrossRef]
  35. Bradley, R. Introduction to Strong Mixing Conditions; Kendrick Press: Heber City, UT, USA, 2007; Volume 1. [Google Scholar]
  36. Laïb, N.; Louani, D. Nonparametric kernel regression estimation for functional stationary ergodic data: Asymptotic properties. J. Multivariate Anal. 2010, 101, 2266–2281. [Google Scholar] [CrossRef]
  37. Bouzebda, S.; Didi, S.; ElHajj, L. Multivariate wavelet density and regression estimators for stationary and ergodic continuous time processes: Asymptotic results. Math. Methods Statist. 2015, 24, 163–199. [Google Scholar] [CrossRef]
  38. Delecroix, M.; Rosa, A.C. Nonparametric estimation of a regression function and its derivatives under an ergodic hypothesis. J. Nonparametr. Statist. 1996, 6, 367–382. [Google Scholar] [CrossRef]
  39. Bouzebda, S.; Didi, S. Additive regression model for stationary and ergodic continuous time processes. Comm. Statist. Theory Methods 2017, 46, 2454–2493. [Google Scholar] [CrossRef]
  40. Bouzebda, S.; Didi, S. Multivariate wavelet density and regression estimators for stationary and ergodic discrete time processes: Asymptotic results. Comm. Statist. Theory Methods 2017, 46, 1367–1406. [Google Scholar] [CrossRef]
  41. Bouzebda, S.; Didi, S. Some asymptotic properties of kernel regression estimators of the mode for stationary and ergodic continuous time processes. Rev. Mat. Complut. 2021, 34, 811–852. [Google Scholar] [CrossRef]
  42. Bouzebda, S.; Didi, S. Some results about kernel estimators for function derivatives based on stationary and ergodic continuous time processes with applications. Comm. Statist. Theory Methods 2022, 51, 3886–3933. [Google Scholar] [CrossRef]
  43. Bouzebda, S.; Chaouch, M. Uniform limit theorems for a class of conditional z-estimators when covariates are functions. J. Multivar. Anal. 2022, 189, 104872. [Google Scholar] [CrossRef]
  44. Bouzebda, S.; Chaouch, M.; Didi, S. Asymptotics for function derivatives estimators based on stationary and ergodic discrete time processes. Ann. Inst. Statist. Math. 2022, 74, 1–35. [Google Scholar] [CrossRef]
  45. Leucht, A.; Neumann, M. Degenerate U- and V-statistics under ergodicity: Asymptotics, bootstrap and applications in statistics. Ann. Inst. Statist. Math. 2013, 65, 349–386. [Google Scholar] [CrossRef] [Green Version]
  46. Neumann, M.H. Absolute regularity and ergodicity of Poisson count processes. Bernoulli 2011, 17, 1268–1284. [Google Scholar] [CrossRef]
  47. Goh, S.S. Wavelet bases for Hilbert spaces of functions. Complex Var. Elliptic Equ. 2007, 52, 245–260. [Google Scholar] [CrossRef]
  48. Kerkyacharian, G.; Picard, D. Density estimation in Besov spaces. Statist. Probab. Lett. 1992, 13, 15–24. [Google Scholar] [CrossRef]
  49. Kerkyacharian, G.; Picard, D. Thresholding algorithms, maxisets and well-concentrated bases. Test 2000, 9, 283–344. [Google Scholar] [CrossRef]
  50. Prakasa, Rao, B. L.S. Nonparametric density estimation for functional data via wavelets. Commun. Stat. Theory Methods 2010, 39, 1608–1618. [Google Scholar] [CrossRef]
  51. Cohen, A.; DeVore, R.; Kerkyacharian, G.; Picard, D. Maximal spaces with given rate of convergence for thresholding algorithms. Appl. Comput. Harmon. Anal. 2001, 11, 167–191. [Google Scholar] [CrossRef]
  52. DeVore, R.A. Nonlinear approximation. In Acta Numerica, 1998; Volume 7 of Acta Numer; Cambridge Univ. Press: Cambridge, UK, 1998; pp. 51–150. [Google Scholar]
  53. Autin, F. Point de vue Maxiset en Estimation Non Paramétrique. Ph.D. Thesis, Université Paris-Diderot-Paris VII, Paris, France, 2004. [Google Scholar]
  54. Gasser, T.; Hall, P.; Presnell, B. Nonparametric estimation of the mode of a distribution of random curves. J. R. Stat. Soc. Ser. B Stat. Methodol. 1998, 60, 681–691. [Google Scholar] [CrossRef]
  55. Donoho, D.L.; Johnstone, I.M.; Kerkyacharian, G.; Picard, D. Density estimation by wavelet thresholding. Ann. Statist. 1996, 24, 508–539. [Google Scholar] [CrossRef]
  56. Peškir, G. The uniform mean-square ergodic theorem for wide sense stationary processes. Stoch. Anal. Appl. 1998, 16, 697–720. [Google Scholar] [CrossRef]
  57. Ferraty, F.; Vieu, P. Nonparametric models for functional data, with application in regression, time-series prediction and curve discrimination. Nonparametric Stat. 2004, 16, 111–125. [Google Scholar] [CrossRef]
  58. Ferraty, F.; Vieu, P. The functional nonparametric model and application to spectrometric data. Comput. Statist. 2002, 17, 545–564. [Google Scholar] [CrossRef]
  59. Ouassou, I.; Rachdi, M. Regression operator estimation by delta-sequences method for functional data and its applications. AStA Adv. Stat. Anal. 2012, 96, 451–465. [Google Scholar] [CrossRef]
  60. Burkholder, D.L. Distribution function inequalities for martingales. Ann. Probab. 1973, 1, 19–42. [Google Scholar] [CrossRef]
  61. de la Peña, V.H.; Giné, E. Decoupling; Probability and Its Applications (New York); From dependence to independence, Randomly stopped processes. U-statistics and processes. Martingales and beyond; Springer: New York, NY, USA, 1999. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

DIDI, S.; AL HARBY, A.; BOUZEBDA, S. Wavelet Density and Regression Estimators for Functional Stationary and Ergodic Data: Discrete Time. Mathematics 2022, 10, 3433. https://doi.org/10.3390/math10193433

AMA Style

DIDI S, AL HARBY A, BOUZEBDA S. Wavelet Density and Regression Estimators for Functional Stationary and Ergodic Data: Discrete Time. Mathematics. 2022; 10(19):3433. https://doi.org/10.3390/math10193433

Chicago/Turabian Style

DIDI, Sultana, Ahoud AL HARBY, and Salim BOUZEBDA. 2022. "Wavelet Density and Regression Estimators for Functional Stationary and Ergodic Data: Discrete Time" Mathematics 10, no. 19: 3433. https://doi.org/10.3390/math10193433

APA Style

DIDI, S., AL HARBY, A., & BOUZEBDA, S. (2022). Wavelet Density and Regression Estimators for Functional Stationary and Ergodic Data: Discrete Time. Mathematics, 10(19), 3433. https://doi.org/10.3390/math10193433

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop