УДК 519.2
A Class of Local Linear Estimators with Functional Data
Sara Leulmi* Fatiha Messaci^
Department of Mathematics University brothers Mentouri Road of Ain El Bey,Constantine, 25017
Algeria
Received 04.12.2018, received in revised form 28.01.2019, accepted 06.03.2019 We introduce a local linear nonparametric estimation for the generalized regression function of a scalar response variable given a random variable taking values in a semi metric space. We establish a rate of uniform consistency for the proposed estimators. Then, based on a real data set we illustrate the performance of a particular studied estimator with respect to other known estimators.
Keywords: locally modelled regression, Nonparametric estimation, Rate of Convergence, Uniform almost complete convergence.
DOI: 10.17516/1997-1397-2019-12-3-379-391.
1. Introduction and preliminaries
Since the pioneer works summarized in [6], several studies have dealt with the nonparametric functional estimation. This research field is motivated by the fact that several data collected in practice, are given in the form of curves. Moreover, the progress of the digital computing tools allows the treatment of such observations. Different kernel type estimators have been studied in the literature, see for example [6]. Then, inspired by the local linear nonparametric method, [1] have introduced a more general and flexible method than the kernel one. It is the so called local modelling approach. They obtain a rate of the pointwise almost-complete convergence for their estimator of the regression function.
But, as pointed out in [4] "the uniform consistency results are indispensable tools for the study of more sophisticated models in which multi-stage procedures are involved". Under uniform convergence, one can make prediction even if the data are not well observed. We also can solve some problems such as data-driven bandwidth choice (see [2]), or bootstrapping (see [5]). Uniform convergence of other local linear nonparametric estimators has been investigated in some papers as [3] and [8]. In this work, our principal aim is to establish the uniform almost complete convergence of the local linear estimator of a generalized regression function (which includes the estimator introduced in [1]) and to focus on a tool of prediction (a conditional quantile estimator). More precisely, Section 2 is devoted to introduce the generalized regression function estimator and to state its pointwise convergence. Section 3 contains the principal results of this work which consist to establish the rate of the uniform almost convergence of the last estimator and to focus on the particular case of the conditional distribution function estimation from which we deduce a rate of the uniform consistency of a conditional quantile estimator. In Section 4 using a real data set, the prediction obtained from this last estimator is compared to those of two other known estimators. Finally, the detailed proofs of some needed lemmas are evoked in the Appendix.
1 [email protected] © Siberian Federal University. All rights reserved
2. The estimation and the pointwise almost-complete convergence
2.1. The model
Let us consider n pairs of random variables (Xi,Yi)i=1,..,n independent and identically distributed as the pair (X, Y) which is valued in Fx R, where (F, d) is a semi-metric space. Our goal is to estimate the generalized regression function, defined for all x in F, by
mv(x) = E(<p(Y )|X = x),,
where p is a known real-valued borel function.
It is clear that mv generalizes the classical regression function (set p(t) = t) as well as the conditional distribution function (set for any y e R,p(t) = 1]-x,,^](t)).
Following [1] who proved the pointwise almost complete convergence of the classical regression function estimator, the local linear estimate of mv is obtained as the solution for a of the following minimization problem
n
min y^ (p(Yi) - a - bp(Xi, x)f K(h-1d(Xi, x)), (a,b)eR2
i=1
where (3(, .) is a known operator from F x F into R such that, Vx e F, /3(x, x) = 0, the function K is a kernel and h := hn is a sequence of strictly positive real numbers which plays a smoothing parameter role.
This approach assumes that a + b3(,x) is a good approximation of mv(.) around x. As 3(x,x) =0, a will be a suitable estimate for mv(x).
By a simple calculus, one's can derive the following explicit estimator
m v(x)= E?J=iWj(x) [o-°) '
where
Wij (x) = 3(Xi,x) (3(Xi,x) - 3(Xj, x)) K (h-1d(Xi, x))K (hr1d,(Xj ,x)). As for l e {0,1}, we have
n
]T Wij(x)pl(Yj) = Y,{(3(Xi,x) - 3(Xj,x)) (3(Xi,x)pl(Yj) - 3(Xj,x)<pl(Y)) x i,j=1 i<j
x K(h-1d(Xi, x))K(h-1d(Xj,x)},
if the denominator of the estimator rnv (x) is zero, it is the same for its numerator. Moreover, under assumptions (H1) and (H3)-(H7), we get E(W12(x)) > 0 (see the proof of Lemma 4.4 in [1]).
Notice that the expression of inv allows fast computational issue and that the choices of 3 and d will be crucial.
2.2. The pointwise almost-complete convergence
Let x be a fixed point in F, for any positive real h, B(x, h) := {y e F/ d(x,y) < h} denotes a closed ball in F of center x and radius h. We also define x(r1,r2) := P(r1 < d(X,x) < r2), where r1 and r2 are two real numbers.
We investigate the asymptotic behaviour of the local linear estimator mn , under the following assumptions.
(H1) For any h> 0, $x(h) := <Px(0, h) > 0. (H2) mv £{f : F^ R, lim f(x') = f (x)}.
d(x,x')^0
(H2') mv £ {f : F ^ R, 3b > 0, Vx' £ F; \ f (x) - f (x')\ < Cxdb(x,x')}, where Cx is a positive constant depending on x.
(H3) The function ¡3(, .) is such that: 30 < Mi < M2, Vx' £ F,
M1d(x, x') < \P(x,x')\ < M2d(x,x').
(H4) The kernel K is a positive and differentiable function on its support [0,1].
(H5) The bandwidth h satisfies: lim„^TO h = 0, and lim„^TO Jh) = 0.
(H6) There exists an integer n0, such that
1 f1 d Vn>no, Vx £F,—— $x(zh,h)—(z2K (z)) >C> 0 &x(h)) J0 dz
and
( / ß2(u,x)dPX (u) I , yJB(x,h) J
hi /3(u,x)dPx (u) = o\ / (u,x)dPx (u)
J B(x,h) \JB(x,h)
where dPX is the distribution of X.
(H7) ym > 2 : x ^ E(\^(Y)\m/X = x) is a continuous operator.
Remark that our hypotheses are very similar to the assumed conditions in [1].
Let us state the pointwise almost-complete convergence (a.co.) of mv(x), along with a rate.
Theorem 1. Assume that assumptions (HI), (H3)-(H7) are satisfied.
(i) Under the additional hypothesis (H2), we have
mv(x) - mv(x) = Oa.co.(l)-
(ii) If in addition (H2') is satisfied, we get
ln n
mh-(x) - m-(x) = O(h ) + Oa.co. , ,, -,
U<P-x(
i rmn~\ yjn&jh))
Notice that the proof of this theorem is based on a standard decomposition given for all x £ F, by
[(m^x) - Emi(x)) - (m- (x) - Em,(x))] - m-(x)(mo(x) ^ mo (x) mo(x)
in-(x) - m-(x) = --- [(mi(x) - Em\(x)) - (m-(x) - Emi(x))]----—-, (1)
where, for l = 0,1
mi(x) = ~(-uEW ( ^WH(x)v\Yj).
n(n — 1)EW12(x)
The study of each term of this decomposition can be carried out exactly as done in the proof of Theorem 4.1 and Theorem 4.2 in [1] with replacing Y by ^(Y), so for the sake of avoiding repetitions, we omit the proof.
Now, we will focus on the uniform consistency.
3. The uniform almost-complete convergence
3.1. The estimator m^
We will establish the uniform almost-complete convergence of fnv on some subset Sf of F which can be covered by a finite number of balls. This number has to be related to the radius of these balls (see hypothesis (U5)).
To this goal, let us recall the following definition.
Definition 1. Let S be a subset of a semi-metric space F, and let e > 0 be given. A finite set of points x1 ,x2,...,xN in F is called an e-net for S if S C |J N=1 B(xk ,e). The quantity 0s(e) = ln(Ne(S)), where Ne(S) is the minimal number of open balls in F of radius e which is necessary to cover S, is called Kolmogorov's e-entropy of the set S.
It is known that the entropy of a set measures its complexity. We refer to [7] and [4] for more details on this topic.
We suppose that x1,... ,xNrn(Sf) is an rn-net for Sf where for all k e {1,..., Nrn (Sf)}, xk e Sf and (rn) is a sequence of positive real numbers. In this study, we need the following assumptions.
(U1) There exist a differentiable function < and strictly positive constants C,C1 and C2 such that
Vx e SF, Vh > 0; 0 < C<(h) < <x(h) < C2<(h) < to
and
3^0 > 0, Vn < no,<'(n) < C, where < denotes the first derivative of < with <(0) = 0.
(U2) The generalized regression function mv satisfies:
3C > 0, 3b > 0,Vx e SF,x' e B(x, h), lmv(x) - mv(x')l < Cdb(x,x').
(U3) The function 3(., ) satisfies (H3) uniformly on x and the following Lipschitz's condition
3C > 0, Vx1 e Sf, x2 e Sf, x e F, l3(x,x1) - 3(x,x2)l ^ Cd(x1, x2). (U4) The kernel K fulfills (H4) and is Lipschitzian on [0,1].
(U5) lim h = 0, and for rn = O (^nn), the function 0Sf satisfies for n large enough:
\2
(ln n)2 (\n n\ n$(h)
< ^Sf ( — <
)
and
n&(h) ^ Y F \ n J ^ ln n
]Texp{(1 - ß)0sA -^J } <
for some 3 > 1. (U6) The bandwidth h satisfies: 3n0 e N, 3C > 0, such that
1 r i d
Vn>no, Vx G ST$x(zh,h)—(z2K (z)) >C> 0 ®x(h) J0 dz
and
h ß(u,x)dPx (u) = o\ ß2 (u,x)dPx (u)
JB(x,h) \JB(x,h) J
uniformly on x.
(U7) 3C > 0 such that Vm > 2 : E(\^(Y )\m/X = x) < Sm(x) <C < to with Sm(.) continuous on Sf .
Roughly speaking, these hypotheses are uniform version of the assumed conditions in the point-wise case and have already been used in the literature. We refer to [8] for conditions (U1), (U3), (U4) and (U6) and to [4] for assumptions (U2), (U5) and (U7). The claimed result is as follows.
Theorem 2. Under assumptions (U1)-(U7), we have
sup \mv(x) — mv(x)\ = O(hb) + Oa
xESf
(f
n$(h)
We can readily deduce the uniform consistency of the estimator studied in [1] for which, to the best of our knowledge, only the pointwise convergence is available.
This result shows that, contrary to the finite case, the rate of convergence obtained may differ from that of the pointwise consistency, it is function of the entropy of the subset on which the uniform convergence states.
It is easy to see that the proof of Theorem 2 is a direct consequence of the decomposition (1) and of the following lemmas for which the proofs are relegated to the Appendix.
Lemma 1. Assume that hypotheses (U1), (U2) and (U4) hold, then:
sup \mv(x) — Em1(x)\ = O(hb).
xESf
Lemma 2. Under assumptions of Theorem 1, we obtain that:
sup \m1 (x) — Em1(x)\ = Oa ' F V n '
(
xeSf \ V n$(h)
Lemma 3. If assumptions (U1),(U3)-(U6) are satisfied, we get:
sup \mo(x) — 1\ = Oa 'Sf[ nJ
\>. j- \ v^a.co. i \i jr./! \
xeSf \ v n®(h)
and
(
VW inf mo(x) < 1 | < to. ^ \xeSf 2 J
n = 1 v 7
3.2. A conditional quantile estimator
Let Fx(y) = P(Y < y\X = x) be the conditional distribution function of Y given X = x where y is real and x is a fixed object in F. To estimate it, we treat this function as a particular
case of mv with ^(t) = 1]-x,,y](t) for y £ R. Thus, we estimate Fx(y) by
F-xiy) = j, (2)
22i,j=1 Wij(x)
where
Wij(x) = 3(Xi,x) (3(Xi,x) — 3(X*,x)) K(h-1d(Xi, x))K(h-1d(Xj,x)).
The conditional quantile of order a (a e (0,1)) is ta(x) = inf{y e R, Fx(y) > a}. So, we deduce from Fnx a natural conditional quantile estimator as,
ta(x)=inf{y e R,Fx(y) > a}. (3)
Notice that t1/2(x) is the so called conditional median.
To investigate the asymptotic convergence of Fx(y), we introduce the following conditions.
(U2)' There exist 5 > 0, C > 0 and b > 0, such that for any x e Sf,x' e B(x,h) and y e [ta(x) - 5,ta(x) + 5], we have
IFx' (y) - Fx(y)l < Cdb(x, x').
(U5)' lim h = 0, and for rn = O (^n-^), the function 0Sf satisfies for n large enough:
(ln n)2 fln n\ n<(h)
<0Sf[— <
n<&(h) F \ n J ln n '
and
J2n(Î+1/2) ex^(1 - ßWsf(\<
for some 3 > 1 and £ > 0. The following result concerns the uniform almost complete convergence of Fx(y). Theorem 3. Under assumptions (U1), (U2)', (US), (U4), (U5)' and (U6), we have
ln
sup sup \Fx(y) - Fx(y) \ = O(hb) + Oa,c
xESf ye[ta(x)-S,ta(x)+S]
i nnn~\
[ynMhj)
To prove this theorem we make use of the decomposition given, for all x and y, by
Fx(y) - Fx(y) = \(FN(y) - EFN(y)) (Fx(y) - EFN(y))l - -^(—o^) - 1), (4)
mo(x) Lv / V /J —o(x)
where FN(y) = —,-N"LTTr—r^r 2 Wij(x)1rYj<v\ and m0(x) is defined in (1). Now, it sufficies
n(n - 1)EW12(x) i=j
to apply Lemma 3 together with the following lemmas.
Lemma 4. Assume that hypotheses (U1), (U2)' and (U4) hold, then
Fx(y) - EFN(y)\= O(hb). Lemma 5. Under assumptions of Theorem S, we obtain that
sup sup
xesr ye[ta(x)-s,ta(x)+s]
sup sup
xesr ye[ta(x)-s,ta(x)+s]
FN(y) - EFxn(y)
(f-
O I J*sf c-n )
The same arguments as in the proof of Lemma 2.8 (resp. Lemma 2.9) in [8] permit us to derive the conclusion of Lemma 4 (resp. Lemma 5).
To obtain the uniform consistency of the conditional quantile estimator, we introduce the following conditions used for example in [8].
(U8) Ve > 0, 3£ > 0 such that for any function ga from Sf into [ta(x) — S,ta(x) + J] we have
sup \ta(x) — ga(x)\ > e implies sup \Fx(ta(x)) — Fx(ga(x))\ >
xesF xesF
(U9) 3j > 1, Vx £ Sf, Fx is j-times continuously differentiable on ]ta(x) — S,ta(x) + S[ with respect to y and satisfies Fx(l) (ta(x)) =0 if 0 < l < j, Fx(j) (ta(x)) >C > 0 and Fx(j) is uniformly continuous on [ta(x) — S, ta(x) + J] where Fx(l) stands for the Ith-order derivative of Fx .
A known method can be applied to derive the following result from Theorem 3, see for example
the proof of Corollary 3.1 in [8].
Corollary 1. Under the hypotheses of Theorem 3 and if (U8)and (U9) are satisfied, we obtain
sup \ta(x) — ta(x)\ = O(h) + Oa.co. h 'SFl( J ^ xeSf \ V n<P(h)
4. A real data application
In this section, we use a real data set to illustrate the efficacy of the studied method through our conditional median estimator t1/2. More precisely, we compare this last estimator to two other conditional median estimators: the first is based on the kernel method (denoted KM) and is studied in [6] and the second is based on the local linear method (denoted LLM) and is introduced in [8].
For this purpose, we use the spectrometric data set which can be found at http ://lib.stat. cmu.edu/datasets/tecator. These data consist of 215 pairs (Xi,Yi)i=1,...,215. For each i, the spectrometric curve Xi is the spectra of a finely chopped meat and Yi is the the corresponding fat content obtained by an analytical chemical process. Our goal is to predict the fat content in a piece of meat from its spectrometric curve. For this, we estimate the median t1/2(x) of the conditional distribution by t1/2 (x).
We split these real data into a learning sample containing the first 160 units used to build the estimator and a test sample containing the last 55 units used to predict the fat content and to make a comparison.
The KM (resp. the LLM) estimator is computed with the same parameters as at Subsection 12.4 in [6] (resp. at section 4 in [8]). For the computation of the estimator t1/2(x), we use the quadratic kernel K(x) = |(1 — x?)1\o011](x), the bandwith h is chosen by a 2-fold cross-validation method, the semi-metric d is based on the derivative described in [6] (see routines "semimetric.deriv" in the website http://www.lsp.ups-tlse.fr/staph/npfda) and 3 = d.
To illustrate the performance of our estimator, we first plot the true values (provided in the test sample) against the predicted ones by means of the three estimators (one in each graph). This is displayed in Fig. 1. Secondly, to be more precise we evaluate their empirical Mean Square Errors (MSE), defined by
MSE :=5k E (ti— Yi)2,
i=1
where Yi (resp. Yi) is the true (resp. the estimated) value.
The obtained results are
MSE(t1/2)=3.22, MSE(LLM)=3.8 and MSE(KM)=4.8.
This shows that the estimator t1/2 performs well and that the local linear method seems to improve the quality of the prediction even for functional data.
Fig. 1. From left to right: the estimator t1/2, the KM estimator and the LLM estimator for the spectrometric data
5. Appendix
In what follows, let C be some strictly positive generic constant and for any x e F, and for all i = 1,..., n:
Ki(x) := K(h-1d(Xi,x)) and 3i(x) := 3(Xi,x).
To treat the uniform convergence of — v(x) , we need to make use of Lemma 4.1 introduced in [8] and stated here as follows.
Lemma 6. Under assumptions (U1),(US),(U4) and (U6), we obtain that: i) V(p,l) e N* x N, supxeSF E (Kpp(x)l3[(x)l) < Chl<(h). ii) infxeSF E (K1(x)32(x)) > Ch2<(h). Proof of Lemma 1. We have
1
Emi(x)
and Emi(x) can also be written as
E(Wi2(x))
E(Wi2 (x)vl (Y2)),
Em1(x) = E (E(m1(x)lX2)) So, we get under assumption (U4)
1
E(Wi2(x))
E (Wi2(x)E(<p(Y2)\X2)).
\mv(x)- Emi(x)
1
mm; f E (Wi2(x)(mv(x) - mV(X2)))\ < sup \mV(x) - m(x'
\E(Wi2 (x))\ x'EB(x,h)
We need to take into account hypothesis (U2) to obtain
sup l—v(x) - E—1(x)l = O(hb).
xESf
Proof of Lemma 2. We use again the following decomposition given in [1]. Namely
—1(x) = Q(x) [S2,1(x)S4,o(x) - S3,1(x)S3,o(x)],
where, for p = 22, 3,4, and l = 0,1,
S ( )= 1 n Ki(x)ßr2(xW(Y)
Sp'l(x) n$x(h)^ hp-2
v ' i=1
and
n2 h2$X(h)
Q(x)
n(n - 1)E (W12(x))'
By following the same steps as in [1], and using lemma 6 instead of lemma A.1 in [1], we obtain under the assumptions (U1)-(U4) and (U6),
Q(x) = O(1), E(Sp,i(x)) = O(1),
uniformly on x, for p = 2, 3,4 , l = 0,1,
sup \E(S2,1(x))E(S4,o(x)) — E(S2,1(x)S4,o(x))\ = o(-^r) , xeSf \n$(h)J
and
sup \E(S3A(x))E(S3,o(x)) — E(S3,1(x)S3,o(x))\ = o(, XeSf \n$(h)J
( li>s (—A
which is, in view of hypothesis (U5), equals to O I V —). We need to check that for p = 2, 3, 4 and l = 0, 1,
sup \SPJ (x) - E (Sp, l(x))\ = Oa 1 V n '
xeSf \ V n$(h)
(f-
To satisfy this aim, let us set
j (x) = arg min d(x,xj),
v ; je{1,2,...,Nrn(Sf)} K 3h
and consider the following decomposition
sup \Sp,i(x) — ESp,i(x)\ < sup \Sp,i(x) — Spi(xj(x))\ +
xesF xesF
+ sup ISp^(xj(x)) — ESp,i(xj(x))\ + XESf
+ sup \ESpii(xj(x)) - ESpi(x)\ := Fpl + F2pl + Fpl.
xESf
Let's, now, study each term F?'1 for k = 1, 2,3.
Study of the terms F?'1 and F?'1.
First, let us analyze the term F?'1. Since K is supported in [0,1] and according to (U1), we
can write for all p = 2,3,4 C
F1p,! ^ nhp-2<(h) xSeUl £lKi(x)3!-2(x)^l(Vi)1B(x,h)(Xi)-
- Ki( xj(x ))3r (xj(x
C<Th) A
n
>p-2
<
<
+
nhp-2<(h) ^S? ^ Ki(x)1B(x,h) (Xi)l^l(Yi)l 3p 2(x) - 3p 2(xj(x) )1B(x3(x),h)(Xi)
C n \ \
+ nhp-2<(h) 3p-2(xj(x))1B(xj{x),h)(Xi)l^l(Yi)l \Ki(x)1B(x,h)(Xi) - Ki(xj(x))\
:= FP:l+ Fp . '2.
Analysis of the term Fp'[.
According to assumption (U3), we get
1B(x,h)(Xi) 3i(x) - 3i (x j(x) )1B(xj(x),h) (Xi ) <
< Crn1B(x,h) n B(x3(x),h)(Xi) + Ch1 B(x,h) n B(xj(x) ,h) (Xi )
and
<
1B(x,h)(Xi) 3i (x) - 3i (xj(x) )1B(xj(3.),h) (Xi)
< Crnh1B(xj(x),h) n B(x,h)(Xi) + Chh 1 B(x,h) n B (xj(x) ,h) (Xi )
By grouping the cases p = 3 and p = 4, we found
1B(x,h)(Xi) 3p 2(x) - 3i \xj(x) )1B(xj(x),h) (Xi)
p-2
<
< CrnhP 1B(x,(x),h) n B(x,h)(Xi) + ChP 1B(x,h)n B(xj(x),h)(Xi).
which gives the following inequality
Cr n
^ sup
h)
i=1
J<h) xS£5F ^ l^l(Yi)lKi(x)1B(x,h) n B(xj(x),h)(Xi) +
Cn
<h) S l^ (Yi)lKi (x)1B(xh)n Bjh)(Xi).
Analysis of the term Fp'2.
Using the following inequality
<
1B(x3 (x),h)(Xi) Ki(x)1B(x,h)(Xi) - Ki(xj(x))1B(x,h) U Bxh)(Xi)
< 1B(x,h) n B(x,x)'h)(Xi )lKi(x) - Ki(xj (x))| + Ki(xj (x))1B(xj (x)h)nBxh)
and by hypotheses (U3) and (U4), we obtain
l3p-2(xj(x))l1B(x,x)h)(Xi) \ Ki (x)1 B(x,h) (Xi) - Ki(xj(x)) l <
(Xi)
< ChP-2
h iB(x'h)nB(xj(x)' h)(Xi)+ Ki(xj(x) )1B(xj(x) ,h)nB(x,h)
(Xi)
(5)
which leads to
Crn
C
+ xfl Ç l^l(Yi)IKi(xj(x))1B(x-h)nB(X3(x),h)(Xi)-
This last inequality combined with (5) allow us to write
Crn "
FP'1 ^ nh&fc) ^ ^Y^BXMrBX^MXi) +
C '
+ mhi XseuSpF £ )IKi(xj(x))1B{x,{x)h)nBXh)(Xi) +
C n
+ rth) g № )\Ki(x)iBWB^(xi)-
Taking into account hypothesis (U4), we find
Cr n
Fp' ^nhm si:\^l(Yi)\iB(x^B{x^h)(xi).
Let
Z _ Crn\^l(Yi)\ (x
Zi _ —h$(h)— ¿S LB(X'h)uB(x3(x)'h)(Xi)-
The assumption (U7) implies that
Crm
E\\ < hm^> (6)
r
so, by applying corollary A.8 in [6], with a? _ —:
n
h$(hy
n ]C Zi _ EZi + Oa-c°- ^
n
n ' \V nh<Ê(h) I
Applying (6) again (for m _ 1), one gets
FP'1 _ O( '-n )+ Oa
I r-nlnn_ )
\\j nh$(h) J
Combining this with assumption (U5) and the second part of the assumption (U1), we obtain
FP'l_ » (T>
Second, since
we deduce that
FP'1 < E sup \SP'i(x) — SP'i(xj(x))\ I
\xGSf J
y
F" _ O ■ -h/ ^nèT I- (8)
Study of the term Fp'1.
For all n > 0, we have that
p(^...x»^(x^-E<
(. \ USr () )
I \Spl(xj(x)) - E(Sp,l (xj(x)))\ > ny ).
< Nrn (Sf ) max P( \SPi(xj{x)) - E(Sp,i (xj{x)))\
Let us set for p = 2, 3,4 that 1
Api hP-2$xAh)
Ki(xj{x))3V\xj{x))vl(Yi) - E(Ki(xj(x))3f~ (xj(x))^l(Yi)) .
Using the binomial Theorem, Lemma 6 and hypothesis (U1), (U2) and (U7), gives for p = 2,3,4,
E lApilm = O (<-m+1(h)) . Therefore, we can apply a Bernstein- type inequality as done in the corollary A-8 in [6], to obtain
p i n
n
E A
i=1
Thus, by choosing 3 such that Crj2 = 3, we get
JW) < 2-hf*S' ("))
P ^ > ' I < CNrn(Sff-f
Then, hypothesis (U5) allows us to write
WS, ) 2 n$(h)
i
*sr C-n1)
F"=o-w "^hr i- (9)
Finally, the result of Lemma 2 follows from the relations (7), (9) and (8). □
Proof of Lemma S. The first part of the claimed results can be directly deduced from the proof of Lemma 2 by taking, for all i, Yi) = 1. For the second part, It comes straightforward that
inf —o(x) < - ^ 3 x e Sf such that 1 - m0(x) > - ^ sup l1 - m0(x)l > -x^Sf 2 2 xesF 2
^ y^P ( mf mo(x) < - < to. ^ KxeSjr ° '
n=0 v F
References
[1] J.Barrientos-Marin, F.Ferraty, P.Vieu, Locally modelled regression and functional data, J.ournal of Nonparametric Statistics, 22(2010), 617-632.
[2] K.Benhenni, F.Ferraty, M.Rachdi, P.Vieu, Local smoothing regression with functional data, Computational Statistics, 22(2007), no. 3, 353-369.
[3] J.Demongeot, A.Laksaci, F.Madani, M.Rachdi, Functional data analysis: conditional density estimation and its application, Statistics, 47(2013), no. 1, 26-44.
[4] F.Ferraty, A.Laksaci, A.Tadj, P.Vieu, Rate of uniform consistency for nonparametric estimates with functional variables. Journal of Statistical planning and inference, 140(2010), 335-352.
[5] F.Ferraty, I.Van Keilegom, P.Vieu, On the validity of the bootstrap in nonparametric functional regression, Scandinavian Journal of Statistics, 37(2008), no. 2, 286-306.
[6] F.Ferraty, P.Vieu, Nonparametric functional data analysis. Theory and Practice. Springer Series in Statistics, New York, 2006.
[7] A.N.Kolmogorov, V.M.Tikhomirov, e-entropy and "-capacity", Uspekhi Matematicheskikh Nauk, 14(1959), 3-86 (in Russian).
[8] F.Messaci, N.Nemouchi, I.Ouassou, M.Rachdi, Local polynomial modelling of the conditional quantile for functional data, Statistics Methods and Applications, 24(2015), no. 4, 597-622.
Класс локальных линейных оценок с функциональными данными
Сара Леулми Фатиха Мессачи
Факультет математики Университет братьев Ментури дорога Айн-эль-Бея, Константин, 25017
Алжир
Введем локальную линейную непараметрическую оценку для обобщенной функции регрессии скалярной переменной отклика для заданной случайной величины, принимающей значения в полуметрическом пространстве. Мы устанавливаем скорость равномерной согласованности для предлагаемых оценок. Затем, основываясь на реальном наборе данных, мы проиллюстрируем эффективность конкретного изученного оценщика по сравнению с другими известными оценщиками.
Ключевые слова: локально моделируемая регрессия, непараметрическая оценка, скорость сходимости, равномерная почти полная сходимость.