a markovian local resampling scheme for nonparametric estimators in ...

9 downloads 64184 Views 295KB Size Report
~kernel! estimator of the conditional expectation m~x;f! E~f~Xt ... Given observations X1, X2,+++, XT the nonparametric estimator of m~x;f!, ...... App+! we have.
Econometric Theory, 17, 2001, 540–566+ Printed in the United States of America+

A MARKOVIAN LOCAL RESAMPLING SCHEME FOR NONPARAMETRIC ESTIMATORS IN TIME SERIES ANALYSIS EF S T A T H I O S PA P A R O D I T I S University of Cyprus

DI M I T R I S N . PO L I T I S University of California, San Diego

In this paper we study the properties of a pth-order Markovian local resampling procedure in approximating the distribution of nonparametric ~kernel! estimators of the conditional expectation m~x;f! 5 E~f~X t11 !6Yt, p 5 x! where $X t , t $ 1% is a strictly stationary process, Yt, p 5 ~X t , X t21 , + + + , X t2p11 ! Á , and f~{! is a measurable real-valued function+ Under certain regularity conditions, asymptotic validity of the proposed resampling scheme is established for a class of stochastic processes that is broader than the class of stationary Markov processes+ Some simulations illustrate the finite sample performance of the proposed resampling procedure+

1. INTRODUCTION Let $X t , t 5 1,2, + + + % be a ~strictly! stationary real-valued stochastic process and denote by FYt, p ~{! the marginal distribution of Yt, p 5 ~X t , X t21 , + + + , X t2p11 ! Á and by FXt11 6Yt, p ~{6{! the conditional distribution of X t11 given Yt, p + In this paper we are interested in estimating the sampling behavior of a nonparametric ~kernel! estimator of the conditional expectation m~x; f! 5 E~f~X t11 !6Yt, p 5 x! where f~{! is a real-valued, measurable function+ Several conditional functionals of the process can be obtained by appropriately specifying the function f+ For instance, for f1~z! 5 z and f 2 ~z! 5 z 2 we have that m~x; f1 ! is the conditional mean, whereas m~x; f 2 ! 2 m 2 ~x; f1 ! is the conditional variance of s 6Yt, p 5 x! X t11 given that Yt, p 5 x+ Apart from the conditional moments E~X t11 the function m~x; f! can also be used to describe some other conditional functionals of the process that may be of interest+ As an example, for f~z! 5 1~2`, y# ~z! and y a real number we have that m~x; f! 5 FXt11 6Yt, p ~ y6x!, that is, the conditional distribution function mentioned previously+ We are grateful to the co-editor Joel Horowitz and to both referees for a number of useful comments+ This research has been supported in part by National Science Foundation Grant DMS-97-03964 and by a University of Cyprus Research Grant+ Address correspondence to: Efstathios Paparoditis, Department of Mathematics and Statistics, University of Cyprus, P+O+ Box 20537, CY-1678 Nicosia, Cyprus; e-mail: stathisp@ucy+ac+cy+

540

© 2001 Cambridge University Press

0266-4666001 $9+50

MARKOVIAN LOCAL RESAMPLING OF KERNEL ESTIMATORS

541

Given observations X 1 , X 2 , + + + , X T the nonparametric estimator of m~x; f!, which we consider in this paper, is given by T21

( f~Xt11 !Kh ~x 2 Yt, p !

m[ h ~x; f! 5

t5p

T21

+

(1.1)

( Kh ~x 2 Yt, p !

t5p

In this notation Kh ~{! 5 h 2p K~{0h! where K : R p r R is a nonnegative kernel satisfying *K~u!du 5 1, *uK~u! du 5 0 and h 5 h~T ! . 0 is the bandwidth used to smooth the observations+ Note that ~1+1! is the common Nadaraya– Watson estimator of m~x; f!+ Nonparametric estimators such as the one considered in this paper are frequently used as an end product in time series analysis for modeling or predictive purposes or, perhaps more important, as a guide in identifying a parametric model to be used in a subsequent stage+ For instance, a general univariate model that contains many of the linear and nonlinear classes discussed in the literature, including the popular nonlinear conditional heteroskedastic ~ARCH! model, is given by X t11 5 g~X t , X t21 , + + + , X t2p11 , «t ! where g : R p11 r R is a measurable function+ Here $«t % is an independent and identically distributed ~i+i+d+! sequence of random variables with mean zero and finite variance and «t is assumed to be independent of X s for s # t ~cf+ Tong, 1990, Ch+ 3!+ One way to gain insight into the nonlinearity features of such models given a set of observations is to estimate nonparametrically functionals such as the conditional mean or the conditional variance and to compare the estimates obtained with those expected under a particular hypothesis about the underlying model+ Such an approach of detecting and modeling nonlinearity in time series analysis has been proposed by several authors; see among others Tong ~1990!, Auestad and Tjøstheim ~1990!, and Tjøstheim and Auestad ~1994!+ Under certain regularity conditions, strong consistency and asymptotic normality of the estimator m[ h ~x; f! have been established by several authors; we mention here among others the papers by Robinson ~1983!, Roussas ~1990!, and Masry and Tjøstheim ~1995!+ The asymptotic normality of the estimator ~1+1! is, for instance, useful in constructing pointwise confidence intervals for m~x; f!+ However, there is an inherent difficulty in using the limiting Gaussian distribution for such purposes that is due to the fact that the mean and the variance of this distribution depend on unknown ~and difficult to estimate! characteristics of the process+ The aim of this paper is to show that a Markovian local bootstrap procedure proposed by Paparoditis and Politis ~1997! and which generates bootstrap replicates X 1* , X 2* , + + + , X T* that “reproduce” correctly the conditional distribution of X t11 given the “past” Yt, p2 leads to an asymptotically valid approximation of the distribution of L T ~x! 5 !Th p ~ m[ h ~x; f! 2 m~x; f!!+

(1.2)

542

EFSTATHIOS PAPARODITIS AND DIMITRIS N . POLITIS

The proposed procedure provides an alternative ~as compared to the normal! approximation of the distribution of L T ~x! and at the same time leads to consistent estimation of quantities such as the mean and the variance of the distribution of interest+ Although the local bootstrap procedure generates a pseudoseries X 1* , X 2* , + + + , X T* with a Markov dependence structure, we show in the present paper that the validity of this resampling scheme for the class of nonparametric estimators considered is not limited to the Markov case only+ To understand this fact intuitively note that the information regarding estimation of FXt11 6Yt, p ~{6{! and its moments typically lies in the “scatterplot” of X t11 vs+ Yt, p for t 5 1,2, + + + , T 2 p; the crucial feature of this scatterplot is the joint distribution of X t11 and Yt, p that is determined by the conditional distribution FXt11 6Yt, p ~{6{! and the marginal distribution FYt, p ~{!+ Therefore, capturing FXt11 6Yt, p ~{6{! and FYt, p ~{! by a bootstrap process may give valid distributional conclusions for general ~not necessarily Markov! stationary processes+ Bootstrapping nonparametric estimators in time series analysis has received considerable interest in recent years+ Franke, Kreiss, and Mammen ~1996! propose some alternative approaches for bootstrapping the ~pointwise! distribution of nonparametric estimators under the assumption that the process considered is generated by the nonlinear autoregressive process X t11 5 f ~X t ! 1 s~X t !«t11 with $«t % an i+i+d+ sequence+ These approaches include a wild bootstrap and a residual-based bootstrap+ Under the assumption that $X t % is a Markov process, Neumann and Kreiss ~1998! show validity of the wild bootstrap for local polynomial estimators of the conditional mean E~X t11 6 X t 5 x! and for supremum type statistics+ Compared to these approaches, our Markovian local resampling scheme avoids any kind of preliminary nonparametric estimation of unknown functions to generate the bootstrap pseudoseries+ Furthermore, its realm of validity includes but is not restricted to the Markov class+ The paper is organized as follows+ In Section 2 the Markovian local resampling scheme is briefly described and the bootstrap approximation to the statistic L T ~x! is given+ Section 3 is devoted to the ability of the Markovian local resampling procedure to approximate the distribution of L T ~x!+ After stating the set of technical assumptions imposed on the class of stochastic process considered, the main result is given+ The finite sample performance of the method proposed is illustrated by some simulated examples in Section 4, and all proofs are collected in Section 5+

2. THE MARKOVIAN LOCAL RESAMPLING SCHEME The Markovian local resampling scheme introduced by Paparoditis and Politis ~1997! generates bootstrap replicates by reshuffling the original data points according to a particular probability mechanism+ This local resampling scheme can be described by the following three steps+

MARKOVIAN LOCAL RESAMPLING OF KERNEL ESTIMATORS

543

1+ Select a resampling width b 5 b~T ! . 0 and a set of p starting values Yp,* p 5 * ~X p* , X p21 , + + + , X 1* !+ In practice, a simple choice is Yp,* p 5 Yp, p + Select further a resampling kernel W that is a probability density on R p satisfying W . 0 and *uW~u!du 5 0+ 2+ For any time point t 1 1 [ $ p 1 1, p 1 2, + + + , T % suppose that Yt,* p has been already generated+ Let J be a discrete random variable taking its values in the set N p, T21 5 $ p, p 1 1, + + + , T 2 1% with probability mass function given by

Y

P~J 5 s! 5 Wb ~Yt,* p 2 Ys, p !

(

Wb ~Yt,* p 2 Ym, p !

(2.3)

m[N p, T21

* * for s [ N p, T21 + The bootstrap replicate X t11 is then defined by X t11 5 X J11 + 3+ If t 1 1 , T then go to step 2+

At completion of the algorithm, a new bootstrap pseudoseries X 1* , X 2* , + + + , X T* is created with distributional properties that “mimic” those of FXt11 6Yt, p of the original series+ Note that the Markovian local resampling scheme described earlier “works” by assigning at each step of the resampling process the resampling probabilities ~2+3! to each of the observed original values X s11 [ $X p11 , X p12 , + + + , X T %+ By this algorithm the probability that the * bootstrap random variable X t11 takes the value X s11 depends on how close Ys, p is to the already generated segment of bootstrap values Yt,* p+ The closer Ys, p is to Yt,* p the larger the resampling probability ~2+3! will be provided the mass of the resampling kernel W is concentrated around its mean value * 5 zero+ It is easily seen that by the positivity of the resampling kernel P~X t11 T21 * * * X s11 6Yt, p ! . 0+ Furthermore, ( s5p P~X t11 5 X s11 6Yt, p ! 5 1+ Our procedure provides a ‘natural’ extension of Efron’s ~1979! classical i+i+d+ bootstrap to the Markovian case+ In effect, Efron’s ~1979! i+i+d+ bootstrap can be considered a special case of our local bootstrap if we allow the Markovian order p to be zero, in which case all observations have the same resampling probability, that is, P~J 5 s! 5 10~T 2 p!+ Given a bootstrap series X 1* , X 2* , + + + , X T* generated by the Markovian local bootstrap, we propose to approximate the distribution of the statistic L T ~x! by the distribution of the bootstrap statistic L*T ~x! 5 !Th p ~ m[ h* ~x; f! 2 m b* ~x; f!!,

(2.4)

where T21

m[ h* ~x; f! 5

* f~X j11 !Kh ~x 2 Yj,* p ! ( j5p T

( Kh ~x 2 Yj,* p !

j5p

(2.5)

544

EFSTATHIOS PAPARODITIS AND DIMITRIS N . POLITIS

and T21

( f~Xt11 !Wb ~x 2 Yt, p !

m b* ~x; f! 5

t5p

(2.6)

+

T

( Wb ~x 2 Yt, p !

t5p

Note that if x [ X p , where X p is the p-times Cartesian product of the set of * observed values X 5 $X 1 , X 2 , + + + , X T %, then m b* ~x; f! 5 E * ~f~X t11 !6Yt,* p 5 x!, * that is, the conditional expectation of the bootstrap processes $X t , t $ 1% is a kernel estimator of m~x; f! with a “smoothing kernel” W and a “smoothing bandwidth” b+ In the following discussion we assume throughout that the set of points x at which the unknown function m~x; f! is estimated is a subset of X p + Note further that in obtaining m[ h* ~x; f! the same formula, that is, the same kernel and the same bandwidth as in ~1+1!, is used with the only difference that the observations X t , t 5 1,2, + + + , T are replaced by the bootstrap replicates X t* , t 5 1,2, + + + , T+ 3. VALIDITY OF THE MARKOVIAN LOCAL BOOTSTRAP FOR KERNEL ESTIMATORS To state the main result of this paper the following set of technical assumptions is needed+ ~A1!+ ~i! FYt, p ~{! and FXt11 6Yt, p ~{6x! are absolutely continuous with respect to Lebesgue measure on R p and R and have bounded densities fYt, p ~{! and fXt11 6Yt, p ~{6x!, respectively+ ~ii! For all x1 , x2 [ R p , and y [ R ø $`%,

*E

y

2`

E

fXt11 , Yt, p ~z, x1 !dz 2

y

2`

*

fXt11 , Yt, p ~z, x2 ! dz # L~ y!7x1 2 x2 7,

where fXt11 Yt, p ~z, x! 5 fXt11 6Yt, p ~z6x! fYt, p ~x!, infy L~ y! . 0 and supy L~ y! 5 L , `+ ~iii! For all y1 , y2 [ R, 6 fXt11 6Yt, p ~ y1 6x! 2 fXt11 6Yt, p ~ y2 6x!6 # C~x!6 y1 2 y2 6 where C~x! . 0 and supx[R p C~x! 5 C , `+ ~iv! A compact subset S of R exists such that X t [ S a+s+ Furthermore, fXt11 6Yt, p ~{6x! . 0 for every x [ S, where S denotes the p-fold Cartesian product of S+

Assumptions ~A1!~ii! and ~iii! are smoothness assumptions on the joint and conditional densities and are common in nonparametric estimation problems such as those discussed here ~cf+ Robinson, 1983; Masry and Tjøstheim, 1995!+ The assumption of compactness of the support of fXt ~{! in ~A1!~iv! is of a rather technical nature+ It is imposed to simplify the technical arguments and

MARKOVIAN LOCAL RESAMPLING OF KERNEL ESTIMATORS

545

can be weakened by simultaneously strengthening other requirements ~for details, cf+ Paparoditis and Politis, 1997!+ ~A2!+ ~i! f is a real-valued, measurable function such that E6 f~ X t !6 m , ` for some m . 4+ ~ii! For all z 1 5 ~ y1 , x1 !, z 2 5 ~ y2 , x2 ! [ R p11 , 6f~ y1 ! fXt11 Yt, p ~ y1 , x1 ! 2 f~ y2 ! fXt11 Yt, p ~ y2 , x2 !6 # C7z 1 2 z 2 7, where 0 , C , `+

The next assumption deals with the dependence properties of $X t %+ Recall the definition of the strong mixing ~a-mixing! coefficient, that is, a~n! 5

sup j

`

6P~A ù B! 2 P~A!P~B!6,

n [ N,

A[B1 , B[Bj1n ` denote the s-algebras generated by the sets of random variwhere B1 and Bj1n ables $X 1 , X 2 , + + + , X j % and $X j1n , X j1n11 , + + + %, respectively+ Based on this definition, $X t , t $ 1% is a-mixing if a~n! r 0 as n r `+ In the subsequent discussion we make the following assumptions+ j

~A3!+ $X t , t $ 1% is a-mixing, and the mixing coefficient satisfies a~n! # Cr n for some positive constants C and r [ ~0,1!+ Remark 3+1+ Note that ~A3! is fulfilled if the process $X t , t $ 1% is a pthorder strictly stationary and geometrically ergodic Markov process and that this Markov class of stochastic processes contains several of the commonly used parametric and nonparametric models in time series analysis ~for several examples, see Tong, 1990, Ch+ 3!+ It is known that if $X t , t $ 1% belongs to this Markov class then it is geometrically absolutely regular ~i+e+, b-mixing!, which implies geometrically a-mixing ~see Doukhan, 1994!+ However, the class of stochastic processes satisfying ~A3! is broader than the Markov class described previously+ The following conditions are imposed on the smoothing and resampling parameters+ ~A4!+ ~i! K~{! is a product kernel; that is, for every x 5 ~ x 1 , x 2 , + + + , x p ! [ R p we have p K~x! 5 ) i51 k~ x i ! where k~{! is a bounded, Lipschitz continuous, and symmetric probability density the support of which is the interval @21,1# + ~ii! The smoothing bandwidth h satisfies h r 0 and Th p r ` as T r `+

546

EFSTATHIOS PAPARODITIS AND DIMITRIS N . POLITIS

~A5!+ ~i! The resampling kernel W is a bounded, Lipschitz continuous, and symmetric probability density on R p satisfying W . 0, *uW~u!du 5 0 and *7u7W~u!du , `+ ~ii! The resampling bandwidth b satisfies b 5 O ~T 2d ! for some 0 , d , 10~2~ p 1 2!!+

The next assumption is needed to get ~asymptotic! expressions for the bias term in estimating m~x; f! using the kernel estimator given in ~1+1!+ ~A6!+ ~i! W is two times continuously differentiable+ Furthermore, for i, j [ $1,2, + + + , p% we have *u is W~u!du 5 0 for s odd and *u i u j W~u!du 5 di, j W2 where di, j is Kronecker’s d and 0 , W2 , ` for all i [ $1,2, + + + , p%+ ~ii! fXt11 Yt, p ~ y, x! is two times continuously differentiable with respect to x, and the ~x ! ~x , x ! functions *c~ y! fXt11l Yt, p ~ y, x!dy and *c~ y! fXt11l1 Yt,l2p ~ y, x!dy are Lipschitz contin~x ! uous for c 5 1 and c 5 f and every l 1 , l 2 [ $1,2, + + + , p%+ Here, fXt11l Yt, p ~ y, x! and ~ x l1 , x l2 ! fXt11 Yt, p ~ y, x! denote first and second order partial derivatives of fXt11 Yt, p ~ y,{! evaluated at x+

Given the preceding set of assumptions, large sample validity of the proposed Markovian local resampling scheme in approximating the distribution of L T ~x! can be established+ By “large sample validity” we mean that the law of L*T ~x! ~conditionally on the data! is close to the law of L T ~x! with high probability, if the sample size T is large enough+ Our main theorem makes the preceding statement precise; here Kolmogorov’s distance d0 ~P, Q! 5 supx[R 6P ~ X # x! 2 Q~ X # x!6 between probability measures P and Q is employed+ THEOREM 3+1+ Suppose (A1)–(A6) hold and let T 102 h ~ p14!02 r Ch $ 0. We then have that, as T r `, d0 $L~L*T ~x!6 X 1 , X 2 , + + + , X T !, L~L T ~x!!% r 0

in probability.

Note that if T 102 h ~ p14!02 r 0 then the bias term !Th p ~Em[ h ~x; f! 2 m~x; f!! is asymptotically negligible, whereas for T 102 h ~ p14!02 r Ch . 0 this term converges weakly to B~x; f! 5 Ch

E

p

H

u 2 k~u!du ( m ~ x l ! ~x; f! fYt, pl ~x! 1 l51

~x !

J

1 ~x x ! m l l ~x; f! fYt, p ~x! , 2 (3.7)

~x ! fYt, pl ~x!

and m ~ x l x l ! ~x; f! denote first and second order parwhere m ~ x l ! ~x; f!, tial derivatives of m~x; f! and fYt, p ~x! with respect to x l ~cf+ Auestad and Tjøstheim, 1990!+ The case T 102 h ~ p14!02 r Ch actually corresponds to optimal smoothing ~minimizing the mean square error of m[ h ~x, f!!, where the case

MARKOVIAN LOCAL RESAMPLING OF KERNEL ESTIMATORS

547

T 102 h ~ p14!02 r 0 corresponds to undersmoothing that is suboptimal+ In most cases where resampling is used to estimate the distribution of a kernel smoothed estimator, suboptimal smoothing is used so that the bias term becomes negligible+ Quite remarkably, the Markovian local bootstrap procedure also approximates correctly the asymptotic mean ~bias term! of the statistic L T ~x!+ 4. NUMERICAL EXAMPLES Two simple numerical examples are presented in this section to demonstrate the finite sample performance of the bootstrap method proposed+ We consider realizations of length T 5 500 from the random coefficient autoregressive model ~cf+ Tong, 1990, p+ 111! X t 5 @0+1 1 bt #X t21 1 «t

(4.8)

and from the moving average model X t 5 «t 1 0+7«t21 ,

(4.9)

where $«t % and $ft % are i+i+d+ sequences independent from each other and satisfying «t ; N~0,1! and ft ; N~0,0+81!+ For model ~4+8! we are interested in estimating the distribution of the kernel estimator of the conditional variance v~ x! 5 Var~X t11 6 X t 5 x! 5 1 1 0+81x 2 + Using a smoothing bandwidth h 5 0+7 and Epanechnikov’s kernel, such an estimator of v~ x! is given by v[ h ~ x! 5 m[ h ~ x; f 2 ! 2 m[ h2 ~ x; f1 !+ For model ~4+9! we are interested in estimating the distribution of the kernel estimator m[ h ~ x! of the conditional mean m~ x! 5 E~X t11 6 X t 5 x!+ Note that under the Gaussianity assumption the conditional mean of X t 5 «t 1 u«t21 is linear and given by m~ x! 5 $u0~1 1 u 2 !%x+ The 5%, 50%, and 95% percentage points of the exact distributions of the statistics considered have been estimated using the corresponding percentage points of the bootstrap statistics !Th~ v[ h* ~ x! 2 vb* ~ x!! and !Th~ m[ h* ~ x! 2 m b* ~ x!! for several values of x in the interval ~23,3!+ Note that v[ h* ~ x! 5 m[ h* ~ x, f 2 ! 2 @ m[ h* ~ x, f1 !# 2 , vb* ~ x! 5 m b* ~ x; f 2 ! 2 @m b* ~ x, f1 !# 2 , m[ h* ~ x! 5 m[ h* ~ x, f1 !, and m b* ~ x! 5 m b* ~ x, f1 !+ The results of this simulation example are presented in Figure 1 for model ~4+8! and in Figure 2 for model ~4+9!+ The bootstrap estimates, shown in these figures by crosses, are mean estimates over 100 independent repetitions of the corresponding models+ For each repetition the distribution of the bootstrap statistic has been evaluated using 1,000 bootstrap replications, the Gaussian resampling kernel, and resampling width b 5 1+0+ The percentage points of the exact distribution, shown in these figures by circles, have been estimated using 10,000 replications of each of the models considered+ As these figures show, the bootstrap procedure proposed gives a very satisfactory estimation of the exact percentage points of the distribution of interest over the whole range of x-values considered+ We mention here that the accu-

548

EFSTATHIOS PAPARODITIS AND DIMITRIS N . POLITIS

Figure 1. Estimated exact and bootstrap estimates of the 5%, 50%, and 95% percentage points of the distribution of !Th~ v[ h ~ x! 2 v~ x!! for model ~4+8!+

racy of the bootstrap estimator decreases as the value of 6 x6 increases+ This can be seen in Table 1, where the estimated standard deviation of the bootstrap estimators is reported for both models considered and for some different values of x+ Table 2 gives ~for both models considered! the empirical coverage probabilities of 90% bootstrap confidence intervals and the corresponding mean lengths

Table 1. Standard errors of the bootstrap estimates of the 5%, 50%, and 95% percentage points of the distributions considered in Figures 1 and 2 Model ~4+8! X 22+5 21+5 0+0 1+5 2+5

Model ~4+9!

5%

50%

95%

X

5%

50%

95%

13+171 3+050 0+545 3+202 12+737

5+245 1+294 0+408 1+372 5+344

15+026 3+411 0+720 3+186 11+028

22+4 21+4 0+0 1+4 2+4

1+048 0+473 0+315 0+406 1+306

0+722 0+296 0+216 0+315 0+801

1+110 0+400 0+287 0+493 1+123

MARKOVIAN LOCAL RESAMPLING OF KERNEL ESTIMATORS

549

Figure 2. Estimated exact and bootstrap estimates of the 5%, 50%, and 95% percentage points of the distribution of !Th~ m[ h ~ x! 2 m~ x!! for model ~4+9!+

of these intervals based on 100 independent repetitions+ For model ~4+8! we calculate the bootstrap confidence intervals for the conditional variance v~ x! and for model ~4+9! the bootstrap confidence intervals for the conditional mean m~ x! using the values x 5 21+5, x 5 0, and x 5 1+5+ Also included in this table is a comparison of the Markovian local bootstrap with the block bootstrap method ~cf+ Künsch, 1989; Liu and Singh, 1992!+ Note that a ~1 2 a!100% bootstrap confidence interval for the conditional variance v ~ x! is defined here by @ v[ h ~ x! 2 q[ 12a02 , v[ h ~ x! 1 q[ a02 # where q[ p denotes the pth quantile of the bootstrap estimate of the distribution of v[ h ~ x! 2 v~ x!+ For the Markovian local resampling scheme this distribution is estimated using the distribution of v[ h* ~ x! 2 vb* ~ x! and for the block bootstrap method using the distribution of v[ h1 ~ x! 2 v[ h ~ x!+ Here v[ h1 ~ x! is the kernel estimator of v~ x! based on the replicate X 11 , X 21 , + + + , X T1 obtained using the block bootstrap method+ The bootstrap confidence intervals for the conditional mean m~ x! have been obtained in an analogous way, that is, by replacing in the preceding formulas v~ x!, v[ h ~ x!, v[ h* ~ x!, vb* ~ x!, and v[ h1 ~ x! by m~ x!, m[ h ~ x!, m[ h* ~ x!, m b* ~ x!, and m[ 1 h ~ x!, respectively, where m[ 1 h ~ x! is the kernel estimate of the conditional mean based on the block bootstrap series+ Finally, and to see the effects of varying the resampling parameters, both methods have been applied using several values of the resampling width b and of the block size B+

550

EFSTATHIOS PAPARODITIS AND DIMITRIS N . POLITIS

Table 2. Empirical coverage probabilities ~C+P+! and mean lengths of 90% bootstrap confidence intervals for the conditional variance of model ~4+8! and for the conditional mean of model ~4+9! x 5 21+5 C+P+

Length

x 5 0+0

x 5 1+5

C+P+

Length

C+P+

Length

Model (4.8) Block bootstrap B 5 10 B 5 15 B 5 20 B 5 25 Markov local bootstrap b 5 0+7 b 5 1+0 b 5 1+2 b 5 1+5

0+86 0+85 0+79 0+79

2+131 2+028 1+888 1+567

0+83 0+96 0+88 0+93

1+475 1+296 1+309 0+771

0+83 0+81 0+79 0+79

2+115 2+049 2+013 1+530

0+82 0+86 0+84 0+87

1+435 1+424 1+419 1+457

0+80 0+92 0+95 0+95

0+408 0+447 0+480 0+518

0+87 0+84 0+84 0+85

1+412 1+429 1+404 1+444

Model (4.9) Block bootstrap B 5 10 B 5 15 B 5 20 B 5 25 Markov local bootstrap b 5 0+7 b 5 1+0 b 5 1+2 b 5 1+5

0+76 0+82 0+85 0+93

0+352 0+349 0+349 0+340

0+85 0+89 0+91 0+89

0+244 0+248 0+240 0+239

0+86 0+89 0+91 0+85

0+345 0+352 0+340 0+345

0+88 0+87 0+93 0+93

0+387 0+386 0+393 0+395

0+90 0+92 0+93 0+93

0+271 0+272 0+271 0+274

0+88 0+92 0+94 0+94

0+384 0+389 0+388 0+388

As this table shows the local bootstrap procedure compares favorably with the block bootstrap procedure regarding the coverage probabilities of the confidence intervals for both classes of models considered+ In fact, both methods seem to behave very similarly taking into account the standard error !p~1 [ 2 p!0100 [ of the estimated coverage probability p+[ Apart from this overall behavior, it seems that in the case of the moving average model ~4+9! the block bootstrap method is slightly more efficient than the local resampling method in terms of the average length of the bootstrap confidence intervals obtained+ On the other hand, and in terms of the same quantity, the local bootstrap method is quite a bit more efficient than the block bootstrap method in the case of the random coefficient autoregressive process ~4+8!; this is not surprising because model ~4+8! is a true first order Markov process+

MARKOVIAN LOCAL RESAMPLING OF KERNEL ESTIMATORS

551

5. DERIVATIONS To prove Theorem 3+1 we make use of some lemmas that are established in what follows+ We begin with two lemmas concerning some asymptotic properties of the bootstrap series $X t* % and that are reproduced from and proved in Paparoditis and Politis ~1997!+ To state these lemmas we fix some additional notation+ * # x2 , + + + , For x 5 ~ x 1 , x 2 , + + + , x p ! [ R p let FT* ~x! 5 P~X t* # x 1 , X t21 * X t2p11 # x p ! be the stationary distribution function of the bootstrap Markov chain and F~x! 5 P~X t # x 1 , X t21 # x 2 , + + + , X t2p11 # x p ! the stationary distribution function of the original process $X t , t $ 1%; that is, we set F~{! 5 FYt, p ~{!+ LEMMA 5+1+ Suppose that Assumptions (A1)–(A3) and (A5) hold. Then conditionally on X 1 , X 2 , + + + , X T we have that, as T r `, sup 6FT* ~x! 2 F~x!6 r 0

a+s+

x[S

LEMMA 5+2+ Under the same assumptions as in Lemma 5.1 and conditionally on X 1 , X 2 , + + + , X T , T0 [ N exists such that for all T $ T0 the bootstrap series $X t* , t $ 1% is strongly mixing where the strong mixing coefficient a * satisfies a * ~n! # C® n for some constants C . 0 and ® [ ~0,1! that are independent of T. LEMMA 5+3+ Suppose that Assumptions (A1)–(A5) hold and let f be a realvalued, continuous function, y a fixed real number, and s a positive integer. We then have that, as T r `,

sup x[S

*

T21

h s21

( f~Xl11 !khs ~ y 2 Xl11 !Wb ~x 2 Yl, p !

l5p

T21

( Wb ~x 2 Yn, p !

n5p

2 f~ y! fXt11 6Yt, p ~ y6x!

E

*

k s ~u!du r 0

in probability. Proof+ Let N~x; f! 5 f~ y! fXt11 Yt, p ~ y, x! *k s ~u!du, NT ~x; f! 5

h s21 T

T21

f~X l11 !khs ~ y 2 X l11 !Wb ~x 2 Yl, p !, ( l5p

552

EFSTATHIOS PAPARODITIS AND DIMITRIS N . POLITIS

fbZ ~x! 5 T 21 ( T21 n5p Wb ~x 2 Yn, p !, and f ~x! 5 fYt, p ~x!+ Note that

*

T21

h s21

( f~Xl11 !khs ~ y 2 Xl11 !Wb ~x 2 Yl, p !

l5p

2 f~ y! fXt11 6Yt, p ~ y6x!

T21

#

1 fbZ ~x!

H

( Wb ~x 2 Yn, p !

n5p

6NT ~x; f! 2 N~x; f!6 1

6N~x; f!6 f ~x!

E

k s ~u!du

J

6 fbZ ~x! 2 f ~x!6 +

*

(5.10)

In the following discussion we show that supx[S 6NT ~x; f! 2 N~x; f!6 r 0 in probability+ Because supx[S 6 fbZ ~x! 2 f ~x!6 r 0 follows by the same arguments, the lemma is established using ~5+10! and ~A1!~iv!+ Consider 6NT ~x; f! 2 N~x; f!6+ For this term we have 6NT ~x; f! 2 N~x; f!6

* E f~z !k ~ y 2 z !W ~x 2 z ! f ~z , z !dz dz 2 f~ y! f ~ y, x! E k ~u!du * 1 1 h * ( f~X !k ~ y 2 X !W ~x 2 Y ! T

# h s21

s h

1

1

b

2

X t11 Yt, p

1

2

1

2

s

X t11 Yt, p T21

s21

l11

s h

l11

b

l, p

l5p

2

E

f~z 1 !khs ~ y 2 z 1 !Wb ~x 2 z 2 ! fXt11 Yt, p ~z 1 , z 2 !dz 1 dz 2

*

5 T1 ~x! 1 T2 ~x! with an obvious notation for T1~x! and T2 ~x!+ Consider first T1~x!+ Using the substitutions z 1 5 y 2 u 1 h and z 2 5 x 2 u 2 b we have T1 ~x! 5

* E~f~ y 2 u h! f 1

X t11 , Yt, p ~ y

2 u 1 h, x 2 u 2 b!

2 f~ y! fXt11 , Yt, p ~ y, x!!k s ~u 1 !W~u 2 !du 1 du 2

#

E*

*

f~ y 2 u 1 h! fXt11 , Yt, p ~ y 2 u 1 h, x 2 u 2 b! 2 f~ y! fXt11 , Yt, p ~ y, x!

*

3 k s ~u 1 !W~u 2 !du 1 du 2 5 O~max$h, b%! uniformly in x by ~A2!~ii!+ Consider next T2 ~x! and observe that because f is continuous and K is bounded and compactly supported, f~X l11 !k s ~~ y 2 X l11 !0h! is bounded+ Fur-

MARKOVIAN LOCAL RESAMPLING OF KERNEL ESTIMATORS

553

s thermore, instead of h s21 T 21 ( T21 l5p f~X l11 !kh ~ y 2 X l11 !Wb ~x 2 Yl, p ! we consider in the following discussion the asymptotically equivalent statistic R T ~ y, x! 5 s h s21 ~T 2 p!21 ( T21 l5p f~X l11 !kh ~ y 2 X l11 !Wb ~x 2 Yl, p !+ Observe that

E~R T ~ y, x!! 5 h s21

E

f~z 1 !khs ~ y 2 z 1 !Wb ~x 2 z 2 ! fXt11 Yt, p ~z 1 , z 2 !dz 1 dz 2

and that sup T2 ~x! # sup 6R T ~ y, x! 2 E~R T ~ y, x!!6 1 oP ~1!+ x[S

x[S

Thus we have to show that supx[S 6R T ~ y, x! 2 E~R T ~ y, x!!6 r 0 in probability+ To deal with this term divide S in a number of NT cubes denoted by I i, T with centers x i and length L T + We then get sup 6R T ~ y, x! 2 E~R T ~ y, x!!6 x[S

sup 6R T ~ y, x! 2 R T ~ y, x i !6

# max

1#i#NT x[SùI i, T

1 max 6R T ~ y, x i ! 2 E~R T ~ y, x i !!6 1#i#NT

1 max

sup 6E~R T ~ y, x i !! 2 E~R T ~ y, x!!6

1#i#NT x[SùI i, T

5 M1, T 1 M2, T 1 M3, T + By the Lipschitz continuity of K it is easily seen that M1, T 5 OP ~L T b 2~ p11! ! and M3, T 5 OP ~L T b 2~ p11! !+ Furthermore, P

S max 6R ~ y, x ! 2 E~R ~ y, x !!6 . L b 1#i#NT

NT

#

T

(P

i51

i

S

T

1 ~T 2 p!hb p

i

*(v

T

T21 l5p

l, T ~ y, x i !

2~ p11!

D

*.L b T

2~ p11!

D

,

where vl, T ~ y, x i ! 5 f~X l11 !k s

F

S

DS D S D S DG

y 2 X l11 x i 2 Yl, p W h b

2 E f~X l11 !k s

y 2 X l11 x i 2 Yl, p W h b

554

EFSTATHIOS PAPARODITIS AND DIMITRIS N . POLITIS

and E~vl,2 T ~ y, x!! # s 2 5 C 2 hb p with C . 0 a generic constant independent of T+ To bound this probability let first vJ l, T ~ y, x i ! 5 C 21 vl, T ~x i ! and LT 5

~log~T 2 p!! 3 !b 21p1e

!T 2 p !h 12e

for some 0 , e , 1+ Applying Proposition 1 of Doukhan ~1994, p+ 33!, with s 5 !hb p and log s 21 5 log!hb p 21 and using h0h 12e . 1 and b 21p1e0 b 21p . 1, we then have that constants a . 0 and b . 0 exist such that NT

(P i51

S

* ( vJ T21

1 ~T 2 p!hb p

l, T ~ y, x i !

l5p

S* ( S* (

*

T21

# O~NT !P

vJ l, T ~ y, x i ! .

l5p

*

T21

# O~NT !P

*. C L b

vJ l, T ~ y, x i ! .

l5p

1

T

2~ p11!

~log~T 2 p!! 3 log!hb p 21 1 C

D

!T 2 p !hb p log!hb p 21

D

~log~T 2 p!! 2 !T 2 p !hb p log!hb p 21

D

# O~NT !a exp $2b log~T 2 p!Y!C% 5 o~1!+

(5.11)

Thus for b 5 O~T T r `+

2d

! and d , 10p we get that L T 0b

~ p11!

r 0 and L T r 0 as

n

The following result deals with a well known property of the kernel estimator m[ h ~x; f!+ Its proof is omitted because it closely follows that of Lemma 5+3+ LEMMA 5+4+ Suppose that Assumptions (A1)–(A4) hold and let w be a realvalued continuous function. Then as T r ` sup 6 m[ h ~x; w! 2 m~x; w!6 r 0 x[S

in probability. Let f ~{6x! denote the conditional density fXt11 6Yt, p ~{6x! and let f Z ~ y6x! be the kernel estimator of f ~ y6x! given by T21

( kh ~ y 2 Xl11 !Wb ~x 2 Yl, p !

f Z ~ y6x! 5

l5p

T21

,

(5.12)

( Wb ~x 2 Yn, p !

n5p

where y [ S+ Note that in the notation of Lemma 5+3 and if we set f~{! [ 1 and s 5 1 we immediately get that supx[S 6 f Z ~ y6x! 2 f ~ y6x!6 r 0 in probability as T r `+

MARKOVIAN LOCAL RESAMPLING OF KERNEL ESTIMATORS

555

To simplify calculations and because the results obtained are asymptotic for T r `, we assume in the following that Yp,* p ; FT* , that is, we assume that the bootstrap series starts with its stationary distribution+ Now, for k [ N let x jk 5 ~ x jk , x jk11 , + + + , x jk1p21 ! [ X p and recall that $X t* , t $ 1% is a discrete Markov chain with state space X p + Using iteratively the relation T

P~Yt,* p 5 x j1 ! 5

* * P~X t* 5 x j 6Yt21, p 5 x j !P~Yt21, p 5 x j ! ( 5p11 1

2

2

jp11

and the transition probability ~2+3! we get the following useful expression for the stationary probability mass function of Yt,* p+ * * * P~X t* 5 x j 6 X t21 5 x j , X t22 5 x j , + + + , X t2p 5 xj ( ( {{{( j

P~Yt,* p 5 x j1 ! 5

1

jp11 jp12

2

3

p11

!

2p

* * * * 3 P~X t21 5 x j 2 6 X t22 5 x j3 , X t23 5 x j4 , + + + , X t2p21 5 x jp12 ! * * 5 x jp 6 X t2p 5 x jp11 , 3 {{{ 3 P~X t2p11 * * 5 x jp12 , + + + , X t22p11 5 x j 2p ! X t2p21 * * * 5 x jp11 , X t2p21 5 x jp12 , + + + , X t22p11 5 x j 2p ! 3 P~X t2p

E) E) p

5

* * * P~X t2i11 5 x ji 6Yt2i, p 5 x ji11 ! dFT ~x jp11 !

i51

Wb ~x ji11 2 Yji 21, p !

p

5

i51

T21

( Wb ~x ji11 2 Yn, p !

dFT* ~x jp11 !+

(5.13)

n5p

To proceed with the proof of Theorem 3+1 we first note that L*T ~x! can be expressed as L*T ~x! 5

1 5

! hT

p T21

1 fhZ * ~x!

1

! hT

p T21

1 fhZ * ~x!

fhZ * ~x!

* * ! 2 E * @f~X j11 !6Yj,*p # !Kh ~x 2 Yj,* p ! ( ~f~Xj11

j5p

* !6Yj,*p # 2 m * ~x; f!!Kh ~x 2 Yj,* p ! ( ~E * @f~Xj11

j5p

$L*1, T ~x! 1 L*2, T ~x!%,

* * where fhZ * ~x! 5 T 21 ( T21 j5p Kh ~x 2 Yj, p ! and an obvious notation for L 1, T ~x! * and L 2, T ~x!+ Using the preceding expression, the assertion of the theorem is established if we show that fhZ * ~x! r f ~x! weakly, that L*1, T ~x! converges to a Gaussian distribution with mean zero and the appropriate variance, and that L*2, T ~x!0fhZ * ~x! is a consistent estimator of the mean ~bias term! of L T ~x!+ This is established in the following three lemmas+

556

EFSTATHIOS PAPARODITIS AND DIMITRIS N . POLITIS

LEMMA 5+5+ Under Assumptions (A1)–(A5) we have conditionally on X 1 , X 2 , + + + , X T that for all x [ S, fhZ * ~x! r f ~x!

in probability.

Proof+ Let x 5 ~ x 1 , x 2 , + + + , x p !+ Using ~5+13! we have E * ~ fhZ * ~x!! 5 5

T2p T T2p T

(

Kh ~x 2 x j1 !P~Yt,* p 5 x j1 !

j1 , j 2 , + + + , jp

E

p

p

( ) ) j , j , + + + , j m51 i51 1

2

kh ~ x m 2 x jm !Wb ~x ji11 2 Yji 21, p ! T21

( Wb ~x j

p

i11

dFT* ~x jp11 !

2 Yn, p !

n5p

5

T2p T

E

p

p

( ) ) f Z ~ x 1 6x j ! j , + + + , j m52 i52

kh ~ x m 2 x jm !Wb ~x ji11 2 Yji 21, p ! T21

2

2

( Wb ~x j

p

i11

dFT* ~x jp11 !

2 Yn, p !

n5p

5

T2p T

E

p

p

( ) ) f ~ x 1 6x j ! j , + + + , j m52 i52

kh ~ x m 2 x jm !Wb ~x ji11 2 Yji 21, p ! T21

2

2

( Wb ~x j

p

i11

dFT* ~x jp11 !

2 Yn, p !

n5p

1 oP ~1!

5

T2p T

E

p

p

( ) )

(j f ~ x 1 6x j !kh ~ x 2 2 x j !Wb ~x j 2

2

3

2 Yj 221, p !

2

T21

( Wb ~x j

j 3 , + + + , jp m53 i53

3

2 Yn, p !

n5p

3

kh ~ x m 2 x jm !Wb ~x ji11 2 Yji 21, p ! T21

( Wb ~x j

i11

dFT* ~x jp11 ! 1 oP ~1!

2 Yn, p !

n5p

5

T2p T 3

E

p

p

( ) ) f ~ x16 x2 , xj , + + + , xj 3

p11

! f ~ x 2 6x j3 !

j 3 , + + + , jp m53 i53

kh ~ x m 2 x jm !Wb ~x ji11 2 Yji 21, p ! T21

( Wb ~x j

i11

dFT* ~x jp11 ! 1 oP ~1!,

2 Yn, p !

n5p

where the preceding oP ~1! terms are due to the uniform ~over x! convergence of f Z ~{6x! given in ~5+12! to f ~{6x! and Lemma 5+3+ Continuing in this way we end up with the expression

MARKOVIAN LOCAL RESAMPLING OF KERNEL ESTIMATORS

E * ~ fhZ * ~x!! 5

T2p T

E

557

f ~ x 1 6 x 2 , + + + , x p , x jp11 ! f ~ x 2 6 x 3 , + + + , x p , x jp11 , x jp12 !

3 {{{ 3 f ~ x p 6 x jp11 , + + + , x j 2p ! dFT* ~x jp11 ! 1 oP ~1! r

E

f ~ x 1 6 x 2 , + + + , x p , x jp11 ! f ~ x 2 6 x 3 , + + + , x p , x jp11 , x jp12 !

3 {{{ 3 f ~ x p 6 x jp11 , + + + , x j 2p ! dF~x jp11 ! 5 f ~x!,

(5.14)

where the last convergence is due to Lemma 5+1 and Helly–Bray’s theorem+ To evaluate the variance of fhZ * ~x!, let Z h,* j 5 Kh ~x 2 Yj,* p ! and note that * Var * @ fhZ * ~x!# 5 ~T 2 p!T 22 Var * ~Z h,1 ! T2p21

1 2T 21

(

t51

* * ~1 2 t0T !Cov * ~Z h,1 , Z h,11t !+

(5.15)

Now, h p Var * @Z h,* j # # h p (j1 , j 2 , + + + , jp Kh2 ~x 2 x j1 !P~Yt,* p 5 x j1 ! and by the same arguments as those leading to ~5+14! we get h p Var * @Z h,* j # r f ~x! *K 2 ~u!du, that is, T 21 Var * @Z h,* j # 5 OP ~T 21 h 2p !+ To handle the second term on the right hand side of ~5+15! note first that by Davydov’s lemma ~Hall and Heyde, 1980, App+! we have * * * n 20n 6Cov~Z h,1 , Z h,11t !6 # 8a * ~t!1220n ~E * 6 Z h,1 6 ! * n for some real n . 2+ Because E * 6 Z h,1 6 5 (j1 , j 2 , + + + , jp Khn ~x 2 x j1 !P~Yt,* p 5 ~12n!p x j1 ! 5 OP ~h ! by the same arguments as in establishing that ~5+14! is true, we get by Lemma 5+2 and arguments similar to those in Masry and Tjøst* * ~1 2 t0T !Cov * ~Z h,1 , Z h,11t !6 5 heim ~1995, pp+ 271–272! that 6T 21 ( T2p21 t51 21 2p oP ~T h !+ n

LEMMA 5+6+ Under Assumptions (A1)–(A5) we have conditionally on X 1 , X 2 , + + + , X T that, as T r `,

! hT

p T21

* ! 2 m b* ~Yj,* p ; f!!Kh ~ x 2 Yj,* p ! n N~0, t 2 ~x; f!! ( ~f~Xj11

in probability,

j5p

where t 2 ~x; f! 5 *K 2 ~u!du f ~x! Var @f~X t11 !6Yt, p 5 x#. Proof+ Define

! hT ~f~X p

* [ Dj11

* j11 !

2 m b* ~Yj,* p ; f!!Kh ~x 2 Yj* !

and denote by Fj * 5 s~X 1* , X 2* , + + + , X j* ! the s-algebra generated by $X 1* , * !6Yt,* p 5 x# we have X 2* , + + + , X j* %+ Because for x [ X p , m b* ~x; f! 5 E * @f~X t11

558

EFSTATHIOS PAPARODITIS AND DIMITRIS N . POLITIS

* by the Markov property that $Dj11 , Fj * % j5p, p11, + + + forms a martingale difference scheme+ Thus to establish the desired result, we apply Theorem 1 of Brown ~1971!+ By this theorem we have to show that T21

* 6Fj * # r ( E * @Dj11 j5p 2

E

K 2 ~u!du f ~x!Var @f~X t11 !6Yt, p 5 x#

in probability (5.16)

and that for every d . 0 the Linderberg condition T21

* * I ~Dj11 $ d!6Fj * # r 0 ( E * @Dj11 j5p 2

2

in probability

(5.17)

is satisfied+ To simplify notation and to stress the essentials, we focus in what follows on the case p 5 1+ The case of general p can be handled along the same lines+ Consider first ~5+16!+ Because 2

* E * @Dj11 6Fj * # 5

h T

T

* 5 X l 6 X j* ! ( ~f~Xl ! 2 m b*~Xj* ; f!! 2 Kh2 ~ x 2 Xj* !P~Xj11 l52

we have T21

( j51

2

* E * @Dj11 6 Fj * # 5

1 Th

T

( K2

j51

S D x 2 X j* h

v[ b* ~X j* ; f!,

where T

v[ b* ~z; f! 5

( ~f~Xl ! 2 m b* ~z; f!! 2 Wb ~z 2 Xl21 ! l52 +

T21

( Wb ~z 2 Xs !

s51

Let v~z; f! 5 E @~f~X t11 ! 2 m~z; f!! 2 6 X t 5 z# and note that as in the proof of Lemma 5+5, ~Th!21 ( Tj51 K 2 ~~ x 2 X j* !0h! 5 OP ~1!+ Then T

1

T

S D D ( S

* 6Fj * # 5 ( E * @Dj11 ( K2 Th j51 j51 2

1

1 Th

x 2 X j*

T

K2

j51

h

v~X j* ; f!

x 2 X j* h

~ v[ b* ~X j* ; f! 2 v~X j* ; f!!+

MARKOVIAN LOCAL RESAMPLING OF KERNEL ESTIMATORS

559

Now, because v[ b* ~z; f! 5 m b* ~z; f 2 ! 2 ~m * ~z; f!! 2 is a Nadaraya–Watson kernel estimator of v~z; f! 5 E @~f~Xt11 ! 2 m~z; f!! 2 6Xt 5 z# we have for Dh 5 @x 2 h, x 1 h# that sup 6 v[ b* ~z; f! 2 v~z; f!6 # sup 6m b* ~z; f 2 ! 2 m~z; f 2 !6 z[Dh

z[Dh

1 sup 6m b* ~z; f! 2 m~z; f!6 sup 6m b* ~z; f! 1 m~z; f!6 z[Dh

z[Dh

r0 in probability by Lemma 5+4+ Thus 1 Th

T21

( K2 j51

S D x 2 X j* h

6 v[ b* ~X j* ; f! 2 v~X j* ; f!6

# sup 6 v[ b* ~z; f! 2 v~z; f!6OP ~1! r 0+ z[Dh

To conclude the proof of ~5+16! we show that VT* ~ x; f! [

1 Th

T

( K2 j51

S D x 2 X j* h

v~X j* ; f! r t 2 ~ x; f!

(5.18)

in probability+ Taking expectation of this term with respect to the bootstrap distribution we get using the same arguments as in establishing ~5+14! that T

E * @VT* ~ x; f!# 5

E

h 21

( v~Xl ; f!~K 2 ~~ x 2 Xl !0h!!Wb ~Xm 2 Xl21 !

l52

T21

( Wb ~Xm 2 Xs !

dFT* ~X m !

s51

r

E

K 2 ~u!duf ~ x!Var @f~X t11 !6 X t 5 x# ,

where the last convergence follows using Lemma 5+3+ Furthermore, along the same lines as in the proof of ~5+15! we get that Var * @VT* ~ x; f!# 5 OP ~T 21 h 2p !+ To prove ~5+17! we use the inequality E @6 X6I ~6 X6 . C!# # C 21 E6 X6 2 where C . 0 is a constant provided the random variable X satisfies E6 X6 2 , `+ We then have for every d . 0

560

EFSTATHIOS PAPARODITIS AND DIMITRIS N . POLITIS

T21

* * I ~Dj11 $ d!6Fj * # ( E * @Dj11 2

2

j51

# 5

1 d

T21

* 6Fj * # ( E * @Dj11 j51 4

T

1 dh 2 T 2

F

* ! 2 m b* ~X j* ; f!% 4 K 4 ( E * $f~Xj11

j51

S D G x 2 X j* h

6Fj *

T

5

1 dh 2 T 2

T21

( K 4 ~~ x 2 Xj* !0h!

( ~f~Xl ! 2 m b* ~Xj* ; f!! 4 Wb ~Xj* 2 Xl21 ! l52 +

T21

(

j51

Wb ~X j*

2 Xs !

s51

Let T

R T ~z; f! 5

~f~X l ! 2 m b* ~z; f!! 4 Wb ~z 2 X l ! ( l52 ,

T21

( Wb ~z 2 Xs !

n51

which is a Nadaraya–Watson estimator of R~z; f! 5 E @~f~Xt11 ! 2 m~z; f!! 4 6Xt 5 x# , and note that supz[Dh 6R T ~z; f! 2 R~z; f!6 r 0 in probability+ Thus T21

(

2

2

* * E * @Dj11 I ~Dj11 $ d!6Fj * # #

j51

1 dT 2 h 2

T21

( K 4 ~~ x 2 Xj* !0h!R T ~Xj* ; f! 1 oP ~1!

j51

5 OP ~d 21 T 21 h 21 !, where the last equality follows by the same arguments as those used to prove n ~5+18!+ LEMMA 5+7+ Under Assumptions (A1)–(A6) and if T 102 h ~ p14!02 r Ch . 0, we have conditionally on X 1 , X 2 , + + + , X T that, as T r `,

! hT

p T21

* !6Yj,* p # 2 m b* ~x; f!!Kh ~x 2 Yj,* p ! ( ~E * @f~Xj11

j5p

r Ch

E

p

H

u 2 k~u!du ( m ~ x l ! ~x; f! fYt, pl ~x! 1 l51

~x !

1 ~x x ! m l l ~x; f! fYt, p ~x! 2

J (5.19)

in probability. * Proof+ Let m b* ~Yt,* p ; f! 5 E * @f~X t11 !6Yt,* p # and x 5 ~ x 1 , x 2 , + + + , x p !+ By a Taylor series expansion we get

MARKOVIAN LOCAL RESAMPLING OF KERNEL ESTIMATORS

! hT

561

p T21

( ~m b* ~Yj,* p ; f! 2 m b* ~x; f!!Kh ~x 2 Yj,* p !

j5p

p

5

( m b*

~ xl !

~x; f!

l51

1

1 2

p

!

p

((

l 151 l 251

m b*

hp T

T21

* 2 x l !Kh ~x 2 Yj,* p ! ( ~Xj2l11

j5p

~ x l1 x l2 !

~ x;I f!

!

hp T

T21

* * 11 2 x l !~X j2l 11 2 x l ! ( ~Xj2l 1

1

2

2

j5p

3 Kh ~x 2 Yj,* p ! for some xI between x and Yj,* p+ Because under the assumptions made !Th p ~Em[ h ~x; f! 2 m~x; f!! r B~x; f! in probability, the assertions of the lemma will be established if we show that, in probability, m b*

!T

hp

~ xl !

~x; f! r m ~ x l ! ~x; f!,

T21

* 2 x l !Kh ~x 2 Yj,* p ! r Ch ( ~Xj2l11

j5p

E

(5.20) ~x !

u 2 k~u!dufYt, pl ~x!,

(5.21)

! hT

p T21

* * * 11 2 x l !~X j2l 11 2 x l !Kh ~x 2 Yj, p ! r dl , l Ch ( ~Xj2l 1

1

2

2

1

j5p

2

E

u 2 k~u!dufYt, p ~x!,

(5.22) and sup

6m b*

~ x l1 x l2 !

~z; f! 2 mK ~ x l

1

x l2 !

~z; f!6 r 0,

(5.23)

z[@x2h, x1h#

where mK ~ x i x j ! ~z; f! 5

H

1 1 ~1 1 di, j ! fYt, p ~z! 2

E

~x x !

~x x !

f~ y! fXt11i Yj t, p ~ y, z!dy 2 m~z; f! fYt, pi j ~z! ~x !

J

~x !

2 m ~ x i ! ~z; f! fYt, pj ~z! 2 m ~ x j ! ~z; f! fYt, pi ~z! + The rest of the proof is devoted to showing that ~5+20!–~5+23! are true+ Consider ~5+20! and note that m b*

~ xl !

~x; f! 5

H

1 1 fbZ ~x! Tb p11

T21

( f~Xj11 !W ~ x ! ~~x 2 Yj, p !0b! l

j5p

2 m b* ~x; f!

1 Tb p11

T21

W ~ x ! ~~x 2 Yj, p !0b! ( j5p l

J

+

(5.24)

562

EFSTATHIOS PAPARODITIS AND DIMITRIS N . POLITIS

Recall that fbZ ~x! r fYt, p ~x! and m b* ~x; f! r m~x; f! in probability+ Furthermore, by standard arguments we have 1 Tb p11

T21

( f~Xj11 !W ~ x ! ~~x 2 Yj, p !0b! r j5p l

E

~x !

f~ y! fXt11l Yt, p ~ y, x! dy

(5.25)

and T21

1

~x ! W ~ x ! ~~x 2 Yj, p !0b! r fY ~x!+ ( j5p

Tb

(5.26)

l t, p

l

p11

~ xl !

Therefore, m b* ~x; f! r m ~ x l ! ~x; f! in probability+ The proof of ~5+21! and ~5+22! is very similar and makes explicit use of ~5+13! and arguments like those applied in the proof of Lemma 5+3+ To avoid tedious manipulations of formulas we focus on the case p 5 1+ Consider ~5+21! so we have E*

F!

h T

G

T21

~X j* 2 x!Kh ~ x 2 X j* ! ( j51

5 ~T 2 1!

!

h T

T

( ~Xl 2 x!Kh ~ x 2 Xl !P~Xt* 5 Xl !

l52

T

! E

h 5 ~T 2 1! T

( ~Xl 2 x!Kh ~ x 2 Xl !Wb ~Xs 2 Xl21 !

l52

dFT* ~X s !+

T

( Wb ~Xs 2 Xn !

n52

Now, to establish the desired result we have to show that

sup y[S

*

T

!Th ( ~Xl 2 x!Kh ~ x 2 Xl !Wb ~ y 2 Xl21 ! l52

~ x!

2 Ch

T21

( Wb ~ y 2 Xn21 !

fXt Xt21 ~ x, y! fXt ~ y!

E

u 2 K~u!du

n51

r 0+

*

(5.27)

By ~A1!~iv! and the uniform convergence of fbZ ~ y! to f ~ y!, it suffices to show that h * ! T ( ~X 2 x!K ~ x 2 X !W ~ y 2 X T

sup y[S

l

l52

r 0+

h

l

b

l21 !

~ x!

2 Ch fXt Xt21 ~ x, y!

E

u 2 K~u!du

*

(5.28)

MARKOVIAN LOCAL RESAMPLING OF KERNEL ESTIMATORS

For this note that E

F!

h T

563

G

T

( ~Xl 2 x!Kh ~ x 2 Xl !Wb ~ y 2 Xl21 !

l52

5 2~T 2 1! 5 ~T 2 1!

! E uK~u!W~r! f h3 T

! E u K~u!W~r! f h5 T

2

X t X t21 ~ x

~ x! I X t X t21 ~ x,

2 uh, y 2 rb!drdu y 2 rb!drdu,

where 6 xI 2 x6 , h+ Thus by the Lipschitz continuity of f ~ x! ~{,{! we get that

! hT E u K~u!W~r! f 5

~T 2 1!

2

~ x!

uniformly in y; that is, we have shown that

h * F ! T ( ~X 2 x!K ~ x 2 X !W ~ y 2 X T

sup E y[S

l

~ x!

E

u 2 K~u!du

~ x!

E

*

~ x,I y 2 rb!drdu r Ch fXt Xt21 ~ x, y!

h

l

b

G

l21 !

l52

r0

2 Ch fXt Xt21 ~ x, y!

u 2 K~u!du

in probability+ Thus to establish ~5+28! it remains to show that h * ! T ( $~X 2 x!K ~ x 2 X !W ~ y 2 X T

sup y[S

l

h

l

b

l21 !

l52

*

2 E~X l 2 x!Kh ~ x 2 X l !Wb ~ y 2 X l21 !% r 0 in probability+ To handle this term, we proceed as in the proof that M1, T r 0 in probability in Lemma 5+3+ In particular, let Z l, T ~ y! 5 ~X l 2 x!K

S

DS D S DS D

x 2 Xl W h

2 E~X l 2 x!K

y 2 X l21 b

x 2 Xl W h

y 2 X l21 b

and note that 6 Z l, T 6 # C1 h and that EZ 2 ~ y! # C22 h 3 b for some generic positive constants C1 and C2 + Using the splitting device applied in the proof of Lemma 5+3 we get by the Lipschitz continuity of W that sup y[S

*

1 !Thb

max * ( Z l, T ~ y! * # 1#i#N ( Z l, T ~ yi ! * 1 O~h 102 L T b22 !+ !Thb l52 l52 T

1

T

T

Along the same lines as in dealing with ~5+11! we get for the first term on the right hand side of the preceding inequality using the notation ZE l, T ~ yi ! 5 h 21 C221 Z l, T ~ yi ! and for 0 , e , 1 and L T 5 ~log~T !! 3 b 3021e h 1021e that

564

P

S

EFSTATHIOS PAPARODITIS AND DIMITRIS N . POLITIS

max 1#i#NT

*

NT

# #

*

T

1 !Thb

L T h 102 b2

( Z l, T ~ yi ! $

S S* (

l52

D

D

( P * l52 ( ZE l, T ~ yi ! * $ i51

L T T 102 C2 b

NT

~log~T !! 2 !T !hb log!hb 21 C2

T

(P i51

*

T

ZE l, T ~ yi ! $

l52

# O~NT !a exp $2b log~T !Y C2 %

D

!

for some a . 0 and b . 0+ Thus by the assumptions made on b and h we get that L T r 0, L T h 1020b 2 r 0, and O~NT !a exp $2b log~T !0C2 % r 0 as T r `+ The proof of ~5+22! follows exactly the same lines as the proof of ~5+21! and is therefore omitted+ To establish ~5+23! verify first that m b*

~ x l1 , x l2 !

5

~z; f!

H

1 1 fbZ ~z! Tb p12 2 m b* 2 m b*

T21

( f~Xj11 !W ~ x

l1 ,

x l2 !

~~z 2 Yj, p !0b!

j5p

~ x l2 !

~ x l1 !

~z; f!

1 Tb p12 ~ x l2 !

~z; f! fbZ

T21

( W ~x

l1 ,

x l2 !

~~z 2 Yj, p !0b!

j5p

~z! 2 m b*

~ x l2 !

~ x l1 !

~z; f! fbZ

J

~z! +

(5.29)

We then have to show that every element on the right hand side of ~5+29! converges uniformly to the corresponding elements on the right hand side of ~5+24!+ We show this for the first term on the right hand side of ~5+29!, the other terms being handled in exactly the same manner+ For this note that sup z[S

*

1 T

T21

( f~Xj11 !Wb~ x j5p

l1 x l2 !

~z 2 Yj, p ! 2

* T ( ~f~X !W T2p 1 sup * Ef~ y!W T

# sup

1

z[S

T21

~ x l1 x l2 ! ~z b

j11

1 ~1 1 dx l 2

1

x l2 !

E

~x 1 x 2!

~ x l1 x l2 !

2 Yj, p ! 2 Ef~X j11 !Wb

j5p

~ x l1 x l2 ! ~z b

z[S

2

1 1 dx l 2

1

x l2

E

* !! *

f~ y! fXt11l Ylt, p ~ y, z! dy ~z 2 Yj, p

2 r! fXt11 Yt, p ~ y, r! dydr

~x 1 x 2!

f~ y! fXt11l Ylt, p ~ y, z! dy

*

5 B1, T 1 B2, T , with an obvious notation for B1, T and B2, T + Now, by the same arguments used to handle the term T1~x! in the proof of Lemma 5+3 and by ~A6!, we get that B2, T 5 O~b!+ Furthermore, by the mean value theorem we have applying the

MARKOVIAN LOCAL RESAMPLING OF KERNEL ESTIMATORS

565

same splitting device as the one used in handling the term T2 ~x! in the lemma mentioned earlier, that sup z[S

* T ( ~f~X !W ~z 2 Y 1 # max * ( ~f~X !W T 1

T21

~ x l1 x l2 ! b

j11

j, p !

~ x l1 x l2 !

2 Ef~X j11 !Wb

~z 2 Yj, p !!

j5p

T21

j11

1#i#NT

~ x l1 x l2 ! ~z i b

*

~ x l1 x l2 !

2 Yj, p ! 2 Ef~X j11 !Wb

j5p

~z i 2 Yj, p !!

*

1 OP ~L T b 2~ p13! !+

To conclude the proof verify that Var~f~X j11 !W ~ x l x l ! ~~z i 2 Yj, p !0b!! # Cb p and let L T 5 ~log T ! 3 b ~ p02!111eY!T for some 0 , e , 1+ Using the notation 1

2

Cj, T ~z i ! 5 f~X j11 !W ~ x l1 x l2 ! ~~z i 2 Yj, p !0b! 2 E~f~X j11 !W ~ x l1 x l2 ! ~~z i 2 Yj, p !0b!!

and applying the exponential inequality for strongly mixing sequences used in the proof of Lemma 5+3 we then get that for some constants a . 0 and b . 0 NT

(P i51

S

1 Tb p12 NT

#

(P

i51

T21

( Cj, T ~zi ! $ L T b2~ p13!

j5p

S(

T21 j5p

C 21 Cj, T ~z i ! $

!T L T Cb

~ p02!11

# O~NT !a exp $2b log~T !Y!C% 5 o~1!+

D

log!b

p02 21

!T b p02 log!b p02 21

D n

REFERENCES Auestad, B+ & D+ Tjøstheim ~1990! Identification of nonlinear time series: First order characterization and order determination+ Biometrika 77, 669– 687+ Brown, B+M+ ~1971! Martingale central limit theorems+ Annals of Mathematical Statistics 42, 59– 66+ Doukhan, P+ ~1994! Mixing: Properties and Examples+ Lecture Notes in Statistics 85+ New York: Springer-Verlag+ Efron, B+ ~1979! Bootstrap methods: Another look at the jackknife+ Annals of Statistics 7, 1–26+ Franke, J+, J+-P+ Kreiss, & E+ Mammen ~1996! Bootstrap of Kernel Smoothing in Nonlinear Time Series+ Preprint+ Hall, P+ & C+C+ Heyde ~1980! Martingale Limit Theory and Its Applications+ San Diego: Academic Press+ Künsch, H+R+ ~1989! The jackknife and the bootstrap for general stationary observations+ Annals of Statistics 17, 1217–1241+ Liu, R+ & K+ Singh ~1992! Moving blocks jackknife and bootstrap capture weak dependence+ In R+ LePage & L+ Billard ~eds+!, Exploring the Limits of the Bootstrap, pp+ 225–248+ New York: Wiley+ Masry, E+ & D+ Tjøstheim ~1995! Nonparametric estimation and identification of nonlinear ARCH time series+ Econometric Theory 11, 258–289+ Neumann, M+ & J+-P+ Kreiss ~1998! Regression-type inference in nonparametric autoregression+ Annals of Statistics 26, 1570–1613+ Paparoditis, E+ & D+N+ Politis ~1997! The local bootstrap for Markov processes+ Journal of Statistical Planning and Inference, forthcoming+ Robinson, P+M+ ~1983! Nonparametric estimation for time series models+ Journal of Time Series Analysis 4, 185–208+

566

EFSTATHIOS PAPARODITIS AND DIMITRIS N . POLITIS

Roussas, G+ ~1990! Nonparametric regression estimation under mixing conditions+ Stochastic Processes and Their Applications 36, 107–116+ Tjøstheim, D+ & B+ Auestad ~1994! Nonparametric identification of nonlinear time series: Projections+ Journal of the American Statistical Association 89, 1398–1409+ Tong, H+ ~1990! Nonlinear Time Series: A Dynamical Approach+ New York: Oxford University Press+

Suggest Documents