Quantile$Based Nonparametric Inference for First ... - Google Sites

1 downloads 136 Views 273KB Size Report
Aug 26, 2008 - The first author gratefully acknowledges the research support of the Social ... when the second author wa
Quantile-Based Nonparametric Inference for First-Price Auctions Vadim Marmer University of British Columbia Artyom Shneyerov CIREQ, CIRANO, and Concordia University August 26, 2008

Abstract We propose a quantile-based nonparametric approach to inference on the probability density function (PDF) of the private values in …rst-price sealedbid auctions with independent private values. Our method of inference is based on a fully nonparametric kernel-based estimator of the quantiles and PDF of observable bids. Our estimator attains the optimal rate of Guerre, Perrigne, and Vuong (2000), and is also asymptotically normal with the appropriate choice of the bandwidth. As an application, we consider the problem of inference on the optimal reserve price. JEL Classi…cation: C14, D44 Keywords: First-price auctions, independent private values, nonparametric estimation, kernel estimation, quantiles, optimal reserve price We thank Donald Andrews, Herman Bierens, Chuan Goh, Christian Gourieroux, Bruce Hansen, Sung Jae Jun, Roger Koenker, Guido Kuersteiner, Isabelle Perrigne, Joris Pinkse, James Powell, Je¤rey Racine, Yixiao Sun, and Quang Vuong for helpful comments. Pai Xu provided excellent research assistance. The …rst author gratefully acknowledges the research support of the Social Sciences and Humanities Research Council of Canada under grant number 410-2007-1998. This version of the paper was completed when the second author was visiting the Center for Mathematical Studies in Economics and Management Science at Kellogg School of Management, Northwestern University. Its warm hospitality is gratefully acknowledged.

1

1

Introduction

Following the seminal article of Guerre, Perrigne, and Vuong (2000), GPV hereafter, there has been an enormous interest in nonparametric approaches to auctions.1 By removing the need to impose tight functional form assumptions, the nonparametric approach provides a more ‡exible framework for estimation and inference. Moreover, the sample sizes available for auction data can be su¢ ciently large to make the nonparametric approach empirically feasible.2 This paper contributes to this literature by providing a fully nonparametric framework for making inferences on the density of bidders’ valuations f (v). The need to estimate the density of valuations arises in a number of economic applications, as for example the problem of estimating a revenue-maximizing reserve price.3 As a starting point, we brie‡y discuss the estimator proposed in GPV. For the purpose of introduction, we adopt a simpli…ed framework. Consider a random, i.i.d. sample bil of bids in …rst-price auctions each of which has n bidders; l indexes auctions and i = 1; : : : ; n indexes bids in a given auction. GPV assume independent private values (IPV). In equilibrium, the bids are related to the valuations via the equilibrium bidding strategy B: bil = B (vil ). GPV show that the inverse bidding strategy is identi…ed directly from the observed distribution of bids: v = (b)

b+

1 n

G (b) ; 1 g (b)

(1)

where G (b) is the cumulative distribution function (CDF) of bids in an auction with n bidders, and g (b) is the corresponding density. GPV propose to use nonparametric ^ and g^. When b = bil , the left-hand side of (1) will then give what GPV estimators G call the pseudo-values v^il = ^ (bil ). The CDF F (v) is estimated as the empirical 1

See a recent survey by Athey and Haile (2007). For example, List, Daniel, and Michael (2004) study bidder collusion in timber auctions using thousands of auctions conducted in the Province of British Columbia, Canada. Samples of similar size are also available for highway procurement auctions in the United States (e.g., Krasnokutskaya (2003)). 3 This is an important real-world problem that arises in the administration of timber auctions, for example. The actual objectives of the agencies that auction timber may vary from country to country. In the United States, obtaining a fair price is the main objective of the Forest Service. As observed in Haile and Tamer (2003), this is a vague objective, and determining the revenue maximizing reserve price should be part of the cost-bene…ts analysis of the Forest Service’s policy. In other countries, maximizing the expected revenue from each and every auction is a stated objective, as is for example the case for BC Timber Sales (Roise, 2005). 2

2

CDF, and the PDF f (v) is estimated by the method of kernels, both using v^il as observations. GPV show that, with the appropriate choice of the bandwidth, their estimator converges to the true value at the optimal rate (in the minimax sense; Khasminskii (1978)). However, the asymptotic distribution of this estimator is as yet unknown, possibly because both steps of the GPV method are nonparametric with estimated values v^il entering the second stage, and because the GPV estimator applies trimming to the observations that lie too close to the boundaries of the support of the bids distribution. The estimator f^ (v) proposed in this paper avoids the use of pseudo-values and does not involve trimming; it builds instead on the insight of Haile, Hong, and Shum (2003).4 They show that the quantiles of the distribution of valuations can be expressed in terms of the quantiles, PDF, and CDF of bids. We show below that this relation can be used for estimation of f (v). Consider the -th quantile of valuations Q ( ) and the -th quantile of bids q ( ). The latter can be easily estimated from the sample by a variety of methods available in the literature. As for the quantile of valuations, since the inverse bidding strategy (b) is monotone, equation (1) implies that Q ( ) is related to q ( ) as follows: Q( ) = q( ) +

(n

1) g (q ( ))

(2)

;

providing a way to estimate Q ( ) by a plug-in method. The CDF F (v) can then be recovered simply by inverting the quantile function, F (v) = Q 1 (v). Our estimator f^ (v) is based on a simple idea that by di¤erentiating the quantile function we can recover the density: Q0 ( ) = 1=f (Q ( )), and therefore f (v) = 1=Q0 (F (v)). Taking the derivative in (2) and using the fact that q 0 ( ) = 1=g (q ( )), we obtain, after some algebra, our basic formula: f (v) =

n n

1 1 g (q (F (v)))

1 n

F (v) g 0 (q (F (v))) 1 g 3 (q (F (v)))

1

:

(3)

Note that all the quantities on the right-hand side, i.e. g (b), g 0 (b), q ( ), F (v) = Q 1 (v) can be estimated nonparametrically, for example, using kernel-based methods. Once this is done, we can plug them in (3) to obtain our nonparametric estimator. 4

The focus of Haile, Hong, and Shum (2003) is a test of common values. Their model is therefore di¤erent from the IPV model, and requires an estimator that is di¤erent from the one in GPV. See also Li, Perrigne, and Vuong (2002).

3

The expression in (3) can be also derived using the following relationship between the CDF of values and the CDF of bids: F (v) = G (B (v)) : Applying the change of variable argument to the above identity, one obtains f (v) = g (B (v)) B 0 (v) = g (B (v)) = 0 (B (v)) =

n n

1 1 g (B (v))

1 n

F (v) g 0 (B (v)) 1 g 3 (B (v))

1

:

Note however, that from the estimation perspective, the quantile-based formula appears to be more convenient, since the bidding strategy function B involves integration of F (see equation (4) below). Also, as we show below, the quantile-based approach eliminates trimming. Furthermore, replacing B (v) with appropriate quantiles has no e¤ect on the asymptotic distribution of the estimator. Our framework results in the estimator of f (v) that is both consistent and asymptotically normal, with an asymptotic variance that can be easily estimated. Moreover, we show that, with an appropriate choice of the bandwidth sequence, the proposed estimator attains the minimax rate of GPV. In a Monte Carlo experiment, we compare the performances of the two estimators, our quantile-based estimator and GPV’s by comparing their …nite sample biases, mean squared errors, and median absolute deviations. Our conclusions is that neither estimator strictly dominates the other. The GPV estimator is more e¢ cient when the PDF of valuations has a positive derivative at the point of estimation and the number of bidders tends to be large. On the other hand, the quantile-based estimator is more e¢ cient when the PDF of valuations has a negative derivative and the number of bidders is small. As an application, we consider the problem of inference on the optimal reserve price. Several previous articles have considered the problem of estimating the optimal reserve price. Paarsch (1997) develops a parametric approach and applies his estimator to timber auctions in British Columbia. Haile and Tamer (2003) consider the problem of inference in an incomplete model of English auction, derive nonparametric bounds on the reserve price and apply them to the reserve price policy in the US 4

Forest Service auctions. Closer to the subject of our paper, Li, Perrigne, and Vuong (2003) develop a semiparametric method to estimate the optimal reserve price. At a simpli…ed level, their method essentially amounts to re-formulating the problem as a maximum estimator of the seller’s expected pro…t. Strong consistency of the estimator is shown, but its asymptotic distribution is as yet unknown. In this paper, we propose asymptotic con…dence sets (CSs) for the optimal reserve price. Our CSs are formed by inverting a collection of asymptotic tests of Riley and Samuelson’s (1981) equation determining the optimal reserve price. This equation involves the density f (v), and a test statistic with an asymptotically normal distribution under the null can be constructed using our estimator. The paper is organized as follows. Section 2 introduces the basic setup. Similarly to GPV, we allow the number of bidders to vary from auctions to auction, and also allow auction-speci…c covariates. Section 3 presents our main results. Section 4 discusses inference on the optimal reserve price. We report Monte Carlo results in Section 5. All proofs are contained in the Appendix.

2

De…nitions

Suppose that the econometrician observes the random sample f(bil ; xl ; nl ) : l = 1; : : : ; L; i = 1; : : : nl g, where bil is the equilibrium bid of bidder i submitted in auction l with nl bidders, and xl is the vector of auction-speci…c covariates for auction l. The corresponding unobservable valuations of the object are given by fvil : l = 1; : : : ; L; i = 1; : : : nl g. We make the following assumption similar to Assumptions A1 and A2 of GPV (see also footnote 14 in their paper). Assumption 1 (a) f(nl ; xl ) : l = 1; : : : ; Lg are i.i.d. (b) The marginal PDF of xl , ', is strictly positive and continuous on its compact support X Rd , and admits at least R 2 continuous derivatives on its interior. (c) The distribution of nl conditional on xl is denoted by N = fn; : : : ; ng for all x 2 X , n 2.

(njx) and has support

(d) fvil : l = 1; : : : ; L; i = 1; : : : ; nl g are i.i.d. and independent of the number of bidders conditional on xl with the PDF f (vjx) and CDF F (vjx). 5

(e) f ( j ), is strictly positive and bounded away from zero on its support, a compact interval [v (x) ; v (x)] R+ , and admits at least R continuous partial derivatives on f(v; x) : v 2 (v (x) ; v (x)) ; x 2 Interior (X )g. (f) For all n 2 N , X.

(nj ) admits at least R continuous derivatives on the interior of

In the equilibrium and under Assumption 1(c), the equilibrium bids are determined by Z vil 1 (F (ujxl ))n 1 du; (4) bil = vil n 1 (F (vil jxl )) v (see, for example, GPV). Let g (bjn; x) and G (bjn; x) be the PDF and CDF of bil , conditional on both xl = x and the number of bidders nl = n. Since bil is a function of vil , xl , and F ( jxl ), the bids fbil g are also i.i.d. conditional on (nl ; xl ). Furthermore, by Proposition 1(i) and (iv) of GPV, for all n = n; : : : ; n and x 2 X , g (bjn; x) has the compact support b (n; x) ; b (n; x) for some b (n; x) < b (n; x) and admits at least R + 1 continuous bounded partial derivatives. The -th quantile of F (vjx) is de…ned as Q ( jx) = F

1

( jx)

inf fv : F (vjx) v

g:

The -th quantile of G, q ( jn; x) = G

1

( jn; x) ;

is de…ned similarly. The quantiles of the distributions F (vjx) and G (bjn; x) are related through the following conditional version of equation (2): Q ( jx) = q ( jn; x) +

(n

1) g (q ( jn; x) jn; x)

:

(5)

Note that the expression on the left-hand side does not depend on n, since, by Assumption 1(d) and as it is usually assumed in the literature, the distribution of valuations is the same regardless of the number of bidders. The true distribution of the valuations is unknown to the econometrician. Our objective is to construct a valid asymptotic inference procedure for the unknown f using the data on observable bids. Di¤erentiating (5) with respect to , we obtain the 6

following equation relating the PDF of valuations with functionals of the distribution of the bids: 1 @Q ( jx) = @ f (Q ( jx) jx) n 1 = n 1 g (q ( jn; x) jn; x)

g (1) (q ( jn; x) jn; x) ; (n 1) g 3 (q ( jn; x) jn; x)

(6)

where g (k) (bjn; x) = @ k g (bjn; x) =@bk . Substituting = F (vjx) in equation (6) and using the identity Q (F (vjx) jx) = v, we obtain the following equation that represents the PDF of valuations in terms of the quantiles, PDF and derivative of PDF of bids: 1 n 1 = f (vjx) n 1 g (q (F (vjx) jn; x) jn; x) 1 F (vjx) g (1) (q (F (vjx) jn; x) jn; x) : n 1 g 3 (q (F (vjx) jn; x) jn; x)

(7)

Note that the overidentifying restriction of the model is that f (vjx) is the same for all n. In this paper, we suggest a nonparametric estimator for the PDF of valuations based on equations (5) and (7). Such an estimator requires nonparametric estimation of the conditional CDF and quantile functions, PDF and its derivative. Let K be a kernel function. We assume that the kernel is compactly supported and of order R. Assumption 2 K is compactly supported on [ 1; 1], has at least R derivatives on R R R, the derivatives are Lipschitz, and K (u) du = 1, uk K (u) du = 0 for k = 1; : : : ; R 1. To save on notation, denote Kh (z) =

1 z K , h h

and for x = (x1 ; : : : ; xd )0 , de…ne K h (x) =

1 x 1 Q xk Kd = d dk=1 K : d h h h h 7

Consider the following estimators: 1X ' ^ (x) = K h (xl L l=1 L

(8)

x) ;

X 1 1 (nl = n) K h (xl ^ (njx) = ' ^ (x) L l=1 L

x) ;

n

l XX 1 1 (nl = n) 1 (bil ^ (njx) ' ^ (x) nL l=1 i=1 n o 1 ^ ^ q^ ( jn; x) = G ( jn; x) inf b : G (bjn; x) ;

L

^ (bjn; x) = G

b) K h (xl

x) ;

b

g^ (bjn; x) =

1 ^ (njx) ' ^ (x) nL n L l XX 1 (nl = n) Kh (bil

b) K h (xl

(9)

x) ;

l=1 i=1

where 1 (S) is an indicator function of a set S R.5;6 The derivatives of the density g (bjn; x) are estimated simply by the derivatives of g^ (bjn; x): (k)

g^

( 1)k (bjn; x) = ^ (njx) ' ^ (x) nL n L l XX (k) 1 (nl = n) Kh (bil

b) K h (xl

x) ;

(10)

l=1 i=1

(k)

1 where Kh (u) = h1+k K (k) (u=h), k = 0; : : : ; R, and K (0) (u) = K (u). Our approach also requires nonparametric estimation of Q, the conditional quantile function of valuations. An estimator for Q can be constructed using the relationship between Q, q and g given in (5). A similar estimator was proposed by Haile, Hong, and Shum (2003) in a related context. In our case, the estimator of Q will be used to construct F^ , an estimator of the conditional CDF of valuations. Since F is 5

We estimate the CDF of bids by a conditional version of the empirical CDF. In a recent paper, Li and Racine (2008) discuss a smooth estimator of the CDF (and a corresponding quantile estimator) obtained by integrating the kernel PDF estimator. We, however, adopt the non-smooth empirical CDF approach in order for our estimator to be comparable with that of GPV; both estimator can be modi…ed by using the smooth conditional CDF estimator. 6 The quantile estimator q^ is constructed by inverting the estimator of the conditional CDF of bids. This approach is similar to that of Matzkin (2003).

8

related to Q through F (vjx) = Q

1

(vjx) = sup f : Q ( jx)

(11)

vg ;

2[0;1]

F^ can be obtained by inverting the estimator of the conditional quantile function. However, since an estimator of Q based on (5) involves kernel estimation of the PDF g, it will be inconsistent for the values of that are close to zero and one. In particular, such an estimator can exhibit large oscillations for near one taking on very small values, which, due to supremum in (11), might proliferate and bring an upward bias into the estimator of F . A possible solution to this problem that we pursue in this paper is to use a monotone version of the estimator of Q. First, we ^ p: de…ne a preliminary estimator, Q ^ p ( jn; x) = q^ ( jn; x) + Q

(n

1) g^ (^ q ( jn; x) jn; x)

Next, pick 0 su¢ ciently far from 0 and 1, for example, monotone version of the estimator of Q as follows. ^ ( jn; x) = Q

(

^ p (tjn; x) ; 0 supt2[ 0 ; ] Q ^ p (tjn; x) ; 0 inf t2[ ; 0 ] Q

0

:

(12)

= 1=2. We de…ne a

< 1; < 0:

(13)

^ ( jn; x) is given The estimator of the conditional CDF of the valuations based on Q by n o ^ ^ F (vjn; x) = sup : Q ( jn; x) v : (14) 2[0;1]

^ ( jn; x) is monotone, F^ is not a¤ected by Q ^ p ( jn; x) taking on small values Since Q ^ ( jn; x) near the near = 1. Furthermore, in our framework, inconsistency of Q boundaries does not pose a problem, since we are interested in estimating F only on a compact inner subset of its support. Using (7), for a given an we propose to estimate f (vjx) by the plug-in method, i.e. by replacing g (bjn; x), q ( jn; x), and F (vjx) in (7) with g^ (bjn; x), q^ ( jn; x), and F^ (vjn; x). That is our estimator f^ (vjn; x) is given by the reciprocal of n n

1 1 g^ q^ F^ (vjn; x) jn; x jn; x 9

F^ (vjn; x) g^(1) q^ F^ (vjn; x) jn; x jn; x

1 n

1

g^3 q^ F^ (vjn; x) jn; x jn; x

:

(15)

While the PDF of valuations does not depend on the number of bidders n, the estimator de…ned by (15) does, and therefore we have a series of estimators for f (vjx): f^ (vjn; x), n = n; : : : ; n. The estimators f^ (vjn; x) ; : : : ; f^ (vjn; x) can be averaged to obtain: n X f^ (vjx) = w^ (n; x) f^ (vjn; x) ; (16) n=n

where the weights w^ (n; x) satisfy

n X

w^ (n; x) !p w (n; x) > 0; w (n; x) = 1:

n=n

In the next section, we discuss how to construct optimal weights that minimize the asymptotic variance of f^ (vjx). We also suggest estimating the conditional CDF of v using the average of F^ (vjn; x), n = n; : : : ; n: n X ^ w^ (n; x) F^ (vjn; x) : (17) F (vjx) = n=n

3

Asymptotic properties

In this section, we discuss uniform consistency and asymptotic normality of the estimator of f proposed in the previous section. The consistency of the estimator of f follows from the uniform consistency of its components. It is well known that kernel estimators can be inconsistent near the boundaries of the support, and therefore we estimate the PDF of valuations at the points that lie away from the boundaries of [v (x) ; v (x)]. The econometrician can choose quantile values 1 and 2 such that 0 < 1 < 2 < 1; in order to cut o¤ the boundaries of the support where estimation is problematic. While v (x) and v (x) are unknown, consider instead the following interval of v’s for 10

selected

1

and

2:

^ (x) =

n h \

n=n

i ^ ( 1 jn; x) ; Q ^ ( 2 jn; x) : Q

(18)

The interval ^ (x) estimates (x) = [Q ( 1 jx) ; Q ( 2 jx)], the interval between the 1 and 2 quantiles of the distribution of bidders’ valuations. As we show below, our estimator of f is uniformly consistent and asymptotically normal when f is estimated at the points from ^ (x). In practice, 1 and 2 can be selected as follows. Since by Assumption 2 the length of the support of K is two, and following the discussion on page 531 of GPV, in the case with no covariates one can choose 1 and 2 such that [^ q ( 1 jn) ; q^ ( 2 jn)]

[bmin (n) + h; bmax (n)

h]

for all n 2 N , where bmin (n) and bmax (n) denote the minimum and maximum bids in the auctions with n bidders respectively. When there are covariates available and f is estimated conditional on xl = x, one can replace bmin (n) and bmax (n) with the corresponding minimum and maximum bids in the neighborhood of x as de…ned on page 541 of GPV. First, we present the following lemma which provides uniform convergence rates for the components of the estimator f^ on appropriate intervals. Since the bidding function is monotone, by Proposition 2.1 of GPV, there is an inner compact interval of the support of the bids distribution, say [b1 (n; x) ; b2 (n; x)], such that [q ( 1 jn; x) ; q ( 2 jn; x)]

[b1 (n; x) ; b2 (n; x)]

b (n; x) ; b (n; x) :

(19)

Note that the knowledge of [b1 (n; x) ; b2 (n; x)] is not required for construction of our estimator. Lemma 1 Under Assumptions 1 and 2, for all x 2 Interior (X ) and n 2 N , (a) ^ (njx) (b) ' ^ (x)

Lhd log L

(njx) = Op ' (x) = Op

Lhd log L

^ (bjn; x) (c) supb2[b(n;x);b(n;x)] jG

1=2

1=2

+ hR .

+ hR .

G (bjn; x) j = Op 11

Lhd log L

1=2

+ hR .

(d) sup (e) sup

2[";1 "]

j^ q ( jn; x)

2[0;1] (limt#

Lhd log L

q ( jn; x) j = Op

q^ (tjn; x)

q^ ( jn; x)) = Op

(f) supb2[b1 (n;x);b2 (n;x)] j^ g (k) (bjn; x)

1=2

+ hR , for all 0 < " < 1=2. 1

Lhd log(Lhd )

. Lhd+1+2k log L

g (k) (bjn; x) j = Op

1=2

+ hR , k =

0; : : : ; R, where [b1 (n; x) ; b2 (n; x)] is de…ned in (19). (g) sup

2[

1

";

such that

2 +"]

1

^ ( jn; x) jQ " > 0 and

(h) supv2 ^ (x) jF^ (vjn; x)

Q ( jx) j = Op 2

Lhd+1 log L

1=2

+ hR , for some " > 0

+ " < 1. 1=2

Lhd+1 log L

F (vjx) j = Op

+ hR , where ^ (x) is de…ned

in (18). As it follows from Lemma 1, the estimator of the derivative of g ( jn; x) has the slowest rate of convergence among all components of f^. Consequently, it determines the uniform convergence rate of f^. Theorem 1 Let ^ (x) be de…ned by (18). Under Assumptions 1 and 2, and for all x 2 Interior (X ), supv2 ^ (x) f^ (vjx)

f (vjx) = Op

Lhd+3 log L

1=2

+ hR .

Remarks. 1. The theorem also holds when ^ (x) is replaced by an inner closed subset of [v (x) ; v (x)], as in Theorem 3 of GPV. 2. One of the implications of theorem is that our estimator achieves the optimal rate of GPV. Consider the following choice of the bandwidth parameter: h = 1=2 and hR are of the same orc (L= log L) . By choosing so that Lhd+3 = log L der, one obtains = 1= (d + 3 + 2R) and the rate (L= log L) R=(d+3+2R) , which is the same as the optimal rate established in Theorem 3 of GPV. Next, we discuss asymptotic normality of the proposed estimator. We make following assumption. Assumption 3 Lhd+1 ! 1, and Lhd+1+2k

1=2

hR ! 0.

The rate of convergence and asymptotic variance of the estimator of f are determined by g^(1) (bjn; x), the component with the slowest rate of convergence. Hence, 12

Assumption 3 will be imposed with k = 1 which limits the possible choices of the bandwidth for kernel estimation. For example, if one follows the rule h = cL , then has to be in the interval (1= (d + 3 + 2R) ; 1= (d + 1)). As usual for asymptotic normality, there is some under smoothing relative to the optimal rate. Lemma 2 Let [b1 (n; x) ; b2 (n; x)] be as in (19). Then, under Assumptions 1-3, for all b 2 [b1 (n; x) ; b2 (n; x)], x 2 Interior (X ), and n 2 N , (a) Lhd+1+2k

1=2

g^(k) (bjn; x)

g (k) (bjn; x) !d N (0; Vg;k (b; n; x)), where

Vg;k (b; n; x) = Kk g (bjn; x) = (n (njx) ' (x)) ; and Kk =

R

K 2 (u) du

d

R

K (k) (u)

2

du.

(b) g^(k) (bjn1 ; x) and g^(k) (bjn2 ; x) are asymptotically independent for all n1 6= n2 , n1; n2 2 N . Now, we present the main result of the paper. By (55) in the Appendix, we the following decomposition: f^ (vjn; x)

f (vjx) =

F (vjx) f 2 (vjx) 1) g 3 (q (F (vjx) jn; x) jn; x)

(n

g^(1) (q (F (vjx) jn; x) jn; x)

g (1) (q (F (vjx) jn; x) jn; x) + op

Lhd+3

1=2

: (20)

Lemma 2, de…nition of f^ (vjn; x), and the decomposition in (20) lead to the following theorem. Theorem 2 Let ^ (x) be de…ned by (18). Under Assumptions 1, 2, and 3 with k = 1, for v 2 ^ (x), x 2 Interior (X ), and n 2 N , Lhd+3

1=2

f^ (vjn; x)

f (vjx) !d N (0; Vf (v; n; x)) ;

where Vf (v; n; x) is given by

n (n

1)2

K1 F 2 (vjx) f 4 (vjx) ; (njx) ' (x) g 5 (q (F (vjx) jn; x) jn; x) 13

and K1 is de…ned in Lemma 2. Furthermore, f^ (vjn; x) ; : : : ; f^ (vjn; x) are independent. Remarks. 1. The theorem also holds for …xed v’s in an inner closed subset of [v (x) ; v (x)]. 2. Our approach can be used for estimation of the conditional PDF of values at quantile , f (Q ( jx)). The estimator, say f^ (Q ( jx) jn; x), is then given by f^ (Q ( jx) jn; x) = and Lhd+3

1=2

n n

1 1 g^ (^ q ( jn; x) jn; x)

f^ (Q ( jx) jn; x)

1 n

g^(1) (^ q ( jn; x) jn; x) 1 g^3 (^ q ( jn; x) jn; x)

1

;

f (Q ( jx) jx) !d N (0; Vf (Q ( jx) ; n; x)).

By Lemma 1, the asymptotic variance Vf (v; n; x) can be consistently estimated by the plug-in estimator which replaces the unknown F; f; '; ; g, and q in the expression for Vf (v; n; x) with their consistent estimators. Using asymptotic independence of f^ (vjn; x) ; : : : ; f^ (vjn; x), the optimal weights for the averaged PDF estimator of f (vjx) in (16) can be obtained by solving the GLS problem. As usual, the optimal weights are inversely related to the variance Vf (v; n; x): w^ (n; x) =

1=V^f (v; n; x) =

n X j=n

!

1=V^f (v; j; x)

n (n 1)2 ^ (njx) g^5 q^ F^ (vjn; x) jn; x jn; x = P ; n 2 5 q ^ (vjn; x) jj; x jj; x j (j 1) ^ (jjx) g ^ ^ F j=n

and the asymptotic variance of the optimal weighted estimator is therefore given by Vf (v; x) = Pn

n=n

n (n

K1 F 2 (vjx) f 4 (vjx) : 1)2 (njx) g 5 (q (F (vjx) jn; x) jn; x)

In small samples, accuracy of the normal approximation can be improved by taking into the account the variance of the second-order term multiplied by h2 . To make the notation simple, consider the case of a single n. We can expand the decomposition

14

in (20) to obtain that Lhd+3 Ff2 Lhd+3 3 (n 1) g

1=2

g^(1)

1=2

f^ (vjx; n)

g (1) + h

3f g

f (vjx) is given by 2nf 2 (n 1) g 2

Lhd

1=2

(^ g

g) + op (h) ;

where, F is the conditional CDF evaluated at v, and g, g (1) , g^, g^(1) are the conditional density (given x and n), its derivative, and their estimators evaluated at q (F (vjx) jn; x). With this decomposition, in practice, one can improve accuracy of asymptotic approximation by using the following expression for the estimated variance instead of V^f alone7 : V~f = V^f + h2

3f^ g^

2nf^2 (n 1) g^2

!2

V^g;0 :

Note that the second summand in the expression for V~f is Op (h2 ) and negligible in large samples.

4

Inference on the optimal reserve price

In this section, we discuss inference on the optimal reserve price given x, r (x). Riley and Samuelson (1981) show that under certain assumptions, r (x) is given by the unique solution to the equation: r (x)

1

F (r (x) jx) f (r (x) jx)

c = 0;

(21)

where c is the seller’s own valuation. One approach to the inference on r (x) is to estimate it as a solution r^ (x) to (21) using consistent estimators for f and F in place of the true unknown functions. However, a di¢ culty arises because, even though our estimator f^ (vjx) is asymptotically normal, it is not guaranteed to be a continuous function of v. We instead take a direct approach and construct CSs that do not require a point estimate of r (x). As discussed in Chapter 3.5 of Lehmann and Romano (2005), a natural CS for a parameter can be obtained by inverting a test of a series of simple hypotheses 7

There is no covariance term because

R

K (u) K (1) (u) du = 0.

15

concerning the value of that parameter.8 We construct CSs for the optimal reserve price by inverting the test of the null hypotheses H0 (v) : r (x) = v. Such hypotheses can be tested by testing the optimal reserve price restriction (21) at r (x) = v. Thus, the CSs are formed by collecting all values v for which the test fails to rejects the null that (21) holds at r (x) = v. Consider H0 (v) : r (x) = v, and the following test statistic:

T (vjx) = Lhd+3

1=2

v

1

F^ (vjx) f^ (vjx)

v u ! u u 1 c =t

F^ (vjx) f^4 (vjx)

2

V^f (v; x);

where F^ is de…ned in (17), and V^f (v; x) is a consistent plug-in estimator of the asymptotic variance of f^ (vjx). By Theorem 2 and Lemma 1(h), T (r (x) jx) !d N (0; 1). Furthermore, due to uniqueness of the solution to (21), for any t > 0, P (jT (vjx)j > tjr (x) 6= v) ! 1. A CS for r with the asymptotic coverage probability 1 is formed by collecting all v’s such that a test based on T (vjx) fails to reject the null at the signi…cance level : CS1

n (x) = v 2 ^ (x) : jT (vjx)j

z1

=2

o

;

where z is the quantile of the standard normal distribution. Asymptotically CS1 (x) has a correct coverage probability since by construction we have that P (r (x) 2 CS1 (x)) = P jT (r (x) jx)j z1 =2 ! 1 , provided that r (x) 2 (x) = [Q ( 1 jx) ; Q ( 2 jx)]. When the seller’s own evaluation c is unknown, using the above approach one can construct conditional CSs for given values of c.

5

Monte Carlo experiments

In this section, we evaluate the accuracy of the asymptotic normal approximation for our estimator f^ established in Theorem 2. In particular, it is interesting to see 8

CSs obtained by test inversion have been used in the econometrics literature, for example, in the context of instrumental variable regression with weak instruments (Andrews and Stock, 2005), for constructing CSs for the date of a structural break (Elliott and Müller, 2007), and in the case of set identi…ed models (Chernozhukov, Hong, and Tamer, 2007); see also the references on page 1268 of Chernozhukov, Hong, and Tamer (2007).

16

whether the boundary e¤ect can create substantial size distortions. We also compare the performance of our estimator with that of GPV in terms of bias, mean squared error (MSE), and median absolute deviation. In our experiment, we consider the case with no covariates (d = 0). The true CDF of valuations is given by 8 > v < 0; < 0; F (v) = v ; 0 v 1; > : 1; v > 1;

(22)

where > 0. Such a choice of F is convenient because, in this case, the bidding strategy is given by 1 v (23) B (v) = 1 (n 1) + 1 and easy to compute. In our simulations, we consider = 1=2; 1, and 2; when = 1, the distribution of valuations is uniform over the interval [0; 1], = 1=2 corresponds to the case of a downward-sloping PDF, and = 2 corresponds to the upward-sloping PDF. The number of bidders n = 2; 3; 4; 5; 6; 7; the number of auctions L, is determined in such a way so that the total number of observations available nL is constant, and therefore any observed di¤erences in the estimation accuracy across n’s are not due to the varying sample size. We consider nL = 4200. We estimate f at the following points: v = 0:2; 0:3; 0:4; 0:5; 0:6; 0:7; 0:8. Each Monte Carlo experiment has 103 replications. Similarly to GPV, we use the tri-weight kernel for all kernel estimators. We choose the bandwidth as follows. The MSE optimal bandwidth for kernel density estimation is of order L 1=5 , and it is L 1=7 for estimation of the density derivative (Pagan and Ullah, 1999, Page 56). Therefore, as in GPV, we use the normal rule-of-thumb bandwidth to estimate g: h1 = 1:06^ b (nL) 1=5 ; where ^ b is the estimated standard deviation of bids, to estimate the PDF g; and the density derivative g (1) is estimated with h2 = 1:06^ b (nL)

17

1=7

:

For each replication, we generate randomly nL valuations, fvi : i = 1; : : : ; nLg, from the distribution described by (22), and then compute the corresponding bids according to (23). Computation of the quantile-based estimator f^ (v) involves several steps. First, we estimate q ( ), the quantile function of bids. Let b(1) ; : : : ; b(nL) i = b(i) . Second, we estimate denote the ordered sample of bids. We set q^ nL g (b), the PDF of bids using (9). To construct our estimator, g needs to be estii mated at all points q^ nL n o : i = 1; : : : ; nL . Given the estimates of g^, we compute ^ p i : i = 1; : : : ; nL using (12), its monotone version according to (13), and Q nL

F^ (v) according to (14). Let dxe denote the nearest integer greater than or equal dnLF^ (v)e . Next, we compute g^ q^ F^ (v) to x; we compute q^ F^ (v) as q^ and nL g^(1) q^ F^ (v) using (9) and (10) respectively, and f^ (v) as the reciprocal of (15). Lastly, we compute the second-order corrected estimator of the asymptotic variance of f^ (v), 0

3f^ (v) V~f (v) = V^f (v) + h22 @ g^ q^ F^ (v) V^f (v) =

K1 F^ 2 (v) f^4 (v) n (n 1)2 g^5 q^ F^ (v)

(n

2nf^2 (v) 1) g^2 q^ F^ (v)

12

A V^g;0 q^ F^ (v)

:

A con…dence interval (CI) with the asymptotic con…dence level 1 f^ (v)

;

z1

=2

q

is formed as

V~f (v) = (Lh32 );

where z is the quantile of the standard normal distribution. Table 1 reports simulated coverage probabilities for 99%, 95%, and 90% asymptotic CIs in the case of Uniform [0; 1] distribution ( = 1). We observe deviation of the simulated coverage probabilities from the nominal values when the PDF is estimated near the upper boundary and the number of bidders is small (n = 2; 3). There is also some deviation of the simulated coverage probabilities from the nominal values for large n and v near the lower boundary of the support. Thus, as one can expect the normal approximation may breakdown near the boundaries of the support. However, away from the boundaries, as the results in Table 1 indicate, the normal approximation works well and the simulated coverage probabilities are close to their 18

nominal values. Similar result have been obtained in the case of = 2 (Table 2) and = 1=2 (Table 3). When = 2, the boundary e¤ect distorting coverage probabilities is somewhat more pronounced near the lower boundary of the support, and less so near the upper boundary. An opposite situation is observed for = 1=2: we see more distortion of coverage probabilities near the upper boundary and less near the lower boundary of the support. This can be explained by the fact that the PDF is increasing in the case of = 2, so there is relatively more mass near v = 1, and it is decreasing when = 1=2, so there is relatively less mass near v = 0. We observe good coverage probabilities away from the boundaries. Next, we compare the performance of our estimator with that of GPV. In their simulations, GPV use the bandwidths of order (nL) 1=5 in the …rst and second steps of estimation. We …nd, however, that using a bandwidth of order (nL) 1=7 in the second step signi…cantly improves the performance of their estimator in terms of bias and MSE. To compute the GPV estimator, we therefore use h1 as the …rst step bandwidth, and h2 at the second step. Similarly to the quantile-based estimator, the GPV estimator is implemented with the tri-weight kernel. To compute the GPV estimator of f (v), in the …rst step we compute nonparametric estimators of G and g, and obtain the pseudo-valuations v^il according to equation (1), with G and g replaced by their estimators. In the second step, we estimate f (v) by the kernel method from the sample f^ vil g obtained in the …rst-step. To avoid the boundary bias e¤ect, GPV suggest trimming the observations that are too close to the estimated boundary of the support. Note that no explicit trimming is necessary for our estimator, since implicit trimming occurs from our use of quantiles instead of pseudo-valuations.9 Table 4 reports bias, MSE, and median absolute deviation of the two estimators for = 1. In the majority of cases, the GPV estimator shows less bias; however neither estimator dominates the other in terms of MSE or median absolute deviation: our quantile-based (QB) estimator appears to be more e¢ cient for small numbers of bidders (n = 2; 3; 4), and GPV’s is more e¢ cient when n = 5; 6, and 7. The GPV is 9 In our simulations, we found that in practice trimming had no e¤ect on the GPV estimator: essentially the same estimates were obtained with or without the trimming. On the other hand, for our quantile-based estimator, monotonization of the estimated quantile function of valuations is very important in order to obtain reasonable estimates of the PDF of valuations near the upper boundary of the support.

19

relatively more e¢ cient when the PDF is upward sloping ( = 2) as the results Table 5 indicate. However, according to the results in Table 6, the QB estimator dominates GPV’s in the majority of cases when the PDF is downward-sloping ( = 1=2). To summarize our …ndings, the GPV estimator shows less bias which can be due to the fact that it is obtained by kernel smoothing of the data, while the QB estimator is a nonlinear function of the estimated CDF, PDF and its derivative. There is no strictly dominating estimator, and the relative e¢ ciency depends on the underlying distribution of the valuations and the number of bidders in the auction. The GPV estimator is more e¢ cient when the number of bidders is relatively large and PDF has a positive slope; on the other hand, the QB estimator is more attractive when the number of bidders is small and the PDF has a negative slope. Tables 4, 5, and 6 also report the average (across replications) standard error for our QB estimator. The variance of the estimator increases with v, since it depends on F (v). This fact is also re‡ected in the MSE values that increase with v. Interestingly, one can see the same pattern for the MSE of the GPV estimator, which suggests that the GPV variance depends on v as well.

Appendix of proofs Proof of Lemma 1. Parts (a) and (b) of the lemma follow from Lemma B.3 of Newey (1994). For part (c), de…ne a function G0 (b; n; x) = n (njx) G (bjn; x) ' (x) ; and its estimator as n

l 1 XX 0 ^ 1 (nl = n) 1 (bil G (b; n; x) = L l=1 i=1

L

b) K h (xl

x) :

Next, ^ 0 (b; n; x) = E EG

1 (nl = n) K h (xl

x)

nl X

1 (bil

i=1

= nE (1 (nl = n) 1 (bil 20

b) K h (xl

x))

!

b)

(24)

= nE ( (njxl ) G (bjn; xl ) K h (xl x)) Z = n (nju) G (bjn; u) K h (u x) ' (u) du Z = G0 (b; n; x + hu) Kd (u) du: By Assumption 1(e) and Proposition 1(iii) of GPV, G (bjn; ) admits at least R + 1 continuous bounded derivatives. Then, as in the proof of Lemma B.2 of Newey (1994), there exists a constant c > 0 such that ^ 0 (b; n; x) G0 (b; n; x) E G Z R ch jKd (u)j kukR du vec DxR G0 (b; n; x)

;

where k k denotes the Euclidean norm, and DxR G0 denotes the R-th partial derivative of G0 with respect to x. It follows then that ^ 0 (b; n; x) = O hR : EG

G0 (b; n; x)

sup

(25)

b2[b(n;x);b(n;x)]

Now, we show that ^0

sup b2[b(n;x);b(n;x)]

jG (b; n; x)

^0

E G (b; n; x) j = Op

Lhd log L

1=2

!

:

(26)

We follow the approach of Pollard (1984). Fix n 2 N and x 2 Interior (X ), and consider a class of functions Z indexed by h and b, with a representative function zl (b; n; x) =

nl X

1 (nl = n) 1 (bil

b) hd K h (xl

x) :

i=1

By the result in Pollard (1984) (Problem 28), the class Z has polynomial discrimination. Theorem 37 in Pollard (1984) (see also Example 38) implies that for any 2 sequences L , L such that L 2L 2L = log L ! 1, Ezl2 (b; n; x) L, 1X zl (b; n; x) L l=1 b2[b(n;x);b(n;x)] L

2

1 L

L

sup

j

21

Ezl (b; n; x) j ! 0

(27)

almost surely. We claim that this implies that sup b2[b(n;x);b(n;x)]

^ 0 (b; n; x) jG

Lhd log L

^ 0 (b; n; x) j = Op EG

1=2

!

:

The proof is by contradiction. Suppose not. Then there exist a sequence and a subsequence of L such that along this subsequence, sup b2[b(n;x);b(n;x)]

^ 0 (b; n; x) jG

^ 0 (b; n; x) j EG

L

Lhd log L

L

!1

1=2

(28)

:

on a set of events 0 with a positive probability measure. Now if we let 2L = hd 1=2 Lhd ) 1=2 L , then the de…nition of z implies that, along the subsequence, and L = ( log L on a set of events 0 , 1X zl (b; n; x) j sup L l=1 b2[b(n;x);b(n;x)] L

2

1 L

L

1=2

=

Lhd log L Lhd log L

1=2

=

Lhd log L

1=2

=

1=2 L

1=2 L

h

d

1=2 L

L

1X sup j zl (b; n; x) L l=1 b2[b(n;x);b(n;x)] sup

b2[b(n;x);b(n;x)] 1=2 L

Ezl (b; n; x) j

Lhd log L

L

^ 0 (b; n; x) jG

Ezl (b; n; x) j

^ 0 (b; n; x) j EG

1=2

! 1;

where the inequality follows by (28), a contradiction to (27). This establishes (26), so that (25), (26) and the triangle inequality together imply that sup b2[b(n;x);b(n;x)]

^ 0 (b; n; x) jG

G0 (b; n; x) j = Op

Lhd log L

1=2

+ hR

!

:

(29)

^ 0 (b; n; x), To complete the proof, recall that, from the de…nitions of G0 (b; n; x) and G G (bjn; x) =

^0 G0 (b; n; x) ^ (bjn; x) = G (b; n; x) ; ; and G (njx) ' (x) ^ (njx) ' ^ (x) 22

^ (bjn; x) so that by the mean-value theorem, G

G (bjn; x) is bounded by

~ 0 (b; n; x) ~ 0 (b; n; x) 1 G G ; ; ~ (n; x) ' ~ (x) ~ 2 (n; x) ' ~ (x) ~ (n; x) ' ~ 2 (x) ^ 0 (b; n; x) G

G0 (b; n; x) ; ^ (njx)

!

(njx) ; ' ^ (x)

' (x)

;

(30)

~ 0 G0 ; ~ ^ 0 G0 ; ^ where G ;' ~ ' G ;' ^ ' for all (b; n; x). Further, by Assumption 1(b) and (c) and the results in parts (a) and (b) of the lemma, with the probability approaching one ~ and ' ~ are bounded away from zero. The desired result follows from (29), (30) and parts (a) and (b) of the lemma. ^ ( jn; x) is monotone by construction, For part (d) of the lemma, since G n ^ (bjn; x) P (^ q ("jn; x) < b (n; x)) = P inf b : G b

^ (b (n; x) jn; x) > " = P G

o " < b (n; x)

= o (1) ; where the last equality is by the result in part (c). Similarly, "jn; x) > b (n; x)

P q^ (1

^ b (n; x) jn; x < 1 = P G

"

= o (1) : Hence, for all x 2 Interior (X ) and n 2 N , with the probability approaching one, b (n; x) q^ ("jn; x) < q^ (1 "jn; x) b (n; x). Since the distribution G (bjn; x) is continuous in b, G (q ( jn; x) jn; x) = , and, for 2 [ 1 ; 2 ], we can write the identity G (^ q ( jn; x) jn; x)

G (q ( jn; x) jn; x) = G (^ q ( jn; x) jn; x)

:

(31)

Using Lemma 21.1(ii) of van der Vaart (1998), 0

1 ; ^ (njx) ' ^ (x) nLhd

^ (^ G q ( jn; x) jn; x)

and by the results in (a) and (b), ^ (^ G q ( jn; x) jn; x) = 23

+ Op

Lhd

1

(32)

uniformly over . Combining (31) and (32), and applying the mean-value theorem to the left-hand side of (31), we obtain q^ ( jn; x) =

q ( jn; x)

^ (^ G (^ q ( jn; x) jn; x) G q ( jn; x) jn; x) + Op g (e q ( jn; x) jn; x)

1

Lhd

;

(33)

where qe lies between q^ and q for all ( ; n; x). Now, according to Proposition 1(ii) of GPV, there exists cg > 0 such that g (bjn; x) > cg for all b 2 b (n; x) ; b (n; x) , and the result in part (d) follows from (33) and part (c) of the lemma. Next, we prove part (e) of the lemma. Fix x 2 Interior (X ) and n 2 N . Let N=

nl L X X

1 (nl = n) Kd (xl ) :

l=1 i=1

Consider the ordered sample of bids b (n; x) = b(0) corresponds to nl = n and Kd (xl ) 6= 0. Then, 0

lim q^ (tjn; x) t#

q^ ( jn; x)

max

j=1;:::;N +1

:::

b(N +1) = b (n; x) that

b(j)

b(j

1)

:

By the results of Deheuvels (1984), max

j=1;:::;N +1

b(j)

b(j

1)

= Op

N log N

1

!

;

and part (e) follows, since N = Op Lhd . To prove part (f), note that by Assumption 1(f) and Proposition 1(iv) of GPV, g ( jn; ) admits at least R + 1 continuous bounded partial derivatives. Let (k)

g0 (b; n; x) =

(njx) g (k) (bjn; x) ' (x) ;

(34)

and de…ne n

l 1 XX (k) 1 (nl = n) Kh (bil nL l=1 i=1

L

(k)

g^0 (b; n; x) =

24

b) K h (xl

x) :

(35)

We can write the estimator g^ (bjn; x) as g^ (bjn; x) =

g^0 (b; n; x) ; ^ (njx) ' ^ (x)

so that

(k)

g^(k) (bjn; x) =

g^0 (b; n; x) ; ^ (njx) ' ^ (x) (k)

By Lemma B.3 of Newey (1994), the estimator g^0 (b; n; x) is uniformly consistent over b 2 [b1 (n; x) ; b2 (n; x)]: (k)

sup b2[b1 (n;x);b2 (n;x)]

j^ g0 (b; n; x)

d+1+2k

(k)

g0 (b; n; x) j = Op

Lh log L

1=2

+ hR

!

:

(36)

By the results in parts (a) and (b), the estimators ^ (njx) and ' ^ (x) converge at the rate faster than that in (36). The desired result follows by the same argument as in the proof of part (c), equation (30). For part (g), let cg be as in the proof of part (d) of the lemma. First, we con^ p ( jn; x) Q ( jx) is ^ p ( jn; x). We have that Q sider the preliminary estimator, Q bounded by j^ g (^ q ( jn; x) jn; x) g (q ( jn; x) jn; x)j g^ (^ q ( jn; x) jn; x) cg jg (^ q ( jn; x) jn; x) g (q ( jn; x) jn; x)j j^ q ( jn; x) q ( jn; x)j + g^ (^ q ( jn; x) jn; x) cg j^ g (^ q ( jn; x) jn; x) g (^ q ( jn; x) jn; x)j + g^ (^ q ( jn; x) jn; x) cg ! supb2[b1 (n;x);b2 (n;x)] g (1) (bjn; x) 1+ j^ q ( jn; x) q ( jn; x)j g^ (^ q ( jn; x) jn; x) cg j^ q ( jn; x)

+

q ( jn; x)j +

j^ g (^ q ( jn; x) jn; x) g (^ q ( jn; x) jn; x)j : g^ (^ q ( jn; x) jn; x) cg

By continuity of the distributions, we can pick " > 0 small enough so that q(

1

"jn; x) > b1 (n; x) and q (

25

2

+ "jn; x) < b2 (n; x) :

(37)

De…ne EL (n; x) = f^ q(

"jn; x)

1

b1 (n; x) ; q^ (

+ "jn; x)

2

b2 (n; x)g :

By the result in part (d), P (ELc (n; x)) = o (1). Hence, it follows from part (f) of the lemma that the estimator g^ (^ q ( jn; x) jn; x) is bounded away from zero with probability approaching one. Consequently, by Assumption 1(e) and part (d) of the lemma that the …rst summand on the right-hand side of (37) is Op L 1 uniformly over [

";

1

2

P

+ "], where

sup 2[

1

2[

1

P

";

L 2 +"]

sup ";

L 2 +"]

L

Lhd+1+2k log L

=

1=2

+h

R

. Next,

j^ g (^ q ( jn; x) jn; x)

g (^ q ( jn; x) jn; x)j > M

j^ g (^ q ( jn; x) jn; x)

g (^ q ( jn; x) jn; x)j > M; EL (n; x)

+P (ELc (n; x)) P

!

sup

L

b2[b1 (n;x);b2 (n;x)]

j^ g (bjn; x)

g (bjn; x)j > M

!

!

(38)

+ o (1) :

It follows from part (f) of the lemma and (38) that sup 2[

1

";

2 +"]

^ p ( jn; x) jQ

Q ( jx) j = Op

Lhd+1 log L

1=2

+ hR

!

:

(39)

^ ( jn; x) Q ^ p ( jn; x) Further, by construction, Q 0 for 0 . We can choose p 0 ^ 2 [ 0 ; ] such that 0 2 [ 1 ; 2 ]. Since Q ( jn; x) is left-continuous, there exists ^ p ( 0 jn; x) = supt2[ ; ] Q ^ p (tjn; x). Since Q ( jx) is nondecreasing, Q 0 ^ ( jn; x) Q ^ p ( jn; x) Q ^ p ( 0 jn; x) Q ^ p ( jn; x) = Q ^ p ( 0 jn; x) Q ( 0 jx) + Q ( jx) Q ^ p ( jn; x) Q ^ p (tjn; x) Q (tjx) + Q ( jx) Q ^ p ( jn; x) sup Q t2[

0;

2

]

sup 2[

1

";

2 +"]

^ p ( jn; x) Q

26

Q ( jx)

= Op

Lhd+1 log L

1=2

+ hR

!

;

where the last result follows from (39). Using a similar argument for < conclude that ! 1=2 d+1 Lh ^ ( jn; x) Q ^ p ( jx) = Op sup Q + hR : log L 2[ 1 "; 2 +"]

0,

we

(40)

The result of part (g) follows from (39) and (40). Lastly, we prove part (h). Let " be as in part(g). By Lemma 21.1(ii) of van der ^ ( jn; x) jn; x , where the inequality becomes strict only at Vaart (1998), F^ Q the points of discontinuity, and therefore ^ ( 1 jn; x) jn; x F^ Q

1

>

1

"

^ ( jn; x) is non-decreasing, for all n. Further, since Q ^ ( 2 jn; x) jn; x < P F^ Q = P

n ^ (tjn; x) sup t : Q

t2[0;1]

^ ( 2 jn; x) < Q ^( P Q

2

2

+"

o ^ Q ( 2 jn; x)

Suggest Documents