The general problem is to draw statistical inference about the unknown F and G based .... with jump size at U j (using our brackets notation) given by. Fn U j. = â.
The Annals of Statistics 1998, Vol. 26, No. 3, 1011–1027
ESTIMATION OF THE TRUNCATION PROBABILITY IN THE RANDOM TRUNCATION MODEL1 By Shuyuan He and Grace L. Yang Peking University and University of Maryland Under random truncation, a pair of independent random variables X and Y is observable only if X is larger than Y. The resulting model is the conditional probability distribution Hx y = PX ≤ x Y ≤ yX ≥ Y For the truncation probability α = PX ≥ Y, a proper estimate is not the sample proportion but αn = Gn s dFn s where Fn and Gn are product limit estimates of the distribution functions F and G of X and Y, respectivly. We obtain a much simpler representation αˆ n for αn . With this, the strong consistency, an iid representation (and hence asymptotic normality), and a LIL for the estimate are established. The results are true for arbitrary F and G. The continuity restriction on F and G often imposed in the literature is not necessary. Furthermore, the representation αˆ n of αn facilitates the establishment of the strong law for the product limit estimates Fn and Gn .
1. Introduction. Let X and Y be two independent random variables having distribution functions Fx and Gx, respectively. Consider an infinite sequence of independent random vectors Xm Ym m = 1 2 where Xm and Ym are independently distributed as X and Y. For each m the pair Xm Ym is observable only when Xm ≥ Ym . Thus the observable random variables are a subsequence of Xm Ym m = 1 2 . It is convenient to denote the observable subsequence by Uj Vj j = 1 2 with Uj ≥ Vj . The random vectors Uj Vj are iid; however, the components of each vector are dependent. Here and after, U V refers to any pair of Uj Vj . The random truncation model is defined by the joint distribution Hx y of U V as (1.1)
Hx y = PU ≤ x V ≤ y = PX ≤ x Y ≤ yX ≥ Y
with marginal distributions, (1.2) (1.3)
1x Gs dFs α −∞ 1x ¯ Fs− dGs G∗ x = PV ≤ x = H∞ x = α −∞ F∗ x = PU ≤ x = Hx ∞ =
where (1.4)
α = PX ≥ Y =
Gs dFs
Received March 1996; revised January 1998. 1 Supported in part by ONR Grant N00014-92-J-1097 and National Natural Science Foundation of China Grant 19571002. AMS 1991 subject classification. 62G05. Key words and phrases. Random truncation, nonparametric estimation, product-limit, truncation probability, strong consistency, LIL, iid representation.
1011
1012
S. HE AND G. L. YANG
b The integral sign a stands for a b . The integral sign without limits refers to integration from −∞ to +∞. The general problem is to draw statistical inference about the unknown F and G based on a sample of n iid random vectors Uj Vj j = 1 2 n from Xm Ym m = 1 2 mn , where the given n ≤ mn and mn is unknown. In the companion paper, He and Yang (1998) in the same issue of this journal, we prove the strong law of large numbers for the product limit estimate Fn given in (2.4). Under the same assumptions used in the companion paper, we address in this paper the estimation of the truncation probability, (1.5) α = PX ≥ Y = Gs dFs The problem is, of course, trivial if we have an iid sample from the original untruncated X Y-data, X1 Y1 X2 Y2 Xmn Ymn , with a known sample size mn . Then the sample proportion of those Xk Yk with Xk ≥ Yk is an optimal nonparametric estimate of α. However, under random truncation, any pair X Y for which X < Y is missing, and it is not known how many pairs are missing in the sample because mn is unknown. Thus it is not at all clear that reasonable estimates for α can be found. Equation (1.4) suggests estimating α by (1.6) αn = Gn s dFn s provided good estimates Fn and Gn for F and G can be obtained. Random truncation restricts the observation range of X and Y. Only F0 x = PX ≤ xX ≥ aG and G0 x = PY ≤ xY ≤ bF can be estimated; see Woodroofe (1985), where aF = inf x Fx > 0
and
bF = sup x Fx < 1
are the lower and upper boundaries of the support of the distribution of X. Let aG and bG be similarly defined. This leads us to the introduction of the following parameter: (1.7) α0 = G0 s dF0 s If aG ≤ aF and bG ≤ bF , then F0 = F G0 = G and α0 = α. Under these conditions and the continuity of F and G, Woodroofe (1985) proved that if Fn and Gn are product-limit estimates [given by (2.4) below], αn converges in probability √ to α as n → ∞. Under similar conditions, the asymptotic normality of nαn − α has been investigated by several authors using different methods. Chao (1987) used influence curves and Keiding and Gill (1990) used finite Markov processes and the δ-method. Since Fn and Gn have complicated product-limit forms, it is generally not easy to study the properties of αn . We propose, instead, to use the relationship (1.8)
¯ Rx = PV ≤ x ≤ U = PY ≤ x ≤ XX ≥ Y = α−1 GxFx−
ESTIMATION OF THE TRUNCATION PROBABILITY
1013
to obtain an estimating equation for α as (1.9)
¯ α = GxFx−/Rx
Replacing F and G by the respective product limit estimates yields another estimate of α as (1.10)
αˆ n = Gn xF¯ n x−/Rn x
for all x such that Rn x > 0. [Defined by (2.3).] An important result of this paper is that if Fn and Gn are the product-limit estimates of F0 and G0 defined by (2.4) below, then αˆ n and αn are equal. In particular, αˆ n is independent of x, provided Rn x > 0 The proof of equivalence is presented in Section 2. It is worth noting that the equivalence is not derived from integration-by-parts. The advantage of αˆ n over αn is its simpler form, which makes the analysis easier and enables us to obtain further properties of the estimate. Using αˆ n , we prove in Section 3 the almost sure convergence of the estimate to α0 and obtain a manageable iid representation for αˆ n and a LIL. The iid representation yields immediately the asymptotic normality of the estimate. The iid representation for αˆ n is deduced from that of Fn and Gn . Several iid representations for Fn (and Gn ) are available in the literature with different remainder terms; see Chao and Lo (1988), Gu and Lai (1990) and Stute (1993). We shall use Stute’s representation, which is derived under the condition that dF dG < ∞ and (1.11) < ∞ G2 F¯ 2 It has a sufficiently higher order remainder term of Oln3 n/n that suits our purpose. This condition can be weakened to dF dG < ∞ and (1.12) < ∞ G F¯ provided the tails of estimates Fn and Gn are properly modified. Under (1.12), the remainder term is of lower order than Olog 3 n/n √ but still good enough to yield the asymptotic normality for Fn at the rate n and a LIL, as shown by Gu and Lai (1990). Based on these, we obtain similar results for a modified αˆ n . Results in Section 3 are established under the continuity of F and G but are true for discrete F and G as well. The generalization to arbitrary F and G is given in Section 4. By construction, αˆ n and αn inherit asymptotic properties of Fn and Gn . Conversely, good behavior of αˆ n induces nice properties in Fn and Gn . As shown in He and Yang (1998), the almost sure convergence of αˆ n to α0 leads to the SLLN for Fn in the sense that ϕ dFn → ϕ dF0 for any integrable ϕ.
1014
S. HE AND G. L. YANG
2. The equivalence of ␣ ˆ n and ␣ n . To avoid triviality, we shall always assume aG < bF , which ensures that α > 0. In what follows, for any real monotone function g, the left continuous version of gs is denoted by gs− or g− s, and the difference gs − gs− by the curly brackets g s . The convergence is “with respect to n → ∞” unless specified otherwise. Lemma 2.1. Let α0 be given by (1.7) and α by (1.5). Then α0 ≥ α. A necessary and sufficient condition for α0 = α is aG ≤ aF
(2.1) Proof.
and
bG ≤ bF
For x ∈ aG bF , we have
G0 x = PY ≤ x/PY ≤ bF
F¯ 0 x− = PX ≥ x/PX ≥ aG
Hence, it follows from (1.8) and Lemma 1 of Woodroofe (1985) that ¯ G −−1 α0 = αGbF Fa
✷
Let IA denote the indicator function of the event A. Let F∗n , G∗n and Rn be the empirical distributions defined by (2.2)
(2.3)
F∗n s = n−1
n i=1
IUi ≤ s
G∗n s = n−1
Rn s = G∗n s − F∗n s− = n−1
n i=1
n i=1
IVi ≤ s
IVi ≤ s ≤ Ui
The well-known product limit estimates of F0 and G0 are defined by F∗n s G∗n s 1− and Gn x = 1− (2.4) Fn x = 1 − Rn s Rn s s≤x s>x where an empty product is set equal to 1. For construction of these estimates, see Woodroofe (1985) or Wang, Jewell and Tsai (1986). The estimates Fn and Gn are step functions. The jumps of Fn occur at the distinct order statistics U1 < U2 < · · · < Ur of the sample U1 U2 Un with jump size at Uj (using our brackets notation) given by F∗n Ui F∗n Uj Fn Uj = 1− (2.5) Rn Ui Rn Uj i 0.
ESTIMATION OF THE TRUNCATION PROBABILITY
1015
Proof. The case αn = 0 is easy. Note that αn = 0 if and only if bFn < aGn or bFn = aGn and 1 − Fn bFn −Gn aGn = 0. This is equivalent to αˆ n = 0 for all x. Now suppose αn > 0. We introduce two independent random variables Z and W which have distributions Fn x and Gn x, respectively. Then αn = PZ ≥ W = Gn dFn The integral (2.6)
x
∞
Gn t dFn t = =
r j=1 r j=1
Gn Uj Fn Uj IUj ≤ x ζn j F∗n Uj IUj ≤ x
where ζn j = Gn Uj Fn Uj /F∗n Uj . In Lemma 2.3 below we show that for n fixed, ζn j is a constant in j, say ζ0 . By setting x = ∞ in (2.6) we obtain ∗ αn = Gn s dFn s = ζ0 (2.7) Fn Uj = ζ0 j
Consequently, the conditional distribution x ∗ ∗ PZ ≤ xZ ≥ W = α−1 Gn t dFn t = α−1 (2.8) n n ζ0 Fn x = Fn x −∞
By symmetry, G∗n x = PW ≤ xZ ≥ W ¯ Therefore, Rn x = Gn∗ x − Fn∗ x− = PW ≤ x ≤ ZZ ≥ W = α−1 n Gn xFn x−; that is, αn =
Gn xF¯ n x− = αˆ n Rn x
for all x such that Rn x > 0. ✷ Remark. Once we have reached (2.8), we could use integration-by-parts to complete the proof. However, it is simpler to use random variables Z and W. Then the result follows immediately by symmetry. Note also that although αn is an MLE, it is not obvious that αˆ n is an MLE, since it is Fn x and not Fn x−, that is, the MLE of F. Therefore we cannot use the MLE argument to claim that αn = αˆ n . Lemma 2.3. Let Fn and Gn be the product-limit estimates given by (2.4). Let ζn j = Gn Uj Fn Uj /F∗n Uj as in (2.6). Then for any fixed n, ζn j = ζn 1 , for j = 2 n.
1016
S. HE AND G. L. YANG
Proof. ζn j
Substituting (2.5) for Fn Uj in ζn j gives q G∗n Vk IVk > Uj F∗n Ui 1 1− 1− = Rn Vk Rn Ui Rn Uj i Uj
F∗n Ui
1− Rn Vk Rn Ui i 0, let us denote by V1 V2 Vh the distinct ordered values of Vj in Uj−1 Uj , that is, Uj−1 < V1 < V2 < · · · < Vh ≤ Uj Then, h G∗n Vl G∗n Vl IUj−1 < Vl ≤ Uj 1− 1− = Rn Vl Rn Vl l l=1 = =
Rn V1 − Rn Vh
G∗n Uj−1 − F∗n Uj−1 G∗n Vh − F∗n Uj−1
This implies that Bn j =
Rn Uj−1 − F∗n Uj−1 Rn Uj−1 −
1 Rn Uj
Rn Uj−1 − F∗n Uj−1 Rn Vh
1 = 0 Rn Uj−1
ESTIMATION OF THE TRUNCATION PROBABILITY
1017
The last equality follows from Rn Uj = Rn Vh . This is because if Vh < Uj , then Rn Uj = G∗n Uj − F∗n Uj − = G∗n Vh − F∗n Vh − = Rn Vh . ✷ Corollary 2.4. αˆ n =
Gn Uj F¯ n Uj − Rn Uj
=
Gn Vj F¯ n Vj − Rn Vj
j = 1 2 n
The next corollary follows either by Theorem 4.1 of He and Yang (1998), or by applying the uniform strong convergence of Fn [see Chen, Chao and Lo (1994)]. Corollary 2.5.
As n → ∞, αˆ n → α0
a.s.
x Remark. If we use = exp− −∞ dF∗n /Rn to estimate 1 − F0 x, ∞ Sn x and Q˜ n x = exp− x dG∗n /Rn to estimate G0 x then, by Corollary 3.2 of He and Yang (1998), for any x such that Rn x > 0, cn =
Q˜ n xSn x− Rn x
is a strong consistent estimate of α0 . 3. The CLT and the LIL for ␣ n . Since αn and αˆ n are equivalent, the √ √ known result of asymptotic normality of nαn − α0 applies to nαˆ n − α0 . On the other hand, the simple form of αˆ n makes it possible to obtain an iid representation from which the asymptotic normality of αˆ n follows immediately. Moreover, an LIL can be obtained. Let Fn and Gn be defined by (2.4). Applying Theorem 2 of Stute (1993) yields the following iid representation for Fn and Gn . This result is needed for deriving the iid representation for αˆ n as given in Theorem 3.2. It is also of independent interest. Lemma 3.1. (3.1)
If F and G are continuous such that ∞ dFs bF dGs < ∞ and x ∞ IVi ≤ s ≤ Ui − dG∗ s RVi R2 s x
are iid random variables with EWi x = 0
VarWi x =
bG∗
x
i = 1 2 n
dG∗ s R2 s
Proof. We prove only (ii) can be proved by symmetry. ∞(i), because 2 It is easy to see that dFs/G s < ∞ if and only if dF0 s/G20 s < aG bF 2 ∞, and −∞ dGs/F¯ s < ∞ if and only if dG0 s/F¯ 20 s < ∞. By Theorem 2 of Stute (1993), log 3 n ¯ a.s., Fn x − F0 x = F0 xLn x + O n with Ln x =
x
aF
n dF∗n s x Rn s 1 Z x − dF∗ s = 2 Rs n i=1 i aF R s
a.s.
Direct computation yields EZi x = 0 and VarZi x =
dF∗ x x EIsIt 1 + dF∗ s 2 dF∗ t 2 2 s R R R t aF∗ aF∗ aF∗ x IU ≤ xIs dF∗ s x dF∗ −2 E = 2 RU R2 s aF∗ aF∗ R x
where Is = IV ≤ s ≤ U. ✷ Remark. Evaluation of integrals similar to the above will be carried out in the proof of the next theorem. √ Theorem 3.2. Under the assumptions of Lemma 3.1, as n → ∞, nαˆ n − 2 α0 converges weakly to the normal distribution N0 σ , and with probability 1, the sequence n/2 log log nαˆ n − α0 n ≥ 7 is relatively compact with its set of limit points −σ σ, where x
1 dF∗ s bG∗ dG∗ s 2 2 + − σ = α0 + 2α0 − 1 2 R2 s Rx x aF∗ R s for x ∈ aG∗ bF∗ , is a positive constant.
1019
ESTIMATION OF THE TRUNCATION PROBABILITY
Proof. Using Lemma 3.1 and the LIL for iid partial sums, we obtain ∀ x ∈ aG∗ bF∗ , with probability 1 for n large: Gn x1 − Fn x− G0 x1 − F0 x − Rn x Rx n n F¯ xRxG0 x 1 1 = 0 Wi x − Z x − Rn xRx n i=1 n i=1 i
αˆ n − α0 =
+O = −α0
log 3 n n
n 1 − IVi ≤ x ≤ Ui − Rx nRx i=1
n 1 log 3 n ζi x + O n i=1 n
a.s.
where (3.2)
ζi x = Wi x + Zi x +
1 IVi ≤ x ≤ Ui − Rx Rx
i = 1 2
is a sequence of iid random variables with mean zero. The theorem follows by the classical CLT and the LIL for partial sums of an iid sequence, if we can show that (3.3)
Varζi x = σ 2
∀ x ∈ aG∗ bF∗
is a positive constant. This requires calculations of the moments and crossproduct moments of Zi , Wi and IVi ≤ s ≤ Ui . The calculation is not hard but tedious. We shall give some key steps only. To proceed, we suppress the subscript i from these variables for simplicity. Put Ts = Is/Rs − 1. Then ETx2 = EZxWx =
1 − 1 Rx x bG∗ EIsIt aF∗
x
R2 t
dG∗ t
1 R2 s
dF∗ s
¯ 1 1 x bG∗ GsFt dG∗ t 2 dF∗ s 2 α aF∗ x R t R s 1 1 1 1 = −α − − ¯ ¯ F∗ GbG∗ Gx Fx Fa =
and
α 1 1 − ¯ ¯ F∗ Gx Fx Fa α 1 1 + − ¯ GbG∗ Gx Fx
EZx + WxTx = −
1020
S. HE AND G. L. YANG
Using the moments of Zi and Wi given in Lemma 3.1, we obtain x dF∗ bG∗ dG∗ 1 Varζx = + 2α0 − 1 + − 2 2 R Rx x aF∗ R ¯ F∗ −1 = αGbF Fa ¯ G −1 [see where we used the fact that α0 = αGbG∗ Fa ∗ ∗ the definition of G and F in (1.2), (1.3)]. Obviously, ζx is not a constant and therefore σ 2 > 0. Then ∀ x y ∈ aG∗ bF∗ , the difference y dR 1 1 Varζx − Varζy = + − = 0 2 Ry Rx x R Hence, σ 2 = Eζx2 for x ∈ aG∗ bF∗ is a positive constant. ✷ Remark. The random variable ζx does not depend on x. This can be seen from the following representation: bG∗ Is 1 ζx = − dG∗ s − 1 a.s. 2 RV aG∗ R s or bF∗ Is 1 ζx = − dF∗ s − 1 a.s. 2 RU aF∗ R s A necessary and sufficient condition for σ 2 < ∞ is ∞ dF bF dG (3.4) < ∞ < ∞ and ¯ −∞ F aG G which is weaker than (3.1). To obtain results similar to Lemma 3.1 and Theorem 3.2 under the weaker condition (3.4) we can use a modified form α˜ n of αn . The modification is necessary to avoid singularities at the boundaries of X. It ˜ n , of Fn , Gn proposed by is constructed based on the modified estimates F˜ n , G Gu and Lai (1990) as given below: IG∗n Ui ≥ nθ−1 ˜ Fn x = 1 − 1− (3.5) nRn Ui i U ≤x i
and (3.6)
˜ n x = G
1−
i Vi >x
IF¯ ∗n Vi − ≥ nθ−1 nRn Vi
for θ ∈ 1/3 1/2. Accordingly, the modified estimate α˜ n for α0 is (3.7)
α˜ n =
˜ n x1 − F˜ n x− G Rn x
˜ n is constructed, put X ˜ = −Y and for any x such that Rn x > 0 To see how G ˜ = −X. Thus X ≥ Y is precisely X ˜ ≥ Y. ˜ Therefore, X ˜ Y ˜ is observable Y ˜ V ˜ = −V −U. Then the corresponding if and only if X ≥ Y, and so U
ESTIMATION OF THE TRUNCATION PROBABILITY
1021
˜ ≤ x based on the data U ˜ j V ˜ j j = 1 n is, modified estimate for PX according to (3.5), 1−
˜j ≤U ˜ i ≥ nθ I j IV 1− ˜ ˜ ˜ j IVj ≤ Ui ≤ Uj ˜ i ≤x i U I j IUj ≥ Vi ≥ nθ 1− =1− j IVj ≤ Vi ≤ Uj i Vi ≥−x IF¯ ∗n Vi − ≥ nθ−1 1− =1− nRn Vi i V ≥−x
i
˜ ≤ x = PY ≥ −x = 1 − G− −x. Therefore, However, by construction PX Gx is estimated by
i Vi >x
IF¯ ∗n Vi − ≥ nθ−1 1− nRn Vi
An iid representation for α˜ n under the weaker condition (3.4) is achieved at the cost of lowering the order of the remainder term. √ However, the order remains high enough to yield the weak convergence of nαˆ n −α0 to normality and a LIL. The proof of this is more involved than that of Theorem 3.2. Let Zi x and Wi x be defined as in Lemma 3.1. Lemma 3.3. Assume that F and G are continuous and satisfy the conditions (3.4). Then for θ ∈ 1/3 1/2, x ∈ aG bF , as n → ∞ we have the following. (i) F˜ n x = F0 x + F¯ 0 x1/n ni=1 Zi x + Oηn a.s. ˜ n x = G0 x − G0 x1/n n Wi x + Oηn a.s. (ii) G i=1 √ where ηn = oφn a.s., φn = 2 log log n/n and ηn = op 1/ n. Proof. We first prove (i) and then show that (ii) can be obtained by means of symmetry. Taking q = θ and c = 1 in Theorem 2 of Gu and Lai (1990), we have, with probability 1 as n → ∞, F˜ n x − F0 x =
mn x 1 1 F¯ 0 x ¯ mn j=1 τn GuFu × d IYj ≤ Xj ≤ u −
+ Onθ−1
u −∞
IXj ≥ s ≥ Yj
dFs ¯ Fs
1022
S. HE AND G. L. YANG
where τn = inf s G∗n s ≥ nθ−1 , and mn is defined in the Introduction. By (1.2) we have
u R s F¯ x x 1 n F˜ n x − F0 x = 0 d nF∗n u − n dF∗ s αmn τn Ru −∞ Rs + Onθ−1
n n 1 ¯ 1 n 1 = F0 x Zi x + − α0 F¯ 0 x Z x n α mn n i=1 i i=1 −
n n ¯ 1 F0 x Z τ + Onθ−1 αmn n i=1 i n
= F¯ 0 x
n 1 Z x + J1 n x + J2 n x + Onθ−1 n i=1 i
According to Corollary 4 of Gu and Lai (1990), for any δ ∈ aG x, n 1/2 δ 1 1 dFs lim sup Zi s ≤ a.s. sup ¯ 2 sGs aG αF n→∞ φn aG ≤s≤δ n i=1 Now
δ
aG
dFs 1 δ dFs ≤ →0 F¯ 2 x aG Gs αF¯ 2 sGs
as δ → aG
n/mn → α0 a.s, and the classical LIL implies that Ji n x = oφn, a.s. for i = 1 2. On the √ other hand, the CLT and the Chebyshev inequality imply that Ji n x = op 1/ n, for i = 1 2. This completes the proof of (i). We now turn to the proof of (ii). As noted earlier, ˜ j ≤ x = 1 − G−x Qx ≡ PX
˜ j ≤ x = 1 − F−x Kx ≡ PY
Thus aQ = −bG , bQ = −aG , aK = −bF , bK = −aF , and aG < bF if and only if −bQ < −aK Therefore, ∞ dQx ∞ −dG−x bF dGx = = < ∞ −∞ 1 − Fx aK Kx −bF 1 − F−x Since aG bF = −bQ −aK so if x ∈ aG bF then −x ∈ aK bQ . Setting y = −x and applying (i), we obtain n 1 ˜ ˜ Qn y = Q0 y + 1 − Q0 y Z y + Oηn n i=1 i ˜ n −y− = 1 − G ˜ n x−, Q0 y = PX ˜ j ≤ yX ˜ j ≥ aK = where Q˜ n y = 1 − G G¯ 0 x− and y IV ˜i ≤s≤U ˜ i ˜ ˜ i y = IUi ≤ y − Z dQ∗ s 2 ˜ ˜ ˜ aK RUi R s
1023
ESTIMATION OF THE TRUNCATION PROBABILITY
˜ ˜i ≤ s ≤ U ˜ i = R−s and Q∗ s = PU ˜ i ≤ s = 1 − G∗ −s. with Rs = PV Hence −x IV ≤ −s ≤ U i i ˜ i y = IVi ≥ x − d1 − G∗ −s = Wi x− Z RVi R2 −s −bF where Wi is given in Lemma 3.1. Therefore n ˜ n x− = G0 x− − G0 x− 1 G W x− + Oηn n i=1 i
a.s.
✷
The following theorem is proved the same way as Theorem 3.2. √ Theorem 3.4. Suppose the assumptions of Lemma 3.3 hold. 2As n → ∞, nα˜ n − α0 converges weakly to the normal distribution N0 σ , and with probability 1, the sequence φ−1 nα˜ n − α0 n ≥ 7 is relatively compact with its set of limit points −σ σ, where σ 2 is defined in Theorem 3.2. 4. Arbitrary F and G. We shall relax the continuity condition on F and G of Theorems 3.2 and 3.4. The results are given in Theorems 4.1 and 4.2. As noted by Woodroofe (1985), the truncation model Hx y is the same if the underlying distributions F and G are replaced by F0 and G0 . Thus for studying H, we may, without loss of generality, assume that F0 = F and G0 = G. It follows that α0 = α. This simplifies the discussion. The proof for arbitrary F and G uses a technique of Major and Rejt¨o (1988). ˆ Y ˆ via a certain specially constructed real Namely, we transform X Y to X function hx. The transformed random variables have continuous distribution ˆ As such, the foregoing theorems apply to functions to be denoted by Fˆ and G. ˆ where the product-limit estimates ˆ G, the product-limit estimates Fˆ n Gˆ n of F ˆ ˆ i Y ˆ i or ˆ Fn Gn are computed, with (2.4), based on the transformed sample, X ˆ ˆ Ui Vi . It is shown in He and Yang (1998) that ˆ Fu = Fhu+ Fn u = Fˆ n hu+
ˆ Gu = Ghu+ Gn u = Gˆ n hu+
for any real number u
By way of this relationship, we show that the results of previous sections hold for arbitrary F and G. To proceed, we need to express Rn and αn in terms of the transformed variables as well. Using the same notation as in He and Yang, we put the symbol ˆ on quantities derived from the transformed data and let A = xj j ≥ 1 be the set of jump points of F and G. If aG∗ < bF∗ , then for x ∈ aG∗ bF∗ − A, with probability 1 for large n we have mn mn 1 1 ˆ i ≤ hx ≤ X ˆ i IYi ≤ x ≤ Xi = IY 0 < Rn x = n i=1 n i=1 (4.1) n 1 ˆ i ≤ hx ≤ U ˆ i ≡ Rˆ n hx = IV n i=1
1024
S. HE AND G. L. YANG
ˆ i ≤ x, Gˆ ∗ x = PV ˆ i ≤ x and Rx ˆ Define Fˆ ∗ x = PU = Gˆ ∗ x − Fˆ ∗ x−. ˆ It follows that Rx = Rhx and (4.2)
αˆ n =
1 − Fn x−Gn x 1 − Fˆ n hx−Gˆ n hx = Rn x Rˆ n hx
Here we have used the fact that hx = hx+ for x ∈ Ac , the continuity set of F and G. Parallel to Theorems 3.2 and 3.4, we arrive at the following general results in Theorems 4.1 and 4.2 for arbitrary F G under condition (3.1) and (3.4), respectively. Note that for continuous F and G, the σ 2 formula in Theorem 4.1 coincides with that in Theorem 3.2. Theorem 4.1. If dFs dGs < ∞ (4.3) 0, we know that hx ∈ ;c and ˆ i = α−1 PY ˆ i ≤ hx ≤ X ˆ i = Rx > 0 ˆ ˆ i ≤ hx ≤ U Rhx = PV
(4.5)
For s ∈ ;j 2 , ˆ αRs = Gxj 1 − Fxj − − 2j2 s − hxj − 1F xj ˆ Fs = Fxj − + 2j2 s − hxj − 1F xj We show that the two integrals in σ12 equal the corresponding ones in σ 2 . The first integral is hx hx dFˆ ∗ ˆ ≤ hx dFˆ IX (4.6) =E ≡ B + D = ˆ − F ˆ ˆ X1 ˆ ˆ X ˆ −∞ −∞ R1 Rˆ 2 R − F where
and
ˆ ≤ hx X ∈ Ac IhX ≤ hx X ∈ Ac IX =E B=E ˆ X1 ˆ ˆ X ˆ ˆ ˆ R − F RhX1 − FhX x IX ≤ x X ∈ Ac dFs =E = IA c ¯ RX1 − FX −∞ RsFs
ˆ ≤ hx X ∈ A IX D=E ˆ X1 ˆ ˆ X ˆ R − F ˆ ∈ ;k IX dFˆ = E = E ˆ X1 ˆ ˆ X ˆ ˆ − F ˆ ;k R1 R − F k xk