Statistical Inference for the Extreme Value Distribution ...

1 downloads 0 Views 826KB Size Report
Statistical Inference for the Extreme Value Distribution under. Adaptive Type-II Progressive Censoring Schemes. Z.S. Ye1; P.S. Chan2†; M. Xie1,3 and H.K.T. ...
Statistical Inference for the Extreme Value Distribution under Adaptive Type-II Progressive Censoring Schemes Z.S. Ye1; P.S. Chan2†; M. Xie1,3 and H.K.T. Ng4 1

Department of Industrial & Systems Engineering, National University of Singapore 2

3

Department of Statistics, Chinese University of Hong Kong

Department of Systems Engineering and Engineering Management, City University of Hong Kong 4

Department of Statistical Science, Southern Methodist University

[E-mail address: [email protected]; [email protected]; [email protected]; [email protected].]

ABSTRACT Adaptive Type-II progressive censoring schemes have been shown to be able to strike a balance between the estimation efficiency and the total testing duration. In this paper, some general statistical properties of an adaptive Type-II progressive censoring scheme are first investigated. A bias correction procedure is proposed to reduce the bias of the maximum likelihood estimators. We then focus on the extreme value distributed lifetimes and derive the Fisher information matrix for the maximum likelihood estimators based on these properties. Four different approaches are proposed to construct confidence intervals for the parameters of the extreme value distribution. Performance of these methods is compared via an extensive Monte Carlo simulation.

Key words: Adaptive Type-II progressive censoring, confidence interval, extreme value distribution, information matrix, the bootstrap.



Corresponding author.

1

1 INTRODUCTION Before a manufacturer launches a newly-developed product to the market, a life-testing experiment is often required to assess its reliability. During the test, censoring technique is usually adopted to obtain the lifetime information within a reasonable timeframe. Two common censoring schemes are the Type-I and the Type-II censoring, where the experiment terminates, respectively, at a pre-determined time T and upon certain number of failures. To allow for more flexibility in removing un-failed units from the tests, more general censoring approaches are called for. The progressive Type-II right censoring scheme is an appealing one and has attracted much attention in the literature, e.g., Wang (2008), Ismail and Aly (2010), Maturi et al. (2010), Wu (2010), Balakrishnan et al. (2011) Burkschat (2008) and Lin et al. (2011), to name a few. A book length treatment on this topic is found in Balakrishnan and Aggarwala (2000). See also Balakrishnan (2007) for a recent comprehensive review on progressive censoring. A typical progressive Type-II censoring scheme works as follows. Suppose that n identical units are placed on a lifetime test and upon the ith failure, Ri functioning units are randomly withdrawn from the remaining n  i  R1   Ri 1 surviving units. The experiment continues until the mth failure is observed. Similar to the traditional Type-II censoring approach, a problem associated with the progressive scheme is that with positive probability, the total testing duration might be unacceptably long. To address this issue, Kundu and Joarder (2006) proposed a hybrid variant of the progressive censoring scheme by imposing a time limit T on the test. Under this scheme, the test stops at Xm:m:n if the mth progressively censored observed failure occurs before time T. Otherwise, the experiment terminates at time T with all remaining items censored at T. However, it is possible that the effective sample size is very small or even zero, indicating possible loss of efficiency under this scheme. To strike a balance between the total testing time and the estimation efficiency, Ng et al. (2009) proposed an adaptive Type-II progressive censoring scheme, which works as follows. If Xm:m:n < T, the experiment proceeds with the pre-specified progressive censoring scheme  R1 , , Rm  . Otherwise the censoring scheme switches to  R1 , , RJ , 0,...0, Rm  , where J < m is the number of failures observed before T and Rm  n  m   k 1 Rk . The basic idea of this J

scheme is to speed up the test as much as possible when the test duration exceeds a pre2

determined threshold T. When T = ∞, the adaptive variant reduces to a regular progressive TypeII censoring scheme. On the other hand if T = 0, this variant leads to a traditional Type-II censoring scheme. It is noted that Cramer and Iliopoulos (2010) studied the general adaptive progressive Type-II censoring schemes that allow experimenters to choose the next censoring number. After proposing the adaptive Type-II progressive censoring scheme, Ng et al. (2009) applied it to the exponential distributions. Nevertheless, compared with the exponential distribution, the Weibull model is more general and receives more applications. The Weibull distribution is one of the most commonly-used models in the parametric analysis of lifetime data. For example, Ye et al. (2011) used it to fit the interface fatigue data of an electronic device, Hong et al. (2009) used it to fit the transformer failure data. More applications of this distribution can be found in the book by Murthy et al. (2004). Logarithm of a Weibull random variable follows an extreme value distribution, which belongs to the location-scale family. Due to the good properties and wellestablished theorems for this family (e.g. Lawless 2003 Chapters 5 & 6), it is often more convenient to work with the equivalent extreme value distribution instead of directly dealing with the Weibull distribution. In view of the advantages of the adaptive Type-II progressive censoring scheme and the popularities of the Weibull model, it is of particular interest to study the performance of this scheme when applying to the extreme value distribution. Maximum likelihood estimation and the expected total time on test have been investigated by Lin et al. (2009). However, the from the tables given by Lin et al. (2009), it is easily seen that the biases of the MLE is significant when the sample size is not large enough. This motivates us to use the bootstrap bias correction technique to calibrate the point estimation. In addition, this study will also focus on the investigation of general properties of this scheme and on the construction of confidence intervals for the scale and location parameters of the extreme value distribution. We consider four approaches to construct confidence intervals of the unknown parameters. Specifically, we consider the confidence interval constructions based on (i) the Fisher (expected) information matrix, (ii) the observed information matrix, (iii) the parametric percentile bootstrap and (iv) the parametric bootstrap-t method. Compared with Ng et al. (2009) and Lin et al. (2009), we also introduce a notation system that is able to greatly simplify the expressions. This paper is organized as follows. In Section 2, marginal distribution of the order statistics 3

under the adaptive censoring scheme is derived. Section 3 discusses how to use the bootstrap method to correct the estimation biases. Section 4 confines to the extreme value distribution and discusses four approaches in constructing confidence intervals for the parameters. In Section 5, an extensive simulation study is carried out to examine the performance of these four approaches. Section 6 concludes the paper.

2 NOTATIONS AND GENERAL RESULTS Let the random variable X be the lifetime of a product with PDF f  x  , CDF F  x  and survival function F  x  . To assess the lifetime distribution, n units are chosen from the product population with corresponding lifetimes X1, X2, …, Xn, which are n independent copies of X. Before the life test, the effective sample size m ≤ n, the progressive censoring scheme

 R1 , , Rm  and the scheme switching time T are specified. When t < T, the test is carried on as a regular progressive censoring scheme  R1 , , Rm  , where Ri functioning units are randomly removed upon the ith failure. Denote the total number of failures observed before time T as J, where J is a random variable with support on 0,1,..., m . If J < m, then after T, the test is switched to the scheme  R1 , , RJ , 0,...0, Rm  where Rm  n  m   k 1 Rk . After the experiment, J

m ordered lifetimes, denoted by X   X1:m:n ,..., X m:m:n  , are observed. In addition, the number of observed failures before T, i.e. J, is also recorded. The joint distribution of X and J can be expressed as f X, J  x1:m:n ,..., xm:m:n , j   cm I  x 

m

 F r  xk :m:n  f  xk :m:n   , T     k

j:m:n , xk 1:m:n

(1)

k 1

where

R , k  j k m  k rk  0, j  k  m , ck   s and k    rs  1 . s k s 1  j n  R   k 1 k The distribution of J is of special interest because it reflects the effectiveness of this adaptive 4

scheme. When J equals 0 with a high probability, the adaptive scheme would proceed with a conventional Type-II censoring scheme most of the time. On the other hand if J assigns a large probability to m, this scheme would proceed with a regular progressive Type-II censoring scheme for most of the time. The probability mass function of J can be computed based on (1) as follows. The event  J  0 is equal to the case where lifetimes of all units are greater than T. therefore, Pr  J  0   F n T 

(2)

It can also be shown (in the appendix) that when j > 0,  bk , j  F j1 T   F k T  

j

Pr  J  j   c j 

k   j 1

k 1

,

(3)

j 1 b  where k , j     and by convention, m1  0 . s 1 s k sk

Next, we shall look at the marginal distribution of Xi:m:n given J. When i  j , the marginal distribution can be derived by direct integration as



dx j:m:n

xi:m:n





dx j 1:m:n

T

dx1:m:n

0



x j 1:m:n

T

x j:m:n

x2:m:n

1 f X i:m:n , J  xi:m:n | J  j   Pr  J  j 

 dx

m:m:n



x j 2:m:n

dxi 1:m:n

0



xi:m:n

dxi 1:m:n .

f X, J  x1:m:n ,..., xm:m:n , j 

T

After some routine deduction, it can be shown that

f X i:m:n , J  xi:m:n | J  j  

c j I  0,T   xi:m:n 

j 1

i

a Pr  J  j  h 1 k i 1

k , h ,i , j

F k T  F h k 1  xi:m:n  f  xi:m:n  ,

where i

ak ,h,i , j   s 1 sh

j 1 1 1  s  h s i 1 s  k . s k

Similarly when i > j, direct calculation shows that 5

(4)

f X i:m:n , J  xi:m:n | J  j  

ci I T ,  xi:m:n 

j 1

i

d Pr  J  j  k 1 h  j 1

k , h ,i , j

F k h T  F h 1  xi:m:n  f  xi:m:n  ,

(5)

where j 1

d k , h ,i , j   s 1 sk

i 1 1  s  k s  j 1 s  h . s h

Based on the above marginal distribution, the expected test duration can be computed as m 1



T

j 0

T

0

E  X m:m:n    Pr  J  j   xf X m:m:n , J  x | J  j  dx  Pr  J  m   xf X m:m:n , J  x | J  m  dx .

(6)

To carry out the statistical inference, the log-likelihood function based on the observed data is required. Based on (1), The log-likelihood function can be expressed as m

l   | X  x, J  j   log cm   rk log F  xk:m:n   log f  xk:m:n  ,

(7)

k 1

where Θ is the unknown parameter to estimate. From this equation, it is readily seen that point estimation of Θ is the same as that under the progressive Type-II right censoring scheme

 R , , R , 0,...0, R  . It is observed that when the sample size is small, the bias of the estimates 1

j

m

is large. This motivates the next section on bias correction.

3 THE BOOTSTRAP BIAS CORRECTED ESTIMATOR ˆ the MLE. Generally speaking, ˆ is a biased Let Θ be the true value of the parameters, and  estimator. When the sample size is small, the bias is often very significant. The bias can be corrected based on the bootstrap principle. Let u be an additive adjustment such that ˆ  u     0 . Here u can be estimated via the bootstrap method (Hall 1997). The procedure E  

is as follows. [1]. Collect the adaptive progressive Type-II censoring data (X, J) and compute the MLE of Θ,

ˆ , by maximizing the log-likelihood function. denoted as 

6

[2]. Based on the adaptive censoring scheme  R1 , , Rm  and the switching time T, generate an adaptive Type-II progressive censoring dataset from the extreme value distribution

ˆ. with distribution parameter  ˆ*. [3]. Obtain the MLE based on this simulated dataset. Denote this bootstrap estimate by  ˆ *,  ˆ * ,...,  ˆ* . [4]. Repeat Steps [2] – [3] B times and obtain  1 2 B ˆ [5]. Then u can be estimated as uˆ  

1 B ˆ*  k . B k 1

The bias-corrected estimate of Θ can then be given by ˆ  uˆ  2 ˆ 

1 B ˆ*  k B k 1

Performance of this bias-corrected estimator will be demonstrated via simulation in Section 5.

4 INTERVAL ESTIMATION FOR THE EXTREME VALUE DISTRIBUTION This section is confined to the special case of the extreme value distribution. The results presented in Section 2 will be utilized here. Note that the extreme value distribution has CDF and PDF as

  x    F  x   1  exp   exp   ,     f  x 

(8)

 x  x    exp   exp   ,       1

(9)

respectively. Therefore, the log-likelihood function can be specified as m x     x l     l   | X  x, J  j   log cm  m ln     k:m:n   rk  1 exp  k:m:n   , (10)     k 1 

where Θ = (μ, σ) is the parameter vector. To simplify the notation, define Z 

7

X 



and

zk:m:n 

xk:m:n  



. The CDF and PDF of Z are linked to (8) and (9) as FZ  z   1  exp   exp  z   and f Z  x   exp  z  exp  z   .

(11)

The partial derivative of l    with respect to μ and σ are given by  m m 1 l         rk  1 exp  zk:m:n  ,   k 1 

(12)

 m m z l        k:m:n 1   rk  1 exp  zk:m:n   .   k 1 

(13)

MLE of μ and σ can be obtained by simultaneously solving the equations

 l     0 and 

 l    0 .  The second derivatives of l    with respect to μ and σ are given by m 2 1  2 l      2  rk  1 exp  zk:m:n  ,  k 1 



m zk2:m:n 2 m m 2 zk:m:n l     1  r  1 exp z  r  1 exp  zk:m:n  ,           k k :m:n  2  k  2  2 k 1  2  k 1  m z 2 m m 1  l      2   2  rk  1 exp  zk:m:n    k:m2:n  rk  1 exp  zk:m:n  .   k 1  k 1 

(14)

(15)

(16)

Because the MLEs for μ and σ do not have explicit expressions, a basic idea for constructing confidence intervals is to make use of the asymptotic properties of the MLE and consider asymptotically pivotal quantities. To construct confidence intervals based on the asymptotically pivotal quantities, the information matrix is required. Two scenarios are considered in the following.

4.1 Confidence Interval Based on Observed Information 8

Matrix The asymptotic variances and covariance of the maximum likelihood estimators can be obtained by inverting the observed information matrix I, which is given by

 2  2 l  l        2   I   2    2 l  l   2        ˆ Because

  l   0 and l    0 ,By combining the results of Equations (12)   ˆ ˆ  

(16), the elements in the observed information matrix can be simplified as

2 m l   2, 2   ˆ ˆ 

(17)

2 m m rk  1 2 l    zˆ exp  zˆk:m:n  ,    2 ˆ 2 k 1 ˆ 2 k:m:n ˆ 

(18)





2 m m 1  l    2   2 zˆk:m:n .  ˆ k 1 ˆ ˆ 

(19)

Based on the asymptotic theory of the MLE for the extreme value distribution (Meeker and

ˆ is bivariate normal with mean Escobar 1998, pp. 622), the asymptotic distribution of the MLE  vector Θ and variance-covariance matrix I 1 . The approximate confidence intervals for μ and σ are then respectively given by









1 1 ˆ  z /2 I111 and ˆ exp z /2 I 22 / ˆ , ˆ / exp z /2 I 22 / ˆ  ,   1 where z /2 is the 100α/2 percentile of the standard normal distribution, while I ij is the (i, j)th

element of I 1 .Here we apply normal approximation to the log-transformed ˆ , construct confidence interval based on this normal approximation, and then transform back to the confidence interval for σ. Efficiency of the log transformation has been verified by Chan et al. (2008) and Ng et al. (2009). Our simulation results also suggest that the log transformation does 9

improve the coverage probability, especially when the sample size is small.

4.2 Confidence

Interval

Based

on

Fisher

Information

Matrix The asymptotic variances and covariance of the maximum likelihood estimators can also be obtained by inverting the expected Fisher information matrix I, which is given by

 2  2 l  l        2   I  E  2    2 l  l   2       

(20)

     l      0 , Equations (14) - (16) can be simplified as Because E  l      0 and E         2  m  E  2 l     2 ,    

(21)

 2  m m 1  E  2 l     2   2 E  rk  1 Z k2:m:n exp  Z k:m:n   , k 1     

(22)

 2  m m 1 E  l      2   2 E  Z k:m:n  . k 1     

(23)

Equations (22) and (23) can be computed by first conditioning on J and then taking expectation with respect to J. The detailed expressions are as follows. 2   2       m m m r 1  E  2 l       E  E  2 l    | J    2   k 2 E  Z k2:m:n exp  Z k:m:n  | J  j  (24)  j  0 k 1          

 2  m m m 1 E  l      2   2 E  Z k:m:n | J  j  j  0 k 1     

(25)

In order to compute the above expectations, the marginal distribution of Zi:m:n given J should be obtained first. When i  j , the marginal distribution of Zi:m:n can be readily obtained based on (4) 10

and (11). Therefore, by direct integration, it can be shown that the moment generating function of Zi:m:n given J is given by

cj

i

j

E exp  tZi:m:n  | J  j    Pr  J  j  h1 k i

ak ,h,i , j FZk  ZT 

 h  k 

t 1

 L  t  1,  h  k  exp  ZT   , (26)

s

x 1  t where ZT  T    /  , and  L  x, s    t e dt is the lower incomplete gamma function. 0

The first and second derivatives of Equation (26) with respect to t are given by j i cj ak ,h ,i , j FZk  ZT  d E exp  tZi:m:n  | J  j   L  t  1,  h  k  exp  ZT    dt Pr  J  j  h 1 k i  h  k t 1



cj

j

i

Pr  J  j 

 h 1 k i

ak ,h ,i , j FZk  ZT 

 h  k 

t 1

log  h  k   L  t  1,  h  k  exp  ZT  

j i cj ak ,h ,i , j FZk  ZT  d2 E exp tZ | J  j    t  1,  h  k  exp  ZT      i:m:n   Pr J  j  dt 2    h1 k i  h  k t 1 L

 

2c j

Pr  J  j  cj

Pr  J  j 

i

j

 h 1 k i i

j

 h 1 k i

ak ,h ,i , j FZk  ZT 

 h  k 

t 1

ak ,h,i , j FZk  ZT 

 h  k 

t 1

log  h  k  L  t  1,  h  k  exp  ZT   log 2  h  k   L  t  1,  h  k  exp  ZT  

where L  x, s  and L  x, s  are respectively given by L  x, s  

s

s

d d2  L  x, s    log t  t x 1et dt and L  x, s   2  L  x, s    log 2 t  t x 1et dt . dx dx 0 0

Note that these two derivatives can be expressed as a function of the Meijer G-function (Mathai and Haubold 2008), and thus can be easily computed. 2 Based on these two derivatives, the expectations of Zi:m:n and  ri  1 Zi:m:n exp  Zi:m:n  can be

specified as

11

cj

E  Z i:m:n | J  j   

cj

i

Pr  J  j 

i

Pr  J  j 

j



j



h  k

h 1 k i

ak ,h ,i , j FZk  ZT 

h  k

h 1 k i

ak ,h ,i , j FZk  ZT 

L 1,  h  k  exp  ZT  

,

log  h  k   L 1,  h  k  exp  ZT  

(27)

c j  Ri  1 i j ak ,h ,i , j FZk  ZT  E  ri  1 Z i2:m:n exp  Z i:m:n  | J  j   L  2,  h  k  exp  ZT    Pr  J  j  h 1 k i  h  k 2 2c j  Ri  1

i

j

ak ,h ,i , j FZk  ZT 

log  h  k  L  2,  h  k  exp  ZT   2  h  k  c j  Ri  1 i j ak ,h ,i , j FZ  ZT   log 2  h  k   L  2,  h  k  exp  ZT    Pr  J  j  h 1 k i  h  k 2 

Pr  J  j 



.(28)

h 1 k i

k

On the other hand, when i > j, the marginal distribution of Zi:m:n can be readily obtained based on (5) and (11). Parallel to the above deduction, it can be shown that j 1 i FZk h  ZT  ci E  Z i:m:n | J  j   H 1, h exp  ZT     d k , h ,i , j  Pr  J  j  k 1 h  j 1 h j 1 i FZk h  ZT  ci  log  h   H 1, h exp  ZT     d k , h ,i , j  Pr  J  j  k 1 h  j 1 h

,

(29)

j 1 i FZk h  ZT  ci E  ri  1 Z i2:m:n exp  Z i:m:n  | J  j   d   k ,h,i, j  2 H  2, h exp  ZT   Pr  J  j  k 1 h  j 1 h j 1 i FZk h  ZT  2ci    dk ,h,i, j  2 log  h  H  2, h exp  ZT   Pr  J  j  k 1 h  j 1 h

.(30)

j 1 i FZk h  ZT  ci  d k , h ,i , j log 2  h   H  2, h exp  ZT     2 Pr  J  j  k 1 h  j 1 h 

x 1  t In the above expressions, U  x, s    t e dt is the upper incomplete gamma function, while s

U  x, s  and U  x, s  are respectively given by U  x, s  





d d2 U  x, s    log t  t x 1et dt and U  x, s   2 U  x, s    log 2 t  t x 1et dt . dx dx s s

Similarly, these two derivatives can also be expressed in terms of the Meijer G-function. 12

Based on the results of (27) - (30), Equations (24) and (25) can be computed, after which the Fisher information matrix (20) can be specified. Similarly, the approximate confidence intervals for μ and σ are then respectively given by









1 1 ˆ  z /2 I111 and ˆ exp z /2 I22 / ˆ , ˆ / exp z /2 I22 / ˆ  .  

4.3 Parametric Percentile Bootstrap Approach The normal approximations based on the asymptotic pivots are adequate when the effective sample size is large enough. However, when the number of failures is not sufficient, the normal approximations may not work well. Simulation techniques such as the bootstrap approaches may be able to provide more accurate approximate intervals. A widely used bootstrap method is the parametric percentile bootstrap described by Efron and Tibshirani (1993). To construct the confidence intervals for Θ, a typical parametric percentile bootstrap algorithm works as follows.

[1]. Collect the adaptive progressive Type-II censoring data (X, J) and compute the MLE of Θ,

ˆ , by maximizing (10). denoted as  [2]. Based on the adaptive censoring scheme  R1 , , Rm  and the switching time T, generate an adaptive Type-II progressive censoring dataset from the extreme value distribution ˆ   ˆ , ˆ  . with distribution parameter 

[3]. Obtain the MLE based on this simulated dataset. Denote this bootstrap estimate by ˆ *   ˆ * , ˆ *  . 

ˆ *   ˆ * , ˆ *  . ˆ *,  ˆ * ,...,  ˆ * , where  [4]. Repeat Steps [2] – [3] B times and obtain  i i i 1 2 B * * * [5]. Arrange ˆ1* , ˆ 2* ,..., ˆ B* and ˆ1* , ˆ 2*..., ˆ B* in ascending orders and obtain ˆ1 , ˆ 2 ,..., ˆ B

and ˆ 1 , ˆ  2 ..., ˆ  B  . *

*

*

A pair of two-sided 100(1 – α)% percentile bootstrap confidence intervals for μ and σ are then * * * * given by  ˆ B /2 , ˆ B1 /2  and ˆ  B /2 , ˆ  B1 /2  . It is noted that the bootstrap estimates

13

ˆ *,  ˆ * ,...,  ˆ * can also be used for bias correction. We can also construct the confidence interval  1 2 B

for the bias-corrected estimator. But this requires the double bootstrap which is quite timeconsuming, and thus is not discussed here.

4.4 Studentized Bootstrap Approach An advantage of the percentile bootstrap method is its simplicity. However, when the sample size is small, the percentile approach is generally not as accurate as the studentized bootstrap approach, abbreviated as bootstrap-t, with an appropriate transformation (Meeker and Escobar 1998). To obtain the bootstrap-t confidence intervals, the following algorithm can be applied.

[1] – [3]. The same as the steps in the percentile bootstrap above. [4]. Compute the variance-covariance matrix I 1* based on the observed information matrix (alternatively, the Fisher information matrix can be used). Obtain the t-statistics





log   log ˆ *  log ˆ / T    ˆ *  ˆ  / I111* and T



1* I 22 / ˆ *



[5]. Repeat Steps [2] – [4] B times and obtain T1 , T2 ,..., TB and T1log , T2log ,..., TBlog . [6]. Sort T1 , T2 ,..., TB and T1log , T2log ,..., TBlog in ascending orders and obtain the ordered    log  log  log  sequences T1 , T 2 ,..., T B  and T1 , T 2 ,..., T B .

A pair of two-sided 100(1 – α)% bootstrap-t confidence intervals for μ and σ are then given by









1 log  1 1  1  ˆ / exp T log  ˆ  T  ˆ ˆ ˆ  ˆ  B /2 I11 ,   T B1 /2 I11  and   B1 /2 I 22 /  ,  / exp T B /2 I 22 /   .  

5 SIMULATION RESULTS To evaluate the performance of the approaches based on the observed information matrix (OIM), the Fisher information matrix (FIM), the percentile bootstrap (PB) and the bootstrap-t (TB), a Monte Carlo simulation study is conducted in this section. Because of the lack of invariance properties of the estimated parameters, some representative settings in the simulation study 14

should be chosen rather than exhausting all the possible settings. For illustrative purpose, we consider the standard extreme value distribution with μ = 0 and σ = 1. The following censoring schemes are considered. Scheme 1: R1  R2  ...  Rm1  1 ; Scheme 2: R1  R2  ...  Rm/2  2 , Rm/2 1  ...  Rm1  0 ; Scheme 3: R1  R2  ...  Rm/2  0 , Rm/2 1  Rm/2 2  ...  Rm1  2 . Here  denotes the integer part of the number in the bracket. The biases, the coverage probabilities and the expected widths of the confidence intervals for these different censoring schemes, different sample sizes and T = – 1, 0 and 1 are presented in Tables 1 – 6.

15

Table 1. Coverage probabilities and expected width of the 95% confidence intervals for µ based on different methods for Scheme 1: The number outside the bracket is the coverage probability and the number in the bracket is the expected width.

(n, m)

T

(25, 10)

(50, 10)

(50, 20)

(50, 25)

bias

OIM

FIM

PB

TB

before corr.

after corr.

-1

-0.129

-0.0181

87.3 (1.23) 87.1 (1.23) 78.1 (1.32)

75.3 (1.79)

0

-0.108

0.0039

87.1 (1.23) 87.0 (1.22) 80.0 (1.31)

77.9 (1.78)

1

-0.110

-0.0005

87.1 (1.23) 87.0 (1.22) 80.6 (1.32)

78.1 (1.79)

-1

-0.196

-0.0159

83.9 (1.88) 83.1 (1.82) 78.1 (1.91)

66.8 (2.61)

0

-0.206

-0.0283

84.0 (1.88) 83.0 (1.81) 80.6 (1.92)

68.9 (2.67)

1

-0.247

-0.0710

83.9 (1.88) 83.0 (1.81) 79.7 (1.92)

68.1 (2.64)

-1

-0.056

-0.0027

91.1 (.907) 91.0 (.906) 88.1 (.973)

85.0 (1.12)

0

-0.060

-0.0013

90.9 (.904) 90.9 (.903) 87.4 (.967)

85.4 (1.12)

1

-0.066

-0.0053

90.9 (.916) 90.9 (.904) 88.4 (.967)

85.6 (1.12)

-1

-0.036

0.0011

93.1 (.766) 93.0 (.764) 92.3 (.799)

92.1 (.879)

0

-0.045

-0.0076

92.9 (.765) 92.9 (.764) 89.9 (.784)

89.9 (.858)

1

-0.044

-0.0050

92.9 (.764) 92.9 (.764) 91.1 (.786)

91.3 (.863)

-0.293

-0.0510

83.1 (2.69) 81.8 (2.57) 80.1 (2.63)

68.4 (3.83)

0

-0.265

-0.0193

83.9 (2.72) 82.5 (2.59) 78.5 (2.59)

66.7 (3.80)

1

-0.257

-0.0127

83.2 (2.71) 81.7 (2.57) 78.4 (2.58)

67.5 (3.77)

-0.060

-0.0120

91.4 (.947) 91.2 (.938) 88.5 (.947)

84.1 (1.05)

0

-0.058

-0.0093

91.4 (.946) 91.2 (.937) 90.2 (.946)

85.9 (1.05)

1

-0.048

0.0085

91.5 (.946) 91.3 (.937) 90.4 (.943)

85.6 (1.05)

-0.024

-0.0032

93.8 (.569) 93.8 (.568) 90.8 (.567)

90.1 (.594)

0

-0.025

-0.0057

93.5 (.559) 93.4 (.559) 93.2 (.564)

93.6 (.589)

1

-0.020

0.0009

93.6 (.558) 93.6 (.558) 92.8 (.561)

92.8 (.588)

(100, 10) -1

(100, 30) -1

(100, 50) -1

16

Table 2. Coverage probabilities and expected width of the 95% confidence intervals for σ based on different methods for Scheme 1: The number outside the bracket is the coverage probability and the number in the bracket is the expected width.

(n, m)

T

(25, 10)

(50, 10)

(50, 20)

(50, 25)

Bias

OIM

FIM

PB

TB

before corr.

after corr.

-1

-0.085

-0.0082

90.3 (1.03) 90.0 (1.05) 82.3 (.947)

94.7 (1.36)

0

-0.081

-0.0040

90.3 (1.03) 90.8 (1.05) 81.7 (.931)

95.0 (1.34)

1

-0.082

-0.0060

90.4 (1.03) 91.0 (1.05) 82.7 (.940)

95.4 (1.35)

-1

-0.082

0.0044

90.2 (1.12) 89.6 (1.10) 81.5 (1.00)

92.8 (1.41)

0

-0.094

-0.0096

90.2 (1.12) 89.6 (1.10) 81.8 (1.01)

93.7 (1.44)

1

-0.105

-0.0224

90.0 (1.12) 89.4 (1.10) 81.2 (1.01)

93.2 (1.43)

-1

-0.041

0.0050

92.9 (.742) 93.8 (.770) 87.9 (.718)

94.1 (.854)

0

-0.041

0.0053

92.9 (.742) 93.8 (.770) 88.5 (.711)

95.3 (.847)

1

-0.036

0.0049

92.5 (.741) 93.5 (.769) 88.1 (.710)

95.3 (.845)

-1

-0.027

-0.0017

93.0 (.613) 94.4 (.649) 90.4 (.630)

95.5 (.717)

0

-0.031

-0.0026

93.0 (.613) 94.4 (.649) 89.7 (.593)

95.2 (.675)

1

-0.032

-0.0021

93.3 (.608) 94.6 (.645) 90.0 (.591)

95.5 (.671)

-0.104

-0.0075

90.0 (1.16) 89.1 (1.11) 81.4 (1.04)

93.9 (1.53)

0

-0.093

-0.0053

90.0 (1.16) 89.1 (1.11) 80.7 (1.02)

94.9 (1.52)

1

-0.096

-0.0092

90.2 (1.16) 89.1 (1.12) 79.5 (1.02)

93.8 (1.51)

-0.033

-0.0047

93.5 (.642) 94.0 (.656) 90.6 (.622)

94.6 (.700)

0

-0.031

-0.0021

93.5 (.642) 94.0 (.656) 89.8 (.620)

95.3 (.700)

1

-0.029

-0.0005

93.5 (.641) 94.1 (.657) 90.0 (.619)

95.5 (.698)

-0.015

0.0009

94.2 (.436) 95.8 (.470) 91.1 (.452)

94.7 (.482)

0

-0.016

-0.0010

94.2 (.436) 95.8 (.470) 92.2 (.432)

94.8 (.460)

1

-0.018

-0.0024

94.0 (.431) 95.7 (.467) 91.6 (.426)

94.6 (.454)

(100, 10) -1

(100, 30) -1

(100, 50) -1

17

Table 3. Coverage probabilities and expected width of the 95% confidence intervals for µ based on different methods for Scheme 2: The number outside the bracket is the coverage probability and the number in the bracket is the expected width.

(n, m)

T

(25, 10)

(50, 10)

(50, 20)

(50, 25)

bias

OIM

FIM

PB

TB

before corr.

after corr.

-1

-0.124

-0.0209

87.3 (1.23) 87.1 (1.23) 83.3 (1.26)

83.8 (.908)

0

-0.129

-0.0297

87.1 (1.23) 87.0 (1.22) 81.8 (1.26)

81.3 (1.65)

1

-0.120

-0.0218

87.1 (1.23) 87.0 (1.22) 83.1 (1.27)

82.5 (1.66)

-1

-0.187

-0.0115

83.9( 1.88) 83.1 (1.82) 80.5 (1.87)

82.9 (.995)

0

-0.186

0.0024

84.0 (1.88) 83.0 (1.81) 79.7 (1.89)

71.8 (2.65)

1

-0.187

-0.0172

83.9 (1.88) 83.0 (1.81) 78.8 (1.87)

69.4 (2.64)

-1

-0.056

-0.0013

91.1 (.907) 91.0 (.906) 87.2 (.917)

89.2 (.674)

0

-0.056

-0.0007

90.9 (.904) 90.9 (.903) 87.9 (.914)

87.0 (1.04)

1

-0.055

0.0038

90.9 (.906) 90.9 (.904) 88.1 (.918)

87.5 (1.04)

-1

-0.037

-0.0010

93.1 (.766) 93.0 (.764) 90.4 (.769)

90.8 (.562)

0

-0.031

0.0022

92.9 (.765) 92.9 (.764) 92.1 (.777)

92.8 (.832)

1

-0.040

-0.0101

92.9 (.764) 92.9 (.764) 90.7 (.778)

92.3 (.833)

-0.266

-0.0526

83.0 (2.66) 81.7 (2.53) 78.6 (2.56)

80.9 (1.02)

0

-0.264

-0.0462

83.1 (2.66) 81.9 (2.54) 79.8 (2.57)

67.7 (3.74)

1

-0.266

-0.0090

83.1 (2.66) 81.8 (2.53) 80.3 (2.58)

68.1 (3.73)

-0.046

-0.0002

91.6 (.895) 91.5 (.889) 88.9 (.899)

90.5 (.560)

0

-0.043

0.0017

91.6 (.895) 91.5 (.889) 90.4 (.896)

86.6 (.990)

1

-0.045

0.0052

91.6 (.894) 91.5 (.888) 90.2 (.903)

87.2 (.999)

-0.019

0.0066

94.1 (.548) 94.0 (.547) 93.9 (.550)

92.9 (.398)

0

-0.015

0.0029

93.9 (.548) 93.9 (.547) 92.7 (.551)

93.6 (.568)

1

-0.015

-0.0039

94.0 (.548) 94.0 (.548) 92.8 (.551)

93.3 (.569)

(100, 10) -1

(100, 30) -1

(100, 50) -1

18

Table 4. Coverage probabilities and expected width of the 95% confidence intervals for σ based on different methods for Scheme 2: The number outside the bracket is the coverage probability and the number in the bracket is the expected width.

(n, m)

T

(25, 10)

(50, 10)

(50, 20)

(50, 25)

bias

OIM

FIM

PB

TB

before corr.

after corr.

-1

-0.076

-0.0058

90.7 (.970) 93.0 (1.08) 82.4 (1.64)

95.2 (1.24)

0

-0.077

-0.0081

90.7 (.963) 92.8 (1.06) 83.6 (.895)

95.3 (1.24)

1

-0.080

-0.0054

91.1 (.967) 93.4 (1.07) 83.4 (.903)

94.6 (1.25)

-1

-0.089

-0.0044

90.3 (1.10) 90.4 (1.10) 70.3 (2.59)

94.3 (1.41)

0

-0.088

-0.0003

90.3(1.10)

90.4 (1.10) 82.9 (1.00)

93.8 (1.45)

1

-0.089

-0.0043

90.1 (1.10) 90.1 (1.10) 81.6 (.992)

94.4 (1.44)

-1

-0.037

-0.0041

92.7 (.697) 95.3 (.783) 86.3 (1.04)

96.0 (.781)

0

-0.038

-0.0037

92.7(.693)

95.0 (.773) 88.1 (.668)

94.7 (.780)

1

-0.036

-0.0028

92.7 (.693) 95.0 (.773) 89.5 (.669)

95.0 (.781)

-1

-0.026

0.0012

93.6 (.569) 96.6 (.693) 91.5 (.820)

94.8 (.621)

0

-0.024

0.0012

93.4 (.547) 96.6 (.653) 91.4 (.539)

95.6 (.597)

1

-0.025

0.0022

93.4 (.547) 96.8 (.649) 90.8 (.540)

95.2 (.600)

-0.110

-0.0064

89.8 (1.15) 89.1 (1.13) 66.4 (3.70)

94.2 (1.49)

0

-0.094

-0.0082

89.9 (1.15) 89.3 (1.12) 81.0 (1.02)

93.4 (1.51)

1

-0.093

-0.0075

89.8 (1.15) 89.0 (1.12) 82.3 (1.02)

94.1 (1.50)

-0.027

0.0045

93.7 (.617) 95.1 (.658) 86.4 (.993)

94.8 (.672)

0

-0.027

-0.0058

93.6 (.617) 95.1 (.658) 91.2 (.598)

95.1 (.670)

1

-0.028

0.0077

93.6 (.617) 95.0 (.657) 91.3 (.602)

95.1 (.674)

-0.015

-0.0014

94.3 (.401) 97.7 (.497) 94.4 (.565)

95.8 (.418)

0

-0.013

0.0024

94.1 (.374) 98.0 (.470) 93.4 (.372)

95.2 (.391)

1

-0.012

0.0056

94.1 (.374) 98.0 (.468) 92.0 (.372)

93.5 (.391)

(100, 10) -1

(100, 30) -1

(100, 50) -1

19

Table 5. Coverage probabilities and expected width of the 95% confidence intervals for µ based on different methods for Scheme 3: The number outside the bracket is the coverage probability and the number in the bracket is the expected width.

(n, m)

T

(25, 10)

(50, 10)

(50, 20)

(50, 25)

bias

OIM

FIM

PB

TB

before corr.

after corr.

-1

-0.131

-0.0073

85.2 (1.38) 84.7 (1.35) 80.9 (1.37)

76.9 (1.84)

0

-0.145

-0.0191

85.2 (1.37) 84.7 (1.34) 79.3(1.35)

76.1 (1.86)

1

-0.142

-0.0138

85.2 (1.37) 84.7 (1.34) 80.2 (1.36)

77.3 (1.89)

-1

-0.202

0.0103

83.4 (2.01) 82.2 (1.92) 78.5 (1.94)

64.2 (2.48)

0

-0.203

-0.0594

83.4 (2.01) 82.3 (1.92) 79.4 (1.95)

66.5 (2.57)

1

-0.204

-0.0095

83.3 (2.00) 82.2 (1.92) 78.6 (1.96)

64.7 (2.54)

-1

-0.054

0.0097

90.1 (1.03) 89.8 (1.01) 85.7 (1.03)

83.2 (1.20)

0

-0.067

-0.0058

90.0 (1.02) 89.7 (1.00) 85.9 (1.01)

83.4 (1.19)

1

-0.071

-0.0091

89.9 (1.02) 89.6 (1.00) 87.3 (1.02)

84.5 (1.20)

-1

-0.045

-0.0004

91.5 (.831) 91.4 (.827) 90.1 (.839)

87.9 (.939)

0

-0.043

0.0007

91.6 (.810) 91.5 (.806) 88.7 (.813)

88.0 (.904)

1

-0.035

0.0088

91.6 (.808) 91.5 (.804) 89.5 (.811)

88.9 (.906)

-0.259

-0.0084

82.8 (2.73) 81.5 (2.60) 77.4 (2.60)

65.8 (3.83)

0

-0.304

-0.0568

82.7 (2.73) 81.2 (2.59) 79.5 (2.62)

66.3(3.83)

1

-0.308

-0.0530

82.7 (2.73) 81.4 (2.60) 79.9 (2.62)

67.5 (3.84)

-0.059

-0.0087

91.2 (.995) 90.9 (.983) 89.6 (.991)

84.0 (1.11)

0

-0.069

-0.0182

91.2 (.995) 90.9 (.983) 89.3(.985)

85.1 (1.10)

1

-0.058

-0.0062

91.2 (.994) 91.0 (.982) 88.0 (.986)

84.4 (1.10)

-0.023

0.0041

93.3 (.601) 93.3 (.599) 93.2 (.600)

92.6 (.634)

0

-0.022

0.0004

93.4 (.585) 93.3 (.584) 92.3 (.586)

91.9 (.618)

1

-0.023

0.0002

93.4 (.584) 93.3 (.583) 92.8 (.582)

92.6 (.614)

(100, 10) -1

(100, 30) -1

(100, 50) -1

20

Table 6. Coverage probabilities and expected width of the 95% confidence intervals for σ based on different methods for Scheme 3: The number outside the bracket is the coverage probability and the number in the bracket is the expected width.

(n, m)

T

(25, 10)

(50, 10)

(50, 20)

(50, 25)

bias

OIM

FIM

PB

TB

before corr.

after corr.

-1

-0.095

-0.0128

89.8 (1.09) 88.8 (1.05) 81.5 (.969)

95.0 (1.43)

0

-0.104

-0.0219

90.0 (1.08) 89.2 (1.04) 81.0 (.948)

95.4 (1.40)

1

-0.094

-0.0117

89.8 (1.08) 88.9 (1.04) 81.5 (.956)

95.1 (1.41)

-1

-0.090

-0.0024

89.9 (1.14) 88.6 (1.10) 80.7 (1.00)

91.4 (1.33)

0

-0.109

-0.0039

90.0 (1.15) 88.8 (1.09) 81.4 (1.01)

92.3 (1.38)

1

-0.091

-0.0035

90.0 (1.15) 88.9 (1.10) 81.4 (1.02)

91.6 (1.36)

-1

-0.041

0.0060

92.5 (.789) 92.0 (.776) 87.8 (.748)

95.4 (.903)

0

-0.051

-0.0077

92.4 (.780) 92.0 (.776) 87.0 (.736)

95.2 (.891)

1

-0.047

-0.0035

92.6 (.780) 92.3 (.767) 88.4 (.741)

94.7 (.896)

-1

-0.035

-0.0004

93.2 (.693) 92.8 (.686) 90.3 (.670)

95.0 (.777)

0

-0.032

0.0009

93.0 (.653) 92.6 (.646) 88.6 (.630)

95.1 (.728)

1

-0.038

-0.0032

93.2 (.653) 93.0 (.645) 88.4 (.625)

95.7 (.725)

-0.091

-0.0017

90.1 (1.16) 89.0 (1.12) 79.8 (1.02)

94.0 (1.53)

0

-0.103

-0.0148

89.8 (1.17) 88.4 (1.11) 81.1 (1.03)

93.8 (1.53)

1

-0.109

-0.0219

90.0 (1.17) 88.7 (1.12) 80.9 (1.03)

93.9 (1.53)

-0.036

-0.0060

93.6 (.664) 93.3 (.655) 90.1 (.637)

94.4 (.723)

0

-0.040

-0.0107

93.2 (.663) 93.0 (.654) 89.5 (.635)

95.0 (.723)

1

-0.032

-0.0014

93.4 (.664) 93.1 (.655) 88.8 (.635)

94.4 (.722)

-0.018

0.0072

94.4 (.498) 94.2 (.496) 92.5 (.487)

95.0 (.524)

0

-0.018

-0.0042

94.2 (.471) 94.0 (.469) 93.5 (.460)

95.4 (.494)

1

-0.018

0.0050

94.0 (.468) 93.9 (.465) 91.9 (.451)

94.6 (.486)

(100, 10) -1

(100, 30) -1

(100, 50) -1

21

A simple comparison between columns 3 and 4 in Tables 1-6 reveals that the bias is significantly reduced after the bootstrap bias correction, especially when the sample size is small. From these tables, we can also see that the coverage probabilities based on the observed information matrix are almost the same to those based on the expected information matrix. Therefore, the approach based on observed information matrix is recommended because of its simplicity. It is also interesting to observe that performance of the parametric bootstrap approach is much lower than performance of the approaches based on the information matrices. The bootstrap-t confidence intervals for µ are also not satisfactory. This might be due to the fact that the MLE ˆ and ˆ are negatively biased. When bootstrap estimates are obtained based on these biased estimators, they are again negatively biased. This makes the bootstrap estimates highly biased, leading to low coverage probabilities. Based on this argument, we conduct another simulation (not shown) that simply moves the bootstrap-p confidence interval by uˆ . The results indicate that after the simple shifting, the coverage probabilities have been improved, and are almost the same as those based on the information matrixes. On the other hand, the bootstrap-t confidence intervals for σ are quite close to 95% even when the effective sample size is small. This indicates the effectiveness of the log-transformation. But when the sample size is large enough, say (100, 50), the confidence intervals for σ based on observed/expected information matrices are also satisfactory. Based on this simulation, the bootstrap bias correction is strongly recommended when the sample size is small. We also recommend the approach based on the observed information matrix for constructing confidence intervals for µ regardless of the sample size, the bootstrap-t approach for σ when the sample size is small, and the approach based on observed information matrix for σ when the sample size is large. A fringe benefit of using the bootstrap approach is that the bias correction can be fulfilled using the same bootstrap samples.

6 CONCLUSION This study investigated some general properties of the adaptive Type-II censoring schemes and proposed the bootstrap bias correction method to cut down the bias of the point estimators. It was found that the bias-corrected estimators perform much better than the MLE when the sample size 22

is not large enough. This paper also studied the construction of confidence intervals for the two parameters of the extreme value distribution. Both the observed and the expected information matrices were derived. The performance of different methods in constructing confidence intervals was compared via a Monte Carlo simulation. Some recommendations about construction of the confidence intervals were made.

Acknowledgement: Mr. Ye’s work is partially supported by the Global Scholarship Programme for Research Excellence 2011 provided by the Chinese University of Hong Kong. Prof. Chan and Prof Ng’s work is supported by the GRF grant CUHK 410410 of HKRGC. Prof. Xie’s work is supported by a grant from City University of Hong Kong (Project No.9380058).

APPENDIX

Distribution of J For the lifetime distribution F  x  , consider the regular progressive Type-II censoring scheme

R1 ,..., Rm  with samples Y1 ,..., Ym  . When 0