Extremes 2:3, 245±268, 1999 # 2000 Kluwer Academic Publishers. Manufactured in The Netherlands.
Non-Parametric Estimation of the Limit Dependence Function of Multivariate Extremes B. ABDOUS Dept. matheÂmatiques et informatique, Universite du QueÂbec aÁ Trois-RivieÁres, Trois-RivieÁres, QC, Canada G9A 5H7 E-mail:
[email protected] K. GHOUDI Dept. matheÂmatiques et informatique, Universite du QueÂbec aÁ Trois-RivieÁres, Trois-RivieÁres, QC, Canada G9A 5H7 E-mail:
[email protected] A. KHOUDRAJI Dept. matheÂmatiques, Universite Cadi Ayyad, Marrakech, Morocco [Received May 13, 1998; Revised August 30, 1999; Accepted November 16, 1999] Abstract. This paper presents a new estimation procedure for the limit distribution of the maximum of a multivariate random sample. This procedure relies on a new and simple relationship between the copula of the underlying multivariate distribution function and the dependence function of its maximum attractor. The obtained characterization is then used to de®ne a class of kernel-based estimates for the dependence function of the maximum attractor. The consistency and the asymptotic distribution of these estimates are considered. Key words. multivariate extremes, kernel estimation, dependence function, Gaussian process, regular variation AMS 1991 Subject Classi®cations.
1.
PrimaryÐ62G32; SecondaryÐ62G05.
Introduction
In many statistical applications the study of extreme values is of great importance. In hydrology, for example, the maximum ¯ow of a river is essential in the design of a dam. For pollution data, it is usually the highest level of pollutants that need be watched. Most of these studies reduce to the estimation of the joint distribution of the extremes of several characteristics. Though, results of this work are valid for multivariate extremes, for simplicity of the presentation, we only focus on the bivariate case. Sklar (1959) de®ned a copula or dependence function as any multivariate distribution function with uniform marginals. Given a bivariate random vector
X; Y with distribution function H having continuous marginals H1 and H2 , there exists a unique copula C associated to H satisfying H
x; y C
H1
x; H2
y, for any x and y. The dependence
246
ABDOUS, GHOUDI, AND KHOUDRAJI
structure between X and Y is perfectly characterized by C. Next, let
X1 ; Y1 ; . . . ;
Xn ; Yn be a sequence of independent random vectors drawn from H
? ; ? and let
Xn ; Yn
max
X1 ; . . . ; Xn ; max
Y1 ; . . . ; Yn : H is said to belong to the max-domain attraction of an extreme value distribution H with non degenerate marginals, if there exist sequences a1n 40, a2n 40, b1n and b2n such that lim P
n??
Xn ÿ b1n Y ÿ b2n x; n y a1n a2n
lim H n
a1n x b1n ; a2n y b2n n??
H
x; y:
1
Let C denote the copula associated to H and let H1 and H2 be the marginals of H . From Galambos (1978), H belongs to the max-domain of attraction of H if and only if H1 and H2 belong to the domain of attraction of H1 and H2 respectively and if C satis®es C
x; y lim C
x1=m ; y1=m m ; m??
for all 0 x; y 1:
2
The marginals H1 and H2 are univariate extreme value distributions. Their characterization can be found in Galambos (1978). As for the copula C , Geoffroy (1958), Tiago de Olivera (1958), Sibuya (1960) and Pickands (1981) gave the following representation. ln
y ; for 0 x; y 1; C
x; y exp ln
xyA ln
xy
3
where A is a convex function on [0,1] satisfying max
t; 1 ÿ t A
t 1. The function A is often called the dependence function associated to C . Clearly, the estimation of H may be accomplished via the estimation of its marginals and the dependence function A
? . In this work, we deal with A
? only, because the problem of estimating univariate extreme value distributions was considered by several authors (see, e.g. Davis and Resnick, 1984; Dekkers and de Haan, 1989). Note also that estimation of multivariate extreme-value distributions has been the subject of several papers, to name a few: Pickands (1981), Tiago de Oliveira, 1989, Tawn (1988), Smith, Tawn and Yuen (1990), Joe, Smith and Weissman (1992), Deheuvels (1991) and CapeÂraaÁ, FougeÁres and Genest (1997). Unfortunately, most of these authors assume that they dispose of a random sample from the distribution H itself. In practice, such samples are hard to come by and if any, they usually suffer from the smallness of their sizes. To circumvent these dif®culties, de Haan and Resnick (1993) and Einmahl, de Haan and Sinha (1997) proposed estimates of H based on a sample from H. However, their
MULTIVARIATE EXTREME DEPENDENCE FUNCTIONS
247
procedures need estimates of the normalizing sequences a1n ; a2n ; b1n and b2n in (1). Unlike all these previous works, we propose a non-parametric estimate of the dependence function A
? which does not require any auxiliary estimates and which is suitable for samples drawn from either H or H . The following characterization constitutes a key element in our estimation procedure. From now on, we shall say that a copula C belongs to the domain of attraction of an extreme value copula C if and only if (2) holds. Theorem 1: Let C be an extreme value copula with associated dependence function A. Then a copula C belongs to the domain of attraction of C if and only if h i 1 ÿ C
1 ÿ u1ÿt ;
1 ÿ ut A
t; Vt [ 0; 1:
4 lim u?0 u Remark: More generally, in the d dimensional case
d 2, replaces
t; 1 ÿ t by Pone d the vector
t1 ; . . . ; td , where 0 tk 1 for k 1; . . . ; d and k1 tk 1, and de®nes A from 0; 1d ÿ 1 into 0; 1. The key idea of the new estimation scheme consists in exploiting Theorem 1 to reduce the estimation of A
? to that of a quantity depending only on C the copula associated to the original data. To be speci®c let
U; V be a random vector with distribution function C, and for each t [ 0; 1, let 8 1=
1 ÿ t ; V 1=t g < 1 ÿ maxfU Wt 1 ÿ U : 1ÿV
for t [
0; 1; for t 0;
5
for t 1:
Observe that Wt takes values in [0,1] and its distribution function Gt satis®es 1ÿt t Gt
u 1 ÿ C
1 ÿ u ;
1 ÿ u . Consequently, C belongs to, d
C , the max domain of attraction of C if and only if lim
u?0
Gt
u A
t; u
Vt [ 0; 1:
6
From this equation one can view A
t as a parameter of regular variation of the distribution function Gt at zero or as the value of its right-hand-side derivative at zero. In this paper, we shall adopt this last point of view and estimate A
t by studying the local behavior of the distribution function Gt near zero. The next Section introduces a class of kernel-based non-parametric estimates of the dependence function A
t. The presentation is divided into two subsections. The ®rst Subsection deals with the estimation procedure when the marginal distributions are assumed to be known. The second Subsection presents the case of unknown marginals.
248
ABDOUS, GHOUDI, AND KHOUDRAJI
Asymptotic behavior of these estimates is provided for each case. Practical issue on bandwidth selection and an example of application are given in Section 3. Proofs are provided in Section 4.
2.
Non-parametric estimates of A( ? )
Let
X1 ; Y1 ; . . . ;
Xn ; Yn be a random sample drawn from a bivariate distribution function H with associated copula C and marginals H1 and H2 . Let Hn be the joint empirical distribution function and let H1n and H2n be the empirical marginal distribution functions. ÿ1 ÿ1
u; H2n
vg. Denote by Cn the empirical copula function given by Cn
u; v Hn fH1n ÿ1 where F
? stands for the left continuous inverse function of any distribution function F. Let also Cn
u; v Hn fH1ÿ 1
u; H2ÿ 1
vg. By Theorem 1, the copula C is in d
C if and only if the function 1ÿt t Gt
v 1 ÿ Cf
1 ÿ v ;
1 ÿ v g has a linear behavior in the neighborhood of zero. Thus a natural estimate of A
t can be obtained by ®tting locally a linear polynomial to the empirical estimate of Gt . To simplify the introduction of such estimator, it is useful to start by assuming that marginals are known. The more realistic and practical case of unknown marginal will be considered later.
2.1.
Known marginals
Assume that the marginals H1 and H2 are given. For each t; v [ 0; 1, we propose to use a weighted least squares criterion to ®t a local linear polynomial to the function Gt;n
v 1 ÿ Cn f
1 ÿ v1ÿt ;
1 ÿ vt g. This is achieved by minimizing with respect to a the following criterion Z
1 0
2 Kh
u ÿ v Gt;n
u ÿ au du;
7
where h (the smoothing parameter) controls the size of the neighborhood of v, K denotes a weight function (or kernel) and Kh
u K
u=h=h. The above integral is restricted to 0; 1, the support of Gt;n , in order to avoid the well-known boundary effects problem. The solution of (7), given by ^ v A
t;
R1 0
uKh
u ÿ vGt;n
u du ; R1 2 0 u Kh
u ÿ v du
provides an estimate of the local behavior of the function Gt
? near v. Now, an estimate of A
t is provided by
249
MULTIVARIATE EXTREME DEPENDENCE FUNCTIONS
^ A
t; ^ 0 A
t R1 0
R1 0
uKh
uGt;n
u du R1 2 0 u Kh
u du
uKh
u
1 ÿ Cn f
1 ÿ u1 ÿ t ;
1 ÿ ut g du : R1 2 0 u Kh
u du
8
Remark: As noted earlier, A
t can be viewed as a parameter of regular variation of the function Gt at zero. Indeed, recall that a special class of regularly varying distribution functions of order ÿ a at zero, is given by the set of distributions F satisfying F
0 0 and for which there exists a constant b such that limx?0 F
x=xa b. In fact from (6), one sees that A(t) plays the role of the constant b in the regularly varying distribution function Gt at zero. Here the index of regular variation a is equal to 1. Estimation of the parameters of regularly varying distribution functions received considerable attention in the literature, see for instance Hill (1975), Hall (1982), CsoÈrgoÈ, Deheuvels and Mason (1985), Haeusler and Teugels (1985), Dekkers and de Haan (1989), and Beirlant, Vynckier and Teugels (1996) and the references therein. Most of these references focus on the estimation of the parameter a, however, Hill (1975) and Hall (1982) proposed an estimate of the parameter b, which in our context becomes A^H
t
r ; nWt;
r
9
where 1 r n ÿ 1 and Wt;
r is the rth order statistic of the sequence h i Wt;i 1 ÿ max H1
Xi 1=
1 ÿ t ; H2
Yi 1=t ;
i 1; . . . ; n:
This estimate is in fact a special case of (8) obtained by choosing K as the Dirac mass function at zero and taking v Wt;
r . The following notation will be used in the sequel R1
uKh
uGt
u du R1 2 0 u Kh
u du R1 1ÿt t uK
u 1 ÿ Cf
1 ÿ u ;
1 ÿ u g du h 0 : R1 2 0 u Kh
u du
m
t; h
0
10
^ ? . The next The rest of this Subsection is devoted to the asymptotic behavior of A
Theorem states the consistency results. Theorem 2: If K is a bounded positive function satisfying Z 05 v2 K
v dv5?:
250
ABDOUS, GHOUDI, AND KHOUDRAJI
If C is in the max-domain of attraction of C and h hn goes to 0 as n tends to in®nity, then the following hold. ^ ÿ A
tj 0 in probability. (i) If limn?? nh2 ?, then limn?? sup0 t 1 jA
t (ii) If ? X
exp
ÿ gnh2 5?;
for every g40;
11
n1
^ ÿ A
tj 0 almost surely. then, lim sup jA
t n?? 0 t 1
Remark: The conditions on h are not optimal and are used for simplicity of the proof. The optimal conditions can be obtained using the results on the rate of growth of the weighted multivariate empirical process (see for instance Alexander (1987)) instead of the rate of growth of the empirical process. Unfortunately, this will lengthen the proof considerably. Since t lives in 0; 1, one gets, as an immediate consequence, the following Lp convergence results. Corollary 1: Under the assumption (i) and the assumption (ii) of Theorem 2, Z lim
n??
1
0
^ ÿ A
tjp dt jA
t
1=p 0;
in probability and almost surely respectively, for any p 1. ^ ? is established in the next Theorem where it is shown that The weak convergence of A
the process an
t
p ^ ÿ m
t; h nh A
t
12
converges weakly to a mean zero Gaussian process. Theorem 3: sup
If K is a bounded positive function satisfying
ÿ ?5v5?
v4 K
v5? and
Z 05
jvj3 K
vdv5?:
If C is in the max-domain of attraction of C and lim n??
hn
nhn
13 ÿ1
0, then an
t
MULTIVARIATE EXTREME DEPENDENCE FUNCTIONS
251
converges weakly to a continuous Gaussian process with covariance function G
? ; ? satisfying R? R? G
s; t
0
0
xyK
xK
yY
x; y; s; tdx dy ; ÿR ? 2 2 0 x K
xdx
where Y
x; y; s; t
xA
t yA
s
txVsy ÿ
1 ÿ txV
1 ÿ sy
txVsy A
1 ÿ txV
1 ÿ sy
txVsy
with xVy denoting the maximum of x and y. This result could be used to construct con®dence regions for A
? . Pointwise con®dence intervals are also provided by using the above Theorem and the following estimate of the ^ variance of A
t ^ ^ t A
t G
t;
R? R? 0
0
xy min
x; yK
xK
ydx dy ; ÿR ? 2 2 0 x K
xdx
^ in the expression of G
t; t. which is obtained by substituting A
t by A
t Remark: In general, the estimate A^ is not convex and does not respect ^ 1. This problem is not special to the estimator presented here, max
t; 1 ÿ t A
t but it is common to most of the estimator of A found in the literature, see for example Deheuvels (1991). A possible solution to bypass this drawback is achieved by modifying (7) as follows Z Minimize
1 0
2 Kh
u ÿ v Gt;n
u ÿ a
tu du
Subject to: a
t is convex max
t; 1 ÿ t a
t 1; for t [ 0; 1: Smoothing under constraints has been successfully considered by many authors, see for instance Delecroix, Simioni and Thomas-Agnan (1996) and Mammen and ThomasAgnan (1999). We intend to tackle the minimization problem above in a forthcoming paper. Theorem 3, still holds if one replaces m
t; h by A
t in (12) provided p nh
A
t ÿ m
t; h converges to zero as n goes to in®nity. This is achieved by conditions on the smoothness of the copula function C and on the speed of convergence of h to zero.
252 2.2.
ABDOUS, GHOUDI, AND KHOUDRAJI
Unknown marginals
If the marginals are unknown, then the same procedure can be repeated with Cn and Gt;n
v 1ÿt t replaced by Cn and Gt;n
v 1 ÿ Cn f
1 ÿ v ;
1 ÿ v g respectively. This yields the following class of estimates. ~ v A
t;
R1 0
uKh
u ÿ vGt;n
udu : R1 2 0 u Kh
u ÿ vdu
~ A
t; ~ 0. Again the estimate of A
t is given by A
t
Remark: The analogous of (9) is given by A~H
t
r ; ~ nWt;
r
~t;
r is the rth order statistic of the sequence where 1 r n ÿ 1 and W h i ~t;i 1 ÿ max H1n
Xi 1=
1 ÿ t ; H2n
Yi 1=t ; W
i 1; . . . ; n:
~ ? are summarized in the next Theorem. To simplify its The asymptotic properties of A
statement, set C
x; y; s; t 3fxA
t yA
s x yg ÿ f
1 ÿ txV
1 ÿ syg ÿ
txVsy tx ÿ f
1 ÿ txV
1 ÿ syg txA f
1 ÿ txV
1 ÿ syg tx
txVsy ÿ
1 ÿ tx
txVsyA
1 ÿ tx
txVsy sy ÿ f
1 ÿ txV
1 ÿ syg syA f
1 ÿ txV
1 ÿ syg sy
txVsy ÿ
1 ÿ sy
txVsyA
1 ÿ sy
txVsy sy ÿ
1 ÿ tx syA
1 ÿ tx sy
253
MULTIVARIATE EXTREME DEPENDENCE FUNCTIONS
tx ÿ tx
1 ÿ syA tx
1 ÿ sy ÿ f
1 ÿ txV
1 ÿ syg
txVsyA
txVsy ; f
1 ÿ txV
1 ÿ syg
txVsy
for 0 x; y; s; t 1. Theorem 4: Let K be a bounded positive function satisfying (13). Assume that C is in the max-domain of attraction of C and that it admits continuous ®rst order partial derivatives. If h hn goes to 0 as n tends to in®nity, then the following hold. ~ ÿ A
tj converges to zero in probability. (i) If limn!? nh2 ?, then sup0t1 jA
t ~ ÿ A
tj 0 almost surely. (ii) If (11) is satis®ed then limn!? sup0t1 jA
t p 2 2 ~ ÿ m
t; h converges in (iii) If limn!? nh =
ln n
ln ln n ?, then nh A
t distribution to a continuous Gaussian process with covariance function given by R? R? xyK
xK
yC
x; y; s; tdx dy : G
s; t 0 0 ÿR ? 2 2 0 x K
xdx
3.
Selection of the bandwidth h
We give in this Section some guidelines to select the smoothing parameter h. We deliberately omit details because we will present them together with a thorough simulation study in a forthcoming paper. Since our estimation problem reduces essentially to a local density estimation problem, we can adopt any of the many existing local bandwidth selectors (see Hazelton (1996) for a ~ and review). To see how this can be achieved, recall that for each ®xed t [ 0; 1, both A
t ^ A
t estimate the probability density of the r.v. Wt , de®ned by (5), at the support's endpoint 0. Thus, a natural procedure of selecting the bandwidth h should be based on a local criterion. The most popular and tractable measure of local error is the Mean Squared Error (MSE), namely ÿ A
t2 ; M
h EA
t stands for either A
t ~ or A
t. ^ where A
t Since A
? is unknown, the optimal parameter h argminh M
h cannot be evaluated in practice. One has to use Taylor series expansion of M
h and then estimate the unknowns by means of plug-in or cross-validation techniques. Also, instead of appealing to asymptotic approximations of M
h, we can
254
ABDOUS, GHOUDI, AND KHOUDRAJI
Figure 1.
estimate the criterion M
h by means of the smoothed local cross-validation or bootstrap methods. Finally, note that these selectors will deliver a bandwidth which varies with t. To reduce computations, we can consider minimizing a global criterion. To this end, we propose to integrate M
h with respect to t, i.e. we seek for a global bandwidth h which minimizes Z IM
h
1 0
ÿ A
t dt: EA
t 2
Here again, we will face the problem of estimating IM
h. But this dif®culty can be easily bypassed by adopting one of the various global bandwidth selectors, see, e.g. Devroye (1997) for a complete and recent review. In the following simulation study, the L1 -double kernel method (Devroye (1997)) has been used to select the bandwidths. The following ®gures show the graphs of the theoretical curve of dependence function A, Hill's estimate and the estimate AÃ discussed in this paper with h selected using the double kernel method. Figure 1 (a) shows the result for a sample of 500 observations of independent uniform random vectors. For Figure 1 (b), a sample of size 500 was generated from the Gumbel-Hougaard family (Gumbel (1960)) with parameter y 2. The optimal choice for h was 0.9346 for the ®rst simulation and 0.3316 for the second.
4.
Proofs
Before dealing with proofs of the results stated earlier, we begin by considering the next Lemma which will be used repeatedly.
MULTIVARIATE EXTREME DEPENDENCE FUNCTIONS
255
Lemma 1: Let C be an extreme value copula with associated dependence function A. Then a copula C belongs to the domain of attraction of C if and only if 1 ÿ C
1 ÿ u; 1 ÿ v A
t u;v?0 uv lim
14
holds for any sequence
u; v in
0; 12 satisfying limu;v?0 v=
u v t. Proof of Lemma 1: The proof is achieved by showing that (14) is equivalent to (2). First, let us show that (14) implies (2). Clearly, (2) holds for x 0; 1 or y 0; 1. Next, for each 05x; y51 and any integer m40, let u 1 ÿ x1=m and v 1 ÿ y1=m . Note that as m goes to in®nity u and v go to zero and v=
u v goes to t ln
y= ln
xy. Now, observe that h i lnC
x1=m ; y1=m 1 ÿ C
x1=m ; y1=m m
2 ÿ x1=m ÿ y1=m 1=m 1=m ln
xy: m ln C
x ; y ln
xy 1 ÿ C
x1=m ; y1=m 2 ÿ x1=m ÿ y1=m The ®rst and the last fractions go to ÿ 1 as m goes to in®nity, since limz?1
1 ÿ z= ln
z ÿ 1 and limz?0 1 ÿ exp
z=z ÿ1. By applying (14), one sees that the second fraction converges to Afln
y= ln
xyg. Taking the exponential of both sides yields (2). To complete the proof it remains to show that (2) implies (14). Assume that (2) holds, then for any sequence
um ; vm
1 ÿ x1=m ; 1 ÿ y1=m , where 05x; y51 and m is an arbitrary integer, one has 1 ÿ C
1 ÿ um ; 1 ÿ vm 1 ÿ C
1 ÿ um ; 1 ÿ vm 1 m lnfC
1 ÿ um ; 1 ÿ vm g: um vm lnfC
1 ÿ um ; 1 ÿ vm g m
um vm Quite similar arguments to those used previously give lim
um ;vm ?0
1 ÿ C
1 ÿ um ; 1 ÿ vm A
t; um v m
15
where t limm?? vm =
um vm . Next, it will be shown that for any a40 and 0 u; au 1 one has 1 ÿ C
1 ÿ u; 1 ÿ au a A : u?0
1 au a1 lim
16
For, observe that 0
1 a ÿ 1; ln
1 ÿ u ln
1 ÿ au
17
256
ABDOUS, GHOUDI, AND KHOUDRAJI
for 0 a 1 and 0 u 1 and ÿa
1 a ÿ 0; ln
1 ÿ u ln
1 ÿ au
18
for a 1 and 0 au 1. Let x exp
ÿ 1 and let y xa then for any 05u; au51 there exists an integer m such that um k u um and vm k au vm , where um 1 ÿ x1=m , vm 1 ÿ y1=m and k is the integer part of a 2. To see this it suf®ces to show that there exists an m such that ÿ1 ÿ1 ÿk m ln
1 ÿ u ln
1 ÿ u and ÿa ÿa ÿk m : ln
1 ÿ au ln
1 ÿ au That is
ÿ1 ÿa ÿ1 ÿa ; ; ÿ k m min : max ln
1 ÿ u ln
1 ÿ au ln
1 ÿ u ln
1 ÿ au Such m exists if ÿ1 ÿ a ÿ 1: kÿ ln
1 ÿ u ln
1 ÿ au By (17) and (18), the above holds if k ÿ max
a; 1 1, which is true by the choice of k. Note that such m goes to in®nity as u goes to zero. Moreover, limm?? vm =
um vm a=
a 1 and limm??
umk vmk =
um vm 1. To complete the proof one uses the monotonicity of C to obtain 1 ÿ C
1 ÿ um k ; 1 ÿ vm k um k vm k 1 ÿ C
1 ÿ u; 1 ÿ v um v m um k v m k 1 au
1 ÿ C
1 ÿ um ; 1 ÿ vm um vm ; um k vm k um v m
taking the limit as u goes to zero (i.e. m going to in®nity) yields the result. Next, recall the following property of copula function jC
x1 ; y1 ÿ C
x2 ; y2 j
jx1 ÿ x2 j jy1 ÿ y2 j;
19
257
MULTIVARIATE EXTREME DEPENDENCE FUNCTIONS
for 0 x1 ; x2 ; y1 ; y2 1. To complete the proof of the Lemma, let
u; v be a sequence such that limu;v?0 v=
u v t. If t 1, then (19) yields 1 ÿ C
1 ÿ u; 1 ÿ v ÿ 1 lim u;v?0 uv
2 lim
u;v?0 u
u 0: v
Whereas, when t51 then by letting a t=
1 ÿ t one gets 1 ÿ C
1 ÿ u; 1 ÿ v 1 ÿ C
1 ÿ u; 1 ÿ au
1 au uv
1 au uv
C
1 ÿ u; 1 ÿ au ÿ C
1 ÿ u; 1 ÿ v : uv
Now (14) follows from (19) and (16).
&
Proof of Theorem 1: From Lemma 1, it suf®ces to show that (4) is equivalent to (14). Clearly, (14) entails (4). Conversely, assume that (4) holds. Consider any sequence
u; v such that limu;v?0 v=
u v t. The case t 1 is treated in the proof of Lemma 1, thus for t51, one has 1ÿt
1 ÿ C
1 ÿ u; 1 ÿ v 1 ÿ C
1 ÿ x uv x
t
;
1 ÿ x
x uv
C
1 ÿ x1 ÿ t ;
1 ÿ xt ÿ C
1 ÿ u; 1 ÿ v ; uv
1=
1 ÿ t
where x 1 ÿ
1 ÿ u . As u goes to zero, the ®rst fraction on the right-hand-side converges to A(t) by (4), the second fraction converges to 1 by de®nition of x and t and from (19), the last fraction is bounded by j
1 ÿ ut=
1 ÿ t ÿ 1 vj=
u v, which goes to zero. & Proof of Theorem 2:
Use notation (10) to write
^ ÿ A
tj sup jm
t; h ÿ A
tj sup jA
t ^ ÿ m
t; hj: sup jA
t
0t1
0t1
0t1
The ®rst term on the right-hand-side of the above inequality can be bounded as follows R? jm
t; h ÿ A
tj
0
IIf05v51=hgjf
t; hvjv2 K
v dv R? ; 2 0 IIf05v51=hgv K
v dv
258
ABDOUS, GHOUDI, AND KHOUDRAJI
where II
? denotes the indicator function and ( f
t; u
Gt
u u
0
ÿ A
t; for 0 t 1 and 05u 1 for 0 t 1 and u 0.
In order to use the dominated convergence Theorem, we have to bound f
? ; ? . To this end, note that for any copula function one has max
0; u v ÿ 1 C
u; v min
u; v; for 0 u; v 1:
20
One also has
1 ÿ ut 1 ÿ tu;
for 0 u; t 1;
21
1 ÿ ut 1 ÿ 2tu;
for 0 t 1 and 0 u 1=2:
22
Now, by using (20)±(21) and the well-known fact max
t; 1 ÿ t A
t 1;
for all t [ 0; 1;
one gets max
t; 1 ÿ t ÿ 1 f
t; u;
for all 0 t; u 1:
23
Furthermore, since C lies between 0 and 1, then when u41=2 f
t; u 2 ÿ max
t; 1 ÿ t;
for all 0 t 1:
24
This inequality is still valid when u 1=2 by virtue of (20) and (22). Finally, by combining (23) and (24) we end up with jf
t; uj 3=2;
for all t [ 0; 1:
Next, since C and A are continuous and C is in d
C , then f
t; u is continuous on 0; 12. Consequently, C is uniformly continuous. The uniform continuity and the fact that f
t; 0 0 for all t in 0; 1 imply that sup0 t 1 jf
t; uj converges to zero as u goes to zero. Upon noting that by hypothesis u2 K
u is integrable, and calling on the dominated convergence Theorem twice, one obtains R? lim sup jm
t; h ÿ A
tj
n?? 0 t 1
0
lim sup jf
t; hvjv2 K
v dv
h?0 0 t 1 R? 2 0 v K
v dv
0:
259
MULTIVARIATE EXTREME DEPENDENCE FUNCTIONS
^ ÿ m
t; hj. Observe that It remains to study sup0 t 1 jA
t
^ ÿ m
t; hj h sup jA
t
0t1
ÿ1
sup
R 1h
0 u;v 1
jCn
u; v ÿ C
u; vj R 10
uK
u du
2 0 u K
u du h
;
and that R 1h 0
uK
u du
0
u2 K
u du
R 1h
R?
0 ÿ? R ? 0
uK
u du u2 K
u du
;
as n??:
^ ÿ m
t; hj Hence, by applying Theorem 1 of Kiefer (1961), we see that sup0 t 1 jA
t 2 converges to zero in probability if nh goes to in®nity, and it converges to zero almost P 2 exp
ÿ gnh is ®nite. & surely if for every g40 the ? n1 Next, before dealing with Theorem 3, we need to consider the following technical Lemma. Lemma 2: Z
1=h 0
If K is a bounded positive function satisfying (13) then for any a; b40 a
1 ÿ
1 ÿ hu uK
u du c1 ah;
25
1 ÿ
1 ÿ hua
1 ÿ
1 ÿ hub uK
u du c2 abh2 ;
26
and Z
1=h 0
where c1 and c2 are positive constants depending only on K. Proof of Lemma 2: a
The proof of this Lemma rests on the following inequality,
1 ÿ
1 ÿ u a maxf2u; ÿ ln
1 ÿ u= ln
2g; for all a40 and 0 u 1:
27
260
ABDOUS, GHOUDI, AND KHOUDRAJI
Note that Z
1=h
0
a
1 ÿ
1 ÿ hu uK
udu Z
2ah
1=2h
0
Z 2ah
? 0
Z 2ah
? 0
Z
a u K
udu ÿ ln
2 2
4ah2 u K
udu ÿ ln
2 2
1=h
1=2h
Z
1=h
1=2h
ln
1 ÿ huuK
udu
ln
1 ÿ huu3 K
udu
4ah 3 sup
juj K
u u K
udu ÿ ln
2 u40 2
Z
1
1=2
ln
1 ÿ udu:
This concludes the proof of (25) since supu40
juj3 K
u is ®nite by (13). As for (26), similar arguments give Z
1=h 0
1 ÿ
1 ÿ hua
1 ÿ
1 ÿ hub uK
udu
4abh2
Z
? 0
u3 K
udu
8abh2 ln
2
sup
u4 K
u 2 u40
Z
1
1=2
ln
1 ÿ u2 du;
which implies (26). Proof of Theorem 3:
& In the sequel, we shall denote
u; t pn
C
1 ÿ u1 ÿ t ;
1 ÿ ut ÿ C
1 ÿ u1 ÿ t ;
1 ÿ ut ; b n n and Z Djh
0
1=h
uj K
udu;
for 1 j 3:
Use these notations to rewrite (12) as follows 1 an
t ÿ p hD2h
Z 0
1=h
hu; tdu: uK
ub n
The weak convergence of an
? , is obtained by establishing its tightness and the convergence of its ®nite dimensional distributions. The tightness of an
? is proved in three steps. Namely,
261
MULTIVARIATE EXTREME DEPENDENCE FUNCTIONS
Step 1. There exists k40 such that for all 1=nh t1 5t5t2 1 ÿ 1=nh R
t1 ; t; t2 Efan
t ÿ an
t1 g2 fan
t2 ÿ an
tg2 k
t2 ÿ t1 2 : Step 2. lim
sup
n?? 0 s 5 t 1=nh
jan
t ÿ an
sj 0;
in probability:
Step 3. lim
sup
n?? 1ÿ1=nh s 5 t 1
jan
t ÿ an
sj 0;
in probability:
Hereafter, we outline the proof of steps 1 and 2, we omit the proof of Step 3 because it is similar to that of Step 2. First consider Step 1. To simplify the expression of R
t1 ; t; t2 set 1ÿt
zi
u; t IIfUi
1 ÿ u
t
; Vi
1 ÿ u g ÿ Cf
1 ÿ u
1ÿt
t
;
1 ÿ u g;
and Zi
u; s; t zi
u; t ÿ zi
u; s: Then, by using jZi
u; s; tj 2; it follows that 1 R
t1 ; t; t2 2 4 h D2h
Z
4 Y
0;1=h4 i 1
ui K
ui
n 1 X EZk1
hu1 ; t1 ; t n2 k ;...;k 1 1
4
6Zk2
hu2 ; t1 ; tZk3
hu3 ; t; t2 Zk4
hu4 ; t; t2 du1 du2 du3 du4 R1
t1 ; t; t2 R2
t1 ; t; t2 ; where 4D2 R1
t1 ; t; t2 2 1h4 nh D2h
Z
0;1=h2
EjZ1
hu1 ; t1 ; tZ1
hu2 ; t; t2 j
2 Y i1
ui K
ui dui ;
and R2
t1 ; t; t2
12D21h h2 D42h
Z
0;1=h2
EjZ1
hu1 ; t1 ; tjEjZ2
hu2 ; t; t2 j
2 Y i1
ui K
ui dui :
262
ABDOUS, GHOUDI, AND KHOUDRAJI
An upper-bound of R2
t1 ; t; t2 is given by 48D21h h2 D42h
Z
0;1=h2
fP
A1 P
B1 gfP
A2 P
B2 g
2 Y i1
ui K
ui dui ;
where A1 f
x; y :
1 ÿ hu1 1 ÿ t1 x
1 ÿ hu1 1 ÿ t and 0 y
1 ÿ hu1 t g; B1 f
x; y : 0 x
1 ÿ hu1 1 ÿ t1 and
1 ÿ hu1 t y
1 ÿ hu1 t1 g; A2 f
x; y :
1 ÿ hu2 1 ÿ t x
1 ÿ hu2 1 ÿ t2 and 0 y
1 ÿ hu2 t2 g; B2 f
x; y : 0 x
1 ÿ hu2
1ÿt
t
t
and
1 ÿ hu2 2 y
1 ÿ hu2 g:
For i 1; 2, an application of (19) yields ÿ maxfP
Ai ; P
Bi g 1 ÿ
1 ÿ hui t2 ÿ t1 :
28
Consequently, by (25), there exists a constant k0 such that R2
t1 ; t; t2 k0
D21h 2
t ÿ t1 : D42h 2
Next, straightforward computations show that the expectation involved in R1
t1 ; t; t2 is less than P
A1 \ A2 P
A1 \ B2 P
B1 \ A2 P
B1 \ B2 12
P
A1 P
B1
P
A2 P
B2 : Note that
A1 \ A2 is not empty only if u1 u2
1 ÿ
1 ÿ hu1
1 ÿ t1 =
1 ÿ t2 =h. Therefore, by (28), ÿ
P
A1 \ A2 1 ÿ
1 ÿ hu1
t2 ÿ t1
8