On the distribution of a statistic in multivariate inverse ... - Project Euclid

3 downloads 0 Views 730KB Size Report
tion produces no effect toS andy since they are independent of 5. Following ... (l + v'v)~1β'β = λ and ζ2ζ2=μ anc * by using the result (Anderson [1], p. 319) on.
HIROSHIMA MATH. J.

14 (1984), 215-225

On the distribution of a statistic in multivariate inverse regression analysis Yasunori FUJIKOSHI and Ryuei NISHII

(Received September 17, 1983)

§ 1. Introduction

In a multivariate inverse regression problem we are interested in making inference about an unknown q x 1 vector jc = (x l 5 ..., xq)' from an observed px 1 response vector y = {yw-> yPϊ' Brown [3] has summarized various aspects of the problem. We assume that y is random, x is fixed and (1.1)

y = a + B'x + e = Θ'

+ e x

where Θ' = [α, £ ' ] : px(l + q) is the matrix of unknown parameters and e is an error vector having a multivariate normal distribution Np[0,11]. Further, suppose that the N informative observations on y and x have been given. When p>q, it is possible to obtain a natural point estimate for JC, and to construct a confidence region for JC, based on a statistic, which is a quadratic form of the estimate. For an application of the confidence region it is required to give the upper percentage point of the statistic. The purpose of this paper is to study the distribution of the statistic mentioned above. We shall derive an asymptotic expansion for the distribution function of the statistic up to the order N~2 and hence for the upper percentage point of the statistic. In Section 3 we treat the distribution problem in the situation where Θ is known and Σ is unknown. We note that the distribution of the statistic in this case is essentially the same as one of a statistic in growth curve model. The distribution has been studied by Rao [6] and Gleser and Olkin [4]. The numerical accuracy of our asymptotic approximations is checked by comparing with exact results of Gleser and Olkin [4]. In Section 4 we treat the distribution problem in the situation where Θ and Σ are unknown. In this case a reduction of the distribution problem is given. By using the reduction and perturbation method we shall obtain the asymptotic expansion of the distribution function of the statistic. Some formulas used in deriving the asymptotic expansions are summarized in Section 5.

216

Yasunori FUJIKOSHI and Ryuei NISHII

§ 2.

A distribution problem

Suppose that the N independent observations on y and x following the model (1.1) have been given, and let these observations be denoted by

Then the observations satisfy (2.1) where j N = (l,...9 1)' and E is an N x p error matrix whose rows are independently distributed as Np[0, Σ~\. The observation y and x for a new object satisfy the model (1.1). We assume that y is observed, but x is unknown. Since x is a fixed variate we may without loss of generality assume that (2.2)

X'jN = 0.

Further, we put on the usual assumptions; (2.3)

rank(Z) = g and n = N - q - 1 > p.

If Θ and Σ are known, from (1.1) we may estimate x by (2.4)

x0

which is obtained by the maximum likelihood method based on (1.1) or by minimizing

with respect to x. Since x is distributed as Np[x, (BΣ'12Γ)"1], use a confidence region for JC, based on (2.5)

Q0 =

this suggests to

(x0-x)'BΣ-'B'{x0-x).

The confidence region for x of confidence 1—α is given by the ellipsoid {x\Q0< Xβ(α)}> where χ^(α) is the upper α point of a χ2-distribution with g degrees of freedom. If Θ and Σ are unknown, it is natural to replace them by their estimates. The usual estimates of Θ and Σ based on (2.1) are given by

ί =s = ^

Multivariate inverse regression analysis

217

We use the following statistics according as Θ is known or not: (i) The case when Θ is known and Σ is unknown (2.6)

β 1 = (S1-x)'*S'-**'(al-x) X

where xt = (JBS"*£')" BS-*(y-a). (ii) The case when Θ and Σ are unknown; (2.7)

Q2 =

(ϊ2-xyβS-iβ'(

Suggest Documents