Variability measures of positive random variables Supporting ... - PLOS

0 downloads 0 Views 38KB Size Report
ch = σh. E(T). = Γ(k) k exp 1k + (1 - k)Ψ(k) - 1l . (S-2). By substitution k = 1/c2 ... k - 2 k . (S-9). In the parametrization in terms of cv, it can be expressed as relation. cJ = √ ... 0. (. ∂ ln f(t). ∂ t. )2 f(t) dt = 2+9c2 v + 21c2 v + 21c6 v. 2 µ2 c2 v . (S-18).
1

Variability measures of positive random variables Supporting information S1 Lubomir Kostal1,∗ , Petr Lansky1 , Ondrej Pokora1 1 Institute of Physiology AS CR, v.v.i., Videnska 1083, Praha 4, Czech Republic ∗ E-mail: [email protected] Note: We refer to the equations in the main text by their number, e.g., Eq. (1), while equations in this material are numbered as Eq. (S-1), Eq. (S-2), etc.

Calculation of ch and cJ Gamma distribution with shape parameter k and scale parameter θ has probability density function given by Eq. (11). The differential entropy (2) of this distribution is equal to [1, p.661] h(f ) = ln θ + ln Γ(k) + (1 − k)Ψ(k) + k ,

(S-1)

d where Ψ(z) = dz ln Γ(z) is the digamma function [2]. Then, using Eqns. (3) and (4) and relation E (T ) = k θ, the entropy-based dispersion coefficient is directly calculated,

ch =

σh Γ(k) = exp {k + (1 − k)Ψ(k) − 1} . E (T ) k

By substitution k = 1/c2v , ch can be expressed in terms of the coefficient of variation, cv , as { [ ]} (1 − c2v ) 1 − Ψ(1/c2v ) 2 2 ch = Γ(1/cv ) cv exp . c2v

(S-2)

(S-3)

The log-density of the gamma distribution is equal to ln f (t) = (k − 1) ln t −

t − ln Γ(k) − k ln θ θ

(S-4)

and its derivative has simple expression θ(k − 1) − t ∂ ln f (t) = . ∂t θt The Fisher information (7) becomes ] ∫ ∞[ (k − 1)2 1 tk−1 exp(−t/θ) J(f ) = − 2(k − 1) + dt . t2 θ2 Γ(k) θk 0

(S-5)

(S-6)

To calculate this integral, it is convenient to divide the expression into three integrals and employ the relations ∫ ∞ tk−l exp(−t/θ) dt = θk−l+1 Γ(k − l + 1) for l = 3, 2, 1 . (S-7) 0

In this case, the resulting Fisher information evaluates to J(f ) =

θ2

1 (k − 2)

(S-8)

2

provided k > 2, otherwise the integral does not converge. Substituting J(f ) into Eqns. (8) and (9), the Fisher information-based dispersion coefficient is obtained, √ k−2 cJ = . (S-9) k In the parametrization in terms of cv , it can be expressed as relation √ (S-10) cJ = c2v (1 − 2c2v ) , √ which is valid for cv < 2/2. Inverse Gaussian distribution with mean µ and scale parameter σ has probability density function f (t) given by Eq. (15). Its log-density evaluates to ) (t − µ)2 1 ( ln f (t) = − ln 2 π σ 2 t3 − . 2 2 µ2 σ 2 t

(S-11)

Substitution into Eq. (2) gives the following expression for differential entropy. Is is convenient to separate it into several integrals in accordance with the power of variable t, ∫ ∞ 1 h(f ) = ln(2 π σ 2 ) f (t) dt + 2 0 ∫ ∫ ∞ 3 ∞ 1 + f (t) ln t dt + f (t) t dt + (S-12) 2 0 2 µ2 σ 2 0 ∫ ∞ ∫ ∞ 1 1 f (t) + dt − f (t) dt . 2 σ2 0 t µ σ2 0 All the integrals in (S-12) can be directly calculated, except the second one which needs special treatment. By [3], the density f (t) can be written in terms of the modified Bessel function of the second kind, K(ν, z), with ν = −1/2, √ ( ) µ µ2 + t2 1 ( ) exp − . (S-13) f (t) = t3 2 µ2 σ 2 t 2 K −1, 1 2 2 µσ

The advantage of such a representation lies in the calculation of the integral √ ∫ ∞ ( ) ( ) 2 f (t) ln t dt = ln µ − exp µ 1σ2 K (1,0) − 12 , µ 1σ2 , 2 πµσ 0

(S-14)

∂ which employs per partes only. Here, K (1,0) (ν, z) = ∂ν K(ν, z) denotes the derivative with respect to the 2 2 first argument. After substitution µ σ = cv for parametrization by cv , the resulting expression for the entropy becomes 3 exp(1/c2v ) (1,0) ( 1 1 ) 1 1 (S-15) h(f ) = ln µ + + ln(2 π c2v ) − √ K − 2 , c2 v 2 2 2 π c2v

and the corresponding dispersion coefficient (Eqns. (3) and (4)) is equal to { } √ σh 3 exp(1/c2v ) (1,0) ( 1 1 ) 2 π c2v ch = = exp − √ − 2 , c2 . K v E (T ) e 2 π c2v

(S-16)

The derivative of the log-density ln f (t) is equal to ( 2 ) ∂ ln f (t) 1 = µ − t2 − 3 µ2 σ 2 t . ∂t 2 µ2 σ 2 t2

(S-17)

3

By direct computation in terms of cv , the Fisher information can be calculated as ∫



J(f ) = 0

(

∂ ln f (t) ∂t

)2 f (t) dt =

2 + 9c2v + 21c2v + 21c6v . 2 µ2 c2v

(S-18)

Hence, employing Eqns. (8) and (9), the Fisher information-dispersion coefficient results in relation √ 2 c2v cJ = . (S-19) 2 + 9c2v + 21c2v + 21c6v The probability density function, f (t), of the lognormal distribution with parameters µ and σ is given by Eq. (19). By [1, p.662], differential entropy of the distribution is equal to 1 ln(2 π e σ 2 ) . (S-20) 2 ( ) Hence, employing Eqns. (3) and (4) and relation E (T ) = exp µ + σ 2 /2 , the entropy-based dispersion coefficient is evaluated as √ σh 2 π σ2 ch = = . (S-21) E (T ) exp(1 + σ 2 ) h(f ) = µ +

In the parametrization by the coefficient of variation, cv , by substituting σ 2 = ln(1 + c2v ), it changes to √ √ 2 π ln(1 + c2v ) . (S-22) ch = e 1 + c2v The derivative of the logarithm of the density f (t) is equal to ∂ ln f (t) ln t − µ + σ 2 = . ∂t σ2 t

(S-23)

Direct calculation of the Fisher information gives ∫



(

J(f ) = 0

∂ ln f (t) ∂t

)2

( ) ( ) 1 f (t) dt = 1 + 2 exp −2 µ + 2 σ 2 . σ

(S-24)

After substitution into Eqns. (8) and (9), expressions for the Fisher information-based dispersion coefficient are obtained: √( ) ( ) 1 3 2 cJ = 1− exp − σ (S-25) 1 + σ2 2 for the µ, σ parametrization, and √( cJ = for the cv parametrization.

1−

1 ln(1 + c2v )

)

1 (1 + c2v )3

(S-26)

4

References 1. Cover TM, Thomas JA (1991) Elements of Information Theory. New York: John Wiley and Sons, Inc. 2. Abramowitz M, Stegun IA (1965) Handbook of Mathematical Functions, With Formulas, Graphs, and Mathematical Tables. New York: Dover. 3. Kawamura T, Iwase K (2003) Characterization of the distributions of power inverse gaussian and others based on the entropy maximization principle. J Japan Statist Soc 33: 95–104.