Multiple Shrinkage and Subset Selection in Wavelets - CiteSeerX

3 downloads 0 Views 213KB Size Report
The resulting estimator for the wavelet coe cients is a multiple shrinkage estimator .... George & Foster (1997 Technical Report, University of Texas at Austin).
Multiple Shrinkage and Subset Selection in Wavelets By MERLISE CLYDE, GIOVANNI PARMIGIANI and BRANI VIDAKOVIC

Institute of Statistics and Decision Sciences, Duke University, Durham, NC 27708-0251, U.S.A.

e-mail: [email protected] [email protected] [email protected] ISDS Discussion Paper 95-37 Revised September 15, 1997

Summary

This paper discusses Bayesian methods for multiple shrinkage estimation in wavelets. Wavelets are used in applications for data denoising, via shrinkage of the coecients towards zero, and for data compression, by shrinkage and setting small coecients to zero. We approach wavelet shrinkage by using Bayesian hierarchical models, assigning a positive prior probability to the wavelet coecients being zero. The resulting estimator for the wavelet coecients is a multiple shrinkage estimator that exhibits a wide variety of nonlinear shrinkage patterns. We discuss fast computational implementations, with a focus on easy-to-compute analytic approximations as well as importance sampling and Markov chain Monte Carlo methods. Multiple shrinkage estimators prove to have excellent mean squared error performance in reconstructing standard test functions. We demonstrate this in simulated test examples, comparing various implementations of multiple shrinkage to commonly used shrinkage rules. Finally, we illustrate our approach with an application to the so-called \glint" data. Some key words: Gibbs Sampling, Importance Sampling, Model Averaging

1. Introduction Wavelets are families of basis or basis-like functions that can be used to approximate other functions. They combine powerful properties such as orthonormality, compact support, varying degrees of smoothness, localization in time and scale (frequency), and fast implementation. Daubechies and Mallat sparked interest in wavelets in the statistics community when they connected wavelets with discrete data processing (Daubechies, 1988; Mallat, 1989). Donoho & Johnstone showed that wavelet shrinkage had desirable statistical optimality properties in problems concerning elimination of noise (Donoho & Johnstone 1994, 1995, Donoho et al., 1995). The wavelet regression model can be described as follows. Suppose the function

f () is sampled at N equally spaced points, f = ff (x ); : : : f (xN )g0, but is observed 1

with additive white noise . If Y = (Y ; : : : ; YN )0 is the vector of observations, then 1

the model is Y = f + , or equivalently, in a wavelet regression form, Y = W + ; where f = W , is the discrete wavelet transformation (DWT) of f , and W 0 is the N  N orthogonal matrix corresponding to the discrete wavelet transformation (Daubechies, 1992). The least squares estimator of is ^ = W 0Y and, in fact, is the DWT of Y . In practice, there is no need to compute W , as there are fast algorithms to compute the empirical wavelet coecients, ^, directly in only O(N ) operations. Shrinkage of empirical wavelet coecients is particularly e ective when many of the coecients represent noise rather than signal. Wavelet shrinkage can be 1

described by a three step procedure: (i) Noisy observations Y are transformed by a DWT to obtain ^; (ii) The coecients ^ are shrunk towards zero or possibly set to 0; (iii) The shrunk coecients are returned to the Y domain by the inverse DWT. The resulting vector is a wavelet shrinkage estimator f^ of the unknown vector f: In this paper, we present methods for multiple shrinkage of wavelet coecients based on a Bayesian hierarchical model. Our main focus is on incorporating model uncertainty about which of the coecients are zero. The resulting posterior mean of the coecients combines linear shrinkage to the prior mean (zero), with additional nonlinear shrinkage to zero resulting from averaging over models where the coecient is zero. As the posterior probability that a coecient is zero is typically not available in closed form, unless the error variance is known, Bayesian model averaging for constructing multiple shrinkage estimators is more computationally intensive compared to many of the alternative approaches. We present two analytic approximations that can be calculated in closed form, and compare these with Monte Carlo methods for obtaining posterior means and variances. In x2, we discuss the Bayesian hierarchical regression model. In x3 we introduce the multiple shrinkage estimators and analytic approximations for calculating approximate posterior means and variances. These approximations are used to implement importance sampling and Gibbs sampling algorithms. These lead to more accurate estimates of posterior means and variances, but at an additional compu2

tational cost. In x4, we compare our multiple shrinkage estimators to conventional alternatives, using simulation with standard test functions. We consider both normal and Student{t errors. In x5, we present an application illustrating the importance of prior information in signal denoising. 2. Bayesian Hierarchical Model The statistical model is described by the following distributional assumptions. First, we assume that the errors i are independent normals with mean 0 and variance

 so that 2

Y j ;   N (W ;  IN ): 2

2

This assumption is standard in many wavelet applications (Donoho & Johnstone, 1995). In x4, we explore the robustness of our estimators to that assumption. We introduce the N dimensional vector , which is a sequence of binary random variables, to represent which elements of are zero. We will identify individual elements of vectors such as , ^ and , by the indices j and k in subscripts jk, corresponding to basis elements

jk .

The

jk

are dilations at level or scale j and location translations

by k2j of the mother wavelet function . In the prior distribution, the jk 's given

and  are independently distributed as jk j jk;   N (0; jk cjk  ); 2

which is normal with variance cjk  when jk = 1, and degenerate at zero when 2

3

jk = 0. We use a conjugate distribution for  , 2

=   ; 2

2

where  and  are xed hyperparameters. The choice  =  = 0 is commonly adopted to represent lack of prior knowledge. We use this in the simulation studies in x4. In x5, we illustrate a case in which prior information is used to determine appropriate values of  and . Finally, the jk are independently distributed as Bernoulli random variables,

jk  Bernoulli(jk ): It can be useful to set jk = j and cjk = cj , so that the hyperparameters are constant within level j , resulting in monotonic shrinkage of coecients within each level. George & Foster (1997 Technical Report, University of Texas at Austin) propose a method for calibrating c so that, conditional on , model selection based on posterior model probabilities corresponds to classical model selection methods. The value of c is chosen by solving 



F (c; ) = 1 +c c log(1 + c) + 2 log 1 ? 



;

where the value of F depends on the model selection criterion. For example,

F (c; ) = log N corresponds to model selection by the BIC (Bayesian Information Criterion, Akaike, 1978), and F (c; ) = 2 log N corresponds to the RIC (Risk In ation Criterion, Foster & George (1994) and Donoho & Johnstone (1994)). The 4

latter choice corresponds to selecting a model where all the jt-statisticsj are greater than (2 log N ) = , as in universal thresholding. This can be applied seperately for 1 2

each level. Mixtures of normals and point masses have been used in Bayesian variable selection (reviewed in George & McCulloch, 1997). A related approach is George & McCulloch's (1993) stochastic search variable selection (SSVS) where the prior distribution on the coecients is a mixture of two normal components, one concentrated around zero and the other suitably dispersed. Chipman, Kolaczyk & McCulloch (1997) apply this method to wavelets. 3. Posterior Inference 31. The multiple shrinkage estimator The posterior mean of jk conditional on is E ( jk j Y ) = jk ^jk =(1+ c?jk ): In turn, 1

the posterior distribution of is

( jY ) = q( )=

X all

0

q( 0)

(1)

where

q( ) =

"

Y jk

jk (1 + c )? = jk 1 ? jk

1 2

 jk # (

 + Y 0 Y ?

X jk

jk ^jk =(1 + c?jk ) 2

)?(N + )=2

1

Posterior probabilities in (1) can be used for model selection under various loss functions. 5

:

As an alternative to selecting a corresponding to the \best" model, we can make inferences by averaging over all possible 's. The posterior mean of jk is

E ( jk jY ) =

X

( jY )E ( jk jY; ) = E ( jk jY ) ^jk =(1 + c?jk ):

(2)

1

This is a multiple shrinkage estimator, in which regression coecients are shrunk linearly towards their respective prior mean (zero in this case), and are further nonlinearly shrunk towards zero as a result of uncertainty about whether the jk is zero. The use of Bayesian model averaging in other contexts has resulted in improved predictive performance (Draper, 1995; Raftery, Madigan & Volinski, 1996). George (1986) discusses desirable Bayes and minimax properties of multiple shrinkage estimators in a general context. The appeal of this shrinkage approach is its

exibility and adaptation to the observed data. The diculty in implementing this in real-time is that the expectations, E ( jk jY ), cannot be calculated analytically. We present four approaches for approximating the posterior means and variances. 32. Analytic Approximation given  { S Closed form approximations for the posterior mean and variance can be obtained based on conditioning on . Given Y and , the jk are independent Bernoulli random variables with parameters:









S pjk () = 1 +ajka(Y;(Y;)) ; ajk (Y; ) = (1 + cjk )? = 1 ?jk exp 21 jk ; (3) jk jk where Sjk = ^jk =(1 + c?jk ). The conditional posterior mean is 2

1 2

2

2

2

1

E ( jk jY; ) = pjk () ^jk =(1 + c?jk ) 1

6

(4)

which can be computed in real time. As  is typically unknown, we can use an estimate, such as ^ = median(j ^ k j)= 06745, proposed by Donoho et al. (1995). 1

The conditional posterior variance of given  is available in closed form: var( jk jY; ) = pjk (?)  + pjk () f1 ??pjk ()g ^jk : 1 + cjk (1 + cjk )

(5)

2

2

1

1 2

and can be used to obtain an approximation of the posterior variance by evaluating (5) at ^ . Let V be the diagonal matrix with var( jk jY; ^ ) on the diagonal and zero elsewhere. The covariance terms are all zero due to the prior independence of the

jk 's and orthogonality of W . The posterior covariance matrix of f given Y and  is (W 0V W ) = W 0(W 0V )0 . An ecient method to compute the variance of f^ is suggested by Chipman, Kolaczyk & McCulloch (1997). This involves applying the inverse DWT to the N columns of V and then successively to the N columns of (W 0V )0 , requiring 2N inverse wavelet transformations. 33. Analytic Approximation { A If the posterior distribution for  is not concentrated, the estimator in (4) may not be very accurate. An alternative approach is to approximate (1) by a model of independence

~ ( jy) = where

n

Y jk pjk (1 ? pjk )1? jk

(6)

jk

P

o

jk exp ? log(1 + cjk ) + (N +  )L ll l S Y 0 Y n o pjk = 2 P 1 ? jk + jk exp ? log(1 + cjk ) + (N +  )L ll l S Y 0Y 1 2

1 2

1 2

=1 (

1 2

7

2 )l ( jk +

)l

( jk )l =1 ( + )l

(7)

and Sjk = ^jk =(1 + c?jk ): The quantity L is a calibration constant de ned by: 2

1

2



L=

log( + Y 0Y ) ? log  + Y 0 Y ?

P

Pl

P

jk Sjk

S l=1 l(+Y 0 Y )l 2 l jk ( jk )

2



:

The approximation is based on a Taylor's series expansion the logarithm of the posterior model probabilities (Clyde, DeSimone & Parmigiani, 1996). The pjk 's can be used to obtain a direct approximation to the multiple shrinkage estimator:

E ( jk jY )  pjk ^jk =(1 + c?jk );

(8)

1

which takes into account uncertainty . 34. Importance Sampling { I Importance sampling is based on generating a sample of 's from (6), by drawing each jk as an independent Bernoulli random variable with P ( jk = 1) = pjk . An importance sampling estimator for is then

E ( jY ) 

X

w( ) jk ^jk =(1 + c?jk ) 1

(9)

where w( ) is the importance sampling weight,

w( ) = P n(n () q0)(q () =0~)=( ~)( 0)

0

and n( ) is the number of times model is sampled. All sums are over the set of sampled models. Alternatively, one can use weights ^ based on renormalizing the

q( )'s obtained from the sampled models (Clyde, DeSimone & Parmigiani, 1996), w( ) = ^  q( )=

X

0

q( 0):

This is preferable if the variance of the importance sampling weights is large. 8

(10)

35. Gibbs Sampling { G In the speci c context of wavelets and other orthogonal bases, we can implement an ecient block Gibbs sampler that alleviates some of the slow mixing problems noted in MCMC samplers that generate from the full conditional distributions of

, and . We obtain samples from the joint posterior distribution of  and by drawing from the full conditional distributions in two blocks, as follows. First, given and Y ,  has an inverse Gamma distribution, or,  j ; Y  ( + Y 0Y ? 2

P

2

jkSjk )= N : Second, conditional on , the distribution of factors as a product 2

2

+

of independent Bernoulli random variables with probabilities pjk () using (3). This has the advantage over other MCMC methods that sample from (1) in that the jk's are conditionally independent, reducing the computations to generate . The main advantage is that the posterior expectation of given Y can be estimated using a Rao-Blackwellized estimate based on (4), M X 1 E ( jk jY )  M pjk (m ) ^jk =(1 + c?jk ): m

(11)

1

=1

where m , m = 1; :::; M are the sampled values. Similarly, a Rao-Blackwellized estimate of the posterior variance is M M X X 1 1 var( jk jY )  M m var( jk jY; m) + M m fE ( jk jY ) ? E ( jk jY; m)g (12) 2

=1

=1

and can be evaluated based on expressions (4), (5), and (11). This is used to compute the posterior variance of f , as discussed earlier. 9

4. Performance with Standard Functions In this section we compare multiple shrinkage rules to existing shrinkage strategies. We use the four test functions \blocks", \bumps", \doppler", \heavisine", proposed by Donoho & Johnstone, and now commonly used in this type of assessment. Each experiment consisted of 100 replicates of N = 1024 observation from the test functions plus independent normal errors with  = 1. The signal-to-noise ratio SNR = 7 and the wavelet bases were chosen to match Donoho & Johnstone (1995). We compared our four implementations of the multiple shrinkage Bayes estimate: S,A,G, and I from x3, to three commonly used shrinkage strategies: HARD, hard thresholding with the universal rule (Donoho & Johnstone, 1994), SOFT, soft thresholding with the universal rule (Donoho & Johnstone, 1994), and SURE, SureShrink adaptive shrinkage rule as implemented in S+Wavelets (Bruce & Gao, 1994). Rules S, HARD, SOFT and SURE all use ^ = median(j ^ k j)= 06745, where 1

^ k are the empirical wavelet coecients at the nest level of detail. To assess 1

performance, we calculated the average mean squared error (MSE) from the 100 simulations:

N 1 XX (fi ? f^il ) ; MSE = 100 N 100

2

l=1 i=1

where fi is the true signal and f^il is the estimate of the function from simulation l. We used seven di erent priors summarized in Table 1. In each case,  and  , the hyperparameters of the distribution of , were set to zero (a noninformative 10

distribution). The parameters j , representing the probability of inclusion of a level

j coecient, vary across priors. Priors (a), (b), and (g) have a constant probability of inclusion. Priors (d) and (e) are based on a geometric decay. The value of c, controlling the prior precision of the wavelet coecients, is determined based on the proposal of George & Foster discussed in x2, assuming =05 and using the RIC criterion; c remains constant across priors, for ease of comparison. An exception is prior (g) where we derive c using the RIC criterion assuming =09. Model selection using universal thresholding gives the same result as maximizing the posterior model probability under priors (b) and (g), given  = ^ : Thus the comparison between universal thresholding and multiple shrinkage using priors (b) and (g) with method S can be seen as a comparison between model selection and model averaging. Figure 1 summarizes the results of the simulations. The performances of the Gibbs sampler G and the analytic approximation S are very close to each other and consistently outperform all alternatives. Accounting for uncertainty about  does not appear to be essential at this level of SNR and N , because the posterior distribution of  is very concentrated. The performance of the importance sampling method I and the related analytic approximation A is consistently worse than the performance of G and S in this scenario. Inspection of the importance sampling weights reveals a high variance. A further factor contributing to the better performance of G compared to I is the Rao-Blackwellization estimator (11) for posterior probabilities. Because the model space is large, 2 11

1024

models, convergence of the im-

portance sampling estimates is slow compared to the Rao-Blackwellized estimator. Posterior probabilities may be estimated as zero, resulting in hard thresholding. In addition we generated data from the same functions using a SNR=3 and Student-t errors with 5 degrees of freedom to generate data with outliers. For brevity, only results for prior (c) will be discussed. All methods show a higher MSE compared to normal errors with the same SNR, but their relative standings remain roughly unchanged, with the multiple shrinkage estimators doing better overall than the standard approaches. For both \heavisine" and \doppler" the Gibbs estimator G had the overall best MSE, 0125 and 0234, respectively compared to SOFT, which had the best MSE of the standard methods with 0131 and 0364, respectively. In \blocks", G still did better than the best standard method, (0312 for G versus 0446 for SURE), but both S and I had better MSE's than G. For \bumps" the analytic approximation was the only method that did better than SURE (0551 compared to 0568), however, S, G, and I, all did better than HARD or SOFT. 5. Glint Data Example To further illustrate features of the multiple shrinkage approach proposed here we used the glint data presented by Bruce & Gao (1994). The data series includes 512 equally spaced observations. The true signal is a low frequency oscillation about zero, resulting from rotating an airplane model. As the model rotates, the center of mass will shift slightly. The measurements, given in angles, are subject to large errors, and can be o the truth by as much as 100 degrees. 12

Throughout, we used the least asymmetric Daubechies 8-tap mother wavelet \s8", which is the default in S+WAVELETS. This basis provides a sensible balance between the compactness of support and smoothness of the wavelet. In Figure 2 we illustrate the t resulting from universal thresholding. The result may not be satisfactory if it is believed that the true signal should be smoother, and that many of the large spikes represent noise. In order to get more realistic estimates, Bruce & Gao (1994) use simple smoothing techniques before applying wavelet shrinkage. In our multiple shrinkage estimator, we can incorporate the prior information about the noise. For the multiple shrinkage estimator, the prior hyperparameters for  are given by  = 1000 and  = 22, re ecting the belief that errors can be as large as 100. The choice of c is based on the George & Foster RIC rule discussed in

x2, leading to c = 2621305. The prior distribution for was based on j = 2j? for 8

j = 1; : : : ; 6 and  =09999, so that coecients from the smooth space are included 7

with a probability near one. This prior distribution penalizes inclusion of coecients at the nest level of detail, j = 1. Figures 3, 4, and 5 shows the resulting posterior estimates from methods A, I, S and G. The two Monte Carlo approaches, G and I were run for 50; 000 iterations, although estimates of the posterior means stabilize much earlier. There was little qualitative di erence between the estimates based on I and G, which are both computationally more intensive than method A. Method A, however, results in less shrinkage than the other two. The analytic approximation S gives a t that 13

includes many more coecients than that of the other three methods, primarily, because it does not incorporate any of the prior information or uncertainty about

, resulting in a t more similar to hard thresholding. The other methods all take into account that  is unknown. Ignoring uncertainty about  also leads to much narrower prediction intervals for S compared to G (Figures 4 and 5). Finally, Figure 6 considers the shrinkage curves resulting from two prior speci cations; prior 1 being the one previously considered. The second prior is based on selecting c based on the BIC criterion with a uniform prior distribution for ,

jk = 1=2. This leads to c = 392. The prior distribution for  is the same as in the rst prior. The shrinkage curves displays nonlinearities from multiple shrinkage. The nonmonotonicities in the overall shrinkage curve for the rst prior arise from the level dependent prior distribution on , which allows a stronger shrinkage at higher level of detail. Within a speci c level, however, shrinkage is monotonic. 6. Discussion We conclude by discussing some potential developments of our approach. The model can be extended from a single xed basis to include wavelet packets and other over-complete libraries that can produce a variety of orthonormal bases. For example, the wavelet packet library (packet table) contains many di erent orthonormal bases. Wavelet packets are often preferred to wavelets for representing signals that exhibit oscillatory or periodic behavior. One can add the choice of basis to the model hierarchy and use the stochastic selection methods described to select the best 14

subspace from the collection of orthonormal bases or to provide multiple shrinkage estimators, where one is now additionally averaging over di erent bases. Nonwhite noise is a problem in many applications. Modeling the error terms as Student-t errors or other heavy tailed distributions can be achieved by using scale mixtures of normals and extending the hierarchical model. While computationally more intensive, Gibbs sampling and model averaging may provide superior performance to the current alternatives. Acknowledgment

Our work has been partially supported by the U.S. National Science Foundation. We thank Dean Foster and Ed George for helpful discussions and for kindly sharing unpublished results. We thank Andrew Bruce and Hong-ye Gao from StatSci, who provided background information about the glint data set. References Akaike, H. (1978). A new look at the Bayes procedure. Biometrika 65, 53{9. Bruce, A. & Gao, H-Y. (1994). S+Wavelets, Users Manual. Seattle: StatSci. Chipman, H., Kolaczyk, E., & McCulloch, R. (1997). Adaptive Bayesian

wavelet shrinkage. J. Am. Statist. Assoc. forthcoming. Clyde, M., DeSimone, H. & Parmigiani, G. (1996). Prediction via orthogo-

nalized model mixing. J. Am. Statist. Assoc. 91, 1197-208. 15

Daubechies, I. (1988). Orthonormal bases of compactly supported wavelets. Com-

munications on Pure Applied Mathematics 41, 909-96. Daubechies, I. (1992). Ten Lectures on Wavelets. Philadelphia: Society for In-

dustrial and Applied Mathematics. Donoho, D. & Johnstone, I. (1994). Ideal spatial adaptation by wavelet shrink-

age. Biometrika 81, 425{55. Donoho, D. & Johnstone, I. (1995). Adapting to unknown smoothness via

wavelet shrinkage. J. Am. Statist. Assoc. 90, 1200-24. Donoho, D., Johnstone, I., Kerkyacharian, G., & Picard, D. (1995).

Wavelet shrinkage: Asymptopia? J. R. Statist. Soc. B 57, 301-69. Draper, D. (1995). Assessment and propagation of model uncertainty (with Dis-

cussion). J. R. Statist. Soc. B 57, 45-98. Foster, D. & George, E. (1994). The risk in ation criterion for multiple regres-

sion. Ann. Statist. 22, 1947{75. George, E. (1986). Combining minimax shrinkage estimators. J. Am. Statist.

Assoc. 81, 437{45. George, E.I. & McCulloch, R. (1993). Variable selection via Gibbs sampling.

J. Am. Statist. Assoc. 88, 881{89.

16

George, E.I. & McCulloch, R. (1997). Fast Bayes variable selection. Statist.

Sinica, forthcoming. Mallat, S. (1989). A theory for multi-resolution signal decomposition: the wavelet

representation. IEEE Transactions on Pattern Analysis and Machine Intelligence 11, 674{93. Raftery, A.E., Madigan, D.M., & Volinsky C.T. (1996). Accounting for

model uncertainty in survival analysis improves predictive performance (with discussion). In Bayesian Statistics 5, Ed. J.M. Bernardo, J.O. Berger, A.P. Dawid and A.F.M. Smith. pp 323{50. Oxford: Oxford University Press. prior

c 











a

1048561 09 09

09

09

09

09

09

b

1048561 05 05

05

05

05

05

05

c

1048561 09 09

05

05

05

05

005

d

1048561 09 0556 0343 0212 0131 0081 005

e

1048561 05 0341 0232 0185 0108 0073 005

f

1048561 09 09

05

05

05

05

05

g

84934641 09 09

09

09

09

09

09

7



6

5

4

3

2

1

Table 1: The seven alternative sets of prior hyperparameters used in the simulation example of Section 4. 17

BLOCKS

BUMPS A 1.6

A A

1.4

0.8

A

I

A A 0.6

I

I

A

I

1.0

MSE

MSE

I A

A

I

I

A

0.8

0.4

A I

I

I

1.2

A

A

I

A I

I 0.6

I

S G

G S

G S

a

b

c

G S

G S

d

e

0.4

0.2

S

G S

G S

G S

f

g

a

S G

S G

b

c

G

d

G

e

PRIOR

PRIOR

DOPPLER

HEAVYSINE

0.20

A

0.4

f

g

A A

0.18

A

A

I

A

I

MSE

I

I

I

A I

0.14

MSE

I

0.16

A

I

0.3

S G

A

A A

S G

A

A A

0.5

S

I

I

I

I

I 0.12

I

S G

S G

S G

S G

0.10

0.2

S G

S G

S G

S G

S G

S G S G

S G

G S

G S

a

b

c

d

e

f

g

a

PRIOR

b

c

d

e

f

g

PRIOR

Figure 1: Comparison of algorithms for the four test functions. The horizontal lines correspond to the thresholding rules: the continuous line is HARD; the dotted line is SOFT; the dashed line is SURE . 18

50 0 -100

-50

GLINT SERIES

0

100

200

300

400

500

TIME

Figure 2: Glint data and tted signal based on the universal thresholding rule

19

50 0 -100

-50

GLINT SERIES

0

100

200

300

400

500

TIME

Figure 3: Glint Data: the estimates resulting from the analytic approximation (A) (dashed line) and the importance sampling (solid line).

20

100 50 0 -50 -100

GLINT SERIES

0

100

200

300

400

500

TIME

Figure 4: Glint Data: posterior mean and prediction intervals about f^ ( two standard errors) using the analytic approximation S.

21

100 50 0 -50 -100

GLINT SERIES

0

100

200

300

400

500

TIME

Figure 5: Glint Data: posterior mean and prediction intervals about f^ ( two standard errors), using Gibbs sampling.

22

••

•• •

100

100



••

•• ••

• ••

•• •• •• •• • •• • • •• •• •••••••••••• •••• ••••••• ••••• •• •• ••• •• •••• •••• •• ••• ••• ••• •••• ••• •• ••• ••• •••••• ••••• ••• ••••••• ••••• ••• • •••



0

• ••

• ••

• •• •

-100

POSTERIOR MEANS



-100







-200

-200

POSTERIOR MEANS

0

•• •••••••••••••••• ••• •• •• •••• •••••••••••••••••••••• ••••• •••••• ••• ••• •• ••• •••• ••• •• ••• ••• ••• •• •••• •••• ••••• •••••••• ••••••• •••• •••••••••• • •









-200

-100

0

100

-200

ORIGINAL COEFFICIENTS

-100

0

100

ORIGINAL COEFFICIENTS

Prior for the glint example

BIC prior

Figure 6: Glint Data: e ect of the choice of prior distributions on shrinkage.

23

Suggest Documents