Bayesian optimum designs for discriminating between models with any distribution
The Bayesian KL-optimality criterion is useful for discriminating between any two statistical models in the presence of prior information. If the rival models are not nested then, depending on which model is true, two different Kullback-Leibler distances may be defined. The Bayesian KL-optimality criterion is a convex combination of the expected values of these two possible Kullback-Leibler distances between the competing models. These expectations are taken over the prior distributions of the parameters and the weights of the convex combination are given by the prior probabilities of the models. Concavity of the Bayesian KL-optimality criterion is proved, thus classical results of Optimal Design Theory can be applied. A standardized version of the proposed criterion is also given in order to take into account possible different magnitudes of the two Kullback-Leibler distances. Some illustrative examples are provided.
Year of publication: |
2010
|
---|---|
Authors: | Tommasi, C. ; López-Fidalgo, J. |
Published in: |
Computational Statistics & Data Analysis. - Elsevier, ISSN 0167-9473. - Vol. 54.2010, 1, p. 143-150
|
Publisher: |
Elsevier |
Saved in:
Saved in favorites
Similar items by person
-
An optimal experimental design criterion for discriminating between non-normal models
López-Fidalgo, J., (2007)
-
Integral approximations for computing optimum designs in random effects logistic regression models
Tommasi, C., (2014)
-
Estimation for finite population variance in double sampling
Diana, G., (2004)
- More ...