Premium
Innovations in Bayes and empirical Bayes methods: estimating parameters, populations and ranks
Author(s) -
Louis Thomas A.,
Shen Wei
Publication year - 1999
Publication title -
statistics in medicine
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 1.996
H-Index - 183
eISSN - 1097-0258
pISSN - 0277-6715
DOI - 10.1002/(sici)1097-0258(19990915/30)18:17/18<2493::aid-sim271>3.0.co;2-s
Subject(s) - bayes' theorem , computer science , statistical inference , bayesian probability , prior probability , ranking (information retrieval) , bayes factor , bayesian inference , statistics , posterior probability , data mining , econometrics , machine learning , artificial intelligence , mathematics
Abstract By formalizing the relation among components and ‘borrowing information’ among them, Bayes and empirical Bayes methods can produce more valid, efficient and informative statistical evaluations than those based on traditional methods. In addition, Bayesian structuring of complicated models and goals guides development of appropriate statistical approaches and generates summaries which properly account for sampling and modelling uncertainty. Computing innovations enable implementation of complex and relevant models, thereby substantially increasing the role of Bayes/empirical Bayes methods in important statistical assessments. Policy‐relevant statistical assessments involve synthesis of information from a set of related components such as medical clinics, geographic regions or research studies. Typical assessments include inference for individual parameters, synthesis over the collection of components (for example, the parameter histogram) and comparisons among parameters (for example, ranks). The relative importance of these goals depends on the context. Bayesian structuring provides a guide to valid inference. For example, while posterior means are the ‘obvious’ and optimal estimates for individual components under squared error loss, their empirical distribution function (EDF) is underdispersed and never valid for estimating the EDF of the true, underlying parameters. Effective histogram estimates result from optimizing a loss function based in a distance between the histogram and its estimate. Similarly, ranking observed data usually produces poor estimates and ranking posterior means can be inappropriate. Effective estimates should be based on a loss function that caters directly to ranks. Using examples of ‘borrowing information’, shrinkage and the variance/bias trade‐off we motivate Bayes and empirical Bayes analysis. Then, we outline the formal approach and discuss ‘triple‐goal’ estimates with values that when ranked produce optimal ranks, for which the EDF is an optimal estimate of the parameter EDF and such that the values themselves are effective estimates of co‐ordinate‐specific parameters. We use basic models and data analysis examples to highlight the conceptual and structural issues. Copyright © 1999 John Wiley & Sons, Ltd.