Constructing informative model priors using hierarchical methods |
| |
Authors: | Wolf Vanpaemel |
| |
Affiliation: | Department of Psychology, University of Leuven, Tiensestraat 102, B-3000 Leuven, Belgium |
| |
Abstract: | Despite their negative reputation, informative priors are very useful in inference. Priors that express psychologically meaningful intuitions damp out random fluctuations in the data due to sampling variability, without sacrificing flexibility. This article focuses on how an intuitively satisfying informative prior distribution can be constructed. In particular, it demonstrates how the hierarchical introduction of a parameterized generative account of the set of models under consideration naturally imposes a non-uniform prior distribution over the models, encoding existing intuitions about the models. The hierarchical approach for constructing informative model priors is made concrete using a worked example, the Varying Abstraction Model (VAM), a family of categorization models including and expanding the exemplar and prototype models. It is shown how psychological intuitions about the relative plausibilities of the models in the VAM can be formally captured in an informative prior distribution over these models, by specifying a theoretically informed process for generating the models in the VAM. The smoothing effect of the informative prior in estimation is demonstrated by considering ten previously published data sets from the category learning literature. |
| |
Keywords: | Bayes Hierarchical Estimation Priors Informative Subjective Abstraction Category learning |
本文献已被 ScienceDirect 等数据库收录! |
|