Abstract
The existing models of Bayesian learning with multiple priors by Marinacci (Stat Pap 43:145–151, 2002) and by Epstein and Schneider (Rev Econ Stud 74:1275–1303, 2007) formalize the intuitive notion that ambiguity should vanish through statistical learning in an one-urn environment. Moreover, the multiple priors decision maker of these models will eventually learn the “truth.” To accommodate nonvanishing violations of Savage’s (The foundations of statistics, Wiley, New York, 1954) sure-thing principle, as reported in Nicholls et al. (J Risk Uncertain 50:97–115, 2015), we construct and analyze a model of Bayesian learning with multiple priors for which ambiguity does not necessarily vanish in an one-urn environment. Our decision maker only forms posteriors from priors that survive a prior selection rule which discriminates, with probability one, against priors whose expected Kullback–Leibler divergence from the “truth” is too far off from the minimal expected Kullback–Leibler divergence over all priors. The “stubbornness” parameter of our prior selection rule thereby governs how much ambiguity will remain in the limit of our learning model.
Original language | English |
---|---|
Pages (from-to) | 409-447 |
Number of pages | 39 |
Journal | Economic Theory |
Volume | 64 |
Issue number | 3 |
DOIs | |
Publication status | Published - 1 Oct 2017 |
Keywords
- Ambiguity
- Bayesian learning
- Berk’s Theorem
- Ellsberg paradox
- Kullback–Leibler divergence
- Misspecified priors