Abstract
An observer of a process (xt) believes the process is governed by Q whereas the true law is P. We bound the expected average distance between P (xt | x1, ..., xt - 1) and Q (xt | x1, ..., xt - 1) for t = 1, ..., n by a function of the relative entropy between the marginals of P and Q on the n first realizations. We apply this bound to the cost of learning in sequential decision problems and to the merging of Q to P.
Original language | English (US) |
---|---|
Pages (from-to) | 24-32 |
Number of pages | 9 |
Journal | Journal of Mathematical Economics |
Volume | 44 |
Issue number | 1 |
DOIs | |
State | Published - Jan 1 2008 |
Keywords
- Bayesian learning
- Entropy
- Repeated decision problem
- Value of information
ASJC Scopus subject areas
- Economics and Econometrics
- Applied Mathematics