Bayesian and Adaptive Optimal Policy under Model Uncertainty
We study the problem of a policymaker who seeks to set policy optimally in an economy where the true economic structure is unobserved, and policymakers optimally learn from their observations of the economy. This is a classic problem of learning and control, variants of which have been studied in the past, but little with forward-looking variables which are a key component of modern policy-relevant models. As in most Bayesian learning problems, the optimal policy typically includes an experimentation component reflecting the endogeneity of information. We develop algorithms to solve numerically for the Bayesian optimal policy (BOP). However the BOP is only feasible in relatively small models, and thus we also consider a simpler specification we term adaptive optimal policy (AOP) which allows policymakers to update their beliefs but shortcuts the experimentation motive. In our setting, the AOP is significantly easier to compute, and in many cases provides a good approximation to the BOP. We provide a simple example to illustrate the role of learning and experimentation in an MJLQ framework.
Year of publication: |
2007
|
---|---|
Authors: | Williams, Noah ; Svensson, Lars E.O. |
Institutions: | Society for Economic Dynamics - SED |
Saved in:
Saved in favorites
Similar items by person
-
Monetary Policy with Model Uncertainty: Distribution Forecast Targeting
Williams, Noah, (2005)
-
Optimal Monetary Policy Under Uncertainty in DSGE Models: A Markov Jump-Linear-Quadratic Approach
Svensson, Lars E.O., (2008)
-
Optimal monetary policy under uncertainty: a Markov jump-linear-quadratic approach
Svensson, Lars E.O., (2008)
- More ...