Learning generates Long Memory - Archive ouverte HAL Access content directly
Other Publications Year : 2013

Learning generates Long Memory

(1) , (2)
1
2

Abstract

We consider a prototypical representative-agent forward-looking model, and study the low frequency variability of the data when the agent's beliefs about the model are updated through linear learning algorithms. We find that learning in this context can generate strong persistence. The degree of persistence depends on the weights agents place on past observations when they update their beliefs, and on the magnitude of the feedback from expectations to the endogenous variable. When the learning algorithm is recursive least squares, long memory arises when the coefficient on expectations is sufficiently large. In algorithms with discounting, long memory provides a very good approximation to the low-frequency variability of the data. Hence long memory arises endogenously, due to the self-referential nature of the model, without any persistence in the exogenous shocks. This is distinctly different from the case of rational expectations, where the memory of the endogenous variable is determined exogenously. Finally, this property of learning is used to shed light on some well-known empirical puzzles.
Fichier principal
Vignette du fichier
WP1113_update.pdf (1.42 Mo) Télécharger le fichier
Origin : Publisher files allowed on an open archive
Loading...

Dates and versions

hal-00661012 , version 1 (18-01-2012)
hal-00661012 , version 2 (15-10-2013)

Identifiers

  • HAL Id : hal-00661012 , version 2

Cite

Guillaume Chevillon, Sophocles Mavroeidis. Learning generates Long Memory. 2013, pp.49. ⟨hal-00661012v2⟩

Collections

ESSEC
112 View
459 Download

Share

Gmail Facebook Twitter LinkedIn More