Abstract

I analyze the effects of memory limitations on the endogenous learning behavior of an agent in a standard two-armed bandit problem. An infinitely lived agent chooses each period between two alternatives with unknown types, to maximize discounted payoffs. The agent can experiment with each alternative and receive payoffs that are partially informative about its type. The agent does not recall past actions or payoffs. Instead, the agent has a finite number of memory states as in Wilson (2004): he can condition his actions only on the memory state he is currently in, and he can update his memory state depending on the payoff received.I find that the inclination to choose the currently better alternative does not constrain learning in the limit as discounting vanishes. Even though uncertainties are independent, the agent optimally holds correlated beliefs across memory states. Optimally, memory states reflect the magnitude of the relative ranking of alternatives. After a high payoff from one of the alternatives, the agent optimally moves to a memory state with more pessimistic beliefs on the other, even though no information about the latter alternative is received. For the case where one alternative is substantially more informative than the other, he chooses the latter only for myopic exploitation purposes, and ignores any information about it, suggesting specialization in learning.For the special case with one known (safe) alternative, a sufficiently patient agent never ceases experimentation and tries the unknown alternative at least occasionally after any history. Furthermore, he chooses the safe alternative with more optimistic beliefs than the optimal full memory cutoff belief, suggesting under-experimentation. Both are counter to what theory predicts with full memory, but in agreement with experimental findings.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.