Experience-weighted attraction learning in normal form games

被引:831
作者
Camerer, C [1 ]
Ho, TH
机构
[1] CALTECH, Div Social Sci, Pasadena, CA 91125 USA
[2] Univ Penn, Wharton Sch, Dept Mkt, Philadelphia, PA 19104 USA
关键词
learning; behavioral game theory; reinforcement learning; fictitious play;
D O I
10.1111/1468-0262.00054
中图分类号
F [经济];
学科分类号
02 ;
摘要
In 'experience-weighted attraction' (EWA) learning, strategies have attractions that reflect initial predispositions, are updated based on payoff experience, and determine choice probabilities according to some rule (e.g., logit). A key feature is a parameter delta that weights the strength of hypothetical reinforcement of strategies that were not chosen according to the payoff they would have yielded, relative to reinforcement of chosen strategies according to received payoffs. The other key features are two discount rates, phi and rho, which separately discount previous attractions, and an experience weight. EWA includes reinforcement learning and weighted fictitious play (belief learning) as special cases, and hybridizes their key elements. When delta = 0 and rho = 0, cumulative choice reinforcement results. When delta = 1 and rho = phi, levels of reinforcement of strategies are exactly the same as expected payoffs given weighted fictitious play beliefs. Using three sets of experimental data, parameter estimates of the model were calibrated on part of the data and used to predict a holdout sample. Estimates of delta are generally around. 50, phi around. 8-1, and rho varies from 0 to phi. Reinforcement and belief-learning special cases are generally rejected in favor of EWA, though belief models do better in some constant-sum games. EWA is able to combine the best features of previous approaches, allowing attractions to begin and grow flexibly as choice reinforcement does, but reinforcing unchosen strategies substantially as belief-based models implicitly do.
引用
收藏
页码:827 / 874
页数:48
相关论文
共 63 条
[1]  
Anderson S.P., 1992, DISCRETE CHOICE THEO
[2]  
[Anonymous], 1962, LOGICAL FDN PROBABIL
[3]  
[Anonymous], RECHERCHES PRINCIPLE
[4]  
ARTHUR WB, 1991, AM ECON REV, V81, P353
[5]  
BATTALIO R, 1997, RISK DOMINANCE PAYOF
[6]  
Ben-Akiva M., 1985, Discrete choice analysis: theory and application to travel demand
[7]  
BEREBYMEYR Y, 1997, LEARNING SUCCESSFUL
[8]  
BORGERS T, 1996, NAIVE REINFORCEMENT
[9]   FICTITIOUS PLAY - A STATISTICAL STUDY OF MULTIPLE ECONOMIC EXPERIMENTS [J].
BOYLAN, RT ;
ELGAMAL, MA .
GAMES AND ECONOMIC BEHAVIOR, 1993, 5 (02) :205-222
[10]  
BRANDTS J, IN PRESS J EC BEHAV