首页 | 本学科首页   官方微博 | 高级检索  
     


Learning to trust in the competence and commitment of agents
Authors:Michael J Smith  Marie desJardins
Affiliation:(1) University of Maryland Baltimore County, 1000 Hilltop Circle, Baltimore, MD 21250, USA
Abstract:For agents to collaborate in open multi-agent systems, each agent must trust in the other agents’ ability to complete tasks and willingness to cooperate. Agents need to decide between cooperative and opportunistic behavior based on their assessment of another agents’ trustworthiness. In particular, an agent can have two beliefs about a potential partner that tend to indicate trustworthiness: that the partner is competent and that the partner expects to engage in future interactions. This paper explores an approach that models competence as an agent’s probability of successfully performing an action, and models belief in future interactions as a discount factor. We evaluate the underlying decision framework’s performance given accurate knowledge of the model’s parameters in an evolutionary game setting. We then introduce a game-theoretic framework in which an agent can learn a model of another agent online, using the Harsanyi transformation. The learning agents evaluate a set of competing hypotheses about another agent during the simulated play of an indefinitely repeated game. The Harsanyi strategy is shown to demonstrate robust and successful online play against a variety of static, classic, and learning strategies in a variable-payoff Iterated Prisoner’s Dilemma setting.
Keywords:Trust  Reputation  Learning  Game theory  Decision theory  Bayesian updating  Q-learning  Harsanyi transformation  Prisoner’  s dilemma  IPD  Evolutionary games  Pavlov  Bayesian Model Averaging
本文献已被 SpringerLink 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号