By Topic

A Survey of Multi-Agent Trust Management Systems

Sign In

Full text access may be available.

To access full text, please use your member or institutional sign in.

The purchase and pricing options are temporarily unavailable. Please try again later.
5 Author(s)
Han Yu ; Sch. of Comput. Eng., Nanyang Technol. Univ., Singapore, Singapore ; Zhiqi Shen ; Leung, C. ; Chunyan Miao
more authors

In open and dynamic multiagent systems (MASs), agents often need to rely on resources or services provided by other agents to accomplish their goals. During this process, agents are exposed to the risk of being exploited by others. These risks, if not mitigated, can cause serious breakdowns in the operation of MASs and threaten their long-term wellbeing. To protect agents from the uncertainty in the behavior of their interaction partners, the age-old mechanism of trust between human beings is re-contexted into MASs. The basic idea is to let agents self-police the MAS by rating each other on the basis of their observed behavior and basing future interaction decisions on such information. Over the past decade, a large number of trust management models were proposed. However, there is a lack of research effort in several key areas, which are critical to the success of trust management in MASs where human beings and agents coexist. The purpose of this paper is to give an overview of existing research in trust management in MASs. We analyze existing trust models from a game theoretic perspective to highlight the special implications of including human beings in an MAS, and propose a possible research agenda to advance the state of the art in this field.

Published in:

Access, IEEE  (Volume:1 )
Comment Policy
comments powered by Disqus