Credibility: How Agents Can Handle Unfair Third-Party Testimonies in Computational Trust Models

  • Authors:
  • Jianshu Weng;Zhiqi Shen;Chunyan Miao;Angela Goh;Cyril Leung

  • Affiliations:
  • Singapore Management University, Singapore;Nanyang Technological University, Singapore;Nanyang Technological University, Singapore;Nanyang Technological University, Singapore;The University of British Columbia, Vancouver

  • Venue:
  • IEEE Transactions on Knowledge and Data Engineering
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

Usually, agents within multiagent systems represent different stakeholders that have their own distinct and sometimes conflicting interests and objectives. They would behave in such a way so as to achieve their own objectives, even at the cost of others. Therefore, there are risks in interacting with other agents. A number of computational trust models have been proposed to manage such risk. However, the performance of most computational trust models that rely on third-party recommendations as part of the mechanism to derive trust is easily deteriorated by the presence of unfair testimonies. There have been several attempts to combat the influence of unfair testimonies. Nevertheless, they are either not readily applicable since they require additional information which is not available in realistic settings, or ad hoc as they are tightly coupled with specific trust models. Against this background, a general credibility model is proposed in this paper. Empirical studies have shown that the proposed credibility model is more effective than related work in mitigating the adverse influence of unfair testimonies.