Filtering trust opinions through reinforcement learning

In open online communities such as e-commerce, participants need to rely on services provided by others in order to thrive. Accurately estimating the trustworthiness of a potential interaction partner is vital to a participant's well-being. It is generally recognized in the research community that third-party testimony sharing is an effective way for participants to gain knowledge about the trustworthiness of potential interaction partners without having to incur the risk of actually interacting with them. However, the presence of biased testimonies adversely affects a participant's long term well-being. Existing trust computational models often require complicated manual tuning of key parameters to combat biased testimonies. Such an approach heavily involves subjective judgments and adapts poorly to changes in an environment. In this study, we propose the Actor-Critic Trust (ACT) model, which is an adaptive trust evidence aggregation model based on the principles of reinforcement learning. The proposed method dynamically adjusts the selection of credible witnesses as well as the key parameters associated with the direct and indirect trust evidence sources based on the observed benefits received by the trusting entity. Extensive simulations have shown that the ACT approach significantly outperforms existing approaches in terms of mitigating the adverse effect of biased testimonies. Such a performance is due to the proposed accountability mechanism that enables ACT to attribute the outcome of an interaction to individual witnesses and sources of trust evidence, and adjust future evidence aggregation decisions without the need for human intervention. The advantage of the proposed model is particularly significant when service providers and witnesses strategically collude to improve their chances of being selected for interaction by service consumers. An actor-critic model (ACT) is proposed to filter out biased testimonies in online review systems.ACT learns key parameter values for existing reputation models to reduce subjectivity.ACT significantly outperforms existing works especially under ballot-stuffing attacks.

[1]  Paul A. Pavlou,et al.  Evidence of the Effect of Trust Building Technology in Electronic Markets: Price Premiums and Buyer Behavior , 2002, MIS Q..

[2]  Richard S. Sutton,et al.  Introduction to Reinforcement Learning , 1998 .

[3]  Michael Rovatsos,et al.  Using trust for detecting deceitful agents in artificial societies , 2000, Appl. Artif. Intell..

[4]  Chunyan Miao,et al.  Dynamic witness selection for trustworthy distributed cooperative sensing in cognitive radio networks , 2011, 2011 IEEE 13th International Conference on Communication Technology.

[5]  Chunyan Miao,et al.  A Survey of Trust and Reputation Management Systems in Wireless Communications , 2010, Proceedings of the IEEE.

[6]  Lik Mui,et al.  A Computational Model of Trust and Reputation for E-businesses , 2002 .

[7]  Mihaela Ulieru,et al.  The State of the Art in Trust and Reputation Systems: A Framework for Comparison , 2010, J. Theor. Appl. Electron. Commer. Res..

[8]  G. G. Stokes "J." , 1890, The New Yale Book of Quotations.

[9]  Jianqiang Shi,et al.  Dealing with Recommendations in a Statistical Trust Model , 2005 .

[10]  K. Suzanne Barber,et al.  Dynamically learning sources of trust information: experience vs. reputation , 2007, AAMAS '07.

[11]  Chunyan Miao,et al.  iCLUB: an integrated clustering-based approach to improve the robustness of reputation systems , 2011, AAMAS.

[12]  Vijay R. Konda,et al.  OnActor-Critic Algorithms , 2003, SIAM J. Control. Optim..

[13]  L. Mui,et al.  A computational model of trust and reputation , 2002, Proceedings of the 35th Annual Hawaii International Conference on System Sciences.

[14]  Stephen Marsh,et al.  Examining Trust, Forgiveness and Regret as Computational Concepts , 2009, Computing with Social Trust.

[15]  Catholijn M. Jonker,et al.  Formal Analysis of Models for the Dynamics of Trust Based on Experiences , 1999, MAAMAW.

[16]  Chunyan Miao,et al.  An Entropy-Based Approach to Protecting Rating Systems from Unfair Testimonies , 2006, IEICE Trans. Inf. Syst..

[17]  Audun Jøsang,et al.  AIS Electronic Library (AISeL) , 2017 .

[18]  Paul Resnick,et al.  Eliciting Informative Feedback: The Peer-Prediction Method , 2005, Manag. Sci..

[19]  Munindar P. Singh,et al.  Detecting deception in reputation management , 2003, AAMAS '03.

[20]  Gerald Tesauro,et al.  Temporal difference learning and TD-Gammon , 1995, CACM.

[21]  K. Suzanne Barber,et al.  Soft Security: Isolating Unreliable Agents from Society , 2002, Trust, Reputation, and Security.

[22]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[23]  Audun Jøsang,et al.  A survey of trust and reputation systems for online service provision , 2007, Decis. Support Syst..

[24]  Chunyan Miao,et al.  A Survey of Multi-Agent Trust Management Systems , 2013, IEEE Access.

[25]  Chunyan Miao,et al.  Trust-based web service selection in virtual communities , 2011, Web Intell. Agent Syst..

[26]  A. Jøsang,et al.  Filtering Out Unfair Ratings in Bayesian Reputation Systems , 2004 .

[27]  H. Chernoff A Measure of Asymptotic Efficiency for Tests of a Hypothesis Based on the sum of Observations , 1952 .

[28]  Boi Faltings,et al.  Towards Incentive-Compatible Reputation Management , 2002, Trust, Reputation, and Security.

[29]  Chunyan Miao,et al.  Credibility: How Agents Can Handle Unfair Third-Party Testimonies in Computational Trust Models , 2010, IEEE Transactions on Knowledge and Data Engineering.

[30]  Jordi Sabater-Mir,et al.  Reputation and social network analysis in multi-agent systems , 2002, AAMAS '02.

[31]  Jie Zhang,et al.  A Trust-Based Incentive Mechanism for E-Marketplaces , 2008, AAMAS-TRUST.