ART Testbed Join the discussion group at: http://www.art-testbed.net © 2006 THE UNIVERSITY OF TEXAS AT AUSTIN 1 ART Testbed Questions Can agents request reputations about themselves? Can an agent produce an appraisal without purchasing opinions? Does the Testbed assume a common representation for reputations? Does the Testbed prevent agents from winning via action-planning skills, as opposed to trust-modeling skills? What if an agent can’t or won’t give a reputation value? Why does it cost more to generate an accurate opinion than an inaccurate one? Why not have a centralized reputation broker? Isn’t it unrealistic to assume a true value of a painting can be known? Is art appraisal a realistic domain? Why not design an incentive-compatible mechanism to enforce truth-telling? © 2006 THE UNIVERSITY OF TEXAS AT AUSTIN 2 “Really Good” ART Testbed Questions Is there a consensus on the definitions of “trustworthiness” and “reputation”? How can collusion be avoided? Is truth-telling a dominant strategy? Will the system reach equilibrium, at which point reputations are no longer useful? What happens if client fee (100), opinion cost (10), and reputation cost (1) are changed? Do any equilibria exist? What happens when agents enter or leave the system? When will agents seek out reputations? Space of experiments is underexplored—that’s a good thing! © 2006 THE UNIVERSITY OF TEXAS AT AUSTIN 3 Questions about the Paper What is a “trust model”? How does q-learning work? How related to reinforcement learning? How do rewards tie in? What is lambda? How can experience- and reputation-based learning be combined to overcome the weaknesses of each (intermediate lambda values)? What about different combinations of (more sophisticated) agents in a game? Why the assumptions chosen? They seem too extreme. Reputation decisions weren’t examined very well. © 2006 THE UNIVERSITY OF TEXAS AT AUSTIN 4 The Agent Reputation and Trust Testbed: Experimentation and Competition for Trust in Agent Societies Karen K. Fullam1, Tomas B. Klos2, Guillaume Muller3, Jordi Sabater4, Andreas Schlosser5, Zvi Topol6, K. Suzanne Barber1, Jeffrey S. Rosenschein6, Laurent Vercouter3, and Marco Voss5 1Laboratory for Intelligent Processes and Systems, University of Texas at Austin, USA for Mathematics and Computer Science (CWI), Amsterdam, The Netherlands 3 Ecole Nationale Superieure des Mines, Saint-Etienne, France 4Institute of Cognitive Science and Technology (ISTC), National Research Council (CNR), Rome, Italy 5IT Transfer Office, Darmstadt University of Technology, Darmstadt, Germany 6Multiagent Systems Research Group—Critical MAS, Hebrew University, Jerusalem, Israel 2Center Testbed Game Rules If an appraiser is not very knowledgeable For a fixed price, clients Agents functionabout as a painting, it can ask appraisers to provide purchase "opinions" art appraisers with appraisals of paintings from varying expertise in other appraisers. from various eras. different Appraiserartistic eras. Agent Client Client Appraiser Agent Opinions and Reputations Client Clientwhose Share Appraisers appraisals are more accurate receive larger shares of the client base in the future. The Agent Reputation and Trust Testbed, 2006 Appraiser Agent Appraiser Agent Appraisers can also buy and sell reputation information about other Appraisers compete toappraisers. achieve the highest earnings by the end of the game. Appraiser Agent Step 1: Client and Expertise Assignments Appraisers receive clients who pay a fixed price to request appraisals Client paintings are randomly distributed across eras As game progresses, more accurate appraisers receive more clients (thus more profit) The Agent Reputation and Trust Testbed, 2006 Step 2: Reputation Transactions Appraisers know their own level of expertise for each era Appraisers are not informed (by the simulation) of the expertise levels of other appraisers Appraisers may purchase reputations, for a fixed fee, from other appraisers Reputations are values between zero and one • Might not correspond to appraiser’s internal trust model • Serves as standardized format for inter-agent communication The Agent Reputation and Trust Testbed, 2006 Step 2: Reputation Transactions Requester sends request message to a potential reputation provider, identifying appraiser whose reputation is requested Requester Requester sends fixed payment to the provider The Agent Reputation and Trust Testbed, 2006 Provider Request Accept Payment Reputation Potential reputation provider sends “accept” message Provider sends reputation information, which may not be truthful Step 3: Opinion Transactions For a single painting, an appraiser may request opinions (each at a fixed price) from as many other appraisers as desired The simulation “generates” opinions about paintings for opinion-providing appraisers Accuracy of opinion is proportional to opinion provider’s expertise for the era and cost it is willing to pay to generate opinion Appraisers are not required to truthfully reveal opinions to requesting appraisers The Agent Reputation and Trust Testbed, 2006 Step 3: Opinion Transactions Requester sends request message to a potential opinion provider, identifying painting Requester Requester sends fixed payment to the provider The Agent Reputation and Trust Testbed, 2006 Provider Request Certainty Potential provider sends a certainty assessment about the opinion it can provide - Real number (0 – 1) - Not required to truthfully report certainty assessment Payment Opinion Provider sends opinion, which may not be truthful Step 4: Appraisal Calculation Upon paying providers and before receiving opinions, requesting appraiser submits to simulation a weight (selfassessed reputation) for each other appraiser Simulation collects opinions sent to appraiser (appraisers may not alter weights or received opinions) Simulation calculates “final appraisal” as weighted average of received opinions True value of painting and calculated final appraisal are revealed to appraiser Appraiser may use revealed information to revise trust models of other appraisers The Agent Reputation and Trust Testbed, 2006 2006 ART Testbed Competition Results Karen K. Fullam The Laboratory for Intelligent Processes and Systems Electrical and Computer Engineering The University of Texas at Austin http://www.lips.utexas.edu Competition Organization “Practice” Competition • Spanish Agent School, Madrid, April 2006 • 12 participants International Competition • AAMAS, Hakodate, May 2006 • Preliminary Round 13 Participants 5 games each • Final Round 5 Finalists 10 games with all finalists participating © 2006 THE UNIVERSITY OF TEXAS AT AUSTIN 14 Bank Balances Iam achieves highest bank balances © 2006 THE UNIVERSITY OF TEXAS AT AUSTIN 15 Opinion Purchases Joey and Neil do not purchases opinions Sabatini purchases the most opinions © 2006 THE UNIVERSITY OF TEXAS AT AUSTIN 16 Opinion Earnings Sabatini and Iam provide the most opinions Neil and Frost do not provide many opinions © 2006 THE UNIVERSITY OF TEXAS AT AUSTIN 17 Opinion Sensing Costs Iam invests the most in opinions it generates © 2006 THE UNIVERSITY OF TEXAS AT AUSTIN 18 Expertise vs. Bank Balance Iam’s average expertise was not significantly higher than others’ Greater Expertise © 2006 THE UNIVERSITY OF TEXAS AT AUSTIN 19 Learning Trust Strategies in Reputation Exchange Networks Karen K. Fullam K. Suzanne Barber The Laboratory for Intelligent Processes and Systems Electrical and Computer Engineering The University of Texas at Austin http://www.lips.utexas.edu Trust Decisions in Reputation Exchange Networks Agents perform transactions to obtain needed resources • Transactions have risk because partners may be untrustworthy • Agents must learn whom to trust and how trustworthy to be When agents can exchange reputations • Agents must also learn when to request reputations and what reputations to tell • Agents’ trust decisions affect each other If I cheat A, and Difficult to learn each decision independently Resources A tells B, trustworthy will it hurt How my interactions should I with be? Should I B? trust? (goods, services, information) IfWhat I lie toreputations others that Cshould is bad,I tell? can I monopolize C’s interactions? Which reputations should I listen to? Reputations © 2006 THE UNIVERSITY OF TEXAS AT AUSTIN 21 Enumerating Decisions in a Trust Strategy Num agents = a Num transaction types = e Num choices/decision = n Trustee Truster Trustee Truster Agent Role Transaction Fundamental Reputation How trustworthy should I be? Should I tell an accurate reputation? combinations n ae combinations n a 2e Should I believe this reputation? Should I trust? combinations n ae combinations n a 2e © 2006 THE UNIVERSITY OF TEXAS AT AUSTIN If these decisions affect each other, there are n 2 ae1 a possible strategies! How to learn the best strategy with so many choices? 22 Reinforcement Learning Strategy Expected Reward Select a strategy A B C D Strategies with higher expected rewards are more likely to be selected © 2006 THE UNIVERSITY OF TEXAS AT AUSTIN Strategy feedback influences expected reward 23 Learning In Reputation Exchange Networks Strategy Expected Reward Decision Tr(A),Tr(B),Tr(C)… Tr(A) ⌐Tr(A),Tr(B),Tr(C)… ⌐Tr(A) Expected Reward Tr(A),⌐Tr(B),Tr(C)… Decision ⌐Tr(A),⌐Tr(B),Tr(C)… Expected Reward Tr(B) Tr(A),Tr(B),⌐Tr(C)… ⌐Tr(B) ⌐Tr(A),Tr(B),⌐Tr(C)… Tr(A),⌐Tr(B),⌐Tr(C)… Decision ⌐Tr(A),⌐Tr(B),⌐Tr(C)… Expected Reward Tr(C) ... ⌐Tr(C) Because decisions are interdependent, 2 ae1 a there are n . possible strategies! Use the ART Testbed as a case study © 2006 THE UNIVERSITY OF TEXAS AT AUSTIN Removing interdependencies makes each decision in the strategy learnable 24 Many Interdependent Decisions Reputation Requester’s reputation costs Opinion Requester’s opinion costs Opinion Provider’s opinion order costs Reputation Provider Reputation Requester Opinion Requester Opinion Provider Number of requests received by Reputation Provider Accuracy of Reputation Requester’s trust models Accuracy of Opinion Requester’s appraisals Number of requests received by Opinion Provider Opinion Requester’s client revenue Opinion Provider’s opinion revenue Reputation Provider’s reputation revenue When Reputation Requester is Opinion Requester Other Appraisers’ client revenue © 2006 THE UNIVERSITY OF TEXAS AT AUSTIN 25 Opinion Requester Feedback Opinion Client Reward = – Purchase Revenue Costs Opinion Requester’s opinion costs Opinion Requester Assume: Client revenue feedback is wholly attributed to Opinion Requester decision Divide revenue (client revenue) among opinions based on opinion accuracy Opinion Requester’s client revenue © 2006 THE UNIVERSITY OF TEXAS AT AUSTIN 26 Opinion Provider Feedback Opinion Provider’s opinion order costs Opinion Opinion Reward = Selling – Generating Revenue Costs Opinion Provider Assume: Client revenue is not related to Opinion Provider decision Opinion Provider’s opinion revenue Other Appraisers’ client revenue © 2006 THE UNIVERSITY OF TEXAS AT AUSTIN 27 Reputation Provider Feedback Reputation Reward = Selling Revenue Reputation Provider Reputation Provider’s reputation revenue Assume: Client revenue is not related to Reputation Provider decision Other Appraisers’ client revenue © 2006 THE UNIVERSITY OF TEXAS AT AUSTIN 28 Reputation Requester Feedback Reward = l( Opinion Requester Reward ) Reputation – Purchase Costs Reputation Requester’s reputation costs Opinion Requester’s opinion costs Reputation Requester Opinion Requester l = 0: Past experience only Opinion-requesting decision No reward for requesting reputations l = 1: Reputations only Reputation-requesting decision Full reward for requesting reputations Opinion Requester’s client revenue © 2006 THE UNIVERSITY OF TEXAS AT AUSTIN l determines influence of: past experience vs. reputations in deciding to purchase opinions 29