Trust Course CS 6381 -- Grid and Peer-to-Peer Computing Gerardo Padilla Source • Part 1: A Survey Study on Trust Management in P2P Systems • Part 2: Trust-χ: A Peer-to-Peer Framework for Trust Establishment 2 Outline • What is Trust? • What is a Trust Management? • How to measure Trust? – Example • Reputation-based Trust Management Systems – DMRep – EigenRep – P2PRep • Frameworks for Trust Establishment – Trust- χ 3 What is Trust? • Kini & Choobineh trust is: "a belief that is influenced by the individual’s opinion about certain critical system features" • Gambetta " …trust (or, symmetrically, distrust) is a particular level of the subjective probability with which an agent will perform a particular action, both before [the trustor] can monitor such action (or independently of his capacity of ever to be able to monitor it) • The Trust-EC project (http://dsa-isis.jrc.it/TrustEC/) trust is: "the property of a business relationship, such that reliance can be placed on the business partners and the business transactions developed with them''. • Gradison and Sloman trust is: "the firm belief in the competence of an entity to act dependably, securely and reliably within a specified context". . 4 What is Trust? Some Basic Properties of Trust Relations • Trust is relative to some business transaction. A may trust B to drive her car but not to baby-sit. • Trust is a measurable belief. A may trust B more than A trusts C for the same business. • Trust is directed. A may trust B to be a profitable customer but B may distrust A to be a retailer worth buying from. • Trust exists and evolves in time. The fact that A trusted B in the past does not in itself guarantee that A will trust B in the future. B’s performance and other relevant information may lead A to re-evaluate her trust in B. 5 Reputation, Trust and Reciprocity • Reputation: perception that an agent creates through past actions about its intentions and norms. • Trust: a subjective expectation a peer has about another's future behavior based on the history of their encounters. • Reciprocity: mutual exchange of deeds reputation Increase pi’s reputation Increase pj’s trust of pi trust reciprocity Increase pi’s reciprocating actions 6 Outline • What is Trust? • What is a Trust Management? • How to measure Trust? – Example • Reputation-based Trust Management Systems – DMRep – EigenRep – P2PRep • Frameworks for Trust Establishment – Trust- χ 7 What is a Trust Management? • “a unified approach to specifying and interpreting security policies, credentials, relationships [which] allows direct authorization of security-critical actions” – Blaze, Feigenbaum & Lacy • Trust Management is the capture, evaluation and enforcement of trusting intentions. • Other areas: Distributed Agent Artificial Intelligence/ Social Sciences 8 What is a Trust Management? Trust Management Policy-Based Trust Systems Social NetworkBased Trust Systems ReputationBased Trust Systems 9 What is a Trust Management? Trust Management Policy-Based Trust Systems Social NetworkBased Trust Systems ReputationBased Trust Systems • Example: PolicyMaker • Goal: Access Control • Peers use credential verification to establish a trust relationship • Unilateral, only the resource-owner request to establish trust 10 What is a Trust Management? Trust Management Policy-Based Trust Systems Social NetworkBased Trust Systems ReputationBased Trust Systems • Example: Marsh, Regret, NodeRanking, … • Based on social relationships between peers when computing trust and reputation values • Form conclusions about peers through analyzing a social network 11 What is a Trust Management? Trust Management Policy-Based Trust Systems Social NetworkBased Trust Systems ReputationBased Trust Systems • Example: DMRep, EigenRep, P2PRep, XRep, NICE, … • Based on measuring Reputation • Evaluate the trust in the peer and the trust in the reliability of the resource 12 Outline • What is Trust? • What is a Trust Management? • How to measure Trust? – Example • Reputation-based Trust Management Systems – DMRep – EigenRep – P2PRep • Frameworks for Trust Establishment – Trust- χ 13 How to measure Trust? An example of Computation Mode A computational Model of Trust and Reputation (Mui et al,2001) • Let’s assume a social network where no new peers are expected to join or leave – (i.e. the social network is static) cooperate a cooperate b Action Space = {cooperate, failing} 14 How to measure Trust? An example of Computation Mode • Let’s assume a social network where no new peers are expected to join or leave – (i.e. the social network is static) cooperate a failing b Action Space = {cooperate, failing} 15 How to measure Trust? An example of Computation Mode • Reputation: perception that a peer creates through past actions about its intentions and norms – Let θji(c) represents pi’s reputation in a social network of concern to pj for a context c. – This value measures the likelihood that pi reciprocates pj’s actions. 16 How to measure Trust? An example of Computation Mode • θab : b’s reputation in the eyes of a. a Context c b • Xab(i): the ith transaction between a and b. 1 Xab(i) 0 if b' s action is cooperate otherwise • After n transactions. We obtained the history data – History: Dab = {Xab(1), Xab(2), … , Xab(n)} 17 How to measure Trust? An example of Computation Mode • θab : b’s reputation in the eyes of a. a Context c b • Let p be the number of cooperations by peer b toward a in the n previous encounters. – b’s reputation θab for peer a should be a function of both p and n. – A simple function can be the proportion of cooperative action over all n encounters (or transactions) • From statistics, a proportion random variable can be modeled as a 18 Beta distribution How to measure Trust? An example of Computation Mode • NOTE: Beta Distribution Shape Parameters 19 How to measure Trust? An example of Computation Mode • Beta distribution: p( ) = Beta(α, β) – : estimator for θ – α and β: α = β = 1 (by prior assumptions) • A simple estimator for θab p ab n • b’s reputation in the eyes of a as the proportion of cooperation in n finite encounters. 20 How to measure Trust? An example of Computation Mode • Trust is defined as the subjective expectation a peer has about another’s future behavior based on the history of encounters. T(c) = E[ θ(c) | D(c)] – The higher the trust level for peer ai, the higher the expectation that ai will reciprocate peer aj’s actions. 21 How to measure Trust? An example of Computation Mode • Assuming that each encounter’s cooperation probability is independent of other encounters between a and b, the likelihood of p cooperations and (n – p) failings can be modeled as: – The likelihood for the n encounters: • Combining the prior and the likelihood, the posterior estimate for becomes (the subscripts are omitted): p ( | D) Beta ( p, n p ) 22 How to measure Trust? An example of Computation Mode • Trust towards b from a is the conditional expectation of given D. Tab = p(xab(n+1)|D) Then p T ab E[ | D] n 23 Outline • What is Trust? • What is a Trust Management? • How to measure Trust? – Example • Reputation-based Trust Management Systems – DMRep – EigenRep – P2PRep • Frameworks for Trust Establishment – Trust- χ 24 Reputation-based Trust Management Systems Introduction • Examples of completely centralized mechanism for storing and exploring reputation data: – Amazon.com • Visitors usually look for customer reviews before deciding to buy new books. – eBay • Participants at eBay’s auctions can rate each other after each transaction. 25 Reputation-based Trust Management Systems P2P Properties • • • • No central coordination No central database No peer has a global view of the system Global behavior emerges from local interactions • Peers are autonomous • Peers and connections are unreliable 26 Reputation-based Trust Management Systems Design Considerations • The system should be self-policing – The shared ethics of the user population are defined and enforced by the peers themselves and not by some central authority • The system should maintain anonymity – A peer’s reputation should be associated with an opaque identifier rather with an externally associated identity • The system should not assign any profit to newcomers • The system should have minimal overhead in terms of computation, infrastructure, storage, and message complexity • The system should be robust to malicious collectives of peers who know one another and attempt to collectively subvert the system. 27 Reputation-based Trust Management Systems Design Considerations : DMRep Managing Trust in a P2P Information System (Aberer,Despotovic,2001) • P2P Facts: – No central coordination or DB (e.g. not eBay) – No peer has global view – Peers autonomous and unreliable • Importance of trust in digital communities, but information dispersed and sources are not unconditionally trustworthy • Solution: reputation as decentralized storage of replicated & redundant transaction history – Calculate binary trust metric based on history of complaints. 28 Reputation-based Trust Management Systems DMRep Notation • Let P denote the set of all peers. • The behavioral data B are observations t(q,p) that a peer q makes when he interacts with a peer p. • The behavioral data of p, B(p) B(p) = { t (p, q) or t (q, p) | q P} B(p) B In a decentralized system, how to model, store, and 29 compute B? Reputation-based Trust Management Systems DMRep • In the decentralized environment, if a peer q has to determine trustworthiness of a peer p – It has no access to global knowledge B and B(p) – 2 ways to obtain data: • Directly by interactions Bq(p) = { t (q, p) | t (q, p) B} • Indirectly through a limited number of referrals from witnesses r Wq P Wq(p) = { t (r, p) | r Wq, t (r, p) B} 30 Reputation-based Trust Management Systems DMRep • Assumption: – The probability of cheating or having malicious within a society is comparably low • In case of a malicious behavior of q, a peer p can file a complaint c(p,q) • Complaints are the only behavioral data B used in this model 31 Reputation-based Trust Management Systems DMRep • Let us look a simple situation • p and q interact, • later r wants to determine the trustworthiness of p and q. – Assume p is cheating, q is honest – After their interaction, • q will file a complaint about p • p will file a complaint about q in order to hide its misbehavior. – r can not detect that p is cheating, – If p continues to cheat with more peers, r can conclude that it is very probable that p is the cheater by observing the other complaints about p 32 Reputation-based Trust Management Systems DMRep • Based on the previous simple scenario, the reputation T(p) of a peer p can be computed as the product • T(p) = |{c(p,q)| qP}| x |{c(q,p)| qP}| – High value of T(p) indicate that p is not trustworthy – |{c(p,q)| qP}|: number of complains made by p – |{c(q,p)| qP}|: number of complains about p Problem: The reputation was determined based on the global knowledge on complains which is very difficult to obtain. How to store the complaints? 33 Reputation-based Trust Management Systems DMRep • The storage structure proposed in this approach uses P-Grid (other can be used, such as CAN or CHORD) • P- Grid is a peer-to-peer lookup system based on a virtual distributed search tree. • It stores data items for which the associated path is a prefix of the data key. – For the trust management application this are the complaints indexed by the peer number. 34 Routing Tables Data Stores 35 Reputation-based Trust Management Systems DMRep • The same data can be stored at multiple peers and we have replicas of this data improve reliability • As the example shows, collisions of interest may occur, where peers are responsible for storing complaints about themselves. We do not exclude this, as for large peer populations these cases will be very rare and multiple replicas will be available to double-check. 36 Reputation-based Trust Management Systems DMRep – Problem: The peers providing the data could themselves be malicious – Assume that the peers are only malicious with a certain probability π ≤ π max <1. • If there are r replicas satisfies on average π rmax < ε, where ε is an acceptable fault-tolerance. – Problem Solution: If we receive the same data about a specific peer from a sufficient number of replicas we need no further checks, otherwise continue search. 37 Reputation-based Trust Management Systems DMRep • How it works? P-Grid has two operations for storage-retrieve information – insert(p; k; v), • where p is an arbitrary peer in the network, k is the key value to be searched for, and v is a data value associated with the key. – query(r; k) : v, • where r is an arbitrary peer in the network, which returns the data values v for a corresponding query k. 38 Reputation-based Trust Management Systems DMRep • How it works? – Every peer p can file a complaint about q at any time. It stores the complaint by sending messages insert(a1; key(p); c(p; q)) and insert(a2; key(q); c(p; q)) to arbitrary peers a1 and a2. 39 Reputation-based Trust Management Systems DMRep – Query Results • Assume that a peer p query for information about q (p evaluates the trustworthiness of q) – p submits messages query(a; key(q)) to arbitrary peers a. – This process is performed s times. 40 Reputation-based Trust Management Systems DMRep – Query Results – The result of these queries, called W, such that • w: number of witness found • cri(q): number of complaints that q received according witness ai • cfi(q): number of complaints q filed according witness ai • fi: the frequency with which ai is found (non-uniformity of the P-Grid structure) 41 Reputation-based Trust Management Systems DMRep – Variability • Different frequencies fi indicate that not all witnesses are found with the same probability due to the nonuniformity of the P-Grid structure. – Thus witnesses found less frequently will probably also not receive as many storage messages when complaints are filed. Thus the number of complaints they report will tend to be too low. – Problem: We need to compensate the information contribution from every witness. – Problem solution: Normalize values by using the frequencies . • High contribution (high fi), high probability • Low contribution (low fi), low probability 42 Reputation-based Trust Management Systems DMRep – Variability the probability of not finding witness i in s attempts. 43 Reputation-based Trust Management Systems DMRep – Trust • This model proposed to decide if a peer p considers peer q trustworthy (binary decision) based on tracking the history and computing T. • Thus p keeps a statistics of the average number of complaints received and complaints filed, aggregating all observations it makes over its lifetime. • Using the following heuristic approach: 44 Reputation-based Trust Management Systems DMRep – Trust if an observed value for complaints exceeds the general average of the trust measure too much, the agent must be Dishonest. 45 Reputation-based Trust Management Systems DMRep - Discussion • Strength – The method can be implemented in a fully decentralized peer-to-peer environment and scales well for large number of participants. • Limitations – environment with low cheating rates. – specific data management structure. – Not robust to malicious collectives of peers. 46 Outline • What is Trust? • What is a Trust Management? • How to measure Trust? – Example • Reputation-based Trust Management Systems – DMRep – EigenRep – P2PRep • Frameworks for Trust Establishment – Trust- χ 47 Reputation-based Trust Management Systems Design Considerations : EigenRep The Eigen Trust Algorithm for Reputation Management in P2P Networks (Kamvar, Schossler,2003) • Goal: To identify sources of inauthentic files and bias peers against downloading from them. • Method: Give each peer a trust value based on its previous behavior. 0.9 0.1 48 Reputation-based Trust Management Systems EigenRep: Terminology Peer 3 c • Local trust value: ij. The opinion that peer i has of peer j, based on past experience. Peer 1 • Global trust value: ti. The trust that the entire system places in peer i. t3=.5 C12=0.3 t1=.3 C23=0.7 C21=0.6 C14=0.01 t2=.2 Peer 2 t4=0 49 Peer 4 Reputation-based Trust Management Systems EigenRep: Normalizing Local Trust Values • All cij non-negative • ci1 + ci2 + . . . + cin = 1 Peer 1 Peer 1 C12=0.9 Peer 2 Peer 4 C14=0.1 Peer 2 50 Peer 4 Reputation-based Trust Management Systems EigenRep: Local Trust Vector • Local trust vector ci: contains all local trust values cij that peer i has of other peers j. 0 0 Peer 2 Peer 1 C12=0.9 Peer 4 C14=0.1 c1 Peer 2 Peer 4 0 0 .9 0 0 .1 Peer 1 51 Reputation-based Trust Management Systems EigenRep: Local Trust Values • Model Assumptions: – Each time peer i downloads an authentic file from peer j, cij increases. – Each time peer i downloads an inauthentic file from peer j, cij decreases. How to quantify these assumptions? 52 Reputation-based Trust Management Systems EigenRep: Local Reputation Values • Local Reputation Values = own experience – sat(i, j): number of satisfactory transactions (downloads) peer i has had with peer j. – unsat(i, j): number of unsatisfactory transactions (downloads) peer i has had with peer j local – Reputation value: sij=sat(i, j) - unsat(i, j). 53 Reputation-based Trust Management Systems EigenRep: Normalizing Local Reputation Value • Normalize Local Reputation Values -> Local Reputation Vector cij max( sij ,0) j ci max( sij ,0) • Note: 1 1, i.e. N c j 1 ij 1 T – Local reputation vector: ci (ci1 ,..., ciN ) • Most cijare 0 – 0 cij 1 54 Reputation-based Trust Management Systems EigenRep: Normalizing Local Reputation Value • Issues: – Advantages of normalizing • Reduce the problem where “malicious peers can assign arbitrarily high local trust values to other malicious peers, and arbitrarily low local trust values to good peers, easily subverting the system.” – Disadvantages of normalizing • “the normalized trust values do not distinguish between a peer with whom peer i did not interact and a peer with whom peer i has had poor experience.” • “these cij values are relative, and there is no absolute interpretation. That is, if cij = cik, we know that peer j has the same reputation as peer k in the eyes of peer i, but we don’t 55 know if both of them are very reputable, or if both of them are mediocre.” Reputation-based Trust Management Systems EigenRep: Local Reputation Values • Problem: – The peers have limited own experience. • Solution: – Get information from other peers who may have more experience about other peers. How? 56 Reputation-based Trust Management Systems EigenRep: Combining information by asking others • Ask for the opinions of the people who you trust. 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 57 Reputation-based Trust Management Systems EigenRep: Aggregating Local Reputation Values • Peer i asks its friends about their opinions on peer k. t ik cij c jk What they think of peer k. j Ask your friends j And weight each friend’s opinion by how much you trust him. 58 Reputation-based Trust Management Systems EigenRep: Aggregating Local Reputation Values t ik cij c jk j • Peer i asks its friends about their opinions on all peers. t i1 c11 ... ... t c ik 1k ... ... t c iN 1N ... c k1 ... c kk ... c kN ... c N 1 ci1 ... ... ... c Nk cik ... ... ... c NN ciN T t i C ci 59 Reputation-based Trust Management Systems EigenRep: Aggregating Local Reputation Values • Peer i asks its friends about their opinions about other peers again. (It seems like asking his friends’ friends) T 2 i i t (C ) c • Continues in this manner, T n ti (C ) ci • If n is large, will converge to the same vector por every peer i (left principal eigenvector of C for every peer i if C is irreducible and aperiodic) 60 Reputation-based Trust Management Systems EigenRep: Global Reputation Vector, • We call this eigenvector vector. t t , the global reputation t – i , an element of , quantifies how much trust the system as a whole places peer j. How to Estimate t? 61 Reputation-based Trust Management Systems EigenRep: EigenTrust Algorithm (non-dist) • Basic EigenTrust Algorithm – Non-distributed Algorithm – Assume that some central server knows all the cij values and performs the computation. (0) t e; repeat ( k 1) T (k ) t C t ; t ( k 1) t ( k ) until 62 Reputation-based Trust Management Systems EigenRep: EigenTrust Algorithm (non-dist) • Basic EigenTrust Algorithm – Issues to consider: • A priori notions of trust – There are some peers in the network that are known to be trustworthy (pre-trusted peers). Good idea to incorporate this information. • Inactive peers or new peers – Peers which do not download files from other peers or do not know other peers. • Malicious collectives – A malicious collective is a group of malicious peers who know each other, who give each other high local trust values and give all other peers low local trust values in an attempt to63 subvert the system and gain high global trust values. Reputation-based Trust Management Systems EigenRep: EigenTrust Algorithm (non-dist) • Pre-trust peers: P is a set of peers which are known to be trusted, p is the pre-trusted vector of P, where, 1 / P , if i P pi 0, othervise • Assign some trust on pre-trust peers ci p • and use this information in new or inactive peers: 64 Reputation-based Trust Management Systems EigenRep: EigenTrust Algorithm (non-dist) • To avoid Malicious collectives ( k 1) T (k ) t (1 a)C t ap • Where a is some constant less than 1. • This strategy breaks collectives by having each peer place at least some trust in the peers P that are not part of a collective. • Strong Assumption: Pre-trusted peers are essential 65 Reputation-based Trust Management Systems EigenRep: EigenTrust Algorithm (non-dist) • Modified Basic EigenTrust Algorithm – non-distributed algorithm (0) t p; repeat t ( k 1) C T t ( k ) ; t ( k 1) (1 a)C T t ( k 1) ap ( k 1) ( k ) t t until Now, let’s consider a distributed environment 66 Reputation-based Trust Management Systems EigenRep: EigenTrust Algorithm (distributed) • All peers in the network cooperate to compute and store the global trust vector. • Each peer stores and computes its own global trust value. • Minimize the computation, storage, and message overhead. ( k 1) T (k ) t (1 a)C t ap ti ( k 1) (1 a)(c1i t1( k ) ... c Ni t N( k ) ) api 67 Distributed Algorithm (cont…) • Ai: set of peers which have downloaded files from peer i. • Bi: set of peers which peer i has downloaded files. ci 2 t i( k 1) 3 2 c1i t1( k ) 4 1 c5i t5( k ) for each peer i do{ query all peers j Ai for t (j0 ) p j ; repeat compute t ( k 1) i (1 a )(c t (k ) 1i 1 ... c t ) ap i ; (k ) Ni N send cijti( k 1) to all peers j Bi ; compute t ( k 1) i t (k ) i 6 tit( ki(k1)) i ... ; 7 12 ( k 1) ji j wait for all peers j Ai to return c t until ; 5 0 c6 i t 6( k ) ci 7 ti( k 1) 8 11 c11i t11( k ) 9 10 ci 9ti( k 1) } Predecessor: Ai (download from i) Successor: Bi (downloaded by i) 68 Reputation-based Trust Management Systems EigenRep: EigenTrust Algorithm (distributed) • Complexity For a network of 1000 peers after 100 query cycles 69 Reputation-based Trust Management Systems EigenRep: EigenTrust Algorithm (dist. secure) • Issue: The trust value of one peer should be computed by more than one other peer. – malicious peers report false trust values of their own. – malicious peers compute false trust values for others. 70 Reputation-based Trust Management Systems EigenRep: EigenTrust Algorithm (dist. secure) • Solution Strategy: – the current trust value of a peer must not be computed by and reside at the peer itself, where it can easily become subject to manipulation. – the trust value of one peer in the network will be computed by more than one other peer. • Use multiple DHTs to assign mother peers, such as CAN or CHORD. • The number of mother peers for one peer is same to all peers. 71 Reputation-based Trust Management Systems EigenRep: EigenTrust Algorithm (dist. secure) 3 2 … 4 1 5 0 Ai, BiAi Ai Bi 0 0 2 ci 1 0 1 2 90.2 1 5 5 9120.5 5 12 0.3 11 #11 11 i 6 Ai ... 7 12 8 11 Bi ci 0 2 0.2 1 9 0.5 5 12 0.3 11 9 10 … Mother: Mi (compute for i) Daughter: Di (computed by i) Predecessor: Ai (download from i) Successor: Bi (downloaded by i) 72 Reputation-based Trust Management Systems EigenRep: EigenTrust Algorithm (dist. secure) 3 2 Ai Bi 4 1 0 2 0.2 1 9 0.5 5 12 0.3 11 5 0 m i 6 ci tAi(1)B ct ( 2 ) i i ii ... t i 0 2Query 0.2all peers j Ai for c jit j c ji p j ; 1 9Repeat 0.5 5 12 0.3 Compute ti( k 1) (1 a )(c1i t1( k ) ... c Ni t N( k ) ) ap i ; 11 (0) ... 7 12 8 11 Send cijti( k 1) to all peers j Bi ; 9 Compute ti( k 1) ti( k ) ; 10 Wait for all peers j Ai to return c jit (j k 1) ; Until ; Mother: Mi (compute for i) Daughter: Di (computed by i) Predecessor: Ai (download from i) Successor: Bi (downloaded by i) 73 Reputation-based Trust Management Systems EigenRep: EigenTrust Algorithm (dist. secure) for each peer i do send Ai , Bi , ci to its mothers M i ; collect Ad , Bd , c d from its daughters Di ; for each d Di do i Hasht (d ); k 0; send c dj t d( k ) c dj p d to all peers Hasht ( j ), for j Bi ; repeat wai t for all peers Hasht ( j ), for j Ad to return c jd t (j k ) ; compute t d( k 1) (1 a )(c1d t1( k ) ... c Nd t N( k ) ) ap d ; send c dj t d( k 1) to all peers Hasht ( j ), for j Bi ; compute t i( k 1) t i( k ) ; k k 1; until ; end end 74 Reputation-based Trust Management Systems EigenRep: Limitation of EigenRep • Cannot distinguish between newcomers and malicious peers. • Malicious peers can still cheat cooperatively – A peer should not report its predecessors by itself. • Flexibility – How to calculate reputation values when peers join and leave, on line and off line. • When to update global reputation values? – According to the new local reputation vector of all peers. • Anonymous? – A mother peer know its daughters. 75 Outline • What is Trust? • What is a Trust Management? • How to measure Trust? – Example • Reputation-based Trust Management Systems – DMRep – EigenRep – P2PRep • Frameworks for Trust Establishment – Trust- χ 76 Reputation-based Trust Management Systems P2PRep: Introduction Choosing reputable servents in a P2P network (Cornelli et al, 2002) • Not focus on computation of reputations • Security of exchanged messages – Queries – Votes • How to prevent different security attacks 77 Reputation-based Trust Management Systems P2PRep: Introduction • Using Gnutella for reference – A fully P2P decentralized infrastructure – Peers have low accountability and trust – Security threats to Gnutella • Distribution of tampered information • Man in the middle attack 78 Reputation-based Trust Management Systems P2PRep: Sketch of P2PRep • To ensure authenticity of offerers & voters, and confidentiality of votes • Use public-key encryption to provide integrity and confidentiality of messages • Require peer_id to be a digest of a public key, for which the peer knows the private key • Votes are values expressing opinions on other peers • Servent reputation represents the “trustworthiness” of a servent in providing files • Servent credibility represents the “trustworthiness” of a servent in providing votes 79 Reputation-based Trust Management Systems P2PRep: Sketch of P2PRep • P select a peer among those who respond to P’s query • P polls its peers for opinions about the selected peer • Peers respond to the polling with votes • P uses the votes to make its decision 80 Reputation-based Trust Management Systems P2PRep: Approaches • Two approaches: – Basic polling • Voters do not provide peer_id in votes – Enhanced polling • Voters declare their peer_id in votes 81 Reputation-based Trust Management Systems P2PRep: Basic Polling • Phase 1: Resource searching. • p sends a Query message for searching resources, and servents matching the request respond with a QueryHit 82 Reputation-based Trust Management Systems P2PRep: Basic Polling • Phase 2: Vote polling. • p polls its peers about the reputation of a top list T of servents, and peers wishing to respond send back a PollReply 83 Reputation-based Trust Management Systems P2PRep: Basic Polling • Phase 3: Voter evaluation. • p selects a set of voters, contacts them directly, and expects back a confirmation message 84 Reputation-based Trust Management Systems P2PRep: Basic Polling • Phase 4: Resource download. • p selects a servent s from which download the resource and starts a challenge-response phase before downloading 85 Reputation-based Trust Management Systems P2PRep: Enhanced Polling • Phase 1: Resource searching. • p sends a Query message for searching resources, and servents matching the request respond with a QueryHit 86 Reputation-based Trust Management Systems P2PRep: Enhanced Polling • Phase 2: Vote polling. • p polls its peers about the reputation of a top list of servents, and peers wishing to respond send back a PollReply 87 Reputation-based Trust Management Systems P2PRep: Enhanced Polling • Phase 3: Voter evaluation. • p selects a set of voters, contacts them directly to avoid servent_id to declare fake IPs 88 Reputation-based Trust Management Systems P2PRep: Enhanced Polling • Phase 4: Resource download. • p selects a servent s from which download the resource and starts a challenge-response phase before downloading 89 Reputation-based Trust Management Systems P2PRep: Comparison: Basic vs Enhanced • Basic polling – all votes are considered equal • Enhanced polling – peer_ids allow p to weight the votes based on v’s trustworthiness 90 Reputation-based Trust Management Systems P2PRep: Security Improvements (1) • Distribution of Tampered Information – B responds to A with a fake resource • P2PRep Solution: – A discovers the harmful content from B – A updates B’s reputation, preventing further interaction with B – A become witness against B in pollings by others 91 Reputation-based Trust Management Systems P2PRep: Security Improvements (2) • Man in the Middle Attack – Data from C to A can be modified by B, who is in the path • A broadcasts a Query and C responds • B intercepts the QueryHit from C and rewrites it with B’s IP & port • A receives B’s reply • A chooses B for downloading • B downloads original content from C, modifies it and passes it to A 92 Reputation-based Trust Management Systems P2PRep: Security Improvements (2) • Man in the Middle Attack – P2PRep addresses this problem by including a challenge-response phase before downloading – To impersonate C, B needs • C’s private key • To design a public key whose digest is C’s identifier – Public key encryption strongly enhances the integrity of the exchanged messages – Both versions address this problem 93 Outline • What is Trust? • What is a Trust Management? • How to measure Trust? – Example • Reputation-based Trust Management Systems – DMRep – EigenRep – P2PRep • Frameworks for Trust Establishment – Trust- χ 94 Frameworks for Trust Establishment Trust- χ: Introduction • Trust establishment via trust negotiation – Exchange of digital credentials • Credential exchange has to be protected – Policies for credential disclosure • Claim: Current approaches to trust negotiation don’t provide a comprehensive solution that takes into account all phases of the negotiation process 95 Frameworks for Trust Establishment Trust- χ: Trust Negotiation model Resource request Client Policy Base Server Policy Base Policies Policies Subject Profile Subject Profile Credentials Credentials 96 Resource granted Slide Slide from: from: http://www.ccs.neu.edu/home/ahchan/wsl/symposium/bertino.ppt http://www.ccs.neu.edu/home/ahchan/wsl/symposium/bertino.ppt Frameworks for Trust Establishment Trust- χ • XML-based system • Designed for a peer-to-peer environment – Both parties are equally responsible for negotiation management. – Either party can act as a requester or a controller of a resource • X-TNL: XML based language for specifying certificates and policies 97 Frameworks for Trust Establishment Trust- χ • Certificates: They are of two types – Credentials: States personal characteristics of its owner and is certified by a CA – Declarations: collect personal information about its owner that does not need to be certified • Trust tickets (X-TNL) – Used to speed up negotiations for a resource when access was granted in a previous negotiation • Support for policy pre-conditions • Negotiation conducted in phases 98 Frameworks for Trust Establishment Trust- χ: Credentials and Declarations a) Credential b) Declaration 99 The basic Trust-X system Alice Policy Database Bob X Profile Tree Manager Policy Database X Profile Compliance Checker Tree Manager Compliance Checker 100 Slide from: http://www.ccs.neu.edu/home/ahchan/wsl/symposium/bertino.ppt Frameworks for Trust Establishment Trust- χ : Message exchange in a Trust-X negotiation Alice Bob Request Prerequisite acknowledge Service request Disclosure policies Disclosure policies Match disclosure policies Credential and/or Declaration Credential and/or Declaration Service granted Preliminary Information exchange Bilateral disclosure of policies Actual credential disclosure INTRODUCTORY PHASE POLICY EXCHANGE CREDENTIAL DISCLOSURE RESOURCE DISCLOSURE 101 Slide from: http://www.ccs.neu.edu/home/ahchan/wsl/symposium/bertino.ppt Frameworks for Trust Establishment Trust- χ: Disclosure Policies • “They state the conditions under which a resource can be released during a negotiation” • Prerequisites – associated to a policy, it’s a set of alternative disclosure policies that must be satisfied before the disclosure of the policy they refer to. 102 Frameworks for Trust Establishment Trust- χ : Logic formalism Disclosure policies are expressed in terms of logical expressions which can specify either simple or composite conditions against certificates. • P() credential type • C set of conditions Policy expressed as P(C) TERM RP1(c), P2(c) Resource which Requested the policy refers to certificates Slide from: http://www.ccs.neu.edu/home/ahchan/wsl/symposium/bertino.ppt 103 Example • Consider a Rental Car service. • The service is free for the employees of Corrier company. Moreover, the Company already knows Corrier employees and has a digital copy of their driving licenses. Thus, it only asks the employees for the company badge and a valid copy of the ID card, to double check the ownership of the badge. By contrast, rental service is available on payment for unknown requesters, who have to submit first a digital copy of their driving licence and then a valid credit card. These requirements can be formalized as follows: 104 Example (2) 105 Trust-X negotiation 106 Frameworks for Trust Establishment Trust- χ: Negotiation Tree • Used in the policy evaluation phase • Maintains the progress of a negotiation • Used to identify at least a possible trust sequence that can lead to success in a negotiation (a view) 107 Frameworks for Trust Establishment Trust- χ : Negotiation Tree (2) 108 Summary • Thanks!, Questions? 109