Social Choice and Computer Science Fred Roberts, Rutgers University 1

advertisement
Social Choice and Computer
Science
Fred Roberts, Rutgers University
1
Social Choice
• How do societies (or groups) reach good
decisions?
• The theory of social choice deals
with this question.
• They argue.
• They find a dictator.
• They vote.
2
What do Mathematics and Computer
Science have to do with Voting?
3
Have you used Google lately?
4
5
Have you used Google lately?
Did you know that Google has something
to do with voting?
6
Have you tried buying a book on online
lately?
7
8
9
Have you tried buying a book on online
lately?
Did you get a message saying: If you are interested
in this book, you might want to look at the
following books as well?
Did you know that has something to do
with voting?
10
Have you ever heard of v-sis?
11
Have you ever heard of v-sis?
•It’s a cancer-causing gene.
Cancer cell
•Computer scientists helped discover how it
works.
•How did they do it?
•The answer also has something to do with
voting.
12
Computer Science and the Social
Sciences
•Many recent applications in CS involve
issues/problems of long interest to social scientists:
preference, utility
conflict and cooperation
allocation
incentives
measurement
social choice or consensus
•Methods developed in SS beginning to be used in
CS
13
CS and SS
•CS applications place great strain on SS methods
Sheer size of problems addressed
Computational power of agents an issue
Limitations on information possessed by
players
Sequential nature of repeated applications
•Thus: Need for new generation of SS methods
•Also: These new methods will provide powerful
tools to social scientists
14
Social Choice and CS: Outline
1.Consensus Rankings
2.Meta-search and Collaborative Filtering
3.Large Databases and Inference
4.Computational Intractability of Consensus
Functions
5.Electronic Voting
6.Software and Hardware Measurement
7.Power of a Voter
15
Social Choice and CS: Outline
1.Consensus Rankings
2.Meta-search and Collaborative Filtering
3.Large Databases and Inference
4.Computational Intractability of Consensus
Functions
5.Electronic Voting
6.Software and Hardware Measurement
7.Power of a Voter
16
How do Elections Work?
• Typically, everyone votes
for their first choice
candidate.
• The votes are counted.
• The person with the most
votes wins.
• Or, sometimes, if no one has
more than half the votes,
there is a runoff.
17
But do we necessarily get the best
candidate that way?
Let’s look back at the 2008 Democratic primaries.
18
Sometimes Having More
Information about Voters’
Preferences is Very Helpful
•Sometimes it is helpful to have voters rank order
all the candidates
•From their top choice to their bottom choice.
19
Rankings
Dennis Kucinich
Bill Richardson
John Edwards
Ties are allowed
Barack Obama Hillary Clinton
20
Rankings
• What if we have four voters and they give us the
following rankings? Who should win?
Voter 1
Clinton
Richardson
Edwards
Kucinich
Obama
Voter 2
Voter 3
Voter 4
Clinton
Obama
Obama
Kucinich Edwards Richardson
Edwards Richardson Kucinich
Richardson Kucinich Edwards
Obama
Clinton
Clinton
21
Rankings
• What if we have four voters and they give us the
following rankings?
• There is one added candidate.
• Who should win?
Voter 1
Clinton
Gore
Richardson
Edwards
Kucinich
Obama
Voter 2
Voter 3
Voter 4
Clinton
Obama
Obama
Gore
Gore
Gore
Kucinich Edwards Richardson
Edwards Richardson Kucinich
Richardson Kucinich Edwards
Obama
Clinton
Clinton
22
Rankings
Voter 1
Clinton
Gore
Richardson
Edwards
Kucinich
Obama
Voter 2
Voter 3
Voter 4
Clinton
Obama
Obama
Gore
Gore
Gore
Kucinich Edwards Richardson
Edwards Richardson Kucinich
Richardson Kucinich Edwards
Obama
Clinton
Clinton
Maybe someone who is everyone’s second choice
is the best choice for winner.
Point: We can learn something from ranking
candidates.
23
Consensus Rankings
•How should we reach a decision in an election if
every voter ranks the candidates?
•What decision do we want?
− A winner
− A ranking of all the candidates that is in some
sense a consensus ranking
•This would be useful in some applications
• Job candidates are ranked by each interviewer
• Consensus ranking of candidates
• Make offers in order of ranking
•How do we find a consensus ranking?
24
Consensus Rankings
Background: Arrow’s Impossibility Theorem:
• There is no “consensus method” that satisfies
certain reasonable axioms about how societies
should reach decisions.
• Input to Arrow’s Theorem: rankings
of alternatives (ties allowed).
• Output: consensus ranking.
Kenneth Arrow
Nobel prize winner
25
Consensus Rankings
• There are widely studied and widely used
consensus methods that violate one or
more of Arrow’s conditions.
• One well-known consensus method:
“Kemeny-Snell medians”: Given set
of rankings, find ranking minimizing
sum of distances to other rankings.
John Kemeny,
pioneer in time sharing
in CS
• Kemeny-Snell medians are having
surprising new applications in CS.
26
Consensus Rankings
These two rankings are very close:
Clinton
Obama
Edwards
Kucinich
Richardson
Obama
Clinton
Edwards
Kucinich
Richardson
27
Consensus Rankings
These two rankings are very far apart:
Clinton
Richardson
Edwards
Kucinich
Obama
Obama
Kucinich
Edwards
Richardson
Clinton
28
Consensus Rankings
•This suggests we may be able to make precise
how far apart two rankings are.
•How do we measure the distance between two
rankings?
29
Consensus Rankings
• Kemeny-Snell distance between rankings: twice the
number of pairs of candidates i and j for which i is
ranked above j in one ranking and below j in the
other + the number of pairs that are ranked in one
ranking and tied in another.
a
b
x
y-z
y
x
z
On {x,y}: +2
On {x,z}: +2
On {y,z}: +1
d(a,b) = 5.
30
Consensus Rankings
• Kemeny-Snell median: Given rankings a1, a2,
…, ap, find a ranking x so that
d(a1,x) + d(a2,x) + … + d(ap,x)
is minimized.
• x can be a ranking other than a1, a2, …, ap.
• Sometimes just called Kemeny median.
31
Consensus Rankings
a1
Fish
Chicken
Beef
a2
Fish
Chicken
Beef
a3
Chicken
Fish
Beef
• Median = a1.
• If x = a1:
d(a1,x) + d(a2,x) + d(a3,x) = 0 + 0 + 2 = 2
is minimized.
• If x = a3, the sum is 4.
• For any other x, the sum is at least 1 + 1 + 1 = 3.
32
Consensus Rankings
a1
Fish
Chicken
Beef
a2
Chicken
Beef
Fish
a3
Beef
Fish
Chicken
• Three medians = a1, a2, a3.
• This is the “voter’s paradox” situation.
33
Consensus Rankings
a1
Fish
Chicken
Beef
a2
Chicken
Beef
Fish
a3
Beef
Fish
Chicken
• Note that sometimes we wish to minimize
d(a1,x)2 + d(a2,x)2 + … + d(ap,x)2
• A ranking x that minimizes this is called a
Kemeny-Snell mean.
• In this example, there is one mean: the ranking
declaring all three alternatives tied.
34
Consensus Rankings
a1
Fish
Chicken
Beef
a2
Chicken
Beef
Fish
a3
Beef
Fish
Chicken
• If x is the ranking declaring Fish, Chicken
and Beef tied, then
d(a1,x)2 + d(a2,x)2 + … + d(ap,x)2 =
32 + 32 + 32 = 27.
• Not hard to show this is minimum.
35
Consensus Rankings
Theorem (Bartholdi, Tovey, and Trick, 1989;
Wakabayashi, 1986): Computing the KemenySnell median of a set of rankings is an NPcomplete problem.
36
Consensus Rankings
Okay, so what does this have to do with practical
computer science questions?
37
Consensus Rankings
I mean really practical computer science
questions.
38
Social Choice and CS: Outline
1.Consensus Rankings
2.Meta-search and Collaborative Filtering
3.Large Databases and Inference
4.Computational Intractability of Consensus
Functions
5.Electronic Voting
6.Software and Hardware Measurement
7.Power of a Voter
39
40
Google Example
• Google is a “search engine”
• It searches through web pages and rank orders
them.
• That is, it gives us a ranking of web pages from
most relevant to our query to least relevant.
41
Meta-search
• There are other search engines besides Google.
• Wouldn’t it be helpful to use several of them and
combine the results?
• This is meta-search.
• It is a voting problem
• Combine page rankings from several search engines to
produce one consensus ranking
• Dwork, Kumar, Naor, Sivakumar (2000): KemenySnell medians good in spam resistance in meta-search
(spam by a page if it causes meta-search to rank it too
highly)
• Approximation methods make this computationally
42
tractable
43
Collaborative Filtering
• Recommending books or movies
• Combine book or movie ratings by various
people
• This too is voting
• Produce a consensus ordered list of books or
movies to recommend
• Freund, Iyer, Schapire, Singer (2003):
“Boosting” algorithm for combining rankings.
• Related topic: Recommender Systems
44
Meta-search and Collaborative
Filtering
A major difference from the election situation
• In elections, the number of voters is large,
number of candidates is small.
• In CS applications, number of voters (search
engines) is small, number of candidates (pages)
is large.
• This makes for major new complications and
research challenges.
45
Social Choice and CS: Outline
1.Consensus Rankings
2.Meta-search and Collaborative Filtering
3.Large Databases and Inference
4.Computational Intractability of Consensus
Functions
5.Electronic Voting
6.Software and Hardware Measurement
7.Power of a Voter
46
Have you ever heard of v-sis?
•It’s a cancer-causing gene.
•Computer scientists helped discover how it
works.
•How did they do it?
•The answer also has something to do with
voting.
47
Large Databases and Inference
• Decision makers consult massive data sets.
• The study of large databases and gathering of
information from them is a major topic in
modern computer science.
• We will give an example from the field of
Bioinformatics.
• This lies at the interface between Computer
Science and Molecular Biology
48
Large Databases and Inference
• Real data often in form of sequences
• Here, concentrate on bioinformatics
• GenBank has over 7 million sequences
comprising 8.6 billion bases.
• The search for similarity or patterns has
extended from pairs of sequences to finding
patterns that appear in common in a large
number of sequences or throughout the database:
“consensus sequences”.
• Emerging field of “Bioconsensus”: applies SS
consensus methods to biological databases.
49
Large Databases and Inference
Why look for such patterns?
Similarities between sequences or parts of
sequences lead to the discovery of shared
phenomena.
For example, it was discovered that the sequence
for platelet derived factor, which causes growth
in the body, is 87% identical to the sequence for
v-sis, a cancer-causing gene. This led to the
discovery that v-sis works by stimulating
growth.
50
Large Databases and Inference
Example
Bacterial Promoter Sequences studied by
Waterman (1989):
RRNABP1:
TNAA:
UVRBP2:
SFC:
ACTCCCTATAATGCGCCA
GAGTGTAATAATGTAGCC
TTATCCAGTATAATTTGT
AAGCGGTGTTATAATGCC
Notice that if we are looking for patterns of length
4, each sequence has the pattern TAAT.
51
Large Databases and Inference
Example
Bacterial Promoter Sequences studied by
Waterman (1989):
RRNABP1:
TNAA:
UVRBP2:
SFC:
ACTCCCTATAATGCGCCA
GAGTGTAATAATGTAGCC
TTATCCAGTATAATTTGT
AAGCGGTGTTATAATGCC
Notice that if we are looking for patterns of length
4, each sequence has the pattern TAAT.
52
Large Databases and Inference
Example
However, suppose that we add another sequence:
M1 RNA: AACCCTCTATACTGCGCG
The pattern TAAT does not appear here.
However, it almost appears, since the pattern
TACT appears, and this has only one mismatch
from the pattern TAAT.
53
Large Databases and Inference
Example
However, suppose that we add another sequence:
M1 RNA: AACCCTCTATACTGCGCG
The pattern TAAT does not appear here.
However, it almost appears, since the pattern
TACT appears, and this has only one mismatch
from the pattern TAAT.
So, in some sense, the pattern TAAT is a good
consensus pattern.
54
Large Databases and Inference
Example
We make this precise using best mismatch
distance.
Consider two sequences a and b with b longer than
a.
Then d(a,b) is the smallest number of mismatches
in all possible alignments of a as a consecutive
subsequence of b.
55
Large Databases and Inference
Example
a = 0011, b = 111010
Possible Alignments:
111010
111010
111010
0011
0011
0011
The best-mismatch distance is 2, which is achieved
in the third alignment.
56
Large Databases and Inference
Smith-Waterman
•Let  be a finite alphabet of size at least 2 and  be a finite
collection of words of length L on .
•Let F() be the set of words of length k  2 that are our
consensus patterns. (Assume L  k.)
•Let  = {a1, a2, …, an}.
•One way to define F() is as follows.
•Let d(a,b) be the best-mismatch distance.
•Consider nonnegative parameters sd that are monotone
decreasing with d and let F(a1,a2, …, an) be all those
words w of length k that maximize
S(w) = isd(w,ai)
57
Large Databases and Inference
•We call such an F a Smith-Waterman consensus.
•In particular, Waterman and others use the parameters
sd = (k-d)/k.
Example:
•An alphabet used frequently is the purine/pyrimidine
alphabet {R,Y}, where R = A (adenine) or G (guanine)
and Y = C (cytosine) or T (thymine).
•For simplicity, it is easier to use the digits 0,1 rather than
the letters R,Y.
•Thus, let  = {0,1}, let k = 2. Then the possible pattern
words are 00, 01, 10, 11.
58
Large Databases and Inference
•Suppose a1 = 111010, a2 = 111111. How do we find
F(a1,a2)?
•We have:
d(00,a1) = 1, d(00,a2) = 2
d(01,a1) = 0, d(01,a2) = 1
d(10,a1) = 0, d(10,a2) = 1
d(11,a1) = 0, d(11,a2) = 0
S(00) =  sd(00,ai) = s1 + s2,
S(01) =  sd(01,ai) = s0 + s1
S(10) =  sd(10,ai) = s0 + s1
S(11) =  sd(11,ai) = s0 + s0
•As long as s0 > s1 > s2, it follows that 11 is the consensus
59
pattern, according to Smith-Waterman’s consensus.
Example:
•Let  ={0,1}, k = 3, and consider F(a1,a2,a3) where
a1 = 000000, a2 = 100000, a3 = 111110. The possible
pattern words are: 000, 001, 010, 011, 100, 101, 110, 111.
d(000,a1) = 0, d(000,a2) = 0, d(000,a3) = 2,
d(001,a1) = 1, d(001,a2) = 1, d(001,a3) = 2,
d(100,a1) = 1, d(100,a2) = 0, d(100,a3) = 1, etc.
S(000) = s2 + 2s0, S(001) = s2 + 2s1, S(100) = 2s1 + s0,
etc.
•Now, s0 > s1 > s2 implies that S(000) > S(001).
•Similarly, one shows that the score is maximized by
S(000) or S(100).
• Monotonicity doesn’t say which of these is highest.
60
Large Databases and Inference
The Special Case sd = (k-d)/k
•Then it is easy to show that the words w that maximize
s(w) are exactly the words w that minimize
id(w,ai).
•In other words: In this case, the Smith-Waterman
consensus is exactly the median.
Algorithms for computing consensus sequences
such as Smith-Waterman are important in modern
molecular biology.
61
Large Databases and Inference
Other Topics in “Bioconsensus”
• Alternative phylogenies (evolutionary trees) are
produced using different methods and we need
to choose a consensus tree.
• Alternative taxonomies (classifications) are
produced using different models and we need to
choose a consensus taxonomy.
• Alternative molecular sequences are produced
using different criteria or different algorithms
and we need to choose a consensus sequence.
• Alternative sequence alignments are produced
and we need to choose a consensus alignment. 62
Large Databases and Inference
Other Topics in “Bioconsensus”
• Several recent books on bioconsensus.
• Day and McMorris [2003]
• Janowitz, et al. [2003]
• Bibliography compiled by Bill Day: In
molecular biology alone, hundreds of papers
using consensus methods in biology.
• Large database problems in CS are being
approached using methods of “bioconsensus”
having their origin in the social sciences.
63
Social Choice and CS: Outline
1.Consensus Rankings
2.Meta-search and Collaborative Filtering
3.Large Databases and Inference
4.Computational Intractability of Consensus
Functions
5.Electronic Voting
6.Software and Hardware Measurement
7.Power of a Voter
64
Computational Intractability of Consensus
Functions
• How hard is it to compute the winner of an
election?
• We know counting votes can be difficult
and time consuming.
• However:
• Bartholdi, Tovey and Trick (1989): There are
voting schemes where it can be computationally
intractable to determine who won an election.
65
Computational Intractability of Consensus
Functions
• So, is computational intractability necessarily
bad?
• Computational intractability can be a good thing
in an election: Designing voting systems where
it is computationally intractable to “manipulate”
the outcome of an election by “insincere voting”:
 Adding voters
 Declaring voters ineligible
 Adding candidates
 Declaring candidates ineligible
66
Computational Intractability of Consensus
Functions
• Given a set A of all possible candidates and a set
I of all possible voters.
• Suppose we know voter i’s ranking of all
candidates in A, for every voter i.
• Given a subset of I of eligible voters, a
particular candidate a in A, and a number k,
is there a set of at most k ineligible voters who
can be declared eligible so that candidate a is
the winner?
• Bartholdi, Tovey, Trick (1989): For some
consensus functions (voting rules), this is an NP-67
complete problem.
Computational Intractability of Consensus
Functions
• Given a set A of all possible candidates and a set
I of all possible voters.
• Suppose we know voter i’s ranking of all
candidates in A, for every voter i.
• Given a subset of I of eligible voters, a
particular candidate a in A, and a number k,
is there a set of at most k eligible voters who
can be declared ineligible so that candidate a
is the winner?
• Bartholdi, Tovey, Trick (1989): For some
consensus functions (voting rules), this is an NP-68
complete problem.
Computational Intractability of Consensus
Functions
• Software agents may be more likely to
manipulate than individuals (Conitzer and
Sandholm 2002):
Humans don’t think about manipulating
Computation can be tedious.
Software agents are good at running
algorithms
Software agents only need to have code for
manipulation written once.
All the more reason to develop
computational barriers to manipulation.
69
Computational Intractability of Consensus
Functions
• Stopping those software agents:
70
Computational Intractability of Consensus
Functions
• Conitzer and Sandholm (2002):
 Earlier results of difficulty of manipulation depend
on large number of candidates
 New results: manipulation possible with some
voting methods if smaller number (bounded number)
of candidates)
 In weighted voting, voters may have
different numbers of votes (as in US
presidential elections, where different
states (= voters) have different
numbers of votes). Here, manipulation is harder.
 Manipulation difficult when uncertainty about
71
others’ votes.
Computational Intractability of Consensus
Functions
• Conitzer and Sandholm (2006):
Try to find voting rules for which
manipulation is usually hard.
Why is this difficult to do?
One explanation: under one reasonable
assumption, it is impossible to find such rules.
72
Social Choice and CS: Outline
1.Consensus Rankings
2.Meta-search and Collaborative Filtering
3.Large Databases and Inference
4.Computational Intractability of Consensus
Functions
5.Electronic Voting
6.Software and Hardware Measurement
7.Power of a Voter
73
•Some connections between Computer Science
and Voting are clearly visible.
•Some people are working on plans to allow us to
vote from home – over the Internet.
74
Electronic Voting
• Issues:
Correctness
Anonymity
Availability
Security
Privacy
75
Electronic Voting
Security Risks in Electronic Voting
• Threat of “denial of service attacks”
• Threat of penetration attacks involving a
delivery mechanism to transport a malicious
payload to target host (thru Trojan horse or
remote control program)
• Private and correct counting of votes
• Cryptographic challenges to keep votes private
• Relevance of work on secure multiparty
computation
76
Electronic Voting
Other CS Challenges:
• Resistance to “vote buying”
• Development of user-friendly interfaces
• Vulnerabilities of communication path between
the voting client (where you vote) and the server
(where votes are counted)
• Reliability issues: random hardware and
software failures
77
Social Choice and CS: Outline
1.Consensus Rankings
2.Meta-search and Collaborative Filtering
3.Large Databases and Inference
4.Computational Intractability of Consensus
Functions
5.Electronic Voting
6.Software and Hardware Measurement
7.Power of a Voter
78
Software & Hardware Measurement
• Theory of measurement developed by
mathematical social scientists
• Measurement theory studies ways to combine
scores obtained on different criteria.
• A statement involving scales of
measurement is considered meaningful if its
truth or falsity is unchanged under acceptable
transformations of all scales involved.
79
Software & Hardware Measurement
• A statement involving scales of
measurement is considered meaningful if its
truth or falsity is unchanged under acceptable
transformations of all scales involved.
• Example: Is it meaningful to say that I weigh
more than a cicada?
• Yes.
• That is because if it is true in kilograms, then it
is also true in pounds, in grams, etc.
80
Software & Hardware Measurement
• Even meaningful to say I weigh twice as much as a
cicada.
81
Software & Hardware Measurement
• Is it meaningful to say that the temperature today is
twice as much as it was yesterday?
• No.
• Could be true in Fahrenheit, false in Centigrade.
82
Software & Hardware Measurement
• Measurement theory has studied what statements you
can make after averaging scores.
• Think of averaging as a consensus method.
• One general principle: To say that the average score of
one set of tests is greater than the average score of
another set of tests is not meaningful (it is
meaningless) under certain conditions.
• This is often the case if the averaging procedure is to
take the arithmetic mean: If s(xi) is score of xi, i = 1, 2,
…, n, then arithmetic mean is
is(xi)/n.
• Long literature on what averaging methods lead to
meaningful conclusions.
83
Software & Hardware Measurement
A widely used method in hardware measurement:
Score a computer system on different
benchmarks.
Normalize score relative to performance of
one base system
Average normalized scores
Pick system with highest average.
Fleming and Wallace (1986): Outcome can
depend on choice of base system.
Meaningless in sense of measurement theory
Leads to theory of merging normalized scores
84
Software & Hardware Measurement
Hardware Measurement
P R
R
O
C
M
E
S
S
O Z
R
E
417
BENCHMARK
F
G
H
83
66 39,449
I
772
244
70
153
33,527
368
134
70
135
66,000
369
85
Data from Heath, Comput. Archit. News (1984)
Software & Hardware Measurement
Normalize Relative to Processor R
P R
R
O
C
M
E
S
S
O Z
R
E
417
1.00
BENCHMARK
F
G
H
I
83
66 39,449 772
1.00 1.00
1.00 1.00
244
.59
70
.84
153
2.32
33,527
.85
368
.48
134
.32
70
.85
135
2.05
66,000
1.67
369
.45
86
Software & Hardware Measurement
Take Arithmetic Mean of Normalized Scores
P R
R
O
C
M
E
S
S
O Z
R
E
417
1.00
Arithmetic
BENCHMARK
Mean
F
G
H
I
83
66 39,449 772
1.00
1.00 1.00
1.00 1.00
244
.59
70
.84
153
2.32
33,527
.85
368
.48
1.01
134
.32
70
.85
135
2.05
66,000
1.67
369
.45
1.07
87
Software & Hardware Measurement
Take Arithmetic Mean of Normalized Scores
P R
R
O
C
M
E
S
S
O Z
R
E
417
1.00
Arithmetic
BENCHMARK
Mean
F
G
H
I
83
66 39,449 772
1.00
1.00 1.00
1.00 1.00
244
.59
70
.84
153
2.32
33,527
.85
368
.48
1.01
134
.32
70
.85
135
2.05
66,000
1.67
369
.45
1.07
Conclude that machine Z is best
88
Software & Hardware Measurement
Now Normalize Relative to Processor M
P R
R
O
C
M
E
S
S
O Z
R
E
417
1.71
BENCHMARK
F
G
H
I
83
66 39,449 772
1.19
.43
1.18 2.10
244
1.00
70
1.00
153
1.00
33,527 368
1.00 1.00
134
.55
70
1.00
135
.88
66,000 369
1.97 1.00
89
Software & Hardware Measurement
Take Arithmetic Mean of Normalized Scores
P R
R
O
C
M
E
S
S
O Z
R
E
417
1.71
Arithmetic
BENCHMARK
Mean
F
G
H
I
83
66 39,449 772
1.32
1.19
.43
1.18 2.10
244
1.00
70
1.00
153
1.00
33,527 368
1.00 1.00
1.00
134
.55
70
1.00
135
.88
66,000 369
1.97 1.00
1.08
90
Software & Hardware Measurement
Take Arithmetic Mean of Normalized Scores
P R
R
O
C
M
E
S
S
O Z
R
E
417
1.71
Arithmetic
BENCHMARK
Mean
F
G
H
I
83
66 39,449 772
1.32
1.19
.43
1.18 2.10
244
1.00
70
1.00
153
1.00
33,527 368
1.00 1.00
1.00
134
.55
70
1.00
135
.88
66,000 369
1.97 1.00
1.08
91
Conclude that machine R is best
Software and Hardware Measurement
• So, the conclusion that a given machine is best
by taking arithmetic mean of normalized scores
is meaningless in this case.
• Above example from Fleming and Wallace
(1986), data from Heath (1984)
• Sometimes, geometric mean is helpful as a
consensus method.
• Geometric mean is
n

 is(xi)
92
Software & Hardware Measurement
Normalize Relative to Processor R
P R
R
O
C
M
E
S
S
O Z
R
E
417
1.00
Geometric
BENCHMARK
Mean
F
G
H
I
83
66 39,449 772
1.00
1.00 1.00
1.00 1.00
244
.59
70
.84
153
2.32
33,527
.85
368
.48
.86
134
.32
70
.85
135
2.05
66,000
1.67
369
.45
.84
Conclude that machine R is best
93
Software & Hardware Measurement
Now Normalize Relative to Processor M
P R
R
O
C
M
E
S
S
O Z
R
E
417
1.71
Geometric
BENCHMARK
Mean
F
G
H
I
83
66 39,449 772
1.17
1.19
.43
1.18 2.10
244
1.00
70
1.00
153
1.00
33,527 368
1.00 1.00
1.00
134
.55
70
1.00
135
.88
66,000 369
1.97 1.00
.99
Still conclude that machine R is best
94
Software and Hardware Measurement
• In this situation, it is easy to show that the conclusion
that a given machine has highest geometric mean
normalized score is a meaningful conclusion.
• Even meaningful: A given machine has geometric
mean normalized score 20% higher than another
machine.
• Fleming and Wallace give general conditions under
which comparing geometric means of normalized
scores is meaningful.
• Research area: what averaging procedures make sense
in what situations? Large literature.
• Note: There are situations where comparing arithmetic
means is meaningful but comparing geometric means is
95
not.
Software and Hardware Measurement
• Message from measurement theory to computer
science (Operations Research, and other
disciplines):
Do not perform arithmetic operations on
data without paying attention to whether
the conclusions you get are meaningful.
96
Social Choice and CS: Outline
1.Consensus Rankings
2.Meta-search and Collaborative Filtering
3.Large Databases and Inference
4.Computational Intractability of Consensus
Functions
5.Electronic Voting
6.Software and Hardware Measurement
7.Power of a Voter
97
Power of a Voter
Shapley-Shubik Power Index
• Think of voting situations with many players
• Shapley-Shubik index measures the power
of each player in a multi-player “voting game.”
• Consider a game in which some coalitions
of players win and some lose, with no
subset of a losing coalition winning.
Martin
Shubik
Lloyd Shapley
98
Power of a Voter
Example: Shareholders of Company
Shareholder 1 holds 3 shares.
Shareholders 2, 3, 4, 5, 6, 7 hold 1 share each.
A majority of shares are needed to make a decision.
Coalition {1,4,6} is winning.
Coalition {2,3,4,5,6} is winning.
A coalition is winning iff it has at least 5 of the 1-share
players or it has the 3-share player and at least two
others.
99
Power of a Voter
Shapley-Shubik Power Index
• Consider a coalition forming at random, one
player at a time.
• A player i is pivotal if addition of i throws
coalition from losing to winning.
• Shapley-Shubik index of i = probability i is
pivotal if an order of players is chosen at random.
• Power measure applying to more general games
than voting games is called Shapley Value.
100
Power of a Voter
Example: Shareholders of Company
Shareholder 1 holds 3 shares.
Shareholders 2, 3, 4, 5, 6, 7 hold 1 share each.
A majority of shares are needed to make a decision.
Coalition {1,4,6} is winning.
Coalition {2,3,4,5,6} is winning.
Shareholder 1 is pivotal if he is 3rd, 4th, or 5th.
So shareholder 1’s Shapley value is 3/7.
Sum of Shapley values is 1 (since they are probabilities)
Thus, each other shareholder has Shapley value
(4/7)/6 = 2/21
101
Power of a Voter
Example: Australian Government
• The same game arises in the workings of the
government of Australia.
• There are six states and the federal
government.
• A measure passes if it has the support of at least 5
states or at least 2 states plus the federal government.
• It is not hard to see that the winning coalitions are
exactly the same as in a game in which the federal
government holds 3 votes and each state holds 1 vote
and 5 votes are needed to win.
• Thus, the federal government has power 3/7 and each
state in Australia has power 2/21.
102
.
Power of a Voter
Example: United Nations Security Council
•15 member nations
•5 permanent members
China, France,
Russia, UK, US
•10 non-permanent
•Permanent members
have veto power
•Coalition wins iff it has all 5 permanent members
and at least 4 of the 10 non-permanent members.
103
Power of a Voter
Example: United Nations Security Council
•What is the power of each
Member of the Security
Council?
•Fix non-permanent member i.
•i is pivotal in permutations in
which all permanent members
precede i and exactly 3 nonpermanent members do.
•How many such permutations are there?
104
Power of a Voter
Example: United Nations Security Council
•Choose 3 non-permanent members preceding i.
•Order all 8 members preceding i.
•Order remaining 6 non-permanent members.
•Thus the number of such permutations is:
C(9,3) x 8! x 6! = 9!/3!6! x 8! x 6! = 9!8!/3!
•The probability that i is pivotal = power of nonpermanent member =
9!8!/3!15! = .001865
•The power of a permanent member =
[1 – 10 x .001865]/5 = .1963.
•Permanent members have 100 times power of 105
non-permanent members.
Power of a Voter
•There are a variety of other power indices used
in game theory and political science (Banzhaf
index, Coleman index, …)
•Need calculate them for huge games
•Mostly computationally intractable
106
Power of a Voter: Allocation/Sharing
Costs and Revenues
• Shapley-Shubik power index and more
general Shapley value have been used
to allocate costs to different users in
shared projects.
Allocating runway fees in airports
Allocating highway fees to trucks of
different sizes
Universities sharing library facilities
Fair allocation of telephone calling
charges among users sharing
complex phone systems (Cornell’s
experiment)
107
Power of a Voter: Allocating/Sharing
Costs and Revenues
Allocating Runway Fees at Airports
• Some planes require longer runways.
Charge them more for landings. How
much more?
• Divide runways into meter-long
segments.
• Each month, we know how many
landings a plane has made.
• Given a runway of length y meters,
consider a game in which the players are
landings and a coalition “wins” if the
runway is not long enough for planes in
the coalition.
108
Power of a Voter: Allocating/Sharing
Costs and Revenues
Allocating Runway Fees at Airports
• A landing is pivotal if it is the first
landing added that makes a coalition
require a longer runway.
• The Shapley value gives the cost of
the yth meter of runway allocated to a
given landing.
• We then add up these costs over all
runway lengths a plane requires and
all landings it makes.
109
Power of a Voter: Allocating/Sharing
Costs and Revenues
Multicasting
• Applications in multicasting, e.g., sending
movies out to viewers.
• Unicast routing: Each packet sent from a
source is delivered to a single receiver.
• Sending it to multiple sites: Send multiple
copies and waste bandwidth.
• In multicast routing: Use a directed tree
connecting source to all receivers.
• At branch points, a packet is duplicated as
necessary.
110
Multicasting
111
Power of a Voter: Allocating/Sharing
Costs and Revenues
Multicasting
• Multicast routing: Use a directed tree
connecting source to all receivers.
• At branch points, a packet is duplicated as
necessary.
• Bandwidth is not directly attributable to a
single receiver.
• How to distribute costs among receivers?
• One idea: Use Shapley value.
112
Allocating/Sharing Costs and Revenues
• Feigenbaum, Papadimitriou, Shenker (2001):
no feasible implementation for Shapley value
in multicasting.
• Note: Shapley value is uniquely characterized
by four simple axioms.
• Sometimes we state axioms as general
principles we want a solution concept to have.
• Jain and Vazirani (1998): polynomial time
computable cost-sharing algorithm
 Satisfying some important axioms
 Calculating cost of optimum multicast tree within
factor of two of optimal.
113
Concluding Comment
• In recent years, interplay between CS
and biology has transformed major
parts of Bio into an information science.
• Led to major scientific breakthroughs in
biology such as sequencing of human
genome.
• Led to significant new developments in
CS, such as database search.
• The interplay between CS and SS not
nearly as far along.
• Moreover: problems are spread over
many disciplines.
114
Concluding Comment
• However, CS-SS interplay has already
developed a unique momentum of its own.
• One can expect many more exciting outcomes
as partnerships between computer scientists
and social scientists expand and mature.
115
116
Download