Phishing - A Social Disorder

advertisement
Phishing - A Social Disorder
L. Jean Camp
Alla Genkina
Allan Friedman
www.ljean.com
A tangent
Workshop on the Economics of Information Security
2- 3 June 2005
Reception June 1
Early registration rates end tomorrow
http://www.infosecon.net/workshop
ROI in security investment, economics of spam, economics of identity,
vulnerability markets, empirical investigation on privacy & security
investments
Phishing - A Social Disorder

A lack of context

Human trust behaviors

Social engineering
– Begin with the social problem not with the technical potential
– Solve for human trust behaviors
– Provide unique signals
Human and Computer Trust
Social
sciences
Experiments designed to evaluate how people extend trust
Game theory
Common assumption: information exposure == trust
Philosophy
Macro approach
Examine societies and cultural practices
Computer
Security
Build devices to enable trust
Philosophy Suggests

Trust is necessary to simplify life
¯ People have an innate desire or need to trust
¯ People will default to extending trust

People will continue to extend trust - so create
another source of trust don’t defeat trusting
behaviors
Research on Humans Suggest...

Humans may not differentiate between machines

Humans become more trusting of ‘the network’

Humans begin with too much trust
Confirmed by philosophical macro observation
Confirmed by computer security incidents
Validated
• E-mail based
Scams
Viruses
Hoaxes
Three Observations

Humans respond differently to human or computer
"betrayals" in terms of forgiveness?

People interacting with a computer do not distinguish
between computers as individuals but rather respond to their
experience with "computers”

The tendency to differentiate between remote machines
increases with computer experience
Response to Failure

Humans respond differently to human or computer
"betrayals" in terms of forgiveness
– Attacks which are viewed as failures as ‘ignored’ or forgiven
– Technical failures as seen as accidents rather than design
decisions
» May explain why people tolerate repeated security failures
– May inform the balance between false positives and negatives in
intrusion detection
» Rarely identified malicious behavior taken more seriously
» Technical failures easily forgiven
» Failures expected
Individiation

People interacting with a computer do not
distinguish between computers as individuals but
rather respond to their experience with
"computers”
– People become more trusting
– People differentiate less
– People learn to trust
» First observed by Sproull in computer scientists in 1991
» Confirmed by all future privacy experiments
Differentiation

The tendency to differentiate between remote
machines decreases with computer experience
– Explicit implication of second hypothesis
– Explains common logon/passwords
» along with cognitive limits
Observed Verification of
Hypotheses

Users are bad security managers
 PGP, P3P,….

Security should necessarily be a default

Does end-to-end security maximize autonomy without end-toend human abilities and tendencies?

Data currently being compiled on experiments

Surveys illustrate a continuing confusion of privacy &
security
Computer security is built for
machines

Passwords
 Humans are a bad source of entropy

SSL




Two categories: secure and not secure
Does not encourage differentiation
Every site should include a unique graphic with the lock
Computer security should seek to differentiate machines
Privacy standards are built for
machines

P3P assumes
– All merchants trustworthy w.r.t. their own policies
– Assumes increasingly sophisticated user
– One standard for all transactions

PGP
– Monotonic increase in trust
– No reset
– No decrease in rate of trust extension
» To compensate for increasing trust
– No global or local reset
» E.g. change in status
Key revocation is built for
Machines

CRL tend to be single level

Different levels of revocation are needed
– Falsified initial credential
» All past transactions suspect
– Change in status
» Future transactions prohibited
– Refusal of renewal
» Current systems adequate

CRL should reflect the entire systems in which they work,
including the social system

CRL is too simplistic, depends on active checking
WHAT TO DO?

Computers
–
–
–
–

Process data
Store data
Transmit data
Distinguish
» atomicity, privacy, availability,
Humans
– Understand context
– Evaluate uncertainty
– Make lumping decisions based on context

Begin with the human as the basis of the design
– Examine human interactions
– Signal humans using pre-existing social capital
Context

Trust is contextual

Phillips on Zero Knowledge
– Nyms had to be selected before the person engaged in
interaction
– The interaction in question is entering information
– The information should be available before the interaction
Not Even Communicating with
Users

Identity theft
– unauthorized use of authenticating information to assert identity in the
financial namespace
– Internal process violation - Choicepoint (at least 145k records)
» All access to the Choiepoint database was authorized
» Subsequent misuse was authorized by the information obtained via
Choicepoint
– Security Violation - Berkeley
– Confidentiality information - Bank of American backup data 1.2M
records

Risk profile is similar for individuals in all three cases
Dominant Trust Communication
Equivalent Value
Cradle to Grave ID…. So What

Authentication as what? For what?

Identification as having what attributes?

Scope of namespace
– License to drive
» requires driving test
– SSN
» taxpayer ID to assert right to benefits
– Birth certificate
» proof of age
– Define a credit namespace that allows for large scale illegal employment
– Require that credit and banking agencies handle their own risks and
pay for their own identity schemes for all transactions but cash
– Make video rental agencies accept their own risks
– Cell phone requires that you have an account to pay for it
– DL requires you know how to drive
Perfect Single ID
… for every namespace
… and every context
… for all people
for definitions: http:// www.ljean.com/
… or solve the problem at hand by
enabling contextually rational trust
behavior
Embedding Browsing in Social
Context

First trust challenge
– Enabling trust to allow entry onto the net
– Enabling monetary flows

Second trust challenge
– Providing meaning trust information
» TrustE, BBB, Verisign
– Namespaces for specific trust assertions
» Christian, gay friendly, responsible merchants
– Requires a common understanding of the limits of the namespaces
» Transitivity
» Automated trust decisions
» Consistency across contexts or explicit definition of context
• E.g., purchase a book
– On divorce
– On impotency
– On effective job searching
– On number theory
Enabling Trust Behavior

Signal not to trust

Combine trust perceptions for end users
– Privacy
» Based on personal experience
» Or verification of centralized authority (BBB)
– Reliability
» Personal experience
» Verification (Consumer reports)
– Security
» Is personal experience valuable here?
• Q: what is the value of peer production for security information
» Centralized verification (Counterpane)
– ID theft vs. account misuse is distinguished by the bank but not by the
customer
– Loss of data from privacy or security is the same for the individual
» For whom should we design
Context

Selected context
determines

Social network
display

Shared
information

NOT certificate
authorities

Depends on
homophily
Visual Trust

Verisign will protect you from anyone who will not give
them money
– There has been no business plan of a single trusted root which
aligns that root with the end user.

There are competitive trust sources that align with user

Uses pop-up blocker
Centralized Elements

No hierarchies
– Trust is not inherently transitive
– “Verisign is the trust company”

Certificates
– Signed green list

Signer determines
–
–
–
–
Frequency of update
Business model
Determinant of entry
Potential examples
» FDIC
» Consumer reports
» BBB
» Phishguard
Reputation

Initial reputation of zero

First visit goes to 1 (out of ten)

After second visit it increases

Each visit decreases rate of delay
– Max of 10

Explicit rating
– Stays constant without alteration over time
A New Paradigm for Design

Design technology to conform to user behaviors

Assume individuals will default to trust, then lump, and
forgive

Depends upon
»
»
»
»
»

Usability
Reputation system design
Homophily
Storage capacity
Efficient search
Provide signals NOT to trust
– Do not assume that no signal means no trust.
– No signal will be interpreted as trust
Definitions

Attribute. a characteristic associated with an entity, such as an individual. Examples of persistent attributes
include height, eye color, and date of birth. Examples of temporary attributes include address, employer,
organizational role. A Social Security Number are an example of a long-lived attribute.
and
Some biometrics data
are persistent, some change over time or can be changed, (e.g., fingerprints and hair color).
Personal identifier.
persistent identifiers associated with an individual human
and the attributes that are
difficult or impossible to alter. For example, human date of birth, height, and genetic pattern.
Anonym (as in anonymous). An identifier associated with no personal identifier, but only with a single-use
attestation of an attribute. An anonymous identifier identifies an attribute, once. An anonymous identifier
used more than once becomes a pseudonym.
Pseudonym. An identifier associated with attributes or sets of transactions, but with no permanent identifier
More Definitions

Identification.
Association of a personal identifier with an individual presenting attributes. For example,
accepting the association between a physical person and claimed name; or determining an association with a
medical record and a patient using physical attributes.
Authentication. Proving as association between
an identifier or attribute. For example, the association of an
automobile with a license plate or a person with an account. The automobile is identified by the license plate,
it is authenticated as legitimate by the database of cars that are not being sought for enforcement purposes.
Identity Authentication. Proving as association between an entity and an identifier. For example, the
association of
a person with a credit or educational record.
Attribute Authentication. Proving as association between an entity and an attribute. For example, the
association of an painting with a certificate of authenticity.
association between entity and identifier
established.
This
is usually a two step process, where the
is established, and then the link to identifier and attribute is
Yielding

Authorization. A decision to allow a particular action based on an identifier or attribute. Examples include
the ability of a person to make claims on lines of credit; the right of an emergency vehicle to pass through a
red light; or a certification of a
Identity. That set of
radiation-hardened device to be attached to a satellite under construction.
permanent or long-lived temporal attributes associated with an entity
Download