Privacy for Pervasive Computing Slides based on http://www.cs.cmu.edu/~jasonh/courses/ubicomp-sp2007/ Why Care About Privacy? End-User Perspective • Protection from spam, identity theft, mugging • Discomfort over surveillance – Lack of trust in work environments – Might affect performance, mental health – May contribute to feeling of lack of control over life • Starting over Everyday Risks – Something stupid you did as a kid Friends, Family • Creativity _________________________________ Extreme Risks Government Stalkers, Muggers andEmployers freedom to experiment _________________________________ __________________________ _________________________________ – ProtectionOver-monitoring from total societies Over-protection Civil liberties Well-being Room for each person to develop individually Social – obligations Discrimination Personal safety Embarrassment Reputation • Lack of adoption of ubicomp tech The Fundamental Tension • Ubicomp envisions – lots of sensors for gathering data – rich world models describing people, places, things – pervasive networks for sharing • This data can be used for good and for bad Find Friends Smart Homes Smart Stores Why Care? Designer and App Developer Perspective • Most obvious problem with ubicomp by outsiders Why Care? Designer and App Developer Perspective • “Do I wear badges? No way. I am completely against wearing badges. I don't want management to know where I am. No. I think the people who made them should be taken out and shot... it is stupid to think that they should research badges because it is technologically interesting. They (badges) will be used to track me around. They will be used to track me around in my private life. They make me furious.” • Ubicomp “might lead directly to a future of safe, efficient, soulless, and merciless universal surveillance” – Rheingold What is Privacy? • No standard definition, many different perspectives • Different kinds of privacy – Bodily, Territorial, Communication, Information What is Information Privacy? • Many different philosophical views on info privacy – Different views -> different values -> different designs – Note that these are not necessarily mutually exclusive Principles vs. Common Interest • Principled view -> Privacy as a fundamental right – Embodied by constitutions, longstanding legal precedent – Government not given right to monitor people • Common interest -> Privacy wrt common good – Emphasizes positive, pragmatic effects for society • Examples: – National ID cards, mandatory HIV testing Self-determination vs. Personal Privacy • Self-determination (aka data protection) – Arose due to increasing number of databases in 1970s – “Privacy is the claim of individuals, groups or institutions to determine for themselves when, how, and to what extent information about them is communicated to others” (Westin) – Led to Fair Information Practices (more shortly) – More of individual with respect to government and orgs • Personal privacy – How I express myself to others and control access to myself – More of individual with respect to other individuals • Examples: – Cell phone communication, instant messaging, Facebook Privacy as Solitude • “The right to be let alone” • People tend to devise strategies “to restrict their own accessibility to others while simultaneously seeking to maximize their ability to reach people” – (Darrah et al 2001) • Example: – Spam protection, undesired social obligations • Ubicomp: – Able to turn system off, invisible mode Privacy as Anonymity • Hidden among a crowd • Example: – Web proxy to hide actual web traffic • Ubicomp: – Location anonymity – “a person” vs “Asian person” vs “Jason Hong” Other Views on Privacy • Transparent Society – Multi-way flow of info (vs one-way to govts or corporations) • Don’t care – I’ve got nothing to hide – We’ve always adapted – "You have zero privacy anyway. Get over it." • Fundamentalist – Don’t understand the tech – Don’t trust others to do the right thing • Pragmatist – Cost-benefit – Communitarian benefit to society as well as individual Why is Privacy Hard? • Hard to define until something bad happens – “Well, of course I didn’t mean to share that” • Risks not always obvious – Burglars went to airports to collect license plates – Credit info used by kidnappers in South America • Change in comfort with time and/or experience • Cause and effect may be far in time and space • Malleable depending on situation – Still use credit cards to buy online – Benefit outweighs cost Why is Privacy Hard? • Data getting easier to store – Think embarrassing facts from a long time ago (ex. big hair) • Hard to predict effect of disclosure – Hard to tell what credit card companies, Amazon are doing • Market incentives not aligned • Easy to misinterpret – Went to drug rehabilitation clinic, why? • Bad data can be hard to fix – Sen. Ted Kennedy on TSA watch list Fair Information Practices (FIPs) • US Privacy Act of 1974 (based on the work by Alan Westin) • Based on Self-determination / Data Protection view • Set of principles stating how organizations should handle personal information • Note: many variants of FIPs Fair Information Practices (FIPs) • Openness and transparency – No secret record keeping • Individual participation – Individual can see and correct the records • Collection limitation – Data collection should be proportional (to its purpose) • Data quality – Data should be relevant to their purpose (and up-to-date) • Use limitation – Data should be used for their specific purposed (by authorized) • Reasonable security – Security based on the sensitivity of the data collected • Accountability – Data keepers must be accountable for compliance with the other principles Adapting FIPs for Ubicomp • Presents a method for analyzing ubicomp systems – Assume designers trying to do “the right thing” • Versus evil people actively trying to intrude • Main areas of innovation and system design – – – – – – Notice Choice and consent Anonymity and Pseudonymity Proximity and locality Adequate security Access and Recourse Adapting FIPs for Ubicomp • Notice – Physical beacons beaming out P3P policies • P3P (Platform for Privacy Preferences) developed for Web access privacy notification and configuration – Personal system that logs policies • Issues – Overwhelmed by notifications? – Understandability of notifications? Adapting FIPs for Ubicomp • Choice and consent – Need a way to confirm that a person has consented – Can digitally sign a “contract” notification • Issues – – – – How can people specify their policies? Can policies match what people really want? How to make people aware of auto-accepts? What if people don’t have a real choice? Adapting FIPs for Ubicomp • Anonymity and Pseudonymity – Try to eliminate any trace of identity – Or have a disposable identifier not linked to actual identity • Issues – What kinds of services can be offered anonymously? – Business models for anonymous services? Adapting FIPs for Ubicomp • Proximity – Limit behavior of smart objects based on proximity • Ex. “Record voice only if owner nearby” – Simple mental model, could be hard to implement though – Weakness: could be easy to subvert • Locality – Information tied to places it was collected – Require physical proximity to query – Weakness: limits some utility (ex. Find friend) Adapting FIPs for Ubicomp • Access and Recourse – How to know what the system knows about you? – What mechanisms for recourse? • Use limitation, access, or repudiation, etc. • E.g., privacy aware data mining • Adequate security – Security solves privacy?? Not really – Ubicomp’s challenges (less capable devices) – Principle of proportionality? (what to make secure is proportional to its value) Unpacking “Privacy” for a Networked World Palen & Dourish HCI 2003 Slides based on http://www.cs.cmu.edu/~jasonh/courses/ubicomp-sp2007/ Overview • Palen & Dourish present model for privacy – Based on theory by social psychologist “Irwin Altman” (1975/1977) – Concept of privacy as dynamic, dialectic process • Privacy management as a balancing act – Multiple factors govern these social interactions – Case studies involving technology illustrate model Traditional approach in HCI • Design of pervasive and mobile systems – Interactions with systems impact personal privacy – New technologies introduce novel situations • Relationship between privacy and technology – Consider outside of static, rule-based systems – Draw on earlier concepts to better understand privacy for new situations involving information technology Privacy regulation theory • Altman sees privacy between individuals: – “As a dialectic process, privacy regulation is conditioned by our own expectations and experiences, and by those of others with whom we interact.” – “As a dynamic process, privacy is understood to be under continuous negotiation and management, with the boundary that distinguishes privacy and publicity refined according to circumstance.” (Palen and Dourish, 2003) Privacy management • Privacy as a social negotiation: – “Privacy management is a process of give and take between and among technical and social entities— from individuals to groups to institutions—in everpresent and natural tension with the simultaneous need for publicity. Our central concern is with how this process is conducted in the presence of information technology.” (Palen and Dourish, 2003) Altman’s model: limitations • For managing personal access in interactions: – Circumstance = f(local physical environment, audience, social status, task or objective, motivation and intention, information technology) • Information technology changed the view on disclosure (physical space vs. unknown distance), identity, time boundaries – IT changed the concept of “conventional” circumstance • Privacy outside physicality: when digital information flows outside physical and temporal constraints, it changes the way to regulate privacy. Disclosure boundary Participation in the social world requires selective disclosure of personal info Bumper stickers, letter to the editor, sitting in sidewalk cafes, walking down public streets, etc People seek to maintain a personal life and a public face Managing privacy is to pay attention to both of these desires Enter IT: deliberate vs. non-deliberate disclosure (e.g., online shopping vs. google search) Tension around privacy and publicity is influenced by identity and temporal concerns (next slides) Identity boundary Boundary between self and other Beyond the spatial extent of the body social phenomenon Affiliation and allegiance make “self” complicated In(ex)clusiveness amplified by “self” and “other” is continually enacted in and through one’s actions in the world Recipient design phenomenon: the way that one’s actions and utterances are designed with respect to specific others Different time/others: professionals, students, fellow bus riders, etc. Reflexive interpretability of action (one’s own ability to access how one’s action appears to others ) was driving privacy management in IT world Technologically mediated “interaction” is less effective Representation is impoverished; indicators of boundary between privacy and publicity are not clear Information persistence makes the problem complicated Temporality boundary Tension between past, present, and future Information disclosure could persist Active privacy control and management needs to be seen in the context of temporal sequence Relevance of permanence and impermanence constrain, undermine, or modify regulatory behavior “ The Disclosure (privacy vs. publicity), Identity (self vs. other) and Temporality (past - future) boundaries, and the tensions that occur with their negotiation, are the primary features of our framework. They demonstrate that privacy regulation is a dynamic, dialectic, negotiated affair. Technology itself does not directly support or interfere with personal privacy; rather it destabilizes the delicate and complex web of regulatory practices.” Genres of Disclosure Genres of disclosure: socially-constructed patterns of privacy management Disclosure, identity, temporal boundaries Socially-constructed genre has both structural properties of communication and social patterns of expectation/response Encounters between representational forms (people/action) and social practice (setting expectations around representations) Systems support different understanding Family Intercom (seamless comm using active badge at home): genre mismatch (home vs. work place); e.g., 6 year old kit vs. parents, 16 year old vs. sibling? Shared calendars: could reveal explicit patterning and sequencing of information (e.g., layoff example) – disclosure Active badges: researcher vs. administrative staff (identity) Cell-phones: boundary between self and other is destabilized (as the caller can’t control the receiver’s physical space) Instant messaging: Tension on the temporal boundary about messages that can be stored for future use (e.g., conventional IM + facebook, tweeter) Conclusion • When considering privacy concerns raised by the dev of new tech, the whole of the social and institutional setting (in which tech are deployed) must be considered • Need to pay attention to the historical continuity of practice (privacy regulation) • Privacy management = balancing act (tension resolution) people and/or their internal conflicts • Active process of privacy management with help of technology; need to be as responsible for what we make possible (enable) as for what we make real (use)