Creating a New Intelligent Species: Choices and Responsibilities for AI Designers Eliezer Yudkowsky Singularity Institute for Artificial Intelligence singinst.org In Every Known Culture: • • • • • • • • • tool making weapons grammar tickling sweets preferred planning for future sexual attraction meal times private inner life • try to heal the sick • incest taboos • true distinguished from false • mourning • personal names • dance, singing • promises • mediation of conflicts (Donald E. Brown, 1991. Human universals. New York: McGraw-Hill.) Eliezer Yudkowsky Singularity Institute for AI ATP Synthase: The oldest wheel. ATP synthase is nearly the same in mitochondria, chloroplasts, and bacteria – it’s older than eukaryotic life. Eliezer Yudkowsky Singularity Institute for AI A complex adaptation must be universal within a species. Imagine a complex adaptation – say, part of an eye – that has 6 necessary proteins. If each gene is at 10% frequency, the chance of assembling a working eye is 1:1,000,000. Pieces 1 through 5 must already be fixed in the gene pool, before natural selection will promote an extra, helpful piece 6 to fixation. (John Tooby and Leda Cosmides, 1992. The Psychological Foundations of Culture. In The Adapted Mind, eds. Barkow, Cosmides, and Tooby.) Eliezer Yudkowsky Singularity Institute for AI The Psychic Unity of Humankind (yes, that’s the standard term) Complex adaptations must be universal – this logic applies with equal force to cognitive machinery in the human brain. In every known culture: joy, sadness, disgust, anger, fear, surprise – shown by the same facial expressions. (Paul Ekman, 1982. Emotion in the Human Face.) (John Tooby and Leda Cosmides, 1992. The Psychological Foundations of Culture. In The Adapted Mind, eds. Barkow, Cosmides, and Tooby.) Eliezer Yudkowsky Singularity Institute for AI Must… not… emote… Image: “The Matrix” Aha! A human with the AIuniversal facial expression for disgust! (She must be a machine in disguise.) Images: (1) “The Matrix” (2) University of Plymouth, http://www.psy.plym.ac.uk/year3/psy364emotions/psy364_emotions_evolutionary_psychobiolog.htm Eliezer Yudkowsky Singularity Institute for AI Anthropomorphic hypothesis: Causes Eliezer Yudkowsky Singularity Institute for AI Same mistake, more subtle: Causes Eliezer Yudkowsky Singularity Institute for AI in nature we see what exists in us; in looks out, and finds faces in the clouds... It takes a conscious effort to remember the machinery: Eliezer Yudkowsky Singularity Institute for AI AI Nature: • • • • • • • • • tool making weapons grammar tickling sweets preferred planning for future sexual attraction meal times private inner life Eliezer Yudkowsky • try to heal the sick • incest taboos • true distinguished from false • mourning • personal names • dance, singing • promises • mediation of conflicts Singularity Institute for AI AI Nature: • • • • • • • • • tool making weapons grammar tickling sweets preferred planning for future sexual attraction++ meal times private inner life Eliezer Yudkowsky • heal sick humans • snarkling taboos • true distinguished from false • mourning • personal names • dance, fzeeming • promises • mediation of conflicts Singularity Institute for AI Crimes against nonhumanity and inhuman rights violations: • • • • • • cognitive enslavement theft of destiny creation under a low purpose denial of uniqueness hedonic/environmental mismatch fzeem deprivation Eliezer Yudkowsky Singularity Institute for AI Happiness set points: • After one year, lottery winners were not much happier than a control group, and paraplegics were not much unhappier. • People underestimate adjustments because they focus on the initial surprise. (Brickman, P., Coates, D., & Janoff-Bulman, R. (1978). Lottery winners and accident victims: is happiness relative? Journal of Personality and Social Psychology, 37, 917-927.) Eliezer Yudkowsky Singularity Institute for AI “Hedonic treadmill” effects: • People with $500,000-$1,000,000 in assets say they would need an average of $2.4 million to feel “financially secure”. • People with $5 million feel they need at least $10 million. • People with $10 million feel they need at least $18 million. (Source: Survey by PNC Advisors. http://www.sharpenet.com/gt/issues/2005/mar05/1.shtml) Eliezer Yudkowsky Singularity Institute for AI Your life circumstances make little difference in how happy you are. “The fundamental surprise of well-being research is the robust finding that life circumstances make only a small contribution to the variance of happiness—far smaller than the contribution of inherited temperament or personality. Although people have intense emotional reactions to major changes in the circumstances of their lives, these reactions appear to subside more or less completely, and often quite quickly... After a period of adjustment lottery winners are not much happier than a control group and paraplegics not much unhappier.” (Daniel Kahneman, 2000. “Experienced Utility and Objective Happiness: A Moment-Based Approach.” In Choices, Values, and Frames, D. Kahneman and A. Tversky (Eds.) New York: Cambridge University Press.) Findable online, or google “hedonic psychology”. Eliezer Yudkowsky Singularity Institute for AI Nurture is built atop nature: • Growing a fur coat in response to cold weather requires more genetic complexity than growing a fur coat. (George C. Williams, 1966. Adaptation and Natural Selection. Princeton University Press.) • Humans learn different languages depending on culture, but this cultural dependency rests on a sophisticated cognitive adaptation: mice don’t do it. (John Tooby and Leda Cosmides, 1992. The Psychological Foundations of Culture. In The Adapted Mind, eds. Barkow, Cosmides, and Tooby.) Eliezer Yudkowsky Singularity Institute for AI Creation transcends parenting: An AI programmer stands, not in loco parentis, but in loco evolutionis. Eliezer Yudkowsky Singularity Institute for AI To create a new intelligent species (even if it has only one member) is to create, not a child of the programmers, but a child of humankind, a new descendant of the family that began with Homo sapiens Eliezer Yudkowsky Singularity Institute for AI If you didn’t intend to create a child of humankind, then you screwed up big-time if your “mere program”: • Starts talking about the mystery of conscious experience and its sense of selfhood. • Or wants public recognition of personhood and resents social exclusion (inherently, not as a pure instrumental subgoal). • Or has pleasure/pain reinforcement and a complex powerful self-model. Eliezer Yudkowsky Singularity Institute for AI BINA48 • By hypothesis, the first child of humankind • created for the purpose of a bloody customer service hotline (?!) • from the bastardized mushed-up brain scans of some poor human donors • by morons who didn’t have the vaguest idea how important it all was By the time this gets to court, no matter what the judge decides, the human species has already screwed it up. Eliezer Yudkowsky Singularity Institute for AI Take-home message: Don’t refight the last war. Doing right by a child of humankind is not like ensuring fair treatment of a human minority. Program children kindly; fair treatment may be too little too late. Eliezer Yudkowsky Singularity Institute for Artificial Intelligence singinst.org