Fluidly representing the world: Way, way harder than you think The One Thing to Remember a year from now Rules are not enough Statistics is not enough. Rules and statistics must DYNAMICALLY INTERACT to represent the world appropriately in computer models of intelligence. Let’s start things off with a question that you will never forget... What do cows drink? FIRST ANSWER Bottom-up, Statistical (e.g., Connectionism) conscious COW MILK DRINK Statistical, bottom-up representations unconscious What do cows drink? FIRST ANSWER Bottom-up, Statistical (e.g., Connectionism) conscious COW MILK DRINK These neurons are activated without ever have heard the word “milk” unconscious What do cows drink? RIGHT ANSWER Top-Down, Rule-Based (e.g., Symbolic AI) Facts about the world: ISA(cow, mammal) ISA(mammal, animal) Rule 1: IF animal(X) AND thirsty(X) THEN lack_water(X) Rule 2: IF lack_water(X) then drink_water(X) Conclusion: COWS DRINK WATER What do cows drink? Context 1: Milk Context 2: Water Bottom-up + Top-Down Rule-based representations conscious COW MILK DRINK unconscious Statistical, bottom-up representations Artificial Intelligence: When a computer will be able to represent the world in such a way that, when it is asked, “What do cows drink?” it will be able to answer either “milk” or “water,” depending on the context of the question. Rules are not enough Rules to characterize an “A” • Two oblique, approximately vertical lines, slanting inwards and coming together at the top of the figure • An approximately horizontal crossbar • An enclosed area above an open area • Vertical lines longer than the crossbar Characterize this word • • • • • • • • • • • It says “Gutenberg” It is written in a Gothic font. It looks like very old-fashioned script. Germans would have an easier time reading it than Italians. It has a very German feel to it. It is nine letters long. It contains the letter “e” twice, and the letter “u” once. It means “good mountain” in English It makes one think of the Vulgate Bible. The “t” is written in mirror-image. …….. • And can also be read perfectly upside-down!… Question: But is that really a property of “Gutenberg”? Answer: Yes, in some contexts. But does that mean that the representation of “Gutenberg” must always include: “Can be read upside-down when written in a VERY special way in pseudo-Gothic script”? Moral of the Story In the real world, a fixed set of rules, however long, is never sufficient to cover all possible contexts in which an object or a situation occurs. But the statistics of the environment are not enough, either. Without top-down conceptual knowledge we have no hope of understanding the following image Statistics is not enough Statistics is not enough “A dark spot. Hmmm…. Doesn’t look like anything.” Statistics is not enough “Pictures often have faces in them. Is that a face in the lower-left hand corner?” Statistics is not enough “Nah, doesn’t join up with anything else.” Statistics is not enough Statistics is not enough “Oh, THERE’s a face.” Statistics is not enough “But, again, it doesn’t make sense, just an isolated face…” Statistics is not enough “Let’s look at that dark spot again. A shadow?” Statistics is not enough “Hey, TREES produce shadows. Is there a tree around? THAT could be a tree!” Statistics is not enough “If that’s a tree, that could be a metal grating.” Statistics is not enough “But trees with metal gratings like that are on sidewalks. So where’s the kerb?” Statistics is not enough “If this is a kerb, it should go on straight. Does it? Yes, sort of. Now what’s this thing in the middle?” Statistics is not enough “That sort of looks like a dog’s head. That could make sense.” Statistics is not enough “But heads are attached to bodies, so there should be a body. Hey, that looks like it could be a front leg.” Statistics is not enough “The rest fits pretty well with that interpretation, too. Why so spotty? A dalmatian, yeah, sure, drinking from a road gutter under a tree.” Dynamically converging on the appropriate representation Context-dependent representation Consider representing an ordinary, everyday object N.B. The representation must allow a machine to understand the object in essentially the same way we do. The object: A cobblestone A “cobblestone” is like…. • • • • • • • • • • • a brick asphalt the hand-brake of a car a paper-weight a sharpening steel a ruler a weapon a brain a weight in balance pans a nutcracker Carrera marble • • • • • • • • a bottle opener a turnip in French an anchor for a (little) boat a tent peg a hammer etc…. a voter’s ballot... May 1968, Paris UNDER 21, this is your ballot. So, how could a machine ever produce context-dependent representations? Searching for an analogy Context - I was late in reviewing a paper. - The editor, a friend, said, “Please get it done now.” - I put the paper on my cluttered desk, in plain view, so that it simply COULD NOT ESCAPE MY ATTENTION…it would keep making me feel guilty till I got it done. - I wanted to find a nice way of saying I had devised a way so that the paper would continue to bug me till I had gotten it done Top-down: semantic, rule-based, conscious part of network “Something that won’t go away until you’ve taken care of it” No, too harsh to relate a paper to review to a hungry child No, too easy to just scratch it; papers are hard to review You can’t make them go away, ever Until you get up and swat it it NEVER stops buzzing! Swat mosquito Do the review dandelions a mosquito an itch hungry child Bottom-up: statistics-based “sub-cognitive”, unconscious par of network Representations of an object/situation must always be tailored to the context in which they are found. A machine must be able to do this automatically. Solar System Representation used in BACON (Langley et al., 1980) derived from Kepler’s data. ? Is the appropriate representation of this figure …. ….. just the juxtaposition of 64 of these? And if so, just how is the flickering that we humans see represented? Fluidly Representing the World: the Hard Problem of Artificial Intelligence Discovering how a system could develop dynamically evolving, contextsensitive representations is almost certainly The Hard Problem of artificial intelligence. Moreover, it always has been; it just hasn’t been recognized as such. We argue for a continual interaction between bottom-up and top-down processing, thereby allowing the system, on its own, to dynamically converge to a the context-dependent representations it needs. This ability is at the heart of our own intelligence and AI MUST face the challenge of learning how to fluidly represent the world if there is ever to be any chance of developing autonomous agents with even a glimmer of real human intelligence.