Can a machine be conscious? (How?) Depends what we mean by “machine? man-made devices? toasters? ovens? cars? computers? today’s robots? "Almost certainly not." Why “almost"? (1) “empirical risk.” (2) “other-minds” problem” Descartes on certainty and uncertainty: Certain: (i) mathematics, (ii) my-own-mind (the "cogito"), and Uncertain: everything else other-minds problem haunts robotics "Is any machine we have built to date conscious?" "Can a man-made artifact be conscious?" Genetic or other biological engineering? Toasters: man-made vs. tree-grown Machine = any causal physical system Includes biological organisms Hence we are conscious machines. So the right question is : “What kinds of machines can and cannot be conscious, and How?” (cognitive science) (And How can we tell? The "other-minds problem.”) "forward-engineering" vs. "reverse-engineering" causal explanation How? Explaining a cardiac system (a heart): F-eng: build a mechanism that can do what the heart can do R-eng: F-eng + explain (and/or build) the structure and the function of the biological heart itself: what the heart is (made out of) how it in particular happens to do what hearts can do. Explaining a conscious system (the brain): F-eng: build a mechanism that can do what the brain can do R-eng: F-eng + explain (and/or build) the structure and the function of the biological brain itself: what the brain is (made out of) how it in particular happens to do what brains can do. Either way, the causal explanation is a structural/functional one. The ghost of the other-minds problem: Note: Cardiac research program is completely unproblematic. The “cardiac vitalist” asks: "Can a machine be cardiac?" (“What kind of machine can and cannot be cardiac, and how?”) At no point would the cardiac vitalist have any basis for saying: "But how do we know that this machine is really cardiac?" There is no way left (other than ordinary empirical risk) for any difference even to be defined. Same would be true if our question had been "Can a machine be alive?" When we ask -- of the man-made, R-eng clone: "But how do we know that this machine is really alive?" Two structurally and functionally indistinguishable systems, one natural and the other man-made, their full causal mechanism known and understood: What does it mean to ask: "But what if one of them is really alive, whereas the other is not?" What property is at issue that one has and the other lacks, when all empirical properties have already been captured? Vitalism sounds like the other-minds problem -- and may be the other-minds problem: the animism at the true heart (soul) of vitalism The (nonanimist) vitalist who accepts that plants are not conscious would be in exactly the same untenable position if sceptical about the R-eng plant as the sceptic about the R-eng heart: What vital property is at issue (if it is not consciousness itself)? But the same is most definitely not true in the case of consciousness itself. Forward-engineering the brain: Build an F-eng robot that passes the Turing Test: It can do everything a real human can do, for a lifetime, indistinguishably from a real human (except appearance: we will return to that). Is it really conscious? It is indistinguishable from us in everything it can do. But conscious is something I am, not something I do. In particular, it is something I feel; indeed, it is the fact that I feel. The sceptic wants to say that F-eng is the wrong kind of machine, that it lacks something essential that we humans have. The robot does not feel, it merely behaves -- behaves exactly, indeed Turing-indistinguishably -as if it feels, but without actually feeling a thing. Whether or not a Turing robot feels is an ontic question rather than merely an epistemic question. Having conceded this point regarding certainty, however, only a fool argues with the Turing-Indistinguishable: Indistinguishable? If you prick us, do we not bleed? Perhaps the sceptic about the F-eng machine should hold out for R-eng machine, made out of the right stuff, Turing-indistinguishable both inside and out at both the macro and micro levels. But would we be right to kick robots that don’t bleed, because we infer that they therefore don’t feel? The Other-Minds Barrier TELEPATHY: mind-reading vs. turing-testing "theory of mind" or "mind-reading" in animals and children. (not philosophy or parapsychology but "other-mind perception”) “Mind-reading” is Turing-testing: inferring mental states from behavior. (Language (also a behavior) is the most powerful and direct means of mind-reading!) The only mind we can read other than by Turing-testing is our own! Don’t make ontic/epistemic conflation here: Turing-testing does not mean that all there is to mind is behavior (as the blinkered behaviorists thought)! It means the only way to read others' minds is through their behavior. Indistinguishable? What about R-eng. and rest of the neuromolecular facts about the brain? Which facts? "What kinds of machines can and cannot be conscious?" We know brains can be, but how? What are their relevant properties (if their weight, for example, is not)? If we pare down the properties of the brain to test which ones are and are not needed to be conscious: What will our test be? We are right back to Turing-testing again, the only non-telepathic methodology available to us, because of the other-minds problem. Will "correlations" do instead? Brain imaging to find the areas and activities that covary with conscious states, as the necessary and sufficient conditions of consciousness? But how did we identify those correlates? Because they were correlates of behavior: and of our own feelings i.e., by Turing-testing. What is our basis for favoring R-eng over F-eng then, if they are Turing-Indistinguishable (behaviorally) and the Turing Test is our only face-valid criterion. Conclusion: "What kinds of machines can be conscious (and how)?" The kinds that can pass the Turing Test, and by whatever means are necessary and sufficient to pass the Turing Test. Two consolations for residual worries about Zombies passing the Turing Test: (1) Darwin is no more capable of telepathy than we are. Evolution can’t distinguish sentients from zombies either. (2) the problem of explaining how (and why) we are not zombies (the “mind/body problem”) is “hard” (probably insoluble) Why the Mind/Body Problem is “hard”? Causality The Scylla of TELEKINESIS: Feeling as a Primal Force The Charybdis of EPIPHENOMENALISM: Feeling as an Inert Fact It would be easy if telekinetic powers existed: Then feelings would be physical forces like everything else. But there is no evidence that feelings are causal forces. Hence both F-eng and R-eng can only explain how it is that we can do things, not how it is that we can feel things. And that is why the ghost in the machine is destined to continue to haunt us even after all cognitive science’s empirical work is done. This paper (and linked references) is available at: http://www.ecs.soton.ac.uk/~harnad/Temp/machine.htm