Restructure: PACK STUFF ABOUT DIFFERENT NOTIONS OF

advertisement
Intractability Arguments for Massive Modularity
Introduction
In recent years it has become commonplace to argue for a massively modular conception
of our mental architecture on the grounds that the alternative would be computationally
intractable. Such intractability arguments vary considerably in detail, but they all share the
following pair of commitments. First of all, they assume the classical computational theory of
mind1:
CTM: Human cognitive processes are classical computational ones –roughly,
algorithmically specifiable processes defined over syntactically structured mental
representations.
As has been commonly observed, however, the truth of CTM requires more than mere
computability. What it requires is that mental processes are, in some appropriate sense, tractably
computable; and it is on this point that intractability arguments seeks to undermine amodular
views of cognition. That is, they endorse the following Intractability Thesis:
IT: Non-modular cognitive mechanisms –in particular mechanisms for reasoning— are
computationally intractable.
Given these commitments, however, its appears to follow that:
MM: The mind –including those parts responsible for reasoning— is composed of
modular cognitive mechanisms.
And this is, of course, precisely what the massive modularity hypothesis requires.
As one might expect, the above argument is extremely popular among advocates of
massive modularity.2 Indeed, Peter Carruthers goes so far as to claim that it provides “the most
important argument in support of massive modularity" (Carruthers, this volume & forthcoming).
But an appreciation of the argument extends far more widely than this. Jerry Fodor, for example,
both rejects the claim that reasoning is modular and yet maintains that the “real appeal of
massive modularity is that, if it is true, we can either solve these [intractability] problems or at
1
Though sometimes only tacitly and sometimes only for the sake of argument.
2
See, for example, Buss (1999), Carruthers (forthcoming), Cosmides & Tooby (1987), Plotkin
(2003), Tooby & Cosmides (1992), Gigerenzer (2000), Sperber (1994).
1
least deny them center stage pro tem” (Fodor, 2000, p23). Indeed the influence of the argument
can be found in regions of cognitive science that do not even explicitly address issues about the
architecture of the human mind. Variants of the argument are, for example, familiar in robotics
and a-life (Brooks, 1999).
The central aim of this chapter is to assess the adequacy of intractability arguments for
massive modularity. In doing so, I assume for the sake of argument that CTM is true and focus
instead on the Intractability Thesis (IT). With so many advocates, one might be forgiven for
thinking that the arguments for IT are overwhelming. But what are the arguments; and what
might they reasonably be taken to show? I argue that when one explores this issue with
appropriate care and attention, it becomes clear that a commitment to IT is built on shaky
foundations; and this is because all the main arguments for it are deeply unsatisfactory. A
satisfactory argument would (minimally) need to conform to the following three conditions. First
and most obviously, it would need to show that amodular mechanisms are intractable. Second, it
would need to do so in such a way as to render an amodular conception of reasoning untenable.
As we shall see, however, this is no straightforward matter since contrary to appearances there
are plenty of ways for a mechanism to be intractable that in no way undermine the claim that it
constitutes part of our cognitive architecture. Finally, a satisfactory argument for IT needs to
avoid being so strong as to undermine the classical-cum-modular account of cognition as well.
Unless this condition is satisfied, the argument for IT clearly cannot be recruited as part of the
case for MM. What I propose to show is that none of the main extant arguments for IT satisfy all
these conditions and, moreover, that there are reasons for pessimism about the prospects of
providing such an argument any time soon. I conclude, therefore, that the significance of
intractability considerations for the contemporary debate over massive modularity has been
greatly overestimated.
Here’s a sketch of things to come. In section 1, I describe MM and the attendant notion of
a module in a bit more detail; and in section 2 I explain how modularity is supposed to help
resolve intractability worries. I then turn my attention to the arguments for IT. In section 3, I
consider and reject three ‘quick and dirty’ arguments that fail because they attribute (without
serious argument) commitments that the amodularist simply does not accept. In section 4, I
consider at some length a recent and influential argument by Jerry Fodor and show that it is both
unsound and too strong in the sense that, even if sound, it would undermine modular and
2
amodular theories alike. In section 5, I consider a rather different kind of argument: an inference
to IT from the pattern of failures in robotics. Finally, in section 6, I conclude by arguing that
there are reasons for pessimism about the prospects of providing a satisfactory argument for IT
any time soon.
1. Massive Modularity
A central theme of the forthcoming discussion is that the arguments for IT frequently turn
on –sometimes egregious— misunderstandings of MM and the attendant notion of a module.
We would do well, then, to get clear on these notions before turning to the arguments
themselves.
1.1 A Sketch
Massive modularity is a hypothesis --or more precisely, a broad class of hypotheses-which maintain that our minds are largely or perhaps even entirely composed of highly
specialized cognitive mechanisms or modules. Slightly more precisely, it can be divided into two
parts. The first concerns the shear number of modules that there are. According to advocates of
MM, there are a huge number (Tooby and Cosmides, 1995, p. xiii). But second, and more
importantly for our purposes, MM incorporates a thesis about which cognitive mechanisms are
modular. According to MM —and in contrast to an earlier, well-known modularity thesis
defended by Fodor and others— the modular structure of the mind is not restricted to peripheral
systems — that is, input systems (those responsible for perception, including language
perception) and output systems (those responsible for action and language production). Though
advocates of MM invariably endorse this thesis about peripheral systems, they also maintain,
pace Fodor, that central systems —paradigmatically, those responsible for reasoning— can ‘be
divided into domain-specific modules’ as well (Jackendoff, 1992, p.70). So, for example, it has
been suggested that there are modular mechanisms for such central processes as ‘theory of mind’
inference (Leslie, 1994; Baron-Cohen, 1995) social reasoning (Cosmides and Tooby, 1992) and
folk biological taxonomy (Attran, 199?).
Clearly, there are a variety of ways in which the above rough sketch might be elaborated;
and depending on how this is done, we end up with interestingly different versions of MM
(Samuels, 2000). One important distinction is between what I’ll call strong and weak MM. By
3
assumption, both are committed to the modularity of peripheral systems; but where strong MM
maintains that all central systems are modular, weak MM claims merely that many though not all
are. Both theses are thus more radical than the version of modularity defended by Fodor (1983)
in that they posit the existence of modular central systems. But whereas weak MM is entirely
compatible with the existence of distinctively –indeed radically-- non-modular reasoning
mechanisms, strong MM is committed to the denial of all such systems.
Which of these versions of MM is the intractability argument supposed to justify? This is
far clear. But since IT maintains that all amodular mechanisms are computationally intractable,
it’s natural to interpret the argument as trying to justify the strong version of MM. It is this
reading of the argument that I’ll be most concerned to rebut. Nevertheless, I suggest that extant
versions of the intractability argument fail even to support weak MM.
1.2. What is a Module?
Another important respect in which versions of MM may differ concerns how they
construe the notion of a module; and clearly I need to say something about this before assessing
the arguments for IT. There is good and bad news here. The bad news is that debate over how to
construe the notion of a module has been a point of considerable confusion in recent years and
has resulted in an enormous proliferation of distinct notions.3 The good news is, however, that
we need not survey these options here since we are concerned only with those properties of
modules that might plausibly contribute to the resolution of intractability worries. And it turns
out that the story about how modularity aids in the production of tractability is quite a standard
one. Indeed, of all the various characteristics that have been ascribed to modular mechanisms,4
there are really only two that get invoked –sometimes in tandem and sometimes individually—
in addressing intractability worries: domain-specificity and informational encapsulation. For
current purposes, then, I adopt the following minimal definition of a cognitive module: A
3
See Segal, 1996; Samuels, 2000 and Fodor, 2000 for discussions of the various notions of
modularity currently in play within cognitive science.
4
And the list of putative characteristics really is a rather long one! Usual suspects include:
domain specificity, informational encapsulation, task/functional specificity, autonomy,
innateness, cognitive impenetrability, limited access of external processes to the module’s
internal states, mandatory (or automatic) processing, relative speed, shallow outputs, fixed neural
architecture, susceptibility to characteristic breakdown patterns, characteristic patterns in
ontogeny, products of natural selection and universality.
4
cognitive mechanism is modular just in case it is domain-specific, informationally encapsulated
or both.
1.3 Domain-Specificity and Informational Encapsulation
What do cognitive scientists mean by ‘domain specificity” and “informational
encapsulation”? Both notions get used in a variety of ways and, moreover, both are vague in the
sense that they admit of degree. But when applied to cognitive mechanisms, they are almost
invariably intended to indicate architectural constraints on information flow that (partially)
specify what the mechanism can and cannot do.5
Very roughly, a mechanism is domain specific if it can only take as input a highly
restricted range of representations. To put it rather crudely: If we think of cognitive mechanisms
as ‘black boxes’ into which representations sometimes enter and from which they periodically
depart, then a mechanism is domain-specific if there are heavy restrictions on the class of
representations that are permitted to enter. Consider, for example, a natural language parser.
Such a device can only process inputs that represent phonological properties and, moreover,
represents them in the appropriate format. In contrast, it cannot take as input representations of
color or artifacts or mathematical entities and so on. Thus it is plausible to claim that natural
language parsers are, in the present sense, domain-specific. Much the same is true of many of the
modular central systems that have been posited in recent years. A folk biology module is, for
example, naturally construed as domain-specific since it can only take representations of
biological phenomena as input. In contrast, the kind of central systems for reasoning posited by
Fodor and others are not plausibly domain-specific in any interesting sense since they are
supposed to be able to take an extraordinarily broad range of representations as input –roughly
any conceptual representation whatsoever. Such mechanisms are thus invariably construed as
domain-general as opposed to domain-specific.
What does ‘architectural’ means in the present context? This is no straightforward matter.
(What is?) But for our purposes we will not go far wrong if we assume that the claim that
something is an architectural constraint only if a) it is a relatively enduring feature of the human
mind; b) it is not a mere product of performance factors such as limitations on time, energy etc.
and c) it is cognitively impenetrable in roughly Pylyshyn’s sense. That is: it cannot be changed
solely as a result of alterations in ones beliefs, goals, intentions and other representational states.
5
5
Let’s now turn to the notion of informational encapsulation. According to one standard
definition, a cognitive mechanism is informationally encapsulated just in case it has access to
less than all the information available to the mind as a whole (Fodor, 1983). Taken literally this
would be a very uninteresting notion since it is overwhelmingly plausible that no single
mechanism has access to all the information available to the mind as a whole. All cognitive
mechanisms would thus be encapsulated on the present definition. In practice, however, this
matters little since informational encapsulation is treated as a property that admits of degree; and
what researchers who deploy the notion are really interested in the extent to which a
mechanism’s access to information it interestingly constrained by its architecture. Though there
are various ways in which a mechanism might be so constrained, perhaps the most common
suggestion is that encapsulated mechanisms only have access to the information contained within
a restricted, proprietary database. Suppose, for example, that we possess a folk biology
mechanism that can only utilize the information contained within a restricted database. Such a
mechanism would be unable to access large parts of our general encyclopedic knowledge even if
this knowledge were relevant to the task being performed by the module. Such a mechanism
would thus be informationally encapsulated. Contrast this with a cognitive mechanism that has
access to all the information stored in a person’s memory. Suppose, for example, that we possess
a reasoning mechanism of the kind proposed by Fodor (1983) which can access pretty much all
of one’s encyclopedic knowledge. Such a device would presumably be highly unencapsulated.
Though regularly confounded, it really is important to see that domain-specificity and
informational encapsulation are genuinely different properties of a cognitive mechanism. In a
sense, both concern the access that a mechanism has to representations. (This is I suspect the
source of the confusion.) Yet the kind of access that they involve is very different. Domainspecificity (and generality) concern what we might call input access. They concern the
representations that a mechanism can take as input and process. For this reason, it is not
uncommon to speak of representations that fall within a mechanism's domain as the ones that
'trigger' it or 'turn it on'. More precisely, on the assumption that cognitive mechanisms are
computational --hence characterizable by the function they compute-- the domain of a
mechanism is also, in the technical sense, the domain of the function computed by the
mechanism. In other words, it is the set of representations that the mechanism will map onto
some element in the range of its function. In contrast, the informational (un)encapsulation of a
6
mechanism does not concern input access but what we might call resource access. That is,
encapsulation does not concern the class of representations that can 'turn on' a mechanism --the
ones for which it can compute a solution-- but the ones it can use as a resource once it has been
so activated.
2. Modularity and (In)tractability
How are the domain-specificity and encapsulation of modules supposed to help resolve
intractability worries? In this section I address this issue. But first, I need to say something about
the notion of intractability itself.
2.1. Computational (In)tractability6
Very roughly, a process, task, algorithm or problem is computationally tractable (or,
equivalently, easy or feasible) if it can be performed in a reasonable amount of time using a
reasonable amount of salient resources, such as effort, memory space and information.
Conversely, something is computationally intractable (or hard or unfeasible) if it is not tractable.
Presumably, a satisfactory argument for IT needs to show that amodular reasoning mechanisms
are intractable in this way. There are, however, two importantly different sub-notions of
intractability that we would do well to distinguish: in-principle and in-practice intractability
(Millgram, 1991).
A problem is intractable-in-principal if (very roughly) it cannot be rendered tractable by
the mere addition of new resources. If, for example, a computer the size of the Universe with as
many circuits as there are elementary particles running since the Big Bang won't solve the
problem, then it is intractable-in-principal (Millgram, 1991). This is the notion of intractability
that tends to concern computational complexity theorists. In order to render it amenable to
formal treatment, however, they tend (at least for heuristic purposes) to draw the
feasible/unfeasible distinction in terms of which algorithms or tasks are polynomial and which
super-polynomial. Polynomial algorithms are ones where resource requirements are only
polynomial in the size of the input.7 In contrast, superpolynomial algorithms are those where
6
My approach to characterizing the notion of intractability is heavily indebted to Millgram
(1991)'s excellent discussion of the issue.
7
.That is, the resources required to compute a solution to some input can be expressed as a
polynomial (or better) function of input size --e.g. n2 or n3000000.
7
resource requirements increase exponentially (or worse) as a function of input size and can thus
only be expressed as superpolynomial functions, such as 2n or 100n. It is this latter kind of
algorithm that complexity theorists label "intractable". Such algorithms are also said to be
combinatorially explosive because of the dramatic way in which the resources required to solve a
problem increases as the input gets larger.
In contrast to tasks that are unfeasible in-principle, a problem is intractable in-practice if
the addition of new resources –more processing speed, memory, energy and so on— would
resolve the problem. There are many sorts of in-practice intractability. For instance, one kind
relates to space requirements, such as the amount of memory storage that a system has. But the
most commonly discussed sort of in-practice intractability concerns the idea of real-time
performance –roughly, the capacity of a system to undergo state changes as fast as (or almost as
fast as) some salient feature of the world. So, for example, one might say that a visual system
operates in real-time when the representation of the distal environment that the system produces
changes at the same rate –or close to the same rate— as the states of affairs that the system
purports to represent.
2.2. Massive Modularity as a Solution to Intractability Worries
How is modularity supposed to resolve intractability worries? As mentioned earlier, it is
their domain-specificity and/or informational encapsulation that's supposed to do the work. In
particular, both characteristics are supposed to contribute to solving what is sometimes called the
Frame Problem,8 though is perhaps more accurately (and less contentiously) referred to as the
Problem of Relevance. Very roughly, this is the problem of restricting the options and items of
information that need to be considered to those that are relevant to the task at hand.
Domain-specificity is supposed to engender tractability because it imposes restrictions
on the range of representations that a mechanism can process. As a consequence, it becomes
possible to ‘build into’ the device specialized information about the domain in which it operates
either in the form of ‘domain-specific rules of relevance, procedural knowledge, or privileged
hypotheses’ (Cosmides & Tooby, 1994, p.94). This, in turn, permits the mechanism to ‘ignore’
options that are not relevant to the domain in which it operates. For this reason, domain8
Fodor, Dennett, Glymour, Haughland and others all appear to think that the Frame Problem just is what, following
Eric Lorman, I call the Problem of Relevance. There is a strong case to be made, however, for the claim that this
identification is incorrect. See Lormand (199?) for an excellent discussion.
8
specificity has seemed to many like a plausible candidate for reducing the threat of combinatorial
explosion without compromising the reliability of cognitive mechanisms (Cosmides & Tooby,
1992; Sperber, 1994).
Encapsulation is supposed to work in a complementary fashion. First, since an
encapsulated device can only access information contained within its proprietary and restricted
database, the range of information it can search is highly constrained. In particular, the option of
searching through the entire system of beliefs –relevant or otherwise—simply does not arise. By
virtue of its architecture, the mechanism is simply incapable of doing so. Second, by limiting the
mechanism’s access to only those representations contained in a proprietary database, the
number of relations between items of information that it can compute is severely reduced as well
(Fodor, 1985). To take a simple example, suppose that a given process requires a pairwise
comparison of every item of information contained in a database. This would require
approximately 2n comparisons and so, by the standard measure, would be unfeasible in-principle.
Moreover, if the database to which the mechanism has access is large, then the task would be
utterly impractical. Suppose, for example, that there were 1000 elements in the database, then the
number of pairwise comparisons is a 302-digit number –considerably larger than the number of
protons in the known universe! (Harel, p.156). In contrast, if the mechanism is sufficiently
encapsulated, it might be able to perform the pairwise comparison even though the task is
combinatorially explosive. In effect, then, informational encapsulation provides a way to render
a task tractable in-practice even when the task in unfeasible in-principle.
Suppose that all the above is correct –that domain-specificity and encapsulation can
engender computational feasibility. Then it should be clear why MM is supposed to address the
threat that intractability poses for CTM. What it does is ensure that cognitive mechanisms are
architecturally constrained with respect to what options and items of information they can
consider. But although I accept that is one way of engendering feasible computation, I deny that
intractability provides us with good reason to reject an amodular conception of central systems.
Specifically, I deny that there are any good arguments for IT. It is to these arguments that I now
turn.
3. Three ‘Quick and Dirty’ Arguments for IT
9
Let me start by considering three arguments for IT that are ‘quick and dirty’ in that they
largely turn on attributing (without serious argument) commitments that the amodularist simply
does not accept.
3.1. Argument 1: Informational Impoverishment
The first, and perhaps the most prominent argument that I’ll discuss is one made popular
by the evolutionary psychologists Leda Cosmides and John Tooby (Tooby and Cosmides,1994).
The argument in question proceeds from the assumption that a nonmodular, hence, domaingeneral, mechanism “lacks any content, either in the form of domain-specific knowledge or
domain-specific procedures that can guide it towards the solution of problems” (Tooby &
Cosmides, 1994, p.94). As a consequence, it “must evaluate all the alternatives it can define"
(ibid.). But as Cosmides and Tooby observe, such a strategy is subject to serious intractability
problems since even routine cognitive tasks are such that the space of alternative options tends to
increase exponentially. Amodular mechanisms would thus seem to be computationally
intractable: at best, intolerably slow and, at worst, incapable of solving the vast majority of
problems that they confront.
Though frequently presented as an objection to amodular accounts of cognition (Tooby &
Cosmides, 1992; Buss, 2000), this argument is in fact only a criticism of theories which
characterize our cognitive mechanisms as suffering from a particularly extreme form of
informational impoverished. Any appearance to the contrary derives from the stipulation that
domain-general mechanisms possess no specialised knowledge. But while modularity is one way
of building (specialized) knowledge into a system, it is not the only way. Another is for amodular
devices to have access to bodies of specialized knowledge. Indeed, perhaps the standard view
among advocates of amodular theories is that reasoning mechanisms have access to huge
amounts of such information. This is, I take it, pretty obvious from even the most cursory survey
of the relevant literatures. Fodor (1983), for example, maintains explicitly that domain-general
central systems have access to enormous amounts of specialized information. So for that matter
do Anderson, Gopnik, Newell and many other theorists who adopt a nonmodular conception of
central systems (Anderson, 1993; Gopnik & Meltzoff 199?; Newell, 1990). In each of these
cases, what makes the hypothesized system domain-general is not an absence of specialized
10
information, but the enormous range of representations that the mechanism is capable of
processing. The argument currently under discussion thus succeeds in only refuting a straw man.
3.2. Argument 2: Optimality
Another possible argument against the tractability of amodular reasoning mechanisms
turns on the claim that that they implement optimization processes. In the present context,
“optimization’ refers to reasoning processes that broadly conform to standards of ideal
rationality, such as those characterized by Bayesian accounts of probabilistic reasoning or
standard approaches to decision theory. Such processes are widely recognized as being
computationally very expensive –as requiring memory, speed and informational access that
human beings could not possibility possess.9 And it is precisely because of these excessive
resource demands that they are commonly termed ideal, unbounded or even demonic
conceptions of reasoning (Simon 1957; Gigerenzer, 2001). Thus it would seem that if advocates
of nonmodular reasoning are committed to optimization, then the view that they endorse is
subject to intractability worries as well.
It is not at all clear to me that anyone explicitly endorses the above line of reasoning. But
it is strongly suggested by recent discussions of nonmodular reasoning architectures. Dietrich
and Fields (1996), for example, maintain that Fodor’s amodular conception of central systems
“tries to explain human intelligence as approximating ideal rationality” (p.23). Similarly,
Gigerenzer and his collaborators have a tendency to present their adaptive toolbox version of
MM as if it were a response to the intractability problems confronting a domain-general,
optimization view of reasoning (Gigerenzer, 2001; Gigerenzer & Todd, 1999).
9
To use one well-known example, on standard Bayesian accounts, the equations for assessing
the impact of new evidence on our current beliefs are such that if one's system of beliefs has n
elements, then computing the new probability of a single belief, B, will require 2n additions. Such
methods thus involve an exponential growth in number of computations as a function of belief
system size. To give you some idea of just how expensive this is, on the hyper-conservative
assumption that we possess 100 beliefs, calculating the probability assignment of a belief B on
the basis of new information, will require the performance of more than 1030 additions, which is
considerably more than the number of microseconds that have elapsed since the Big Bang!
11
The argument is not, however, a good one. Though optimal reasoning is (at least in the
general case) intractable,10 it really needs to be stressed that amodularists are in no way
committed to such a view of human reasoning. What is true is that for a mechanism to optimize
it needs to be unencapsulated hence, amodular; and this is because (at least as standardly
construed) optimization demands the updating of all of one’s beliefs in the light of new
information. But the converse is not true: An unencapsulated mechanism need not be an
optimizer. On the contrary, since the inception of AI it has been commonplace to combine an
amodular conception of reasoning with the explicit denial of optimization. Consider, for
example, Newell and Simon’s seminal work on the General Problem Solver. As the name
suggests, GPS was designed to apply across a very wide range of content domains without
architectural constraint on what representations is could use. It is thus not plausibly viewed as
modular. Yet, to use Simon’s famous expression, it was designed to satisfice –to arrive at
solutions that were good enough as opposed to optimal. Much the same could be said of many of
the amodular, classical accounts of reasoning to be found in AI and cognitive science, including
Laird and Newell’s SOAR architecture and Anderson’s ACT theory (Newell, 1990; Anderson,
1993). These are among the paradigm nonmodular approaches to cognition and yet they are in no
way committed to optimization.
3.3. Argument 3: Exhaustive Search
Still, even if optimization as such is not a problem for amodular accounts of reasoning, it
might still be that there are properties of optimal reasoning to which the amodularist is
committed and that these properties are sufficient to generate intractability problems. Exhaustive
search is perhaps the most plausible candidate for this role. The rough idea is that amodular
reasoning mechanisms must perform exhaustive searches over our belief systems. But given even
a conservative estimate of the size of any individual’s belief system, such a search would be
10
Though there is lots of good research which aims to discover tractable methods for applying
ideal standards of rationality to interesting –but restricted—domains. See, for example, the
literature on Bayesian networks (?????)
12
unfeasible in practice.11 In which case, it would seem that amodular reasoning mechanisms are
computationally intractable.
Again, it’s not at all clear to me that anyone really endorses this line of argument.
Nevertheless, some prominent theorists have found it hard not to interpret the amodularist as
somehow committed to exhaustive search. Consider, for example, the following passage in
which Clark Glymour discusses Fodor’s conception of central systems:
Is Fodor claiming that when we set about to get evidence pertinent to a hypothesis we are
entertaining we somehow consider every possible domain we could observe? It sounds
very much as though he is saying that, but of course it is not true. (Glymour, 1985, p. 15)
Glymour is surely correct to claim that no such exhaustive search occurs; and though he is not
wholly explicit on the matter, part of his reason for saying so appears to be that such a process
would be computationally unfeasible. Carruthers (forthcoming) is more explicit:
Any processor that had to access the full set of the agent’s background beliefs (or even a
significant subset thereof) would be faced with an unmanageable combinatorial
explosion. We should, therefore, expect the mind to consist of a set of processing systems
which are isolable from one another, and which operate in isolation from most of the
information that is available elsewhere in the mind (Carruthers, forthcoming)
This really does sound like an argument for IT. Moreover, given the reference to what a
processor “had to access” –rather than merely could access— it really does sound as if the
argument assumes that amodular mechanisms engage is (near) exhaustive search. Interpretative
issues to one side, however, the argument as it stands is not a good one.
Once more, the problem is that it’s very hard to see why the amodularist should accept
the claim that central systems engage in exhaustive search. What the amodularist does accept is
the unencapsulation of reasoning mechanisms which, by definition, have access to huge amounts
of information –we may suppose, all the agent’s background beliefs. But the notion of access is a
modal one. It concerns what information –given architectural constraints— a mechanism can
mobilize in solving a problem. In particular, it implies that any background belief can be used.
But it does not follow from this that the mechanism in fact mobilizes the entire set of background
11
Though not necessarily in-principle. Exhaustive search might only involve a worst-case
running time which is on the order of N—i.e. where time grows linearly with the number of
beliefs in the belief system.
13
beliefs –i.e. that it engage in exhaustive search. This is simply not a commitment that the
advocate of amodular central systems would be much inclined to accept. Indeed, as Fodor has
pointed out, it would be absurd to hold an amodular view of reasoning if it implied such a
commitment (Fodor, 1985).
Of course, the fact that the amodularist does not endorse the claim that central systems
engage is exhaustive search is perfectly consistent with their being an argument which shows
that such processes would need to occur if the amodular theory were true. In the coming section,
I consider a prominent argument that has been widely interepreted as supporting this conclusion..
4. The Globality Argument
In the Minds Doesn’t Work That Way, Fodor develops an argument which is supposed to
show the inadequacy of classicism as an account of our cognitive processes. Nevertheless, it has
been widely viewed by advocates of MM as a way of defending the Intractability Thesis
(Carruthers, forthcoming; Sperber, 2002). In what follows, I show both that Fodor’s argument is
unsound and that, even if sound, it is too strong to figure in the intractability argument for MM.
4.1. The Argument
Fodor’s argument is a complex one whose structure is far from transparent. But in rough
form, it concerns a tension between two prima facie plausible claims. The first is that classical
computational processes are in some sense (soon to be discussed) local. The second is that our
reasoning processes are global in roughly the sense that they are sensitive to context-dependent
properties of the entire belief system. Slightly more specifically, Fodor claims that abductive
reasoning (or inference to the best explanation) is global because it is sensitive to such properties
as simplicity and conservativism; properties which, he maintains, are both context dependent and
somehow determined by the belief system as a whole. The problem, then, is this: If classical
computational operations are local, how can we provide a computationally tractable account of
abduction? Fodor's claim is that we almost certainly cannot and that this shows the inadequacy of
classicism as an account of human reasoning.
Let’s spell out the argument in a bit more detail. One central feature is that it proceeds
from an analysis of how best to understand the classical thesis that cognitive processes are
syntactically driven. The rough idea is a very familiar one: cognitive processes are
14
representational processes in which the representations have their causal role in virtue of their
syntactic properties. But as Fodor observes, there is an ambiguity here between what we might
call an essentialist (or extreme) thesis and a more moderate thesis. According to the essentialist
reading, which Fodor dubs E(CTM): Each mental representation, R, has its causal role solely in
virtue of its essential syntactic properties –i.e. its constituent structure. To put the point in a
slightly different way, on this view, R’s causal role in cognitive processes supervenes12 on R’s
essential syntactic properties. In contrast, according to the more moderate version of classicism,
which Fodor refers to as M(CTM), the causal role of a representation, R, need not depend on its
essential (constituent) syntactic properties. Rather, it need only depend on some set of syntactic
properties. That is, R’s causal role in cognitive processes supervenes on some syntactic facts or
other. Either way, however, Fodor thinks that the classical conception of cognitive processes is
in serious trouble.
First, suppose that E(CTM) is true. Then the syntactic properties that determine R’s
causal role are the essential ones. In which case, the determinants of its casual role must be
context invariant. But this, according to Fodor, is simply not true of representations involved in
abductive reasoning since their causal role is determined, at least in part, by such properties as
simplicity and conservativism; properties which he maintains are context sensitive par
excellence. When deciding between a range of hypotheses one’s selection will be sensitive to
such issues as which of the hypotheses is the simplest or most conservative. And the degree to
which a hypothesis is simple or conservative depends in large measure on what background
beliefs one has –that is, on the relationship between the hypothesis and the theory or epistemic
context in which it is embedded. But facts about R’s relations to other beliefs are not essential to
R’s being the representation that it is. In which case, E(CTM) must false.
As Fodor notes, the above argument does not undermine M(CTM): the thesis that a
mental representation’s casual role is determined by some syntactic facts or other. Even so, he
maintains that this should be of little solace to classical cognitive scientists. First, according to
Fodor, M(CTM) is not in fact a classical view at all since “by definition, which Classical
computations apply to a representation is determined not just by some of its syntactic properties
or other but, in particular, by its constituent structure, that is by how the representation is
12
Terminological note: To say that causal roles supervene on constituent structure means (very
roughly) that representations cannot differ in their casual roles unless they differ in their
15
constructed from its parts”[p.30]. In short: according to Fodor the classicist is committed to
E(CTM).
Second, Fodor claims that M(CTM) only avoids the above objection to E(CTM) “at the
price of ruinous holism”: of “assuming that the units of thought are much bigger than in fact they
could possibly be” [Fodor, p.33]. Indeed, Fodor appears to think that if M(CTM) is true, then the
units of thought will be total theories –entire corpuses of epistemic commitments [p.33]. Though
the argument for this claim is far from transparent, the following strikes me as a charitable
reconstruction. First, suppose, that the conclusions of Fodor’s previous arguments are in force.
That is:
1. E(CTM) is false because the causal role of a representation in abductive reasoning is at
least partially determined by its relation to an embedding theory T (i.e. background
beliefs).
2. The classical computations that apply to a representation are determined solely by its
essential (i.e. constituent) structure. Very crudely: classical processes can only ‘see’ the
internal syntactic structure of the representations that are being processed.
Now as Fodor points out, M(CTM) is compatible with (1) since the relation between a
representation and its embedding theory might be a relational syntactic property S (or at least
depend on S). Nonetheless, if (2) is true, then a classical computational system is only sensitive
to the essential –i.e. constituent-- syntactic properties of a representation and not to relational
properties like S. In which case, a classical system cannot be influenced by S merely as a
consequence of having access to R. Nevertheless, as Fodor points out, it is straightforward to
transform R so that S can have a causal influence: The mechanism might simply rewrite R as the
conjunction of R and whatever parts of T are relevant to determining S. In the worst case, the
shortest expression to which the computer needs access in order to be sensitive to S is the entire
theory –i.e. T including R.
All the above is compatible with classicism. But according to Fodor a serious dilemma
looms large. On one hand the representations over which cognitive processes are typically
defined are much shorter than whole theories [p.31]. But on the other hand, Fodor maintains that
“the only guaranteed way of Classically computing a syntactic-but-global property” is to take
constituent structure as well.
16
‘whole theories as computational domains”; and this clearly threatens to render abduction
computationally intractable. The problem, then, is this:
Reliable abduction may require, in the limit, that the whole background of epistemic
commitments be somehow brought to bear on planning and belief fixation. But feasible
abduction requires in practice that not more than a small subset of even the relevant
background beliefs are actually consulted. [p.37]
Thus it would seem that if Classicism is true, abduction cannot be reliable. But since abduction
presumably is reliable, Classicism is false.
4.2 Problems with the Globality Argument
Though I accept Fodor’s objection to E(CTM), I have three worries about his case against
the more moderate M(CTM). A first and relatively minor concern is that, contrary to what Fodor
appears to think, the classicist in not committed to E(CTM). The reason is that there are two
importantly different versions of M(CTM); and the classicist need only deny one of them.
According to the first version:
M(CTM)1: Though R’s causal role is determined by some syntactic fact or other,
essential syntactic properties need make no determinative contribution
whatsoever.
Classicists are committed to the denial of M(CTM)1 because of the conditions that need to be
satisfied in order for different representations to have distinct casual roles within a classical
system. In brief, a representation, R, can have a causal role that is distinct from those of other
representations only if the mechanism can distinguish it from other representations. But a
classical, Turing-style computational device can distinguish R from R* only if they differ in their
essential syntactic/formal properties. So, a necessary condition on R and R* possessing different
causal roles is that they differ in their essential syntactic properties. In which case, on the
assumption that mental representations do differ in their causal roles, it follows that the causal
role of any representation is at least partially determined by its essential --i.e. constituent—
syntactic properties.
All this is, however, entirely compatible with an alternative version of M(CTM):
M(CTM)2: Though R’s causal role is partially determined by its essential
syntactic properties, other syntactic facts may be partially determinative as well.
17
Since this thesis requires that the causal role of R be partially determined by its essential
syntactic properties it does not conflict with the fact that classical systems distinguish between
representations in virtue of their essential syntactic properties. But M(CTM)2 also permits that
R’s causal role partially depend on inessential or relational syntactic properties. And this is just
as well since many paradigmatic classical systems conform to M(CTM)2 not E(CTM). For
example, programmable Turing machines are organized in such a way that R's causal role is
jointly determined by its essential syntactic properties and the program. (Change the program
and you change R’s causal role as well.) But the program just is a series of representations that
influence the causal role of R in virtue of syntactic properties. So, M(CTM)2 and not E(CTM) is
true of a programmable Turing machine.13 Though I won’t go into the issue here, much the same
is true of pretty much every system from the classical AI literature.
A second problem with Fodor’s argument against M(CTM) is that, in order to go through,
it is not enough that abduction is context sensitive, it must be global as well. Without this
assumption, no intractability worries arise since there is no reason to suppose that abductive
reasoning within a classical system would require (near) exhaustive search of background
beliefs. The problem, however, is that there is no reason whatsoever to accept the claim that
abduction is global in the relevant sense. Recall: on Fodor’s view, abduction is a global process
because it depends on such properties as simplicity and conservativism which, on his view, are
somehow determined by the belief system as a whole. But claims about the globality of these
properties may be given either a normative or a descriptive-psychological reading. On the
normative reading, assessments of simplicity and conservativism ought to be global: that is, a
normatively correct (or at any rate ideal) assessment of the simplicity or conservativism of a
hypothesis ought to take into consideration one’s entire background epistemic commitments. But
of course it is not enough for Fodor’s purposes that such assessments ought to be global. Rather,
13
Nor is the present point especially about programmes as opposed to (other) data-structures. For
the sake of illustration, consider a system that incorporates a semantic network (Quinlan). Such
systems are standard fair in classical AI. Nonetheless, what inferences can be drawn from a
representation within such a system depend not only on its essential syntactic properties of but
also on the arcs that link it to other representations –which are, by assumption, non-essential
syntactic relations.
18
it needs to be the case that our assessments of simplicity and conservatvism are, in fact, global.
And to my knowledge, there is no reason to suppose that this is true.
A comparison with the notion of consistency may help make the point clearer.
Consistency is frequently construed as a normative standard against which to assess our beliefs
(Dennett, 1987). Roughly: all else being equal, our beliefs ought to be consistent with each other.
When construed in this manner, however, it is natural to think that consistency should be a
global property in the sense that a belief system in its entirety ought to be consistent. But there is
absolutely no reason whatsoever to suppose –and indeed some reason to deny-- that human
beings conform to this norm. Moreover, this is so in spite of the fact that consistency really does
play a role in our inferential practices. What I am suggesting is that much the same may be true
of simplicity and conservativism. When construed in a normative manner, it is natural14 to think
of them as global properties –i.e. that assessments of simplicity and conservativism ought to be
made in the light of our total epistemic commitments. But when construed as properties of actual
human reasoning processes, there is little reason to suppose that they accord with this normative
characterization.
A final problem with Fodor’s argument is that even if we suppose that simplicity and
conservativism are, in fact, global properties, the argument still does not go through since it turns
on the implausible assumption that we are guaranteed to make successful assessments of
simplicity and conservativism. Specifically, in arguing for the conclusion that abduction is
computationally unfeasible, Fodor relies on the claim that “the only guaranteed way of
Classically computing a syntactic-but-global property” is to take ‘whole theories as
computational domains” (Fodor, 2000, p.?). But guarantees are beside the point. Why suppose
that we always successfully compute the syntactic-but-global properties on which abduction
relies? Presumably we do not. And one very plausible suggestion is that we fail to do so when
the cognitive demands required are just too great. In particular, for all that is known, we may
well fail under precisely those circumstances the classical view would predict –viz. when too
much of our belief system needs to be consulted in order to compute the simplicity or
conservativism of a belief.
4.3. Does the Globality Argument Support MM?
14
Though by no means mandatory.
19
So far we have seen that Fodor’s argument suffers from a number of deficiencies. But
even if it were sound, the following issue would still arise: Why would anyone think that it
supports MM? After all, if sound, the globality argument appears to show that classical theories
of abduction tout court –not merely amodular ones— are unfeasible. The short answer, I suspect,
is that MM appears to render reasoning sufficiently local to avoid the sorts of feasibility problem
that Fodor raises (Fodor, 2000, p23). But this impression is misleading. Even if sound, the
Fodorian argument would still fail to support MM. In what follows, I argue for this point by
showing how one central premise of the argument— that abduction is a global process— cannot
be squared with both MM and CTM. Hence, there is no stable position which permits the
advocate of a classical-cum-MM account of cognition to claim that Fodor has provided a sound
argument for the IT. Consider the options.
Option 1: Deny Globality. The most obvious strategy would be to deny the globality of
reasoning processes. This is, in fact, pretty much what Fodor takes the advocate of MM to be
doing. Thus he maintains that MM preserves the thesis that mental processes depend on local
properties “by denying –or, anyhow, downplaying—their globality and context sensitivity”
(Fodor, 2000, p.36). In the present context, however, this approach is unsatisfactory. For if one
rejects a central premise of Fodor’s argument, one cannot maintain that it is a sound argument. In
which case, one preserves a commitment to MM only at the cost of giving up the Fodorian
argument for IT.15
Option 2: Explain Globality. An alternative strategy is to accept Fodor’s claim that
abductive reasoning is global in nature and suggest that it can be explained within the massively
modular framework. But it is utterly unclear how this can be done without either compromising
the commitment to MM or rejecting the Fodorian argument for intractability. To see the point,
we need to work through the possible ways in which the modularist might try to explain
globality.
15
I can imagine someone suggesting that the denial of globality is tantamount to a rejection of
nonmodular reasoning mechanisms. But this strikes me as very implausible. Though it may be
true that globality requires nonmodularity, the converse is very obviously untrue. A mechanism
could be radically domain-general and unencapsulated and yet still not be sensitive to global
properties.
20
Version 1: Global Processing Modules. One option would be to claim that some
individual module implements the sorts of global processes allegedly involved in abduction. But
prima facie, this strategy is highly suspect. First, a ‘module’ that implements global processes is
surely no module at all. To speak in such an oxymoronic fashion merely courts confusion and
involves a change of subject matter in which the term ‘module’ is used to describe entities that
would previously have counted as paradigmatically amodular devices. Second, even if we are
prepared to waive such definitional concerns and speak of ‘global processing modules’, it would
seem that the Fodorian argument, if sound, would show such a mechanism to be intractable. The
intractability worries that are allegedly generated by globality are first and foremost worries
about the intractability of abductive processes; and it is only because cognitive mechanisms are
supposed to implement such processes that they too are said to be intractable. But if this is so,
then it’s utterly unclear why globality should pose any less of an intractability problem for the
present suggestion than it does for the explicitly nonmodular alternative. After all, they are just
two approaches to implementing a putatively intractable process. In which case, any mechanism
that implements the process –call it a module, if you like—will be intractable as well. The
present suggestion thus fails to explain how to both preserve a commitment to MM and endorse
the Fodorian argument for IT.
Version 2: Collaborative Activity of Modules. A second way of trying to combine MM
with the acceptance of globality would be to argue that global processes are implemented by
collections of modules acting in concert. Thus while no single module would perform global
operations, the collaborative activity of suites of interconnected modules might subserve global
processes. Here there are two versions that we need to keep an eye on.
Version 2a: Global processes are (classical) computational ones. The idea here is that
the collaborative activity of modules results in a global and context sensitive process that is,
itself, a classical computational one. But, again, it’s hard to see how this could be a solution to
the intractability worries raised by Fodor. To repeat: Fodor’s claims are first and foremost about
the intractability of global computational processes. It is because abductive reasoning
mechanisms implement such processes that they are supposed to succumb to intractability
worries. But if this is so, then whether the process is implemented by a multitude of modular
mechanisms or by a single non-modular device should make no difference to whether or not the
process so implemented is tractable.
21
Version 2b: Global processes are noncomputational. This leads us to an alternative view,
recently defended by Dan Sperber (Sperber 2002). Roughly put, the idea is that although
modules are classical devices, the global, context sensitive processes which they collectively
subserve are noncomputational in character. Slightly more precisely, Sperber suggests that
various forms of global, context sensitivity might arise from non-computational interactions
between modules which govern competition for cognitive resources. Thus Sperber claims that
this constitutes a classical, modularist proposal that evades Fodorian worries about the globality
of abduction:
The general point is that a solution to the problems raised for a computational theory of
mind by context-sensitive inference may be found in terms of some "invisible hand"
cumulative effect of non-computational events in the mind/brain, and that Fodor has not
even discussed, let alone ruled out,16 this line of investigation. (Sperber200?)
This is, I think, an ingenious response to Fodor’s argument that surely deserves further
exploration. Nevertheless, as a way of squaring a classical version of MM with Fodor’s argument
for IT, it suffers from three deficiencies.
First, it is not strictly speaking the classical theory of cognition at all. In order to see this
point, we need to distinguish between two claims:
Classical Theory of Mechanisms: Cognitive mechanisms are classical computational
devices.
Classical Theory of Processes: Cognitive processes are classical computational
processes.
As Sperber presents his view it is supposed to preserve a classical conception of cognitive
mechanisms. This is because he endorses both the claim that all cognitive mechanisms are
modular and that all modules are classical devices. This further implies that all intra-modular
processes are classical. Yet the conjunction of these claims is not quite enough to preserve the
classical computational theory as ordinarily construed. This is because on standard
characterizations classicism is a thesis about cognitive processes, not mechanisms; and what is
more, it’s a claim about all cognitive processes and not merely some or even most of them. But
Sperber maintains that inter-modular processes are non-classical –indeed noncomputational. So,
16
This is not quite true. Though Fodor does not explictly consider Sperber's suggestion, he does
consider --and reject-- a class of suggestions of which Sperber's is a member (Fodor, 2000, p.??).
22
at best, the proposal is an attenuated version of classicism that cleaves to a classical conception
of cognitive mechanisms and a partially classical view of processes. Yet it’s not clear that even
this much is true. For although I won't press the point too hard, it’s utterly obscure why the
cognitive structure that consists of the non-computationally interacting modules should not count
as a (non-computational) cognitive mechanism in its own right.17 In which case, it’s utterly
unclear why Sperber’s strategy not only relinquishes the classical account of cognitive processes
but a classical account of mechanisms as well.
Still, the above may sound all too much like hairsplitting. Surely a more charitable
interpretation is that, in view of Fodorian concerns about globality, Sperber has tried to push the
classical conception of architecture as far as it can go and then suggest an alternative where the
classical view breaks down. This sounds like good scientific methodology: the most conservative
modification of (what Speber sees as) the current best theory that will accommodate the facts
about globality. The spirit of classicism is thus preserved even if the letter must be rejected… or
so it would appear.
But –and this is my second concern with the proposal— for all that Sperber says, his positing
of global, noncomputational processes is unmotivated. As Sperber himself admits, his proposal
on the face of it, is no more than "a vague line of investigation" --one that would not "for
instance, impress a grant-giving agency" (Sperber 200?). Nonetheless, he thinks that here
"unexpectedly, Fodor comes to the rescue" (ibid.). In particular, Sperber seems to think that the
Fodorian objections to global computational processes leaves his brand of MM as a line of
investigation that deserves serious attention. But as we have already seen, the globality
argument is unsatisfactory. In which case, for all that Sperber says, the proposal is still a vague
one in search of a motivation for further investigation.
Finally, Sperber’s proposal is not as consonant with the spirit of classicism as it may initially
appear. Among the most fundamental aims of the classical approach is to provide a mechanistic
17
Certainly, there is no problem with the view that cognitive mechanisms can be nested. On the
contrary, this is very much a standard view among cognitive scientists. It looks, then, as if the
issue will boil down to the (rather vexed) question about how to individuate cognitive
mechanisms.
23
account of cognition. Such an account is supposed to explains intelligence in non-mysterious and
manifestly mechanistic terms: to show how intelligence could result from the operations of a
mere machine. And in the service of this objective, the classical theory purports to explain
cognitive activity in terms of processes that are built up from sets of operations that are primitive
in at least two senses. First, they are computationally primitive: they have no other computational
operations as proper parts. Second, they are supposed to be dumb –that is, obviously mechanical
operations whose execution requires no intelligence whatsoever.18 Moreover, part of what
makes these operations appropriately dumb and unmysterious is that they are extremely local.
Consider, for example, the primitive operations of the read-write head on a Turing machine –
move left one square, move right one square, erase symbol etc. Precisely how such operations
are performed depends only on what is detected on the tape and what state the read-write head is
in. And it is, in large measure, because these operations are so local that they are obviously
executable by a mere mechanism. Much the same is also true of all the main mechanistic
alternatives to classicism. So, for example, within connectionist networks, primitive operations
are local interactions between nodes. The dumbness and locality of primitive operations thus
really seem fundamental to the classical paradigm and, more generally, the mechanistic
enterprise in cognitive science.
What does all this have to do with the globality argument? From the standpoint of the
classical theory, Sperber’s global processes are computationally primitive operations in the sense
that they do not have other computational operations are proper parts. Nonetheless, they are not
plausibly viewed as primitive in the other sense of the term --as being executable without
recourse to intelligence. On the contrary, they are a) nonlocal and b) so myterious that we
currently have (virtually) no idea how they might be implemented in a physical system.
5. The Robot Argument
18
This notion of a computational primitive is central to blocking long-standing homunculi
objections to mechanistic theories of cognition. Such arguments purport to show that mechanistic
explanations either assume precisely what they are supposed to explain (viz. the intelligence of
cognitive activity) or else spiral into an infinite regress of nested ‘intelligences’.
24
A rather different sort of argument for IT consists in an inference from the systematic
failure to model central processes, to a conclusion about the unfeasibility of amodular reasoning.
A full dress version of this argument would demand very careful analysis of the history of
computational modeling. To my knowledge, however, no such argument has ever been provided.
Instead, advocates tend to illustrate their general conclusion –the unfeasibility of amodular
architectures— with one of more choice examples for the history of AI. And of these examples,
the most commonly cited –and most plausible— is the failure of an early approach in robotics
that relied heavily on amodular reasoning processes: the so-called Sense-Model-Plan-Act
paradigm (SMPA) (Bonasso et al, 1998; Brooks, 1999).19 In what follows I briefly sketch this
argument and highlight its deficiencies.
The SMPA paradigm was a general approach to robot construction that dominated the
early history of robotics in AI but is now widely viewed as a dead end. On this approach,
information processing is divided into a number of stages, each of which is assumed to depend
on a distinct set of computational mechanisms. The first stage –preprocessing— involves a
mapping of the information received via sensors –typically cameras— onto representations that
can be deployed by central processing. Stage two –a kind of theoretical reasoning— utilizes the
output of preprocessing in order to update a world model –a ‘unified’ representation of the state
of the robot and the environment that it occupies. This model, along with a specification of the
robot’s goals, then serves as input to a third processing stage –a planner which churns through
possible sets of actions in order to fix upon a plan. Finally, the plan is passed onto the control
level of the robot –i.e. to mechanisms that generate movements of the robot itself.20 The cycle of
sensing, modeling, planning and acting then starts all over.
Perhaps the most obvious problem with the SMPA approach was the failure to get the
cycle from perception via modeling and planning to action to run in real time (Bonasso et al.).
As a consequence, robots had an unfortunate tendency to exhibit disastrously maladaptive
behavior. Here is how the editors of one well-known anthology on robotics illustrate the point:
Now the robot is moving. In a few seconds, its cameras detect in its path a large pothole
that is not predicted by its world model. The sensor information is processed; the world
19
The most famous product of the SMPA paradigm was Shakey the Standford Research Institute
robot (Nilsson, 1984).
20
It may not have escaped the reader's attention that the SMPA resembles Fodor's account of
cognitive architecture to a remarkable degree.
25
model is populated; and at some point while the world model is trying to assert that the
large black blob on the ground is a depression, not a shadow, the robot falls into the
pothole. (Bonasso et al., 1998, p. 5)
For our purposes, the main difficulty is that before the robot can change direction it needs to
complete the modeling and planning stages of the cycle. But these central processes simply could
not be made to run fast enough to keep up with a robot that was moving around the world at the
sort of speed that human beings do. Moreover, despite considerable efforts to address the
difficulty, no satisfactory solution was found: no suitably efficient algorithms for modeling or
planning have been identified and hardware solutions, such as increasing memory size or CPU
speed, only highlight the severity of the problem. In particular, the resulting robots tend to be
laughably slow (Bonnasso, et al, Brooks, Brooks 1999, Gigernezer, 2001, p.43).
What conclusions should be drawn from the failure of SMPA? Given what I have said so
far, it is tempting to conclude that it highlights the unfeasibility of amodular approaches to the
design and understanding of intelligent systems. Rodney Brooks and his collaborators, for
example, famously take it to support the view that “there is no central processor or central
control structure” but instead a large collection of “reactive behaviors” –roughly, modular
devices which generate highly specialized responses to very specific environmental conditions
(Brooks, 1999, p.90). Similarly, Gigerenzer suggests that the failure of SMPA supports the view
that “smart robots need to be … equipped with special-purpose abilities without a centralized
representation and computational control system” (Gigerenzer, 2001, p.43). From this
perspective, then, the failure of SMPA both supports the IT and suggests a more modular
approach to the design of intelligent systems.
But this inference to the intractability of amodular central systems per se is too quick.
One reason is that some of the sources of real-time failure within the SMPA paradigm have little
to do with central processing as such. For example, the preprocessing stage –the perceptual task
of extracting information from the distal environment— was a source of serious combinatorial
problems as well (Bonnasso, et al). But even if we focus exclusively on the difficulties posed by
central processing, I maintain that on closer analysis it's plausible to trace SMPA’s failures not to
amodular processing as such, but to a pair of assumptions about the specific role that such
processes were required to play within the SMPA paradigm: assumptions that the amodularist
need not (and, indeed, should not) endorse.
26
The first of these assumptions is that the planner is required to pass its plans directly to
the control level. As a consequence, it needs to micromanage the activities of the robot: to
specify series of actions in an extremely fine-grained fashion, right down to the actuator level
(Bonnasso, et al). So, for example, it is not enough for the plan to specify that the robot should
go to the refrigerator. Rather, it needs to specify precise movements for getting there, such as:
“turn .56 radians”, “move 122.25 cm” and so on (ibid.). But if plans are individuated so finely,
then there are more plans that can be generated and between which a selection needs to be made;
and this means that the computational demands on planning are very extreme indeed.
Second, if the planner is required to pass onto the control level a very fine-grained
specification of actions, then it also becomes necessary to constantly update the world model.
This is because the planner can only produce successful, fine-grained plans, if it is operating on
the basis of sufficiently detailed and accurate information about the world. In which case, the
demand for fine-grained plans more-or-less mandates that the SMPA cycle be completed before
the robot can take any action at all. And this, in turn, means that the advocate of SMPA must
make central processes run in real-time in order that the robot, itself, operate in a suitably rapid
manner.
The above observations suggest an approach to the design of intelligent systems that does
not deny the existence of amodular central systems but merely assigns them a different role from
the one that they play within the SMPA paradigm. Moreover, it is a proposal –widely known as
the hybrid approach– that has dominated the robotics community in AI for more than a decade
(Gat, Kortenkamp et al.).21 In rough outline, the idea is this: If it’s not possible to get a planner to
run fast enough to govern the fine-grained behavior of a mobile robot, then we should decouple
such central processes from the real time task of generating routine behaviors. On this approach
reactive, modular devices of the sort suggested by Brooks can subserve routine tasks, such as the
avoidance of obstacles. The slower, more deliberative central systems then need only be pressed
into service when the reactive mechanisms run out of routine solutions –for example, to specify a
high level objective or to answer specific ‘queries’ posed to it by the reactive modules. In this
way, amodular systems can play a vital role in governing the gross behavior of the robot without
21
It is also plausible to claim that it is the approach that has resulted in the most successful robot
systems developed so far (Bonnasso, et al).
27
thereby being required to solve the (apparently) insurmountable task of generating effective,
fine-grained plans for even the most routine of tasks.
To summarize: Though the interpretation of SMPA’s failure is no straightforward matter,
the rejection of amodular central systems is clearly too quick since the current state of robotics –
in particular, the success of the hybrid approach— strongly suggests that such mechanisms can
play a vital role, just not the role assigned to them by the SMPA paradigm.
6. Conclusion: Intractability Arguments Are (Probably) Too Hard To Come By
The burden of this chapter has been to show that the extant arguments for IT are
unsatisfactory. Many of them fail because they impose commitments that the amodularist simply
does not accept. Others make implausible assumptions about the nature of human reasoning and
some of them, even if sound, would still be unsatisfactory since they undermine computational
theories of reasoning tout court and not merely the amodular ones. I maintain, therefore, that we
currently have no good reason to accept a version of IT that can figure in a satisfactory
intractability argument for MM.
Of course, nothing I’ve said so far precludes the possibility that a good argument for IT
will be forthcoming. But I do think that there are grounds for thinking it unlikely to happen any
time soon. As mentioned in section 2, there are two broad categories of intractability: inprinciple intractability and in-practice intractability. So, presumably a satisfactory argument for
IT needs to show that amodular mechanisms are intractable in one or both of these senses and,
moreover, do so in such a way as to render the amodularist position untenable as a conception of
our reasoning architecture. But the prospects of an argument for either kind of intractability
appear bleak.
Consider first the prospects of an argument for in-principle intractability –the sort of
argument that forms the stock-in-trade of computational complexity theory. For a variety of
reasons, there are grounds for pessimism here. First, virtually all extant in-principle, complexity
arguments concern worst-case complexity –i.e. the time and effort that’s required for the worst
possible input to a process. But the relevance of such results is questionable since worst-case
intractability is entirely compatible with the task very frequently --indeed normally-- being
significantly less expensive than the worst case. Thus even if an in-principle argument for the
28
worst-case intractability of amodular mechanisms could be had, this alone would not show that
the amodularist view of reasoning should be rejected.
Of course, complexity theorists have good reason to focus primarily on worst-cases,
namely that it readily permits the application of formal methods. But there is an alternative
measure of feasibility sometimes discussed by complexity theorists –viz. what is normally
required for the solution of a task. Such average-case analyses are more plausibly viewed as
relevant to the assessment of claims about cognitive architecture. But now the problem is that inprinciple arguments for average-case intractability are very hard indeed to come by. First, in
order to develop such arguments, one not only needs to specify the possible range of inputs to a
process but their probability of occurrence as well. And this is a kind of information that we
simply do not currently possess for human reasoning processes. A second problem is that the
mathematics of such arguments is extremely complex. As one well-known computer scientist put
it:
Average-case analysis is considerably more difficult to carry out than worst-case
analysis. The mathematics is usually far more sophisticated, and many algorithms exist
for which researchers have not been able to obtain average-case estimates at all. (Harrell,
p.135)
Given that we currently possess few (if any) serious worst-case analyses that bear on the topic of
this chapter, I suspect that the smart money should be firmly against providing relevant averagecase analyses any time soon.
Let's now turn to the prospects of in-practice feasibility arguments. If in-principle
arguments are too much to expect, then perhaps we can hit upon this alternative kind of
argument for IT. But even here the prospects do not seem good. One important feature of inpractice intractability claims is that they need to be relativized to particular systems (or classes of
systems). This is because what resources it is practical or reasonable to use can differ from one
system to another. In particular, if one system has more resources at its disposal –time,
information, processing power, memory etc.— than another, a task may be intractable-inpractice for one but not another. This gives rise to a pair of problems for in-practice arguments.
First, a corollary of the relative nature of practical feasibility is that it’s a largely
empirical matter whether a given task or algorithm is practically feasible. One needs to
determine, in particular, whether or not the target system --the system under consideration-- is
29
capable of solving the task without incurring, what would be for it, unreasonable resource
demands. In the present context, the target is the human mind and the mechanisms from which it
is composed. And this means that an assessment of practical feasibility requires that we possess
substantial amounts of information about the problems that human beings solve, the mechanisms
which permit their solution and the resource limitations that constrain the operation of these
mechanisms. As of now, it is I think fair to say that the psychology of reasoning simply has not
provided us with detailed answers to these questions. In short: the general state of empirical
research into human reasoning imposes a bottleneck on what plausible in-practice feasibility
arguments we are in a position to make.
A second and related issue is that in order to assess whether or not a given computational
task is unfeasible in-practice, one needs to know quite a bit about the role that it plays within the
overall economy of the human mind. This is a generalization of a point made earlier. Recall: in
assessing the Robot Argument, I argued that the computational expense of planning need not
constitute grounds for rejecting the existence of an amodular planner. In particular, following
advocates of the hybrid approach, I argued that even a computationally expensive planner can
play an important role within cognition is it is suitably decoupled from the real-time production
of routine behavior. But the point can be generalized: In order to show that a process or
mechanism is practically unfeasible, it is not enough that one show how, in some context or other
it makes impractical or unreasonable resource demands. Rather, what one needs to show is that
its resource demands are unreasonable in the contexts it in fact occupies within the cognitive
system. And this is, I take it, a kind of knowledge that our cognitive psychology is unlikely to
deliver any time soon.
30
Download