From: AAAI Technical Report FS-96-02. Compilation copyright © 1996, AAAI (www.aaai.org). All rights reserved. Studying the Role of Embodiment in Cognition Maja J Matarid Interaction Lab Volen Center for Complex Systems Computer Science Department Brandeis University Waltham, MA02254-9110 maj a@cs.brandeis.edu http://www.cs.brandeis.edu/,~maj a Abstract Questions The following summarizes the key question behind the synthesis of various fields and lines of inquiry we are pursuing in this paper: This paper raises the question of the connection between embodimentand higher-level cognition which has been eloquently addressed before, but has not yet received muchfocus in the AI community. The paper then proceeds to break the question downinto subparts, and address howeach can be approachedand studied. Finally, the paper briefly overviewstwo directions of our work: group behavior and imitative behavior, and describes their relation to the issue of embodiment and cognition. Is higher-level cognition fundamentally different from lower-level processing involved in spatial and motor (and perhaps even social) interactions? not, then how can the spatial/Zmotor mechanisms be used to construct/evolve and understand the rest of cognition? Motivation The role of physical embodimentin cognition has long been the subject of debate. In Artificial Intelligence (AI) it is largely accepted that embodimenthas strong iinplications on the control strategies for generating purposive and intelligent behavior in the world. However, some theories outside AI have proposed that embodiment not only constrains but may also facilitate certain types of higher-level cognition. Evidence from neuroscience allows for postulating shared mechanisms for low-level control of embodied action (e.g., motor plans for limb movement) and higher-level cognition (e.g., abstract plans). Workin animal behavior has also addressed the potential links between the two systems, and linguistic theories have long recognized the role of physical and spatial metaphors in language. This paper raises the question of the connection between embodiment and higher-level cognition which has been eloquently addressed before, but has not yet received much focus in the AI community. The paper then proceeds to break the question down into subparts, and address how each can be approached and studied. Finally, the paper briefly overviews two directions of our work: group behavior and imitative behavior, and describes their relation to the issue of embodinlent and cognition. 72 ~,’Ve believe that answering this question must involve a combination of analysis of natural cognition in its various forms and synthesis of artificially cognitive systems. To study the critical, functional differences between so-called low-level and higher-level cognition, we must define those, within the context of the systems in question. Next we must somehow compare at least two alternatives by controlling for the cognitive processes in one over the other. Two paths toward studying cognitive alternatives are at our disposal: . 2. analysis of e~sting species and/or developmental stages and/or pathologies, in search of the key differences that address our question. synthesis of different alternatives to cognition, embodied and otherwise, and the study of their comparative properties. Valiant efforts have been made along both lines of pursuit. Wereview some of them next. Natural Systems The study and analysis of natural embodied cognition,is the much older of the two approaches. It has been undertaken by a variety of sciences ranging from biology to psychology, and no-longer-sciences, such as phrenology (Gould 1981). Volumes of data on animal behavior allow us to perform comparative studies across species, including comparisons of close evolutionary relatives such as humans and chimpanzees (Goodall 1971) to more distant cousins like humans and monkeys (Cheney & Seyfarth 1991, Cheney Seyfarth 1990) to many-times-removed family members (Chase 1994, Chase & Rohwer 1987, McFarland 1987, Gould 1982). While these studies can tell us muchabout cognition, they cannot speak directly to the role embodiment within a species. To address that, researchers have endeavored to remove or at least diminish bodily capacities on a particular animal, and observe the results. For obvious reasons, such studies are done on a small set of species, none of which are typically considered to be capable of so-called higher-level cognition. These studies largely address mapping of neural functionality rather than cognitive/behavioral questions even remotely as abstract as the one we have posed. In contrast to animal work, where structure can be ablated in a premeditated and controlled fashion but the subjects’ cognitive capabilities are lacking, work on human pathology offers no such control in the choice of deficits, but provides more than sufficiently complex subjects. Muchof the humandata on highly cognitive deficits are, even today, anecdotal and sparse, but important strides have been madeprecisely in the direction of studying the mind-body connection ?). Finally, humandevelopmental research is an avenue that allows for comparative studies of various deficits and capacities related to the question of cognition. The gradual development and acquisition of both motor and cognitive skills in children provides perhaps the largest corpus of data for addressing our question. Artificial Systems As discussed above, two key problems with biological data are 1) their incompleteness, due to our inability to experiment with arbitrary systems, and 2) their abundance and disconnectedness. The study of artificial systems allows us to deal with both of the problems at once by giving us free reign over what capabilities we include in the system and what biological properties we choose to model. Indeed, we have such unlimited freedom of what we study, that we can be accused of irrelevance. In contrast to the first approach, the study of natural systems, the second one is relatively new, and mired with all the methodological problems of a hopeful nascent science. Post-synthetic analysis, i.e., studying systems one has built, presents unlimited opportunities for bias, subjectivity, and unconscious cheating. As a "science of the artificial" (Si- 73 mon1969), the second pursuit should borrow from the first. To focus in on specific questions to be addressed with computational approaches, we can break down our initial inquiry about the role of embodimentinto the following: ¯ Are the computational and representational structures necessary for motor control and spatial reasoning themselves sufficient for so-called higher-level (more abstract) cognition or are fundamentally different computational capabilities necessary/present? Those who would like to claim a positive answer to this question unfortunately have no solid evidence to support it, but nor do the opposers. Biology provides no straightforward evidence we are currently capable of understanding, although many eloquent arguments for and against the claim have been made (Dennett 1995). Until we are better able to analyze the biological evidence, a synthetic approach is the best first step in attempting to answer at least the question of how possible it is that higher-level reasoning can be a simple extension of lower-level structures. Somebranches of AI have begun to address this problem, as discussed below. ¯ Is humanhigher-level cognition based on so-called lower-level metaphors and constructs including spatial, motor, and social ones? What might human non-spatial/non-motor representations come from and look like? The abundance of spatial and body-based metaphors in human linguistic expressions (Lakoff & Johnson 1980, Johnson 1987, Lakoff 1992) can be used as evidence for the body as the foundational system of reference for higher-level as well as lower-level reasoning. One might ask if any representations, other than the traditional symbolic ones used in AI and its spinoffs, employ encodings that are not fundamentally based on spatio-motor information (Jackendoff 1992). As above, the most productive means of addressing this question is through a synthetic approach, at least until nmch more is knownabout biological cognition. ¯ Do current methods for modeling embodied insect and animal behavior (including humanoid robots) automatically assume a negative answer to the questions above by separating control and cognition? Partially due to AI’s legacy of designer-imposed modularity, almost all of today’s systems separate the low and high levels of reasoning, even if the latter are themselves none too high (Brooks & Stein 1994). The question of this separation has been a topic of hot dispute for the past decade in the robot/agent control community, where reactive, deliberative, hybrid, and behavior-based systems have all been proposed, demonstrated, and debated as alternatives (Matarid 1995, Matarid 1992a, Brooks 1991). Of those, behavior-based systems may present a possible alternative to this separation, but empirical demonstrations of several such systems, combining low and high-level cognition, are still lacking. Meanwhile,the connectionist communityhas wholeheartedly embraced variations of neural-network models that present a monolithic alternative (McNaughton & Nadel 1990, l~umelhart James L. McClelland 1986). Historically not separated into modules, these systems present, to many, the most promising alternative. However,they are also yet to be scaled up to contain both low-level control and highlevel cognition. Interestingly, manyefforts in this direction have resulted in less monolithic, more hierarchical variations that begin to resemble more modular, M-style systems. Thus, while parts of the connectionist community are moving from monolithic to hierarchical, parts of the AI conmmnityare moving form the hierarchical to the monolithic. Not surprisingly, various hybrid alternatives have been implemented, but again largely for specialized systems. In Pursuit ically evaluating behaviors from the basis set (Matarid 1992b, Matarid 1994a). Wehave also extended the basis behavior idea to serve as a foundation for learning in the group domain. Wefirst demonstrated that behaviors, coupled with their triggering conditions, constitute an effectively minimized learning space for multiple robots automatically discovering strategies for group foraging, through the use of a formulation of shaping in reinforcement learning (Matarid 1994c, Matarid 1996@ Wethen expanded the behavior set to include social behaviors, and demonstrated that the same approach to behavior representation can be successfully used to learn a non-greedy social policy that included yielding and information sharing (Matarid 1994b). Finally, demonstrated the ideas on a tightly-coupled cooperative task in which a pair of robots had to learn to cooperate and comnmnicate (Simsarian & Matarid 1995). Our work with embodied agents has shown that the behaviors that served as a stable and effective basis for multi-agent interactions were firmly grounded in the immediate, local, physical interactions between the agents. The behaviors in the basis set included safe-wandering, aggregating, dispersing, following, and homing, all of which were spatial. Even the comparatively higher-level social behaviors, yielding and communicating, involved physical interaction. While it is not surprising that learning to yield involved physical interactions between the agents, it is perhaps more interesting that learning to communicatedid as well. In order to learn to share information about the world, the agents needed repeated physical interactions with each other and with the objects in order for the difficult credit assignment problem to be resolved in the multi-agent domain (Matarid 1996b). Complementary to our efforts with physically embodied multi-robot systems, we are also pursuing work with disembodied multi-agent systems consisting of abstract computational agents. For example, a 4,096agent simulation was developed to study resource sharing between heterogeneous spatially distributed agents to determine under what conditions cooperation would emerge and when it would be the most effficient/stable strategy (Kitts 1996). More abstract information agents were also explored in a number of projects. In one, a US economic model was used as a testbed for exploring local and global multi-agent optimization methods. In another, set of Fortune 500 companies was modeledto study the effects global effects of local trading patterns. The embodied and disembodied nmlti-agent efforts are being pursued in parallel in search of common principles underlying group dynamics. Comparisons of Answers Having considered different lines of pursuit toward answering the question of the role of embodiment,in this section we briefly describe the specific research directions we are focusing on. Social Behavior and Cognition One direction of our work pursues a research program for synthesizing and analyzing group behavior and learning in situated agents. The goal of the research is to understand the types of simple local interactions which produce complex and purposive group behaviors. Wefocus on a synthetic, bottom-up, behaviorbased approach that can be used to structure and simplify the process of both designing and analyzing emergent group behaviors. Our approach utilizes a biologically-inspired notion of basis behaviors as a substrate for control and learning, both at the individual and the collective level, and for both the interactions within an individual agent’s cognitive processes and between multiple agents in a social context. Wehave tested the basis behavior concept by developing a basis behavior set for embodied, mobile agents interacting in complex domains. Wealso introduced methods for selecting, formally specifying, algorithmically implementing, and empir- 74 between the two domains also allow us to study embodiment-specific questions. Imitative Behavior perimental domains: on a dynamical simulation of a human body, on a group of mobile robots, and on a collection of software agents. To keep the model behaviorally relevant, we base and constrain it with data from cognitive neuroscience, psychophysics, and ethology. For the purposes of this paper, one of the most interesting aspects of the imitation work is the role of embodiment. Psychophysical experiments have shown that people tend to a limited part of a. demonstrated behavior and are still capable of successful imitation (Matarid 1996a). For exaple, when shown complex arm movements, subjects fixate on the hand, but can accurately reconstruct the entire posture. The resulted imitated behavior tends to depend more on what the subjects are told (i.e., primed with) than on their observations prior to imitation. This guides us to postulate two testable theories: and Cognition Another active line of our research is into the mechanislns underlying imitalion. As one of the most ubiquitous forms of learning in nature, imitation presents an important research problem in AI and machine learning, as well as in the behavior and neural sciences. As argued above, most adaptive systems involve both cognitive and subcognitive processes. The tension between what is typically referred to as lowlevel control (e.g., motor control, obstacle avoidance, navigation) and high-level cognition (e.g., planning) is ubiquitous in adaptive systems ranging from pathplanning robots to information agents (Maes 1994). One of the key aspects of such systems is the abstraction of data/information between different modules. Fully reactive systems (Brooks 1986, Agre & Chapman 1987, Rosenschein & Kaelbling 1986), behavior-based systems (Matarid 1991, Matarid 1992c, Maes 1989) and hybrid systems (Firby 1987, Arkin 1989, Georgeff Lansky 1987, Payton 1990, Connell 1991) have all been proposed as alternatives to solving the data abstraction problem in control. However, not much work has yet been done in scaling up such systems to problems that require complex enough control and challenging enough planning abilities to demandmultiple interacting representational systems. Weare using the problem of learning by imitation to study the interaction between multiple data representations. Imitation involves the interaction of perception, memory, and motor control, subsystems which typically utilize very different representations, which in this task must interact to produce and learn novel behavior patterns. Gaining insight into the mechanisms of imitation is compelling from the standpoint of AI and behavioral sciences, since the propensity for it appears to be innate (McFarland 1985, McFarland 1987, Gould 1982) and the mechanism is phylogenetically old, found in someinvertebrates (Fiority &Scotto 1992), many birds (Moore 1992), aquatic mammals (Mitchell 1987), and finally primates and humans. all cases, it is a faster and moreefficient form of acquiring new behaviors than its traditional classical conditioning counterpart; it is critical during development and remains an important aspect of social interaction and adaptation throughout life. Consequently, learning by imitation presents a rich domain for studying multi-representational interaction, a crucial aspect of embodiment. Weare developing a model of learning by imitation and testing it empirically on multiple systems in three different ex- 1. The specification of a "task" or "behavior" to be imitated is typically muchmore abstract than the particular demonstrated model, with the exception of specific motor skills, and thus may involve higherlevel cognitive interpretation. This would explain the surprising lack of true imitative behavior in species other than humans, chimpanzees, and dolphins (Tomasello, Kruger & Rather 1993, Mitchell 1987, Davis 1973). 2. The performance of imitation invokes various previously learned and possibly innate capabilities that simplify, and likely guide, the observational process. For example the humancapability to accurately reconstruct posture from only a single saccade implies that complete internal kinematic and dynamic models are involved. At a more abstract level, the mistakes children makein being overly specific in imitation (i.e., imitating superfluous details) and conversely the mistakes adults makein being overly general in imitation (i.e., leaving out what they consider to be superfluous details) both point to imitationspecific computational processes that 1) develop over time, and 2) involve a transition from purely embodied and spatial to the more abstract representations. Toward the end goal of modeling imitation and studying the role of embodimentand its interaction with higher-level aspects of the imitative task, we are beginning to address the many facets of the imitation system: the perceptual component, the nmltirepresentational transforms, the motor system, and the learning component. Just as our nmlti-agent work is organized around the principle of basis behaviors, our imitation work also involves the use of such behaviors, both in the perceptual and in the motor compo- 75 nents of the system. Weare exploring the possibility that the perceptual and motor systems develop in parallel, aided though the imitative process. The motor primitives that provide the complex and growing motor repertoire affect and are affected by perceptual primitives which allow for parsing a complex continuous stream of visual and other sensory inputs. One of the projects currently under way develops methods that automatically extract perceptual primitives using biases imposed by the motor system. Wehope that our parallel efforts in the two domains, group behavior and imitative behavior, will yield results we can use for comparative studies toward the goal of studying embodiment. Chase, I. D. ~’~ Rohwer, S. (1987), ’Two Methods for Quantifying the Development of Dominance Hierarchies in Large Groups with Application to Harris’ Sparrows’, Animal Behavior 35, 1113-1128. Cheney, D. L. L: Seyfarth, 1%. M. (1990), HowMonkeys See the World, The University of Chicago Press, Chicago. Cheney, D. L. ~ Seyfarth, R. M. (1991), 1%eading Minds or 1%eading Behaviour?, in A. Whiten, ed., ’Natural Theories of Mind’, Basil Blackwell. Connell, J. H. (1991), SSS: A Hybrid Architecture Applied to Robot Navigation, in ’IEEE International Conference on Robotics and Automation’, Nice, France, pp. 2719-2724. Summary This paper has explored the pursuit of the role of embodiment in cognition. Wehave proposed a general question, broken it into more specific subquestions, enumerated ways of pursuing research programs aimed at answering parts of the inquiry, and reviewed some work in several related disciplines. In the last part of the paper we described two directions of our research, the study of group behavior and imitative behavior, aimed at the general question relating embodimentand cognition. Damasio, A. R. (1994), set/Putnam, NY. Descartes’ Error, Gros- Davis, J. M. (1973), hnitation: A Review and Critique, in Bateson L: Klopfer, eds, ’Perspectives in Ethology’, Vol. 1, Plenum Press. Dennett, D. C. (1995), Darwin’s Dangerous Idea, Simon & Schuster, NY. Fiority, G. L: Scotto, P. (1992), ’Observational Learning in Octopus vulgaris’, SCI 256, 545-547. Acknowledgments The research reported here was supported in part by the National Science Foundation CAREER Grant IRI9624237 and in part by the National Science Foundation Infrastructure Grant CDA-9512448. Firby, R. J. (1987), Aninvestigation into reactive planning in complex domains, in ’Proceedings, Sixth National Conferenceon Artificial Intelligence’, Seattle, pp. 202-206. Georgeff, M. P. & Lansky, A. L. (1987), Reactive Reasoning and Planning, in ’Proceedings, Sixth National Conferenceon Artificial Intelligence’, Seattle, pp. 677-682. References Agre, P. E. ~ Chapman, D. (1987), Pengi: An Implementation of a Theory of Activity, in ’Proceedings, AAAI-87’, Seattle, WA,pp. 268-272. Arkin, R. C. (1989), Towards the Unification of Navigational Planning and Reactive Control, in ’AAAI Spring Symposiumon Robot Navigation’, pp. 1-5. Goodall, J. (1971), In the Shadow of Man, Houghton Mifflin, Boston. Gould, J. L. (1982), Ethology; The Mechanisms and Evolution of Behavior, ~’V. W. Norton ~= Co., NY. Bermudez, J. L., Marcel, A. & NaomiEilan, e. (1995), The Body and the Self, MIT Press, Cambridge, MA. Gould, S. J. (1981), The Missmeasure of Man, W. W. Norton & Co., NY. Brooks, R. A. (1986), ’A Robust Layered Control System for a Mobile Robot’, IEEE Journal of Robotics and Aulomation P~A-2, 14-23. Brooks, R. A. (1991), Intelligence Without Reason, ’Proceedings, IJ CAI-91’. Jackendoff, 1%. (1992), Languages of the Mind, MIT Press, Cambridge, MA. Brooks, 1%. A. & Stein, L. A. (1994), ’Building Brains for Bodies’, AutonomousRobots 1, 7-25. Kinsbourne, M. (1995), Awareness of One’s Own Body: An Attentional Theory of Its Nature, Development, and Brain Basis, in J. L. Bermudez, A. Marcel ~= N. Eilan, eds, ’The Bodyand the Self’, MITPress, Cambridge, MA. Johnson, M. (1987), The Body in the Mind: The BOdily Basis of Meaning, Imagination, and t~eason, The Chicago Univeristy Press, Chicago. Chase, I. D. (1994), Generating Societies: Social Organization in Humans, Animals and Machines, Harvard University Press, Cambridge, MA. 76 Kitts, B. (1996), The Evolution of Multi-celled Organization, Technical Report working paper, Volen Center for ComplexSystems, Brandeis University. Matarid, M. J. (1996b), Learning in Multi-Robot Systems, in G. Weiss & S. Sen, eds, ’Adaptation and Learning in Multi-Agent Systems’, Vol. Lecture Notes in Artificial Intelligence Vol. 1042, SpringerVerlag, pp. 152-163. Lakoff, G. (1992), Women, Fire, and Dangerous Things, University of Chicago Press, Chicago. Matarid, M. J. (1996c), ’Reinforcement Learning the Multi-Robot Domain’, to appear in Autonomous Robots. Lakoff, G. & Johnson, M. (1980), Metaphors We Live By, University of Chicago Press, Chicago. Maes, P. (1989), The Dynamicsof Action Selection, ’IJCAI-89’, Detroit, MI, pp. 991-997. McFarland, D. (1985), Animal Behavior, Cummings, Menlo Park, CA. Maes, P. (1994), ’Modeling Adaptive Autonomous Agents’, Artificial Life 1(2), 135-162. McFarland, D. (1987), The Oxford Companion to Animal Behavior, in ’Oxford, University Press’. Matarid, M. J. (1991), ’Behavior Synergy Without Explicit Integration’, SIGARTBulletin, the Procceedings of the AAAI Spring Symposium on Integrated Inielligenl Archiieetures 2(4), 130-133. McNaughton, B. L. & Nadel, L. (1990), Hebb-Marr Networks and the Neurobiological Representation of Action in Space, in M. A. Gluck & D. E. Rumelhart, eds, ’Neuroscience and Connectionist Theory’, Lawrence Erlbaum Associates, Hillsdale, NJ. Matarid, M. J. (1992a), Behavior-Based Systems: Key Properties and hnplications, in ’IEEE International Conference on Robotics and Automation, Workshop on Architectures for Intelligent Control Systems’, Nice, France, pp. 46-54. Benjmnin Mitchell, R. W. (1987), A Comparative-Developmental Approach to Understanding Imitation, in Bateson & Klopfer, eds, ’Perspectives in Ethology’, Vol. 7, Plenum Press. Matarid, M. J. (1992b), Designing Emergent Behaviors: From Local Interactions to Collective Intelligence, in J.-A. Meyer, H. Roitblat &S. Wilson, eds, ’From Animals to Animats: International Conference on Simulation of Adaptive Behavior’. Moore, B. R. (1992), ’Avian MovementImitation and a new Form of Mimicry: Tracing the Evolution of a Complex Form of Learning’, Behavior 122, 614-623. Matarid, M. J. (1992c), ’Integration of Representation Into Goal-Driven Behavior-Based Robots’, IEEE Transactions on Robolics and Automation 8(3), 304312. Matarid, M. J. (1994a), Interaction and Intelligent Behavior, Technical Report AI-TR-1495, MITArtificial Intelligence Lab. Payton, D. (1990), Internalized Plans: a representation for action resources, in P. Maes, ed., ’Designing Autonomous Agents: Theory and Practice from Biology to Engineering and Back’, MITPress. Rosenschein, S. J. &Kaelbling, L. P. (1986), The Synthesis of Machines with Provable Epistemie Properties, in J. Halpern, ed., ’Theoretical Aspects of Reasoning About Knowledge’, Morgan Kaufmann, Los Altos, CA, pp. 83-98. Matarid, M. J. (1994b), Learning to Behave SoeiMly, in D. Cliff, P. Husbands, J.-A. Meyer &S. Wilson, eds, ’From Animals to Animats: International Conference on Simulation of Adaptive Behavior’, pp. 453-462. l~umelhart, D. E. & James L. McClelland, e. (1986), Parallel Distributed Processing, MIT Press, Cambridge, MA. Matarid, M. J. (1994c), RewardFunctions for Accelerated Learning, in W. W. Cohen &H. Hirsh, eds, ’Proceedings of the Eleventh International Conference on Machine Learning (ML-94)’, Morgan Kauffman Publishers, Inc., NewBrunswick, NJ, pp. 181-189. Simon, H. (1969), The Sciences of the Artificial, Press. Sacks, O. (1987), The Man Who Mistook His Wife for a Hat, Harper & Row, NY. MIT Simsarian, K. T. & Matarid, M. J. (1995), Learning Cooperate Using Two Six-Legged Mobile Robots, in ’Proceedings, Third European Workshop of Learning Robots’, Heraklion, Crete, Greece. Matarid, M. J. (1995), ’Issues and Approaches In the Design of Collective Autonomous Agents’, Robotics and Aulonomous Systems 16(2-4), 321-331. Tomasello, M., Kruger, A. C. & Rather, H. H. (1993), ’Cultural Learning’, The Journal of Behavioral and Brain Sciences 16(3), 495-552. Matarid, M. J. (1996a), Analyzing Visual Behaviors in Hand Gesture hnitation, Technical Report Computer Science Technical Report CS-96-185, Brandeis University. 77