t for Two: Linear Synergy Advances the Evolution of Directional

advertisement
Action Causes Perception Causes Action:
From Sensory Substitution to Situated
Robots
Lecture 3+4, Unit 5
NUCOG Seminar: Action, Perception, Motivation
Akureyri, Iceland 10.2.-20.2.2006
Marieke Rohde
Centre for Computational Neuroscience and Robotics
University of Sussex
Recapitulation
•
Situated and Embodied View:
– The closed sensorimotor loop
– The rejection of (a priori) internal localisation of cognitive function
– Sensorimotor coordination as reciprocally causal process.
•
Empirical research
– Perceptual suppleance (sensory substitution)
– Change blindness
– Delay experiments
•
Sensorimotor contingencies
– Descriptive concepts for a situated view.
– Can be used to explain cognitive phenomena and faculties (e.g. perceptual
modalities) without localising meaningful cognitive phenomena
This module
Tuesday:
1.
History and Motivation
2.
The Importance of Situatedness: Empirical Evidence
3.
A Sensorimotor Account
Today:
4.
Robotics
5.
The Question of Value
6.
Conclusion
4.) Robotics
Shakey
•
“Shakey was the first mobile robot to
reason about its actions”
•
Shakey implemented the perception,
planning, action approach.
–
Detailed world model
–
Three levels of complexity
http://www.sri.com/about/timeline/sha
key.html
Video:
http://www.ai.sri.com/movies/Shakey.ram
Braitenberg
• Cunningly simple
• No internal state – yet
„cognitive“ behaviour
• One controller –
radically different
behaviours
Braitenberg, V.: „Vehicles. Experiments in Synthetic
Psychology.“ illustrations by Maciek Albrecht, MIT Press, 1984
The Case of Walking
Why is walking so easy for us, and so difficult for
robots?
(see e.g. Honda’s Humanoid ASIMO:
http://world.honda.com/ASIMO/ )
Passive Dynamics
• Passive dynamic walking at Cornell:
http://www-personal.engin.umich.edu/~shc/movies/passive_angle.mov
• Slowly introducing activity (power) in simulation:
http://www.droidlogic.com/sussex/dphil/movies/fullspine3_front.mov
• Take inspiration from airplanes: Gradually add control to gliding.
A Lesson from Robotics
•
In Shakey, a theoretical model was „acid tested“ (with little success)
•
The actual problems had not been recognised.
•
A controller that is no controller at all outperforms Shakey effortlessly
•
Close coupling instead of „dead reckoning“
•
Detail through detailed representation.
•
Classical approaches focus on what people are bad at and computers are
good at (logics, mathematics, chess...). They fail to account for what people
are good at, but computers are bad at.
•
The rise of Behaviour Based Robotics: e.g. Brooks: „Intelligence Without
Reason“, subsumption architectures.
Evolutionary Robotics
Advantages
• Integrated sensorimotor systems
– Close coupling between agent and environment (which is typically
bypassed or modelled very poorly).
• Control (and minimise) prior assumptions (prejudices)
– About what internal structure is necessary to solve a task
– About what kind of functional decomposition underlies the mastery of a
task
– About which strategy is applied to solve a task
– Goes beyond human ingenuity, particularly with respect to complex
nonlinear dynamics
Minimally Cognitive Behaviour
Beer, R.D. (2003).
• Minimally cognitive Behaviour – what and why
– Raises genuine cognitive interest
– Minimal complexity on which we can build up
systematically
– Dynamical Explanation of brain, body, world
interaction
– „Intellectual Warm-Up“, „Frictionless Brains“
Categorical Perception
• Task
– Circular Agent
– Recurrent Neural Network
Controller (CTRNN)
– Move left and right
– Distance sensor array
– Objects fall from the sky
– Catching Circles, avoiding
Diamonds
– (symmetry)
Behavioural Explanation
• Foveate Object
• Scan Left and Right
• Circle: scanning
movements smaller
and smaller
• Diamond: large
avoidance movement
What the agent sees...
„Carrot on a stick“
„Psychophysics“
• Labeling and
Discrimination
• Discrimination Criteria
(Width!)
Psychophysics
• When is the
decision made?
Dynamical Explanations
• Hardcore Mathematics
• Dynamics of agent environment
system
Memory
• Izquierdo-Torres & Di Paolo
(2005):
– Is a reactive agent capable of
only solving reactive tasks?
– Reactive task: what is to do is
immediately obvious from the
sensory data
Memory
•
Izquierdo-Torres & Di Paolo (2005):
– Same Task, same settings (symmetry etc.)
– Feed Forward network without decay
– Perfect mastery, even with respect to „psychophysics“
– An embodied and situated agent is never purely reactive: Agents modify their
position with respect to the objects and thereby partially determine their sensory
perception at the next time step
More Cases
• I will just brush over these...
Active Vision
•
Coevolution of active vision and feature selection (Floreano et Al. 2004)
– A feedforward network masters complex behaviour relying on vision
•
Extension: Development under motor disruption (Floreano et Al. 2004b)
– Analogous to Held‘s Kitten caroussel
Learning Dynamics
•
Tuci et Al. (2002)
– Recurrent neural networks are dynamical
networks, i.e. They have neural state
– Neural network learning is normally thought
to happen through a different process (i.e.
Synaptic learning)
– Learning in a fixed weight neural network
– Floor sensor, light sensor: find out at the
beginning of 14 trials if you are in a
„landmark near“ or „landmark far“
environment
Communication and Social
Interaction
• Matt Quinn (2001):
– Origins of Communication: Making signal only if
it is understood. But: the first time made it will
not be understood.
– Dedicated Channels? No prior assumptions
about how to communicate
– Homogenous population allocates roles
(„leader“, „follower“)
– Minimal sensory and motor equipment (Only
distance, 5 cm range, noisy)
Where‘s the Talking Robots?
• People tend not to be impressed with this
• I am not impressed with AIBO, ASIMO, the Sony Humanoid, ...
(well engineeringwise, I am impressed)
• The biggest issues in current cognitive robotics (e.g. Robocup) are
still the ones that we get for free. Timing issues: Blur, Slip, Delays...
• Manufacturing robots are normally controlled according to dynamical
principles
What Does that Prove?
• Following up on David‘s question
– „When an ER experiments replicates some
cognitive capacity of a human or
animal,typically in simplistic and minimal form,
what conclusions can be drawn from this?“
(Harvey et Al., 2005)
Answers
• Harvey et Al. (2005): Existence proof
– sufficient conditions to generate behaviour x
– catalysing theoretical re-conceptualizations
– Facilitating the production of novel hypotheses
• Di Paolo et Al. (2000): Opaque Thought Experiment
– Results follow from Premises – but in a non obvious way
– Empirical Flavour: must be observed and understood
– Go beyond human ingenuity and thus
• Make a stronger case
• Can uncover novel concepts, relations etc. to be incorporated in a theory
Conclusions so far
• What all these experiments show is that behavioural and cognitive
phenomena that are typically put in distinct boxes in psychology, can
be realised by a system whose mechanic structure is very different
(orthogonal) to the structure of the behaviour space.
• These mechanisms have a tendency to be more efficient (i.e.
computationally cheaper), and exploitation of a close coupling to the
environment is part of this advantage.
• My question: Why would nature/evolution box up functional
mechanisms?
5.) Values
The Problem
• Why is light meaningless to a
Braitenberg vehicle?
• What constitutes genuine purpose,
genuine values, genuine
intentionality?
A Look at Biology
• Living organisms have genuine purposes. They care for their
survival. They have to, otherwise they would not live.
• Survival is not „for free“ as it is in the case of the robot.
• They cannot be reprogrammed
• What is good or bad to them is not down to interpretation.
– Can you redefine what is reward and what is punishment for a living
organism?
– Can there be conventions about what is harmful for a living organism?
I could never put it as nicely…
• “the ill person that cannot express himself anymore,
animals, yes, even a paramecium that cramps before it
is killed by the picric acid dribbled under the cover slip,
the saddening look of a limp plant, the foetus that
defends itself with hands and feet against the
instruments of the doctor - they all present the meaning
of that what happens to them. The meaning is explicitly
evident in the gestures.“ (Weber, 2003. p. 149, my
translation)
Autopoiesis
•
Maturana and Varela (1980): operational definition of
the living.
•
Definition:
a network of processes of production (synthesis and destruction)
of components such that these components:
1.
continuously regenerate and realize the network that produces them,
and
2.
constitute the system as a distinguishable unity in the domain in which
they exist
(Weber & Varela, 2002, p. 115)
Is that enough?
• Autopoiesis just accounts for robustness.
– What is dying? Illness? Stress?
– A merely autopoietic system has no reasons to
improve the conditions for its continued existence.
Adaptivity (Di Paolo, forthcoming)
“a system’s capacity, in some circumstances, to
regulate its states and its relation to the
environment with the result that, if the states are
sufficiently close to the boundary of viability,
1. Tendencies are distinguished and acted upon
depending on whether the states will approach or
recede from the boundary and, as a
consequence,
2. Tendencies of the first kind are moved closer to or
transformed into tendencies of the second and so
future states are prevented from reaching the
boundary with an outward velocity.”
Values
• Metabolic value, as an end, and a criterion for
judgment seems reasonable.
• Maybe it is enough to explain the behaviour of
the most simple organisms.
• However, not all our judgments or all our actions
seem to be measurable against metabolic value.
• Do all values derive from metabolic value?
Our Definition of Value
“We propose to define value as the extent to which a
situation affects the viability of a self-sustaining and
precarious process that generates an identity”
(Di Paolo & Rohde, work in progress)
• Note: There is reciprocal causality!!!!
• Which other values could there be?
• Do non-metabolic values parasite on metabolism?
• Could there be values without a metabolism?
Value System Architectures
• Edelmann‘s Theory of neuronal group selection (e.g. Edelmann
(1989), Sporns & Edelmann (1993))
– Neural circuits are selected due to principles of Darwinian evolution
during lifetime
– Selection through a value signal
– E.g. Reaching: „good“ if hand close to target
– Reinforcement learning with internally generated reinforcement signal
• Very popular with Pfeifer et Al. (e.g. Pfeifer and Scheier 1999)
The Value System
• ``[if] the agent is to be autonomous and situated, it has to
have a means of `judging' what is good for it and what is
not. Such a means is provided by an agent's value
system.'' (Pfeifer and Scheier 1999)
• ``already specified during embryogenesis as the result of
evolutionary selection upon the phenotype'' (Sporns and
Edelmann 1993).
Rephrasing it:
• There is a structure that knows what is good and bad (a priori)
• The rest of the organism/learning mechanism is ignorant and obeys
blindly
• This localised structure, by necessity needs to have dedicated input
and output channels
• Value systems themselves do not learn, they control learning
(functional division)
• Or if, they learn through a „higher level“ value system (regressus ad
infinitum?)
A „VISTIGIAL GHOST IN THE MACHINE“!!!! (Rutkowska 1997)
What is wrong with value systems?
The principle objection:
– Values are arbitrary.
– Values are generated seperately.
– Values are specified a priori.
– Values are not subject to change themselves.
– What happened to the reciprocal causality?
Think about: sensory substitution, goggle experiments...
Think about: social/abstract values and the requirement for
„simple criteria of salience and adaptiveness“ ()
These are not genuine
values!!!
What else is wrong with value
systems?
Are value systems good to model values?
– Investigation of „pseudo values“
– Models and idealisation: To remove gravity from a model of balloon
flight is simply to do away with the original problem we wished to solve.
• So what is wrong?
– Vulnerability
– Generality/Specificity trade-off
– Buckpassing explanatory burden
– False dilemmas: analoguous to nature/nurture divide (Oyama 1985)
– The impossibility of novel values
Damasio
• ``[somatic markers] help us thinking, by illuminating
some (dangerous or beneficial) options in the right way
so they are quickly removed from further reasoning. You
can imagine this as an automatic system for the
evaluation of predictions'' (my translation from German
translation of Descartes error) :$
• ``we are born with the neuronal mechanism necessary to
generate somatic states facing certain classes of stimuli
- the apparatus of primary emotions.”
Emotion systems
• conceived as forming a complementary system to colder or more
detached cognitive processes
• kind of “early warning system” that directly monitors bodily
conditions to generate states that modulate all kinds of interactions
and internal dynamics.
• Again:
– a priori built-in rules
– emotional states
– functional division between the emotion system and other emotion--free
cognitive processes
Just to make this very clear:
• Nobody denies that such mechanisms can and possibly
do work in situations that rely on ontogenetically or
phylogenetically preestablished situations.
• Both value and emotion systems provide the other
cognitive mechanisms with information of the relative
relevance of their activities and future choices.
• This cannot account for the generation of novel values.
How else could you model values
• Reciprocal Causality between value and
value appraising agent:
– Dynamics
– Plasticity
– Situatedness
– No functional separation
Evolutionary Robotics
• First step: The phototactic
homeostatic robot. (Di
Paolo 2000, 2003)
Trying to get a grip on „value
signals“
• The fitness evolving robot
– Evolve robot to perform a task (phototaxis) and a signal that
represents a fitness estimate.
– The fitness estimate is a standard neuron
– Give the robot an environment that requires adaptation (e.g.
Sensor swapping)
How does the behaviour relate to the „value signal“?
How does the „value signal“ relate to the behaviour?
6.) Conclusions
Summary
• The situated and embodied approach to cognition and how it differs
from the „cognition as information processing“ metaphor.
• How empirical research supports the situated and embodied view
– Perceptual perturbations
– Perceptual suppleance (sensory substitution)
– Change blindness
– Delay experiments
• The concepts and language that Noe, O‘Regan and Hurley
contribute to the situated, embodied and dynamical study of
cogntion. In particular, the idea of sensorimotor contingencies and
the mastery of the laws of sensorimotor contingency.
Summary
• How findings from robotics research, particularly from evolutionary
robotics, can inform cognitive science
• How behaviour can be explained dynamically (i.e. Using dynamical
systems theory)
• How true genuine value appraisal cannot be produced by a box of
judgment rules
• What biology teaches us about metabolic (and other?) values.
• How possibly to approach a complex and rich phenomenon like
value appraisal through computational modelling.
Thanks are Due to
• Ezequiel Di Paolo
• Sarah Angliss
• Inman Harvey
• Bill Bigge
Any questions?
References
•
•
•
•
•
•
•
•
•
•
•
•
Beer, R.D. (2003). The dynamics of active categorical perception in an evolved model agent (with commentary and
response). Adaptive Behavior 11(4):209-243.
Braitenberg, V.: „Vehicles. Experiments in Synthetic Psychology.“ illustrations by Maciek Albrecht, MIT Press, 1984
Collins, S.: Passive Dynamic Walking at Cornell University (retrieved 13.2.2006). Information:
http://ruina.tam.cornell.edu/hplab/pdw.html
video: http://ruina.tam.cornell.edu/hplab/downloads/movies/Steve_angle.mov
Damasio, A. R:. Descartes' Irrtum. Fühlen, Denken und das menschliche Gehirn. München: DTV, 2001
Di Paolo, E. A., Adaptive Systems lecture presentations, University of Sussex (UK), Spring Term 2006.
Di Paolo, E. A., (2005). Autopoiesis, adaptivity, teleology, agency. Phenomenology and the Cognitive Sciences.
Forthcoming.
Di Paolo, E. A., (2003). Organismically-inspired robotics: Homeostatic adaptation and natural teleology beyond the
closed sensorimotor loop, in: K. Murase & T. Asakura (Eds) Dynamical Systems Approach to Embodiment and
Sociality, Advanced Knowledge International, Adelaide, Australia, pp 19 - 42.
Di Paolo, E. A., Noble, J., & Bullock, S. (2000). Simulation models as opaque thought experiments. In Bedau, M.
A., McCaskill, J. S., Packard, N. H., & Rasmussen, S. (Eds.), Articial Life VII: Proceedings of the Seventh
International Conference on Articial Life, pp. 497-506. MIT Press, Cambridge, MA.
http://citeseer.ist.psu.edu/dipaolo00simulation.html
Di Paolo, E. A., (2000). Homeostatic adaptation to inversion of the visual field and other sensorimotor disruptions.
Proc. of SAB'2000, MIT Press.
Edelman, G.: The Remembered Present. A Biological Theory of Consciousness. Basic Books, New York 1989.
Floreano, D., Kato, T., Marocco, D. and Sauser, E.: Coevolution of Active Vision and Feature Selection. Biological
Cybernetics, 90(3) 2004. 218-228.
Harvey, I.: Various presentations and lecture material
References
•
•
•
•
•
•
•
•
•
•
•
•
•
•
Harvey, I., Di Paolo, E., Wood, R., Quinn, M. and Tuci, E. A. (2005). Evolutionary Robotics: A new scientific tool for
studying cognition. Artificial Life, 11(1-2):79-98.
Harvey, I., Husbands, P., Cliff, D., Thompson, A. and Jakobi, N. (1996). Evolutionary Robotics at Sussex. In
Robotics and Manufacturing: Recent trends in research and applications (Proc. World Automation Conf. WAC'96),
pages 293-298. New York: ASME Press.
Honda‘s Asimo (retrieved 13.2.2006): http://world.honda.com/ASIMO/
Maturana, H.R. And F.J. Varela „Autopoiesis and cognition: the realization of the living.“ Reidel 1980
Oyama, S.: The Ontogeny of Information. Developmental Systems and Evolution. Cambridge University Press,
Cambridge 1985.
Quinn, M. (2001). Evolving communication without dedicated communication channels. In Kelemen, J. and Sosik,
P., editors, Advances in Artificial Life: Sixth European Conference on Artificial Life: ECAL2001, pages 357-366.
Springer.
Rutkowska, J.: What's value worth? Constraining Unsupervised Behaviour Acquisition. In: Proc. of the Fourth
European Conference on Artificial Life 1997. 290--298.
Shakey the robot (Stanford Research institute) retrieved 13.02.2006 http://www.sri.com/about/timeline/shakey.html
Tuci, E., Quinn, M. and Harvey, I. (2002). Evolving fixed-weight networks for learning robots. In Congress on
Evolutionary Computation: CEC2002, pages 1970-1975. IEEE Press.
Sporns, O., and G.M. Edelman: Solving Bernstein's Problem: A Proposal for the Development of Coordinated
Movement by Selection. Child Dev. 64 (1993) 960--981.
Suzuki, M., Floreano, D. and Di Paolo, E. A., (2005). The contributions of active body movement to visual
development in evolutionary robots. Neural Networks. 18(5/6) pp. 657-666.
Vaughan, E.‘s passive dynamic walking research (retrieved 13.2.2006) http://www.droidlogic.com/
Weber, A., & Varela, F. J. Life after Kant: Natural purposes and the autopoietic foundations of biological
individuality. Phenomenology and the Cognitive Sciences, 1 (2002), 97-125.
Weber, Andreas Natur als Bedeutung.Versuch einer semiotischen Theorie des Lebendigen. Königshausen &
Neumann, 2003.
Download