Intent Recognition as a Basis for Imitation Learning in Humanoid Robots

advertisement
Intent Recognition as a Basis
for Imitation Learning
in Humanoid Robots
Andrew Fagg, Rod Grupen, Mike
Rosenstein, and John Sweeney
UMass Amherst
NEMS 2005
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
Programming a Humanoid is
Hard
• Complex mechanisms: many DOF, many sensor
streams.
• Programming by demonstration:
– Demonstrator performs task, robot extracts salient
knowledge to reproduce across many instances.
• Notables: [Pook & Ballard 93], [Kuniyoshi et al. 94], [Voyles
et al. 99], and [Ijspeert et al. 02]
• Imitation:
– Imitation learning augments stochastic exploration for
acquiring control knowledge.
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
On Imitation
• Assume that demonstrator is performing a
goal-directed behavior.
• Kinematic properties of demonstration are
not important to us.
– Can refine using robot specific objectives.
• Interested in the work conveyed by
demonstration.
– How the objects are manipulated, and what
sequence, etc.
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
Why Intention?
• Infer the goal and recognize the scene,
and the behavior can be successfully
reproduced.
– We have some domain specific knowledge.
• Intention is compatible across
morphologies.
• Recognize more from less.
– More abstract representations of actions.
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
Demonstration by Teleoperation
• Direct access to joint velocities and tactile
information.
• No difficulty with correspondence.
• Difficulty of teleoperation:
– Minimal feedback, communication delays.
– Fatigue.
– Discrepancy between human and robot
observational frames.
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
A Robot Teleoperation Interface
NASA/JSC Telepresence System
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
Video
QuickTime™ and a
Video decompressor
are needed to see this picture.
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
How to infer intent?
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
Mirror Neurons
What this suggests:
Action generation and perception[Rizzolatti et al. 01]
are initimately related: use controller
as sensor!
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
Controller as Sensor
• Set of controllers defined by objects in the
scene.
• Compare what demonstrator does to what
each controller would do.
– “Control Projection”
• Use domain knowledge to inform when
meaningful events occur.
– Pay attention to tactile events!
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
Set of Primitive Controllers
• Each controller represents one identified
affordance in the scene.
– Domain: objects on a table in front of the robot: cans,
beanbags, and targets.
AFFORDANCE
A functional matching between object and
actor; described by particular perceptual
features. [Gibson 77]
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
The Robot’s View
• Recognized Affordances:
– Type of grasp: top, side
– DOF constraints: don’t care about rot. about Z
– Every object has a previous place
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
From Scene to Controllers
Object Models
Controllers
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
Extracting a Sequence of
Actions
• Set of controllers represent hypotheses of
intention.
• Observable variables: controller errors, force
magnitude at fingertips.
• Controller i explains sequence of observations if:
– Each step in sequence reduces error
– Error at end is small
– Finishes with tactile event
• Use Bayesian inference to infer most likely.
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
Controller Primitives
• Define Cartesian controller i:
Reference
Error
Joint Command
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
Determining Likelihood
• Error given by distance between joint
commands:
• Compute likelihood at time t:
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
An Extracted Sequence
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
An Extracted Sequence
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
Playing Back the Sequence
• Most likely controller at tactile event is
recorded.
• Extracted sequence makes reference to
affordances in relation to specific objects.
– Can rearrange scene: just find
correspondence between objects.
– Simple visual models used.
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
Further
• More elaborate model of activity
– Look at controller error and change in error.
• More elaborate representations of task
– Hierarchical
– Example: using a tool, building a structure.
• Identify affordances from interaction.
– Find visual features that predict affordances.
– Categorization
• Relational models describe object interaction
– How objects can interact depend on identity.
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
Related Work
• Teleoperation activity recognition [Pook &
Ballard 93]
• Block stacking imitation [Kuniyoshi et al.
94]
• Gesture-Based Programming [Voyles et al.
99]
• Movement Imitation [Ijspeert et al. 02]
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
• Mirror Neurons
– Area of ventral premotor cortex in primates
disocovered by Rizzolatti et al. [1996] that fire
when monkey performs grasps and observes
others perform grasp.
LABORATORY FOR PERCEPTUAL ROBOTICS • UNIVERSITY OF MASSACHUSETTS AMHERST • DEPARTMENT OF COMPUTER SCIENCE
Download