Obstacle avoidance

advertisement
Smart Wheelchairs
Friday, 4/8/2011
Professor Wyatt Newman
Outline
• What/Why Smart Wheelchairs?
• Incremental Modules
– Reflexive collision avoidance
– Localization, trajectory generation, steering and
smart buildings
– Speech-driven wheelchair control
• Natural language interfaces
Architecture
Natural language/
speech processing
localization/motion
control (or joystick)
sensors
reflexes/local mapping
Wheelchair command
“Otto”
instrumented
wheelchair
*Kinect
*Hokuyo
*“Neato”
*ultrasound
Sensing the world
• All mobile vehicles should avoid collision.
• “Ranger” sensors
– Actively emit energy to detect obstacles
• Cameras
– Passively absorb light and can use machine vision
techniques to estimate obstacle positions.
Rangers
• Simple rangers
– Can be sonar or infrared.
– Limited information arises from wide “cone” emitted by
sensor.
Laser Scanners
• Lidars (LI Detection And Ranging)
– Much better information.
– Many radial points of data.
• Velodyne
– Three dimensional lidar.
– Very expensive.
Laser Scanners
• Neato sensor:
– Low-cost sensor
– 1-deg range values
– Not yet available as
separate unit
Cameras
• Monocular cameras cannot return depth information.
• Stereo cameras do return depth information.
– This requires two sensors and has computational and
calibration overhead.
• Hybrid sensor: Swiss Ranger
– Uses infrared time of flight calculations with a monocular
camera to produce a 3D map.
• Kinect sensor:
– Low-cost, mass-produced camera for computer gaming
– Uses structured light to infer 3-D
Autonomous Mode
• Localization
– Relative frame
– Global frame
• Navigation
– Goal planning
– Path planning
– Path following/Steering
Localization
• Local frame sensors
– Odometry
– Gyros
– Accelerometers
• Fusion with Kalman Filter
• Drifty and unreliable for long term position
estimation
Localization
• Global frame
– SLAM (Simultaneous Localization & Mapping)
– AMCL (Adaptive Monte Carlo Localization)
Navigation
• Rviz (robot’s perception)
• video
Smart Building
• Coordination & Cooperation
– Smart devices work together to improve quality of
life for users
– Multi-robot path planning and congestion control
– Robots invoke services within buildings
• video
Vocal Joystick
• A hands free control system for a wheelchair will
provide restored independence
– Quadriplegics, ALS, MS, Cognitive Disorders, Stroke
• Assistive Technology – High Level of Abandonment
– Comfort
– Difficult interface
– Doesn’t properly fit the problem
– Hard to make small adjustments
Alternative Wheelchair
Control
• Voiced
– Path Selection vs. Goal Selection (“Go to”)
– “Natural” language commands (Left, Right)
• Non-Voiced
– Humming controller
• Mouth-Controlled
– “sip and puff”
– tongue
Alternative Wheelchair
Control
•
•
•
•
Head Joystick
Eye movement (“Gaze”)
Chin Control
EMG
Why not voice?
• Voice is the most natural way to interface
with a wheelchair. Why have we not seen
voice activated wheelchairs in the market?
–
–
–
–
–
Recognition problems
Over simplified
Difficulty in precision control without collision avoidance
Difficult HMI
Hard to make small adjustments
Speech-driven Wheelchair
Control
• A naturalistic “vocal” joystick for a wheelchair (or any
other mobile vehicle).
• Prosodic features will be extracted from the user when
giving a command.
– Pitch, Stress, and Intensity
– Modeled and learned (through training simulations)
• Uses a Small corpus
– Users wont have to manage many commands.
– With added prosodic features could provide a more natural
means and solve the small changes in velocity, a problem
described earlier.
• video
A linguistic interface
• Longer-term research in natural human
interfaces
• There are three ways to think and speak about
space in order to travel through it.
(1) MOTION driving, (2) voyage DRIVING, and (3) goal
driven speech control of motion: (1)–>(2)–>(3)




We control each others’ movements, when it is relevant,
by (1) motor commands, (2) indications of paths, and (3)
volitive expressions of goals. So:
Speaking to a taxi driver, (3) the mention of a goal is
normally enough to achieve proper transportation.
Speaking to a private driver as his navigator, we would
instead give (2) indications for the trajectory by referring
to perceived landmarks.
Speaking to a blindfolded person pushing your wheelchair,
we would finally just use (1) commands corresponding to
simply using a joystick in a videogame.
Interface Architecture:
SPEECH
Rec. &
Prod.
Visual
display
!
?
Sensor
signal
Parsing
& Interpretation
Obstacle avoidance
Local Ontology
Incl. sites and known objects
Motor
action
Future Work
• Wheelchair as personal assistant
– Safety monitoring
– Health monitoring
– Assistive functions
• Wheelchair users focus group input
• User trials
• Add-on modules
–
–
–
–
Automated seat pressure redistribution
Medication reminders/monitoring
BP and weight monitoring
Distress sensing/response
Summary/Q&A
•
•
•
•
•
•
Reflexive collision avoidance—near-term product?
Localization, trajectory generation and steering
Verbal joystick w/ prosody
a priori maps vs. teaching/map-making;
smart buildings/smart products
Natural language processing and human interfaces—
longer term
Download