Image-Based Models with Applications in Robot Navigation

advertisement
Image-Based Models
with Applications in
Robot Navigation
Dana Cobzas
Supervisor: Hong Zhang
Dana Cobzas-PhD thesis
3D Modeling in Computer Graphics
Aquisition
Rendering
Range sensors
Modelers
Real scene
Geometric model
+ texture
New view
[Pollefeys & van Gool]
 Graphics model: 3D detailed geometric model of a scene
 Goal: rendering new views
Dana Cobzas-PhD thesis
Mapping in Mobile Robotics
Map Building
sensors
Navigation
environment
Localization/
Tracking
Robot
Map
 Navigation map: representation of the navigation space
 Goal: tracking/localizing the robot
Dana Cobzas-PhD thesis
Same objective:
How to model existing scenes?
Traditional geometry-based approaches:
= geometric model + surface model + light model
- Modeling complex real scenes is slow
- Achieving photorealism is difficult
- Rendering cost is related to scene complexity
+ Easy to combine with traditional graphics
Alternative approach: image-based modeling:
= non-geometric model from images
- Difficult to acquire real scenes
- Difficult to integrate with traditional graphics
+ Achieving photo-realism is easier if starting from real photos
+ Rendering cost is independent on scene complexity
In this work we combine the advantages of both for
mobile robotics localization and predictive display
Dana Cobzas-PhD thesis
This thesis
Investigates the applicability of IBMR techniques in
mobile robotics.
Questions addressed:
 Is it possible to use an IBM as navigation map for
mobile robotics?
 Do they provide desired accuracy for the specific
applications – localization and tracking?
 What advantages do they offer compared to traditional
geometric-based models?
Dana Cobzas-PhD thesis
Approach
Solution:
 Reconstructed geometric model combined with image
information
 2 models
Model1: calibrated: panorama with depth
Model2: uncalibrated: geometric model with
dynamic texture
 Applications in localization/tracking and predictive
display
Dana Cobzas-PhD thesis
Model1: Panoramic model
Dana Cobzas-PhD thesis
Model1: Overview
Standard panorama: - no parallax, reprojection from
the same viewpoint
Solution – adding depth/disparity information:
1. Using two panoramic images for stereo
2. Depth from standard planar image stereo
3. Depth from laser range-finder
Dana Cobzas-PhD thesis
Depth from stereo
Trinocular Vision System
(Point Gray Research)
Cylindrical image-based panoramic models
+ depth map
Dana Cobzas-PhD thesis
Depth from laser range-finder
 CCD camera
 Laser rangefinder
 Pan unit
180 degrees panoramic mosaic
Corresponding range data (spherical representation)
Data from different sensors: requires data registration
Dana Cobzas-PhD thesis
Model 1: Applications
Absolute localization:
Incremental localization:
Predictive display:
Dana Cobzas-PhD thesis
Input: image+depth
Features: planar patches
vertical lines
Input: intensity image
Assumes: approximate pose
Features: vertical lines
Model 2:
Geometric model with dynamic texture
Dana Cobzas-PhD thesis
Model 2: Overview
Tracking
Geometric model
Dynamic texture
Input images
Dana Cobzas-PhD thesis
Model
Rendering
Applications
Geometric structure
Tracked features
poses
structure
Structure from
motion
algorithm
Dana Cobzas-PhD thesis
Dynamic texture
Input Images
I1
It
Dana Cobzas-PhD thesis
Re-projected
geometry
Texture
Variability
basis
3D SSD Tracking
 Goal: determine camera motion (rot+transl) from image differences
 Assumes: sparse geometric model of the scene
 Features: planar patches
3D Model
differential
motion
initial
past
motion
motion
past warp
current warp
differential warp
Dana Cobzas-PhD thesis
current
motion
Tracking example
Dana Cobzas-PhD thesis
Tracking and predictive display
 Goal: track robot 3D pose along a trajectory
 Input: geometric model (acquired from images) and initial pose
 Features: planar patches
Dana Cobzas-PhD thesis
Thesis contributions
Contrast calibrated and uncalibrated methods for
capturing scene geometry and appearance from
images:
panoramic model with depth data (calibrated)
geometric model with dynamic texture (uncalibrated)
Demonstrate the use of the models as navigation maps
with applications in mobile robotics
absolute localization
incremental localization
model-based tracking
predictive display
Dana Cobzas-PhD thesis
Thesis questions
 Is it possible to use an image-based model as navigation
map for mobile robotics?
 A combination of geometric and image-based model can be used as
navigation map.
 Do they provide desired accuracy for the specific
applications – localization, tracking?
 The geometric model (reconstructed from images) is used for
localization/tracking algorithms. The accuracy of the algorithm
depends on the accuracy of the reconstructed model.
 The model accuracy can also be improved during navigation as
different levels of accuracy are needed depending on the location
(large space/narrow space) – future work.
 What advantages do they offer compared to traditional
geometric based models?
 The image information is used to solve data association problem.
 Model renderings are used for predicting robot location for a remote
user.
Dana Cobzas-PhD thesis
Comparison with current approaches
Mobile Robotics Map
+ Image information for data association
+ Complete model that can be rendered – closer to human perception
- Concurrent localization and matching (SLAM-Durrant-Whyte)
- Invariant features (light, occlusion) (SIFT-Lowe)
- Uncertainty in feature location (localization algorithms)
Graphics Model
(dynamic texture model-hybrid image+geometric model)
+ Easy acquisition: non-calibrated camera (raysets, geometric models)
+ Photorealism (geometric models)
+ Traditional rendering using the geometric model (raysets)
- Automatic feature detection for tracking – larger scenes
- Denser geometric model (relief texture)
- Light-invariance (geometric models, photogrammetry)
Dana Cobzas-PhD thesis
Future work
Mobile Robotics Map





Improve map during navigation
Different ‘map resolutions’ depending on robot pose
Incorporate uncertainty in robot pose and features
Light, occlusion invariant features
Predictive display: control robot’s motion by ‘pointing’ o ‘dragging’ in
image space
Graphics Model (dynamic texture)
 Automatic feature detection for tracking
 Light-invariant model
 Compose multiple models into a scene based on intuitive geometric
constraints
 Detailed geometry (range information from images)
Dana Cobzas-PhD thesis
Download