Imaging, Integration and Visualization of Multi-sensor Data for Road Terrain Mapping

advertisement
Imaging, Integration and Visualization of
Multi-sensor Data for Road Terrain
Mapping
Imaging, Robotics and Intelligent Systems Lab
The University of Tennessee, Knoxville
Sreenivas Rangan
Sijie Yu
David Page
Mongi Abidi
September 2005
ii
Abstract
In this report, we present our experience in building a mobile imaging system that
incorporates multi-modal sensors for road surface mapping and inspection
applications. Our proposed system leverages 3D laser-range sensors, video cameras,
global positioning systems (GPS) and inertial measurement units (IMU) towards the
generation of photo-realistic, geometrically accurate, geo-referenced 3D models of
road surfaces. Based on our summary of the state-of-the-art systems, we address the
need and hence several challenges in the real-time deployment, integration and
visualization of data from multiple sensors. We document design issues concerning
each of these sensors and present a simple temporal alignment method to integrate
multi-sensor data into textured 3D models. The 3D models generated by our mobile
system on the one hand serve as inputs to simulators that study vehicle-terrain
interaction and also possess the required accuracy even for crack detection towards
road surface inspection in airfields and highways.
iii
Contents
1
INTRODUCTION .......................................................................1
1.1
1.2
1.3
1.4
2
LITERATURE REVIEW ...........................................................7
2.1
2.2
2.3
2.4
3
Road Terrain Profiling for Vehicle Terrain Simulators......................................... 1
Airfield/ Highway Pavement Distress Survey ....................................................... 3
Proposed Approach ............................................................................................... 4
Document Organization......................................................................................... 6
Overview of the State-of-the-art ............................................................................ 7
Commercial Systems – Techniques and Methodologies ....................................... 9
Summary ............................................................................................................. 11
Remarks ............................................................................................................... 12
DATA COLLECTION SYSTEM ............................................13
3.1
Hardware Components ........................................................................................ 13
3.1.1
3.1.2
3.2
3.3
4
System Design ..................................................................................................... 18
Hardware and Software Interaction Architecture ................................................ 21
DATA INTEGRATION ............................................................23
4.1
4.2
4.3
4.4
5
Range Profile Alignment ..................................................................................... 23
Direct Pose Measurement .................................................................................... 26
Filtering Redundancy .......................................................................................... 28
Smoothing and Triangulation .............................................................................. 29
ANALYSIS AND RESULTS ....................................................31
5.1
Modeling Sensor Noise ....................................................................................... 31
5.1.1
5.1.2
5.1.3
5.2
Modeling GPS and IMU Errors ......................................................................... 31
Laser Range Sensor Measurement Error ........................................................... 35
3D Modeling Accuracy ..................................................................................... 42
Video Mosaic Results .......................................................................................... 47
5.2.1
5.3
6
Navigation Equipment ....................................................................................... 13
Range and Intensity Imaging ............................................................................. 17
Texture Mapping ............................................................................................... 48
Examples and Discussion .................................................................................... 52
CONCLUSIONS........................................................................61
BIBLIOGRAPHY.............................................................................62
iv
Chapter 1: Introduction
1
1
INTRODUCTION
The need for accurate 3D road terrain models arises from two different yet significant
applications: (a) simulators for vehicle-environment interaction and (b) automated
distress survey in airfields and highways. The generation of such photo-realistic, georeferenced, geometrically accurate 3D terrain models begins with the design of a mobile
mapping system. The mobile acquisition system collects multi-sensor position,
orientation and geometric data which is later processed and integrated into 3D models
suitable for virtual reality simulations and distress survey algorithms. Before getting into
any further details on the construction of our system and real-time data collection, we
very briefly summarize the expectations on our system and the 3D models for each of
the aforementioned applications.
1.1
Road Terrain Profiling for Vehicle Terrain Simulators
The U.S Army has constructed several proving grounds scattered all over the country.
The idea being that army vehicles (tanks and carrier equipment) can be tested on
rough and uncertain terrain before deployment in real world scenarios. The testing is
performed by driving assembled and fully equipped vehicles across different types of
surfaces such as the Belgian blocks, Perryman surfaces etc. and making measurements
for fatigue, damage, wear and tear of different components in the automobile. Such an
assessment gives an idea of the robustness of the vehicle under test. Also, the vehicle
dynamics can be better understood through experimental measurements from surfaces
of different material (grass, concrete etc.), road roughness and topography (hilly,
plain, and rugged). These test drives also help in analyzing the driver behavior and
form the feedback loop towards the improvement of army vehicles and making them
suitable for battlefield conditions. Recently, the U.S army concluded that experiments
using real vehicles were very expensive and proposed to move towards virtual reality
testing using 3D CAD models of automobiles and road terrain. The Virtual Proving
Grounds (VPG) for driver/soldier training, soil-tire interaction, vehicle terrain
interaction and component behavior analysis is an effort in that direction. The terrain
models for these vehicle-terrain simulators are presently generated using
profilometers. In Figure 1.1, we show how profilometers are used to generate
graphical primitives and emphasize the difference between reality testing and virtual
reality testing in Figure 1.2.
Chapter 1: Introduction
Figure 1.1: Generating 3D models for vehicle-terrain interaction simulators using
profilometers.
Figure 1.2: Reality testing vs. virtual reality testing.
2
Chapter 1: Introduction
3
As seen, the models generated from profilometers do not embed the real-world
uncertainty into the simulation. Hence, the need for generating 3D models of dynamic
environments such as hilly terrain, a speed breaker, a gravel road arises. To meet such
a requirement, we propose in this report a data collection system capable of collecting
and processing data of any arbitrary terrain almost real-time and generating models
suitable for vehicle-terrain interaction analysis. The requirement imposed by such
simulators are that the 3D models be in a commonly used format that is easy to
visualize and are also suitable for modeling vehicle dynamics using finite element
analysis.
1.2
Airfield/ Highway Pavement Distress Survey
Traditionally, general aviation airfield pavements are maintained based on the
inspection staff’s judgment and experience [Walker, 2004]. The inspection personnel
walks or drives slowly through asphalt and concrete pavements observing surface
defects and degradation to make recommendations for immediate and long term
maintenance. The manual inspection procedure (as shown in Figure 1.3) is not only
cumbersome, time consuming and expensive but is also susceptible to human error
and inefficiency. With safety of aircrafts and passengers in mind, this functional and
important process of inspection can be significantly improved using a formalized
imaging system that will ease the effort required to inventory airfield conditions
through periodic evaluations and subsequent surface distress management without
compromising safety.
Figure 1.3: Airfield pavement inspection. (a) Manual inspection procedure. (Image
from the FAA conference) (b) Types of distress on the pavement. (c) Damage that
pavement distress can cause on aircrafts.
Chapter 1: Introduction
4
The key to successful road surface evaluation lies in identifying different types of
distress and linking them to the cause. Recognizing the defects and also understanding
their cause based on their appearance helps rate pavement conditions and select cost
effective repair measures. As a first step towards automation, high speed digital
imaging sensors deployed on mobile vehicles have been successfully demonstrated in
combination with image processing algorithms for crack detection. Such video based
vision systems have two major drawbacks in extension to airfield inspection. They do
not provide sufficient depth information and also have ambient illumination
requirements. The depth information is of particular significance in airfields because
the rating scheme for the runway surface is not just dependent on the length and width
of the cracks alone as is the case with pavement distress applications but also on the
depth. Crack depths in the order of a few millimeters require high precision distance
measurements. Hence, the design requirements for a comprehensive airfield data
collection system should address accuracy and precision in three dimensions of
measurement, speed of acquisition, time required for post processing, ease of
visualization and evaluation. To that end, we present a prototype mobile 3D data
acquisition system in this paper. Our system integrates visual range and color data
with position and orientation information through hardware measurements and
provides better accuracy for fast digitization of large scale road surfaces at almost
equal acquisition and processing time. We are able to generate accurate georeferenced 3D models that are compensated for sensor motion caused by the changing
physical environment. The multi-sensor integrated 3D models improve automatic
crack detection and classification. Furthermore, accurate archives of such 3D models
of road surfaces over time can be used for statistical wear and tear analysis.
1.3
Proposed Approach
We are looking to digitize road surfaces as accurately and quickly as possible with
available technology and processing equipment. Our primary goal being able to image
at high accuracy capable of measuring depth of cracks along the road, a 3D range
sensor on a mobile platform directly solves this problem. We have tried to use three
different 3D acquisition methods (triangulation-based, time-of-flight and structured
lighting) for the data collection and document our efforts in this regard. We concluded
that the triangulation based system matched our requirements for high speed and high
accuracy.
Though the 3D information alone is sufficient for crack detection and surface rating,
we need spatial information for crack localization and repair. We collect physical
location information by setting up a GPS base station and placing a receiver on the
mobile platform. The GPS data is accurate up to 4 cm in the motion direction and
gives us 10 samples of position information in one second. The GPS can be thought of
Chapter 1: Introduction
5
as sampling the 3D motion of the mobile platform that houses the sensors. In the
schematic shown in Figure 1.4, we have shown a video camera mounted on a rod,
whose image axis is orthogonal to the road surface. We prefer the orthogonal field-ofview because it makes the registration of range and intensity profiles trivial and
considerably improves integration time without having to consider CCD calibration
and rectification of images. The video data in addition to providing visual cues for
crack detection also helps in estimating the motion. The video data is particularly
useful when the GPS satellite signals are intermittently not available from the satellites
during certain time intervals of the day.
The system components that we have described thus far are sufficient for scanning
planar road surfaces. For roads with sufficient distress, varying terrains and embanked
pavements, the effect of driving on such terrain and roads with bumps needs to be
compensated. The oscillations on the mobile platform caused by the suspension
system also have to be considered. We have hence used the IMU for measuring the
orientation Euler angles (roll, pitch and yaw) of the sensor mount during data
collection. We have used a high performance computer with a Pentium 4 processor
that supports hyper threading with 1GB of RAM and with special high speed serial
interface cards as the processing equipment. Our multi-threaded, multi-document
graphical user interfaces written in C++ are capable of real-time buffer memory
management, storage and initial processing.
Figure 1.4: Schematic of our proposed system.
Chapter 1: Introduction
1.4
6
Document Organization
We have organized this report to address the design challenges in the construction of
the multi-modal integrated imaging system that is capable of real-time data collection
without restriction on the planarity of road surfaces. We have begun by listing the
specific requirements as a problem statement. In Section 2, we summarize
contemporary commercial systems targeting road surface inspection. The literature
survey emphasizes on the design methods implemented thus far and also serves as a
reference study to understand the difficulty in building systems for real-time
deployment. We introduce our prototype system and explain the idea behind using
multi-modal sensors in Section 3. After the data acquisition, we deal with the
integration of multi-modal data in Section 4. The integration involves the
representation of range and visual data into a spatially meaningful form using the
information from position and motion sensors. We show the 3D models generated
using our system driving a van along a test area containing different types of cracks in
Section 5 and conclude with recommendations for possible improvements and
reproducibility of our system in Section 6.
Chapter 2: Literature Review
2
7
LITERATURE REVIEW
In presenting the state-of-the-art and available technology we begin this chapter with a
brief overview of the state-of-the-art on road terrain mapping in Section 2.1 and later
delve into details of commercial systems, and available technologies in the Section 2.2.
We draw conclusive remarks based on our survey in Section 2.3
2.1
Overview of the State-of-the-art
Related work towards pavement distress, especially on airport runways and army
maintained highways dates back to early 1980’s. The pavement management system
(PMS) idea was proposed by the U.S Army [U.S Army, 1984] and has since then
undergone metamorphosis keeping pace with improving imaging technology.
However, transportation departments met with limited real-time success using digital
imaging techniques towards automatic crack detection and filling [NHRCP Report,
2002], until the late nineties. Non-visual sensors and several improvements on imagebased methods were proposed during this period. We summarize these methods in
Figure 2.1 and discuss the advantages and disadvantages of the different types of
sensing methodologies. Analog films have been completely replaced by digital
cameras and digital video systems are preferred to high resolution line scan methods
for the ease of use without special illumination requirements, though line scan
methods offer very high resolution data. Range sensors that directly give depth
measurements have limited field of view while profilometers and acoustic sensors
though inexpensive can only provide low resolution and a low dynamic range.
In 1987, Mendelsohn [Mendelsohn, 1987] listed several of these methods including
acoustic sensors and profilometers and suggested that the imaging modality was a
promising approach. At that time, the processing and image acquisition speeds
challenged the feasibility of a fast and efficient inspection system. Several surveys
were conducted to make an assessment of the feasibility of incorporating image
acquisition and processing methods for both development and implementation of
automated road surface inspection [Howe, 1998]. The conclusions of the survey
encouraged by improving hardware and processing equipment have led to most of the
commercial video-based systems available today that basically consist of an array of
high speed imaging sensors supported with illumination equipment. The video data
Chapter 2: Literature Review
8
from such systems though promises to be sufficient for distress detection [Meignen,
1997], requires spatial information for crack filling after detection and maintenance. A
potential solution AMPIS [Chung, 2003] was proposed that combined GPS
information with video to create GIS-like databases of road surfaces. AMPIS claims
improved road network identification, pavement inspection for better maintenance and
data management over the base framework of PMS.
Taking a robotic approach, Hass et al. [Haas, 1992] proposed a system to overcome
the shortcomings of the video-based system to make depth measurements by
incorporating a laser range sensor. Hass et al. concluded that combining laser range
data and video image data can provide overall better accuracy and speed of crack
detection although due to the time consuming aspect of laser range sensing in 1992,
they demonstrated range imaging for crack verification after an initial pass of the
video based system. Several 3D approaches have then been demonstrated since then.
Laurent et al. propose a synchronized laser scanning mechanism to capture high
precision 3D range and texture profiles. Bursanescu and Blais [Bursanescu, 1997]
reiterate a 3D optical sensor as the answer to high resolution and high accuracy
acquisition and redesign a Biris sensor to meet the specific requirements of the
pavement inspection application. They demonstrate six such sensors mounted on a
mobile platform acquiring data at normal highway speeds. Now that we have given a
brief introduction about methodologies already implemented, we shift into the details
about the sensor and system development in the following subsection.
Figure 2.1: Available technologies for road surface mapping.
Chapter 2: Literature Review
2.2
9
Commercial Systems – Techniques and Methodologies
We begin with imaging approaches, which include using analog and digital imaging to
acquire pavement textures; and then discuss sensing approaches that use acoustics and
infrared laser beams to acquire 3D geometric information. We trace the evolution of
multi-sensor approaches by combining two or more of these sensing methods and
clearly list the advantages over a single sensing strategy.
Analog imaging refers to the obsolete yet simple film photography (usually with 35
mm film) and videotaping (e.g. Super VHS). Analog Photographing or photologging it
is popularly was adopted as the method for Long Term Pavement Performance
(LTPP) program. Photographing using a 35-mm film requires human observation for
identification of distress. “Photologging mobile” as a system is typically a van with a
downward facing camera (to acquire distress data) and one or more facing forward or
in another direction (e.g. to estimate trajectory). “Photologging mobile” is reported to
function at nights using lighted cameras to reduce shadows cast by mobiles, traffic or
roadside features. In most cases photographing consciously samples only a section of
the roadway and with a maximum accuracy archived is cracks that are 1 mm wide
[McGhee, 2004] at speeds up to 60 mph when controlled illumination is available.
Digital imaging is similar to video logging in concept but uses a charge-coupled
device (CCD) for imaging. With such devices, though shadow is still a problem that
needs to be overcome using special lighting, the motivation to switch from the analog
to digital approach is that less or no more human intervention is required in processing
the acquired data. With digital images, computers can be programmed for distress
survey. Yet another convenience that the digital data offers is in terms of storage,
backup, archival and retrieval.
Digital imaging currently is more popular than analog imaging. Their lower price and
higher degree of automation make them the preferred method in most transportation
agencies [McGhee, 2004]. There are two types of cameras currently used: “area scan”
and “line scan”. Area scanning gives a 2D array of pixels which describes a snapshot
of objects. A sequence of snapshots is captured sequentially and periodically. For a
single image it normally covers half or full-lane width and 3-5 m long section of the
road, depending on camera intrinsic parameters (e.g. lens and field-of-view) and
system setup. The distance between successive frames however is dependent on the
motion of the mobile platform. We note that in most attempts using a digital camera, a
downward facing camera orthogonal to the surface of interest is preferred to minimize
pixel distortion.
Line scanning gives a single line of pixels (1D) that can be thought of as an intensity
profile of the road surface. Accumulating such profiles over a pre-determined motion
Chapter 2: Literature Review
10
path can be integrated as a high resolution image. One single large scale 2D image (an
array of pixels) is acquired by aligning pixel lines based on the mobile trajectory.
Each transverse line usually covers one full-lane of a highway, but the width depends
on the height of camera above the surface. Image length is determined by the motion
trajectory. Shadow is still a problem and needs to be overcome in the system design
phase.
Acoustic sensing was demonstrated for measuring smoothness or roughness of the
road surface using the definition of the International Roughness Index. The acoustic
sensing approach emerged in 1981 from the South Dakota department of
transportation (DOT). South Dakota DOT first used it as a roughness measuring
device that became so popular that by 1991, the number increased to 25 DOT agencies
choosing this system for real-time deployment. However, with increasing speed of
imaging-base sensors, a much more recent survey in 2004 for Synthesis of Highway
Practice reports that only 3 agencies are still using it. The South Dakota profiler device
simultaneously collects three ultrasonic profiles, one for each wheel path and another
for the lane center. Three profiles are mathematically integrated to generate roughness
measuring results and rutting at specified intervals along a moving trajectory.
Laser sensing is the rapidly developing and widely used technique for extracting 3D
geometry that has been replacing most of the aforementioned methods. Generally,
laser range scanners are more accurate and are easy to use and require no special
illumination requirements. Laser scanning as the name suggests, leverages an active
light source in the form of a laser to obtain range information, such as height and
distance. The choice of the type of sensor (triangulation, time-of-flight, stereo etc.) is
based on the accuracy and field-of-view required by our problem.
With these sensing modalities, modern approaches have evolved into integrated
sensing techniques which can be described in one word as multi-discipline combined,
multi-sensor integrated, multi-platform compensated, or multi-data fused to mean that
several sensors were involved in the modeling pipeline [Tao, 1998]. A simple example
would be a system that combines laser sensing and digital imaging to yield
photorealistic 3D surface information. Such integrated systems, essentially consist of
three sub systems: navigation, imaging and geometry. The navigation component
usually comprises one or more of accelerometers, GPS, Differential-GPS, or IMU.
Navigation equipment provides a sample version of the actual trajectory in 3D space.
Imaging components, as mentioned before, include digital cameras and video cameras.
It images texture information of the surface of interest. The geometry component uses
one or more of three types of sensors: laser, acoustic, or infrared. It yields 3D
geometric information of the road surface.
11
Chapter 2: Literature Review
Of the several integrated approaches, a summary of the system characteristics and
accuracy that we present in Table 1 reveals more information on how each modality
contributes to the accuracy and photo-realism requirements.
2.3
Summary
The main goal being able to generate large scale terrain models as fast and accurate as
possible, current data collection methods still necessitate integration of several
heterogeneous technologies. We further identify the scope for improvements in system
design targeting the time of acquisition and processing and list the important
characteristics of a real-time deployable system. An ideal road data collection system
must operate in real time gathering and post processing speeds. The duration required
for data analysis should not overwhelm the time required for acquisition. Furthermore,
much of the state-of-the-art appears to be restricted to small section image processing
using visual images when 3D scanning methods can significantly contribute to
accurate and realistic digitization. A single pass data collection strategy for costeffective distress identification and localization for the airfield survey applications,
and high accuracy dynamic terrain modeling applications have not been explored with
critical focus on the accuracy and robustness of the system and its extendibility to
arbitrary terrain. To that end, with all these system requirements in mind we now
present our prototype system in the Section 3.
Table 2.1: Summary of state-of-the-art integrated approaches.
System/
Group
Modalities used
Komatsu
Video, Line scan
WiseCrax
Dual video cameras
GPSVan
Stereo and analog
camera, GPS , INS
Resolution
of
imagery
4
mega
pixel
image.
Detects 3mm wide
cracks
Built for large scale
imaging
3D range
< 0.5 mm
National
Optics
NRC
(Canada)
AMPIS
RoadCrack
Video, GPS
3mm
wide,4mm
deep cracks
0.3 mega pixels
Array of CCD
1mm crack width.
3D range
Special notes
Collects data in the night with
argon lights at 10 km/hr.
Can collect data at 80 km/hr.
Acquires geo-spatial data for urban
planning.
Novel synchronized laser scanning
approach proposed.
An array of Biris 3D sensor used.
Limited field of view.
Can collect 400 km of data in one
day at highway speeds.
Chapter 2: Literature Review
2.4
12
Remarks
The main goal being able to generate large scale terrain models as fast and accurate as
possible, current data collection methods still necessitate integration of several
heterogeneous technologies. We further identify the scope for improvements in system
design targeting the time of acquisition and processing and list the important
characteristics of a real-time deployable system. An ideal road data collection system
must operate in real time gathering and post processing speeds. The duration required
for data analysis should not overwhelm the time required for acquisition. A single pass
data collection should be sufficient for cost-effective distress identification and
localization for the airfield survey applications, the critical aspect being the accuracy
and robustness of the system and its extendibility to arbitrary terrain. With all these
system requirements in mind we now present our prototype system in the Section 3.
Chapter 3: Data Collection System
3
13
DATA COLLECTION SYSTEM
From the state-of-the-art summary, we identified that the integrated sensing approach of
using multiple sensors was the right direction towards a fast mobile scanning system
capable of generating 3D terrain models for simulators providing high accuracy to the
extent of detecting cracks on the road surface. In this chapter, we present our prototype
system that we demonstrate as a potential solution for high accuracy road terrain
mapping. We discuss each component of our system along with its hardware
specifications and limitations in Section 3.1. Later in Section 3.2, we discuss the
software user interface and present a very brief tutorial on how to interface the hardware
with the software to acquire data real time.
3.1
Hardware Components
Our system also leverages the three subsystem approach and includes navigation,
imaging and geometry acquisition equipment. Our navigation equipment consists of a
differential GPS system from Leica Geosystems to measure the position of the sensing
equipment at a given point of time and an inertial measurement unit (Xsens) to track
the orientation of the sensors during the data collection process. Our imaging
components that include a handheld video camera, a triangulation based 3D scanner
(IVP) and a time of flight scanner (SICK) together provide the flexibility to switch
between high accuracy and large field of view. In Figure 3.1, we show the picture of
components that constitute our prototype system and list the hardware specifications in
Table 3.1.
3.1.1
Navigation Equipment
GPS and IMU technologies have been widely used in a variety of positioning and
navigation applications. The usual integration method for GPS with IMU data is
through the implementation of a Kalman filter. In all these integration modes the INS
error states (attitude, rate-of-turn), together with GPS navigation state (position,
velocity) are estimated using a mathematical stochastic model.
14
Chapter 3: Data Collection System
(a)
(b)
(c)
(d)
(e)
(f)
Figure 3.1: Components used in our prototype system. (a) SICK LMS 200 range
sensor (b) SONY digital camera. (c) IVP triangulation-base range sensor (d) Xsens
IMU. (e)Leica GPS-500. (f) Differential GPS system and radio transmission
equipment.
Chapter 3: Data Collection System
Table 3.1: Hardware specifications.
Sensor
SICK LMS200
(Indoor version)
Sony DCR-TRV730
Digital 8™ Camcorder
Leica System 500
SR530 sensor
Xsens MT9
IVP MAPP 2500
Specifications
Sampling frequency: 37 profiles per second
Range: 150 m – 0.2 m
Resolution: 5 mm
Error range: + 15 / - 15 mm, RMS 5mm in mm mode within 8 m
Sampling frequency: up to 30 frames per second, normally work
at 10 frames per second
View angle: 45 degree (max)
Pixel size: 720 x 480
Image format: BMP
Focal length in memory mode: 39 - 702 mm
Sampling frequency: 0.1 s-60 s (for NMEA83 sentences)
3D accuracy : (DGPS) 0.1 cm for North and East, Accuracy in
height = 2 x accuracy in position
Output streaming data format: NMEA 0183 V2.20 and Leica
proprietary
Sampling frequency: 25 Hz – 512 Hz
Angular resolution: 0.05 degree
Angular accuracy: 3 degree RMS
Given configured for a baseline of 70cm and stand off 70cm and
a triangulation angle of 45 degrees,
Sampling frequency: 2000 profiles per second
Resolution: 1 mm
Range: 1m width of view
Accuracy: RMS 1 mm
15
16
Chapter 3: Data Collection System
Differential GPS (DGPS) is the most common approach to significantly improve
measurement accuracy. A DGPS system uses a reference GPS and a rover GPS. The
position information of such a system is computed based on the signal reception from
four different satellites. The absolute location in terms of latitude and longitude at a
particular point is localized as the intersection of four spheres (ellipsoids) associated
with each satellite. Hence, the differential-GPS system requires that there be at least
four satellites to interact with. We have used two Leica GPS-500 station transceivers
that can provide centimeter accuracy on altitude which translates to about two
centimeter accuracy on position. The maximum rate for phase and code measurements
is 10 Hz. The GPS 3D coordinates sampling frequency for our system is one of 1Hz, 5
Hz or 10 Hz. We usually operate the system at 10 Hz data acquisition rate to ensure
high accuracy in position and also reduce uncertainty while modeling.
As shown in figure 4, DGPS involves 2 GPS receivers, one is a base receiver and the
other one serves as a roaming receiver. Base receiver obtains GPS signals and then
generates correction to roaming (or rover) receiver. Then roaming receiver uses
corrections to update and corrects its position. The final output is downloaded from
roaming receiver. Baseline length is an important factor influencing GPS position
accuracy. Based on experiments we proved that the measurement accuracy changes
linearly with respect to baseline length. Therefore, when placing GPS receivers, it is
better to keep them as close as possible.
Satellites
Roaming GPS
Reference GPS
Baseline length
Figure 3.2: Simple schematic diagram to understand the Differential-GPS.
17
Chapter 3: Data Collection System
3.1.2
Range and Intensity Imaging
Range Sensing: The 3D sensing being a significant component in our project, we start
with the description of our laser range sensor SICK LMS 200. A real 3D range image
is obtained by moving the laser relative to the scene. SICK has a adjustable scanning
angle of 100 or 180 degree, sampling resolution of 0.25, 0.5, or 1.0 degree, maximal
range distance of up to 8m with 0.5 degree configuration, average distance error of
+15 / -15 mm, profiling rate of up to 37 profiles/s (statistically estimated) and 5 mm
standard deviation (based on manufactory specifications). The data transfer rate can be
programmed to be 9.6, 19.2, 38.4, or 500K baud. When setting as 500K baud, a
RS422 PCI card is needed for desktop and a PCMCIA card for laptop.
The scanner emits a safe laser beam outside in a fan-shaped patter from right to left.
The emitted laser beam is reflected back when it strikes some object on its path. The
range is computed from the time taken by the beam to travel from the scanner to the
object and back to the scanner. Or, it is based on the computation of time of flight of
the beam. If the beam is not reflected back, it gets lost and the scanner inputs an error
flag. The beam gets lost when it passes through the glass windows or due to a long
range distance. When the beam pass through a glass window, sometimes it is reflected
by inner objects but the reflected beam can not be received by the scanner due to the
reflection in different direction or multiple reflections. Errors are generated when the
sun light directly falls on the scanner mirror. The scanner gets confused by the direct
fall of sunlight and can not compute the true range. This is one of the problems of
outdoor operations. Suppose a single SICK profile line i is generated by connecting a
series of points with distance and angle (r j ,  j ) . The local spatial information ( s i , j ) of
each point (i, j ) can be calculated as below:
S i , j  (0,ri  cos  i ,ri  sin  i )
(1)
where ri is the measured range and  i is calculated by angular range and resolution. If
the
angular
range
is
180
degree
with
resolution
0.5
degree,
then  i  0.5(i  1), i  1,..,361 , that is 361 points are measured for each scan i .
The manufactory specification states Sick has transversal range capability of 8 m.
Because points which are further away from laser mirror center distributes more
sparsely, we use partial points in each scan. Therefore the actual transversal range is
less than 8 m. For example, if the sensor mirror center point is 80 cm away from
object surface, to keep the spatial resolution between 2 closet points within 10 cm, we
use 298 points out of 361 points and the actual transversal range is 6 m.
Chapter 3: Data Collection System
18
Intensity imaging using a SONY video camera: A SONY video camcorder is used
in this prototype. It is used for capturing video images in a sequence when moving
forward. The image frames from the video provide the texture information for the
profiles from the range sensor to generate textured 3D geometric models. Our SONY
video camera is capable of imaging frames with 720 pixels horizontal x 480 pixels
vertical resolution. The CCD in the video camera has a diagonal chip cell size of 5.4
mm and a focal length 4.1 mm. Based on the camera focal length, we decided to
mount the camera at a 194 cm above the road surface (mounted the camera on a van
roof), so that a single image frame covered over a 129 cm wide x 194 cm long
rectangular flat area.
3.2
System Design
The system is configured to be downward looking with 4 equipments mounted on a
rigid flat metal plate. Each sensor has a local coordinate frame centered at its mass
center. Data fusion actually is a set of transformations from sensor’s local coordinate
frame to a global coordinate frame. Each sensor’s body-centered coordinate frames are
explained as below,

IMU local coordinate frame with X going through its body (also parallel to
moving direction), Y following right-handed rule, and Z being upward.

Camera local coordinate frame or image pixel frame with X denoting pixel
column number and Y denoting row number.

Laser local coordinate frame is centered at the body, with X pointing to
moving direction, Y following right-handed rule and Z being upward.

The global coordinate frame which is also the local earth fixed coordinate
frame. The GPS outputs are based on this frame, with X parallel to North, Y
parallel to East and Z being upward. The XY plane is the tangent
approximation of the local earth surface. Range and video measurements are
transformed into this frame.
The camera is placed downward facing with a small incidence angle (from 0 to 10
degrees). Enlarging the angle will obtain bigger field of view; while the sacrifice is the
image distortion and the camera may not capture thinner crack. If the camera is
perpendicularly downward looking at pavement the field of view is 45 degree. We
initially set the orientation angle at 0 degree.
19
Chapter 3: Data Collection System
To achieve high speed acquisition we mount the system on a van, but actually it can
also be mounted on a cart. Thus the outputs will be denser. There are 3 things for
considering the sensor placement, supposing the system is mounted on a van:

The distance from laser mirror center to pavement surface, which affects
significantly on ranging accuracy. Based on our extensive indoor experiments, we
found the optimal distance is between 60 cm and 80 cm. The distance herein is set
as 65 cm. The optimal distance is obtained from indoor calibration experiments,
which will be explained in the Chapter 5.

The distance from video camera focus to pavement surface. This parameter affects
camera field of view. If the distance is set as 194 cm, given camera field of view is
45 degree, a single image can cover a flat area of 129 cm x 194 cm. Larger
distance will generates wider field of view.

Based on the above set up, given the measurement van moves at a speed of 5 mph,
video images are captured at intervals of 22.3 cm, range scan lines are profiles at
intervals of 6.2 cm, and GPS are measured at intervals of 12.4 cm with IMU at
intervals of 0.7 cm. cubic spline interpolation is used for adding missing GSP and
IMU samples.
The sensor placement for the system design is illustrated in Figure 3.3 and the
instrumentation mounted on the van are shown in Figure 3.4
GPS antenna
Video
camera
Camera focal
Sensor mirror
center
194 cm
IMU
Mobile
65 cm
Sick
Surface
Figure 3.3: System design. (a) Sensor placement. (b) Design parameters.
Chapter 3: Data Collection System
Figure 3.4: Our prototype data acquisition system and its close up views.
20
21
Chapter 3: Data Collection System
3.3
Hardware and Software Interaction Architecture
Sick binary data is downloaded from a RS-422 serial communication port on a
QuadTech 2-port PCI serial communication card. GPS/IMU steaming data is
downloaded from two RS-232 serial communication ports respectively (e.g. com1 and
com2). Video images are grabbed from IEEE 1394 frame grabber. IEEE 1394 is a
standard definition for a high speed serial bus. This bus is also named as FireWire by
Apple. It is ideal for consumer electronics audio/video (A/V) appliances, storage
peripherals, and portable devices. The software architecture is shown in Figure
3.5. Our programs are multi-threaded windows architecture based programs that have
built in routines for acquiring and storing data from the IEEE 1394 ports for the video,
COM serial ports for the range sensor and pose estimation hardware. Our processing
computer which houses a hard drive stores the data from all these ports in a format
readily usable for integration. In Figure 3.6, we have shown the snapshots of the GUI
that we have created for real-time data acquisition. These interfaces are extremely easy
to use with quick buttons to trigger and stop data transfer from the sensing device to
the computer. The GPS/ IMU acquisition program outputs data GPS position in the
NEMA standard format that and the IMU data in an easy three column ASCII format.
So far, in this chapter we have listed the hardware components required for such a
road terrain mapping system and also demonstrated the hardware software interaction.
We have used Windows API commands in a Visual C++ environment to create
interfaces to enable easy and real-time data collection.
Imaging
Laser
Video camera
Navigation
I/O
0.027s/p
0.100s/f
0.100s
GPS
0.010s
IMU
Storage
RS-422
Serial
ASCII file
Serial
Comm
class
IEEE 1394
RS-232
Serial
RS-232
Serial
DirectX
show
BMP
Still
images
Serial
Comm.
COM
component
Figure 3.5: Hardware/ Software interaction.
+
ASII
Time
File
ASCII file
22
Chapter 3: Data Collection System
(a)
(b)
(c)
Figure 3.6: GUI built for data collection (a) GPS/IMU navigation data collection
GUI snapshot. (b) Snapshot of the GUI for frame grabbing. (c) Snapshot of the range
data acquisition program.
Chapter 4: Data Integration
4
23
DATA INTEGRATION
In this chapter, we describe the algorithmic procedure to reconstruct multi-sensor
integrated 3D information. We begin with the noise filtering from the range points and
then discuss about the range profile alignment based on position information. We
present the equations governing the alignment.
As mentioned before the SICK range sensor outputs range information. The third
degree information is obtained from measurement vehicle motion (combination of
translation and rotation). Given the moving direction points to positive x direction, y
axis follows right handed rule, and z axis is upward, laser 3D coordinates is computed
as ( x, y, z )T  (0,r cos ,r sin  )T , in which r is the range value and  is the scanning
angle. The laser scans in a constant angular speed therefore for each profile the
distance interval between two closest points is not uniform. To partially compensate
for the non-uniformly scanning of the laser, in each scan, the range values in each
profile are firstly interpolated by cubic splines. This particular interpolation method
can guarantee C 2 continuity along each profile. This allows scan rows have a regular
grid structure. Therefore it is easily to identify the neighbors to right, left, above and
below for each point, as seen later, is essential for the generation of a depth image and
triangulation, or future operations. For the raw range measurements are rather noisy, a
1D median filter is applied to remove noise. After the above operation the resulting
range image has a resolution of 1cm along profile direction (i.e. Y axis).
The measurement vehicle moves in a dynamic mode therefore the distance interval
between two close range profiles is also non-uniform. In our experiment, this scan row
interval varies from 3cm to 7cm. In Figure 4.1 we show the conceptual idea to
generate a grid range image.
4.1
Range Profile Alignment
To obtain a real 3D range image, the individual 2D range scans (or profiles) are
spatially aligned using motion estimated from GPS/IMU. This process can be
explained as an affine transformation for each laser scan. Scans are based on local
laser coordinate system, while alignment process is to transform them into a global
24
Chapter 4: Data Integration
(a)
(b)
Profile Interval
Figure 4.1: Range profile processing. (a) Original range profile. (b) Filtered range
profile (c) Non-uniform accumulation of profiles.
25
Chapter 4: Data Integration
coordinate system (e.g. the local earth-fixed coordinate system). Affine transformation
computation is illustrated in equation (5) and (6) [Zhao, 2003].
To obtain a real 3D range image, the individual 2D range scans (or profiles) are
spatially aligned using motion estimated from GPS/IMU. This process can be
explained as an affine transformation for each laser scan. Scans are based on local
laser coordinate system, while alignment process is to transform them into a global
coordinate system (e.g. the local earth-fixed coordinate system). Affine transformation
computation is illustrated in equation (5) and (6) [Zhao, 2003].
( X ,Y , Z )T  TsgThg ( x, y, z)T
Thg  Tgps  Rimu
(4-1)
(4-2)
Where ( x, y, z )T be the laser measurements in the local laser coordinate frame, which
can be calculated as ( x, y, z )T  (0, r cos ,r sin  )T ,  (i  1)0.5, i  1,...,361.
Thg be the transformation from GPS to the global coordinate frame (e.g. the local
earth-fixed coordinate system).
T sg be the translation from laser to GPS, which is manually measured.
T gps be the translation from GPS to global coordinate frame (e.g. the local earth-fixed
coordinate system), which is also the absolute GPS measurements.
Rimu be the rotation from GPS to global coordinate frame (e.g. the local earth-fixed
coordinate system), which is also the absolute IMU measurements.
Thus by aligning a set of range scans into a global coordinate system a 3D point cloud
is obtained. But because of non-uniformly vehicle motion, the distance interval
between the two closest scans varies. Therefore the range data is scattered and sparse.
In human visualization it looks rather blurring and fails to present road surface
features (e.g. ridge, crease, and jump). In general there is no guarantee that the
sampling density is even sufficient for a correct reconstruction. Moreover measured
range values are noisy, and they may contain outliers. The noise sometimes lies in the
same level of small road features. Therefore post-processing is applied to deal with the
problems described above. Post-processing herein includes redundancy filtering,
interpolation, and spatial filtering. The goal of the processing is to reconstruct an
accurate and eye-appeasing geometry and topology of the scanned surface.
Chapter 4: Data Integration
4.2
26
Direct Pose Measurement
As mentioned in the section 4.1 individual laser scan is spatially transformed into a
global coordinate system. The transformation parameters actually are directly obtained
from GPS and IMU. If we use 6 DoF (Degree of Freedom), the three translation
parameters Thg are measured from GPS and the three orientation parameters Rimu are
measured from IMU. That is from GPS we obtain X, Y and Z and from IMU we
obtain Euler angles (i.e. roll, pitch and yaw). We will discuss the measurement
Actually the raw data downloaded from GPS is National Marine Electrics Association
(NMEA) 0183 v 2.20 sentences. The NMEA is a standard describing a set of
massages that can be output by navigation sensors. The Leica is configured to output
Global Positioning Fix Data, or called GGA sentence for short. The information GGA
provides is listed below:
a.
b.
c.
d.
e.
f.
g.
h.
i.
j.
k.
Time of position (in Universal Time Code, or UTC);
Latitude coordinate;
North or South;
Longitude coordinate;
East or West;
GPS quality;
Number of satellites in use;
Antenna altitude above/below mean sea level;
Geodetic separation;
Age of differential GPS data;
DGPS station ID number.
The latitude and longitude (degree) can be extracted from NEMA sentences. Those
coordinates are based on geodetic WGS84 ellipsoid coordinate system. For fusing
with other sensor data they are transformed into the Earth Centered Earth Fixed
(ECEF) rectangular coordinates (meter) [Guo, 2002]. ECEF rectangular coordinates
are further transformed to the local earth-fixed coordinates. This coordinate frame is a
local approximation of earth surface.
The raw data downloaded from IMU are 3D accelerations and 3D gyroscopes. We use
the development kit provided by Xsens to compute 3D Euler angles. The MT9
software component is implemented in a COM object (.dll) and it supports the
IDispatch interface. The rotation matrix Rimu can be explained as shown in the
following equations.
27
Chapter 4: Data Integration
Rimu  RX RY RZ
4-3
0
0
1
0 cos(ex) sin( ex)
RX  
0  sin( ex) cos(ex)

0
0
0
cos(ey )
 0
RY  
 sin( ey )

 0
0  sin( ey )
1
0
0 cos(ey )
0
0
 cos(ez ) sin( ez )
 sin( ez ) cos(ez )
RZ  
 0
0

0
 0
0
0
1
0
0
0
0

1
0
0
0

1
0
0
0

1
4-4
4-5
4-6
In general the multi-sensory data are loaded into computer asynchronously at multirate. For example, laser scan samples at 0.02s, GPS points samples at 0.2s, and IMU
Euler angles samples at 0.01s. Before registration, or future fusion their data are
needed to be synchronized. The synchronization is performed by an interpolation. That
is interpolating both GPS and IMU for missing data at the laser sampling time t . To
ensure C 2 continuity of vehicle trajectory, we chose 1D Cubic spline to define the
interpolants. Here we assume the moving path is a smooth curve. Cubic spline is a
piecewise interpolation by 3rd polynomials, between n-adjustment points. Therefore
they take into account all trajectory behaviour on each piece of interval, while
respecting continuity conditions on position, velocity and acceleration on each
adjustment point. It can be shown that Cubic spline minimizes, among all interpolation
functions, acceleration on the curve [Gontran, 2005]. Considering these specifications,
such curves are particularly well suited for road axis modeling in fast kinematic
surveys. If we use GPS north measurement as an example, the interpolation is
explained as below,
N (ti )  ati  bti  cti  d
3
2
4-7
Where ti be the GPS time and a, b, c, d be the Cubic Spline coefficients.
One is important that in our synchronization scheme, if a function N(t) has been
selected which, for a given value of t, will produce a value, i.e., north = N(t). If
28
Chapter 4: Data Integration
position is being interpolated then 3 functions are used in the following manner. The
north, west and up coordinates are considered independently. For example, the points
(x, t) are used as control points for the interpolating curve so that x = Nx(t) results,
where N denotes an interpolating function and the subscript x is used to denote that
this specific curve was formed using the north (or x) coordinates of the key positions.
Similarly, y=Wy (t) and z=Uz(t) are formed so that at any time, t, a position (x, y, z) can
be produced.
4.3
Filtering Redundancy
As the van drives forward the range scans are captured in an order with respect to
temporal order. But when turning a big heading angle the order may be reversed,
which is illustrated in Figure 4.2. From figure 4.2 (a) we can see if the range point has
a longer transversal distance (i.e. ri cos  i ) than critical distance (i.e. d crit ), the reverse
order will happens. But if the transversal distance is less than critical distance, shown
in figure 13 (b) the scan order is kept. For a given moving translation distance s and
a heading angle difference  between two scans, the critical distance is computed as
below,
d crit 
s
sin(  )
4-8
Therefore the redundancy filtering can be described as for any range point has longer
transversal distance than critical distance, it will be removed from 3D model.
Point i in scan 2
ri cos  i
ri cos  i
Point i in scan 2
Scan 1
Scan 1
s

Scan 2
Scan 2
d crit
d crit
Moving path
Moving path
Figure 4.2: Scanning real-time (a) Reverse order. (b) Normal serial profile order.
Chapter 4: Data Integration
4.4
29
Smoothing and Triangulation
So far the 3D point cloud still has noise and outliers. To remove those unexpected
range values a 2D median filter is applied. For the points have connectivity
relationship with their neighbors, a median filter is performed in a neighborhood
defined by a window with 3 x 3 point size.
Triangulation is performed on the “clean” 3D range point cloud and 3D models are
obtained by rendering the triangulated results. The triangulation algorithm is very
simple: it divides the grid point cloud into small cells and then divides the cells into 2
triangles. We call this method as uniform triangulation. Actually most of the model
surface patches are flat and with small curvatures. This corresponds to the real
pavement surface which is also mostly flat and only cracks generate big curvatures.
Therefore uniform triangulation can not represent such surface characteristics; we
chose to use multiples sizes of triangles for representing surface area and cracks
separately. This method is called multi-level triangulation with level of detail (LOD).
The algorithm can be described as whenever there is a point with big curvature; we
divide the triangulation which includes the point into 4 smaller triangulations. The
criterion for estimating curvature is defined as “mean distance”, which is the mean of
distances from points to the plane defined by their outer triangle. If the mean distance
is bigger than a threshold value then the triangulation is divided into 4 smaller ones.
This LOD algorithm is firstly introduced in [Sequeira, 1999]. It can be explained as
below,

The point cloud is projected on a 2D big grid. In the previous step we interpolate
the points along the transversal direction (or Y axis); the result should be a grid
with each point at the corner of a small virtual cell inside of the grid.

The big grid is divided into, say 50 x 81 cells. For each cell 2 triangles are
generated by connecting diagonal points.

Then for each triangle “mean distance” is computed for all points falling inside it.
If the mean distance is larger than a predefined value, the triangle is divided into 4
smaller triangles. This step is performed iteratively until the triangulation can not
be divided any more.
The figure 4.3 shows an example of uniform triangulation and multi-level of
triangulation. As a comparison we also applied 2D Delaney triangulation to the point
cloud. The main advantage of uniform triangulation over 2D Delaney is its efficiency;
while the LOD triangulation result reduces significantly the model size. The point
cloud has 65,325 points representing a virtual grid of 201x325 point size.
30
Chapter 4: Data Integration
(a)
(e)
(d)
Figure 4.3: An example of the reconstructed 3D model (a) Uniform triangulation
result. (b) Triangulated with level of details.
Chapter 6: Conclusions
5
31
ANALYSIS AND RESULTS
In this section we begin with the characterization of the sensors that are deployed in the
system, to enable us to use these datasets in a much better fashion.
5.1
Modeling Sensor Noise
Totally eight types of measurements are acquired asynchronously from four sensors
during operation time. They are 3D positions (North, West and Height); 3D
orientations (Euler angles: roll, pitch and yaw), 1 intensity image and 1 range value.
Because of equipment physics and moving dynamics, we expect noise and outliers
come along with true measurements. To precisely create surface 3D models, we need
to model sensor output instead of assuming that we can measure the 8 types of
information exactly. In the following sections DGPS, IMU and TOF laser are
investigated respectively. Their error models are created to estimate their electronic
errors precisely. This work has been ignored by most 3D reconstruction work.
Equipment noise is usually obtained from manufactory specifications.
5.1.1
Modeling GPS and IMU Errors
The original outputs from an Inertial Measurement Unit (IMU) are 3D linear
accelerometers and 3 ring laser gyroscopes. Euler angles (or attitudes) are computed
from raw measurements using a custom program provided by Xsens. The Euler angles
then are used to compute motion orientation matrix (refer to section C). We expect
errors generated from fusion process and equipment electronics. Manufactory
specifications the attitude error is less than 1 degree when sensor being placed
stationary and 3 degree when the sensor is moving. But actually errors are larger than
the specifications.
The DGPS errors are caused by many sources. In [] describes the major sources, be
satellite clock error, ephemeris error, receiver errors, and atmospheric delay. DGPS is
much more accurate than a standard GPS measurement. The table 5.1 lists common
GPS errors.
32
Chapter 6: Conclusions
The GPS outputs are pseudo-range and carrier phase. A dual frequency receiver
outputs range and phase measurements for each carrier frequency. These four outputs
are combined to obtain earth centered coordinates (degree) and mean altitude (m), e.g.
WGS84 coordinates. Then off-line transformation is performed to obtain local earthfixed coordinates (degree). During this process we expect errors are generated from
equipment electronics and from transformations. By using DGPS, Leica system 500
can achieve 2cm accuracy in North and East direction. Altitude accuracy is normally
double worse, i.e. 4cm.
Besides the error source mentioned above there are also other factors influencing GPS
accuracy. For example the satellites can be blocked by high buildings, tunnels and
overpasses. If this happens the GPS will not be able to estimate position. The GPS
updating rate is usually slower than other sensors used in system. Therefore we have
to synchronize GPS measurements with respect to faster sensors, i.e. TOF laser.
Synchronization strategy is basically an interpolation. We use Cubic spline for
estimating missing GPS coordinates. This method can reconstruct a 3D smooth curve
and then corresponding positioning information is extracted from this curve. But if
discrete sampling process is under-sampling, the reconstruction can’t represent the
moving trajectory, and then positioning information is not accurate in which undersampling happens.
But IMU and GPS random error characteristics have been well studies. It is possible to
use a mathematical model to estimate their values. The model is constructed on static
experimental data. The sensor measurement residual can represent their errors if the
measuring process is long enough. In [] the residual is analyzed and then the paper
suggests using an Autoregressive (AR) model to estimate it. The reason of doing this
is that the sensor residual is closely similar to an empirical higher order Gauss-Markov
Process. The paper emphasizes that the 1st GM process can not model most residual
sequence. In figure 15 we give an example of IMU yaw angle residual with its
histogram and autocorrelation and compare it with a simulated 1st GM sequence. We
can see how closely these two sequences similar to each other. For a more general
case higher order GM is used. But paper further points out that an AR model can be
easily constructed instead of higher order GM. The modeling process can be
illustrated as below:

Feed the measurement residual into a inverse z-transformation equation and then
the residual at current time n can be computed as,
p
y(n)   ak y(n  k )  bw(n)
k 1
5-1
33
Chapter 6: Conclusions

Determine AR model parameters ( ak , b, k  1,... p ). This is performed by
minimizing the error between the original signal represented by the AR and the
estimated signal

p
y   a k y ( n  k )
5-2
k 1
The cost function for this minimization problem can be computed as below,
N
E   e 2 ( n)
n 1
N

  [ y (n)  y (n)]
2
5-3
n 1
N
  b 2 w(n)  min
n 1
In which, N be the measurement size; p be the AR model order, could be 1, 2, 3, or 4
(higher ones are not practical for computation); w(n ) be white Gaussian sequence with
zero mean and unity variance.
From equation (9) we can find that the minimization is b 2 . It is also the estimated
variance  w2 of the white noise input the AR model, or more generally, it represents the
prediction Mean Square Error (MSE)  2 .
Using Yule-Walker method the AR parameters are optimally estimated by solving the
following,
E
0
a k
5-4
If assuming we have the autocorrelation computed from residual sequence, which is
represented as R (m ) , where m=-N,…,N.
Then the parameters are computed as below,
Ra  r
1
5-5
a  R r
5-6
a  (a1 ,..., a p )T
5-7
34
Chapter 6: Conclusions
r  ( R(1),..., R( p)) T
5-8
... R( p  1) 

... R( p  2) 
...
... 

...
R(0) 
R( p) 1    2 
R(1)
 R(0)

R(0)
 R(1)
R
...
...

 R( p  1) R( P  2)

R(1)
 R(0)

R(0)
 R(1)
 ...
...

 R( p) R( P  1)

...
  
... R( p  1)  a1   0

...
...  ...   ...
  
...
R(0)  a p   0






5-9
5-10
After the parameters are obtained, we can model the measurement errors as a
correlated sequence. If this is combined with a linear Kalman filter, the data set could
be smoothed correctly.
The Kalman filter can be illustrated as an optimization method such that the following
statistical conditions hold:

On average the estimate of the state will equal the true state;

On average the estimate error is minimized.
There are many alternative to formulate Kalman filter equations. The most important
thing here is to find an appropriate model for this specific application. Optimal
parameters such as measurement noise and state noise are the other critical
consideration for implementation. We initially chose a linear Kalman which is given
in the following equations [Simon, 1998].
xk 1  Axx  Fak
5-11
zk 1  Hxk  vk
5-12
S k  Pk  R
5-13
K K  FPk S k1
5-14
Pk 1  APk AT  Q  APk S k1 Pk AT
5-15



x k 1  A x k  K k ( zk 1  H x k )
5-16
35
Chapter 6: Conclusions
In which, k denotes states and measurements happen at time k; x k  1 be the initial state
estimation from current state with accelerations estimated as a k . z k  1 be the real
measurement with errors vk . Here the error sequence can be represented by a AR
p
model as vk   ai v(k  i)  bw(k ) . As suggested by the [], the order p is optimally 2,
i 1
3, or 4. S k called covariance of the innovation. K k be kalman gain matrix. Pk  1 be

covariance of the prediction error. x k  1 be the optimally estimated state. The
computation is fairly intuitive. The first term means the state estimate at time k+1 is
just state estimate at time k. this term would be the state estimate if we don’t have a
measurement. In other words the state estimate propagates in time just like the state
vector. The second term means corrector. It corrects the propagated estimate using
measurement. This term means that if the measurement noise is much grater than the
process noise, K k would be small, then the measurement will contribute to final
estimate less; otherwise K k is large, then the measurement will be trusted more.
From the error estimation result, we observe that,

Pitch performance is statistically better than roll and yaw. When being absolute
static the measurement accuracy is better than 1 degree. But in a moving mode
accuracy decrease significantly, especially heading angles (yaw). Therefore denoising is necessary before data fusion,

The IMU sensor needs around 3s to initialize. We use the data after initializing
time,

Heading accuracy is worse, which is sensitive to vehicle maneuver.

GPS Z value has larger error than X, and Y values. This phenomenon happens
normally in a GPS receiver.
5.1.2
Laser Range Sensor Measurement Error
The TOF laser range value is computed from the time period during which laser
travels from scanner to target object and then back to scanner. The manufactory
specifications state the scanner can measure up to 8 m transversal range with +/- 15
mm system error and 5 mm standard deviation. Actually range error dynamically
changes with respect to many factors, including target surface properties (reflectivity,
gray level, color), distance from laser center to object, and orientation degrees. Sick
36
Chapter 6: Conclusions
LMS 200 has been calibrated in [Ye, 2002] by extensive indoor experiments. Those
experiments are conducted in an empirical environment (indoor with temperature
24°c). The scanner is configured to work at 1 degree angular resolution and with 500K
Baud rate. The paper concludes that,

When operating at 500kbaud, occasionally data packages may lose, while no
package is lost at the slower 38.6K baud rate. However the lost data packets in the
experiment never exceed 0.08%.

There is noticeable drift until 3 hours after start-up, but that drift related
fluctuations stabilize after 3 hours.

Mixed pixels may occur in range measurements, which is described as “when a
laser spot is located at the very edge of an object, the measured range is that of a
combination of the foreground object and the background object”.

Target surface reflectance properties dramatically influence measurement errors.
The [Sick, 2002] list 9 common object surfaces and their reflectivity. From the list
asphalt pavement surface has medium reflectivity capability.
Incident angle of the laser beam affects on measurement errors. The paper concludes
that when the angle between +/- 20 degrees will not cause a significant error in the
range measurements. A relatively larger error was caused by the angel of 30 degrees
and the error is small again from 30 to 60 degrees. In the experiments scanner is put
parallel to the horizontal floor surface. But in our case we put the scanner downward
and we noticed significant errors happening on the center points (around 20 points get
effects). Even if we orient the scanner a small angle (10 degrees) away from horizontal
plane the center points are still noisy. The only way to get rid of this is to put the
scanner farther away from target object, e.g. 100 cm, but in that case range resolution
dramatically decreases.

The most important contribution of this paper is set up a mathematical model
between true range value and mean measured value. Therefore the model can be
used for estimating measurement errors. The equation is listed below,

r  k  b
where,
k, b are parameters;

r is the true estimation of range;
(24)
37
Chapter 6: Conclusions
 is the mean measured value.
Based on the paper we conduct extensive experiments both indoor and outdoor to
calibrate the TOF laser. The above equation is used to estimate true measurement
values and then roughly evaluate the ranging accuracy. All experiments are conducted
with laser angular resolution of 0.5 degree, transfer
rate of 500 Kbaud and profiling rate of 37 p/sec. If higher angular resolution is
selected the profiling rate will consequently decreases, which is not suitable for high
speed scanning application. Our experiments are illustrated as below,
(a)
(b)
(c)
(d)
(e)
(f)
Figure 5.1: Simulated 1st order Gaussian – Markov function with Yaw measurements. (a) 1st
order Gaussian Markov samples; (b) Histogram; (c) Autocorrelation sequences. (d)Yaw
measurements residual; (e) Histogram; (f) Autocorrelation sequence.
38
Chapter 6: Conclusions
Table 5.1: GPS and IMU Measurement Error Modeling Results
Measurements
Max
error
Min
error
*RMS
X (cm)
Y (cm)
Z (cm)
Roll (deg)
Pitch (deg)
Yaw (deg)
2.5
3.1
2.9
2.7
2.4
3.7
-2.4
-2.8
-3.3
-2.0
-2.3
-3.6
0.9
1.0
1.0
0.9
0.9
1.1
N
 (measurement
*RMS =
i 1
N
2
i
)
2nd AR Model
Parameters
b
a1
a2
1.35 -0.38
0.17
1.31 -0.37
0.28
1.34 -0.41
0.28
1.13 -0.13
0.04
1.06 -0.07
0.05
0.83
0.15
0.89
Manufactory
specifications
<2
<2
<4
<1
<1
<1
Chapter 6: Conclusions
39
40
Chapter 6: Conclusions
(a)
(b)
(c)
(d)
Figure 5.2: Kalman filter is used to smooth measurement errors; (a) Yaw measurements before
smoothing and after (unit: degree); (b) Yaw errors estimated using 2nd AR model (unit: degree); (d) Z
measurements before and after smoothing (unit: cm); (d) Z errors estimated using 2 nd AR model (unit: cm).
41
Chapter 6: Conclusions

An indoor experiment to calibrate optimal beam number (or point number in a
single scan). We mount the system on a cart and push it smoothly, therefore the
motions dynamics will be minimized but still simulate real data acquisition
process. The height from cart to surface is 88.1 cm. According to [Sick, 2002], the
distance from laser back to the mirror center is 9.3 cm. Therefore the true
measurement distance should be 78.8 cm ( d1 ). The laser incident angle is 0 degree
and scanner is place downward looking at the target surface. Totally 8,472 scans
are acquired. Then the scanner is lift up to 94.8 cm ( d 2 ) away from floor surface.
The D which is the difference between d1 and d 2 ( d1  d 2 ) is 16 cm. we compute
mean values for all beams from number 1 to number 361 at distance d1 and d2
respectively; compute their mean value difference. We found that the 171 th beam
(i.e. the 171th measured value of a scan) was the beam that most accurately
represents the true distance ( d1  d 2 = 16.9 cm) and that beam is subsequently
used in the following experiments.

A set of indoor experiments in which the sensor is downward looking and with its
scanning plane parallel to surface plane. The purpose of these experiments is to
calibrate optimal distance from laser to target surface. The sensor firstly is placed
stationary and then is moved by a cart smoothly. The laser is then tilted so that
there is a small incident angle, i.e. 10 degrees. Ye states that if the incident angle is
between – 20 / + 20 degree there will be no significant changes in range
measurements.

Outdoor experiments on a flat asphalt pavement surface and a flat concrete
pavement surface, respectively. We mount the system on a cart and the height
from laser window center to surface is measured as 73 cm. The sensor incident
angle is 0 degree and scanner is place downward looking at the target surface.
We then use the linear model (i.e. equation 24) to estimate true range value.
Parameters values have been calibrated in [Ye, 2002], with k  1.0002, b  3.6 (unit
is mm). Those values are estimated from an empirical indoor environment with data
acquired from a medium grey target (RGB = 127) at 0 degree orientation angle. It
represents the median situation. Then errors are computed respectively and compared

with measured ones. The result is shown in figure 16 with r r vs. the scan number.

Where r and r denote true and measured value respectively. Table 9 summarizes
measurement errors for optimal distance experiment. We can conclude that when the
sensor is place around 80cm from target the range accuracy is maximized. Figure 17
shows experiment results with distance vs. error. The global minimum error is
achieved with the distance of 73cm. Starting from distance of 60cm the error
decreases fast until passing after distance of 100cm. that is we can say the optimal
distance range is 60cm – 100cm.
42
Chapter 6: Conclusions
Table 10 shows the pavement surface properties effects on range accuracy. The indoor
concrete surface is the most flat one therefore its accuracy is the smallest. On the other
hand when scanning a coarse asphalt surface but still flat we expect the accuracy
performance worse. Outdoor scanning has many environmental factors to influence
ranging, that is for almost the same flat concrete surface indoor accuracy is better than
outdoor one.
Table 5.2: Errors Influenced by Surface Properties
Experiment
Indoor concrete flat surface
Outdoor asphalt surface
Outdoor concrete surface

Error upper
bound (mm)
16.8
30.4
19.6
Error lower
bound (mm)
13.2
-19.6
-11.4
Error (mm)
5.6
7.8
6.1
The manufactory specification states the laser has +/- 15 mm system error with 5
mm statistical error when being set as (mm-mode with 1…8m range). Compared
with our experiments, most static experiments generate better accuracy while all
kinematic experiments have worse performance. The motion dynamics add
significant errors into measurement results. Therefore when a van is used as
measurement vehicle the driving speed should not be moderate otherwise errors
will be introduced because of vehicle vibration and tire friction.
Table 11 summarizes the system equipment accuracy performance. The true value of
GPS and IMU are estimated using Kalman filter and AR model. Then the measured
values are used to compute RMS. The range accuracy is estimated from outdoor
asphalt flat surface scanning, which can be the typical pavement surface encountered
outdoor.
5.1.3
3D Modeling Accuracy
We use our system to scan many indoor and outdoor objects. In general indoor objects
are well structured and have well-defined features therefore they are better to test
system capability; while outdoor targets can best simulate real cracks inspection
applications. To compute accuracy we firstly select features from 3D models, such as
point to point distance, maximum or minimum depth, and round object radius. Then
we manually measured features and compare results from estimated values from 3D
models. 3D model manipulation is conducted in Rapidform (www.rapidform.com), a
commercial 3D scan and metrology software. Actually we estimated features 10 times
43
Chapter 6: Conclusions
for each comparison and then compute the mean values as the finally results. Table 10
lists our results and figure 18 explains experimental objects and their features.
One can observe that, the depth accuracy is worse than width and length
measurements. The measurement error is worse than manufactory specifications.
Vehicle dynamics add a large degree of noise into the measuring process.
Table 5.3: Optimal Distance Experimental Results
Experiment
Distance
(cm)
Placement
Mobile
Error upper
bound (mm)
Error lower bound
(mm)
RMS
(mm)

n
 (r
i 1
i
 r)2
n
31+parallel
31
Cart and static
14.7
-6.3
5.2
Cart and static
14.8
-5.1
4.9
Cart and static
13.3
-4.7
4.9
Cart and static
13.1
-7.0
5.1
Cart and static
11.7
-6.3
5.2
Cart and static
14.4
-3.6
5.7
73
Parallel to horizontal
plane
Parallel to horizontal
plane
Parallel to horizontal
plane
Parallel to horizontal
plane
Parallel to horizontal
plane
Parallel to horizontal
plane
Downward
51+parallel
51
76+parallel
76
101+parallel
101
152+parallel
152
203+parallel
203
79+downward
Cart and static
12.8
-5.1
4.1
95+downward
89
Downward
Cart and static
135
-8.5
4.3
Chapter 6: Conclusions
44
Chapter 6: Conclusions
45
Figure 5.3 Optimal distance experiment with distance vs. errors. The distance herein is
measured manually from sensor window center to the target surface. Errors are computed as RMS
(root mean square).
46
Chapter 6: Conclusions
Table 5.4 Laser-range, GPS, and IMU Measurement Error Estimation
Sensor
GPS
Experiment
Measurement
Outdoor static with 3,011
points obtained in 5 minutes
X (cm)
Y (cm)
Z (cm)
X and Y
Z
Roll (degree)
Pitch (degree)
Yaw (degree)
Roll, pitch
and yaw
(degree)
Range (cm)
Manufactory specifications
IMU
Outdoor static with 19,985
points obtained in 5 minutes
Manufactory specifications
Sick
Outdoor asphalt flat surface
Manufactory specifications
n
* RMS

r
i 1
ture
 restimated
n
Error upper
bound
+ 1.8
+ 2.4
+ 4.0
/
/
+ 0.5
+ 0.3
+ 0.5
/
Error lower
bound
- 1.8
- 2.5
- 3.0
/
/
- 0.3
- 0.2
- 0.3
/
*RMS
+ 3.0
+ 1.5
- 2.0
- 1.5
7.8
0.5
0.6
1.0
1.6
2
4
2.4
1.8
5.3
<1
47
Chapter 6: Conclusions
5.2
Video Mosaic Results
The main purpose of this part of work is to prepare a large range texture image
covering whole scanning area for 3D models. According to [Chen, 1998] mosaic is
divided into two parts, image registration and image merging. In this project we
proposed two methods to do image registration. The first registration process select
feature points and intensity difference based registration and this approach requires
selecting a set of feature points from reference image and then exhaustively
calculating position difference between them. According to [Szeliski, 1996], a
homograph transformation matrix can be set up between two images taken at
successive time. This matrix explains spatial translation and rotation from first image
pixel coordinates to the second image ones or vise versa.
Suppose in image frame i pixel p( x, y ) transforms to image frame i  1 with p ' ( x ' , y ' ) ,
the transformation between them can be expressed as below,
p ' ( x ' , y ' )  T ( R( p( x, y)))
(25)
Where
p ' ( x ' , y ' ) be the image projection of pixel p( x, y ) ;
T be the translation matrix;
R be the rotation matrix.
By solving (18) image frame i is resample into the image frame i  1 ’s local coordinate
system. The two successive image frames then is registered as a single large image
frame. Feature points are selected respectively from two frames. Then transformation
is solved to minimize spatial position difference function (19) between points and its
projection using rotation and transformation.
E( R, T )   ( x '  x) 2   ( y '  y) 2
(26)
To minimize the above summation, we could either apply gradient-based optimization
techniques nongradient-based techniques such as down hill Simplex. Here we use
Levenberg-Marquardt interactive nonlinear minimization [Lee, 2002] and thus
compute correspondence ( R, T ) in the target image.
48
Chapter 6: Conclusions
The second registration method is rather straightforward. It uses GPS and IMU
measurement to compute transformation matrix. The advantage of this method is its
efficiency, especially for images which don’t include features. The disadvantage is the
measurement errors generate not so accurate transformation as one computed from
matching features.
5.2.1
Texture Mapping
Texture mapping is to assign a 2D texture image coordinate to each 3D range point.
The texture image coordinate is then interpolated between range points. The
assignment process is to compute a transformation matrix between 3D model points
and image pixels. To do this calibration of the camera with respect to the model is
required. Calibration uses feature points in 3D models and their correspondences in
2D texture images, and then find the transformation matrix parameter. The
transformation projects a model point X laser  ( X , Y , Z ) , in 3D laser local coordinates,
onto a 2D image point xcamera  (u , v) as the below:
X 
u 
    M Y 
v 
 Z 
(27)
Actually transformation matrix is determined by cameral models used. Thus in turn
determines the number of types of parameter in M . If a pin pole camera model of
perspective projection is used, according to [Tsai, 1987], the transformation M has 11
parameters, in which 6 parameters denote the translation and rotation of the camera
with respect to the local laser coordinate frame. 5 parameters denote the optical
properties of the cameras, which are also camera intrinsic parameter. They include a
1st order approximation of a coefficient to correct radial distortion.
If we can find 7 feature points from 3D model and their correspondences in 2D texture
image, then the camera model parameters can be solved. Corresponding information
can be supplied by cracks. When more than the minimum numbers of correspondence
are found, we can use the RANSAC parameter estimation method described in
[Fischler, 1981]. The paper attempts to identify and discard erroneous data, and thus
use only the most accurate data to find a solution.
49
Chapter 6: Conclusions
Table 5.5 3D Modeling Errors
3D Models
Features
Manually
measured

Estimated from
models m (cm)
Error

m  m (cm)
m (cm)
Indoor Flat indoor with a long
wooden log
Log width
8.2
7.8
-0.4
Log thickness
4.0
5.1
+1.1
Outdoor flat asphalt with a rounded
plastic box
Radius
12.5
11.9
-0.6
Outdoor flat asphalt with a handicap
mark
Mark width
12.3
Mark length
12.4
Outdoor flat asphalt with a L shape
cutting
D1
69.3
75.3
+6.0
D2
106.0
110.9
+4.9
Max slope
3.0
8.9
+6.9
Manhole radius
66.4
63.2
+3.2
Grass shape
N/A
Can be measured
N/A
Outdoor grass and cutting
50
Chapter 6: Conclusions
(a) Indoor Flat indoor with a long wooden log
(b) Features: log with and depth
(c) Outdoor flat asphalt with a rounded plastic box
(d) Box radius
(e) Outdoor flat asphalt with a longitudinal cutting
(f) Min and max cutting depth
51
Chapter 6: Conclusions
D2
Slope
D1
(i) Outdoor flat asphalt with a L shape cutting
(j) Corner point to point distance
(k) Outdoor flat rugged surface with a big crack
(l) Min and max depth and width
Slope
(m) Outdoor grass and cutting
Figure 5.4: 3D models and their features for accuracy estimation.
D
(n) A manhole radius
Chapter 6: Conclusions
5.3
52
Examples and Discussion
We conduct 2 outdoor large scale experiments on campus to evaluate system stability
and capability. Post-processing is performed off line for obtaining 3D geometry
models and 2D texture information. The below describes the 2 data collection
processes in details.

The first data set is collected by driving a van along an asphalt roadway and then
turns into a mix surface parking lot. The test site is carefully selected therefore the
data have varied features which can be found on a normal pavement surface. These
features are such as cracks, cuttings, grass, holes etc. The measurement vehicle
ran a course of 419 m at a speed of 3 mph to 5 mph. A GIS map showing testing
site and the vehicle trajectory plot are presented in figure 20. The path covers an
area of 104 m x 321m. Totally 1,291 GPS points, 8,840 laser scans and 2,543
images are measured respectively as the vehicle moved forward. The spacing
between two closest scans is 3 cm to 6 cm. For each scan we select 313 from 361
points and 1D cubic interpolation is used to uniformly resample them up to 601
points. The spacing between 2 closet points is 1cm. The moving path is shown in
figure 19 and modeling results are shown in figure 20.

The second experiment is conducted in a parking lot. The measurement van drives
in a U shape curve. The test area covers a rectangular of 267 m long and 120 m
width. The 3D model is more than 20 Mbytes. We split it into 4 smaller pieces,
with each piece of 7 Mbytes (with redundancy). IVP ranger is also used for crack
modeling. For the ranger data is rather large we just use it as an aid in a small
cracking area. Figure 19 shows the experiment results.

Then we show a set of small models which can be normally found on pavement
surface. As a comparison IVP high resolution models are also shown. These high
resolution models are obtained by moving a cart not a driving a van instead. Their
results are shown in figure 22.

Finally video image results are shown in figure 23, figure 24 and figure 25. Figure
23 shows a large scale mosaic result. Figure 24 and figure 25 otherwise show
“crack slit” which can be used for crack detection and classification. Observing the
results we can notice that light illumination influence image quality significantly.
Choosing an optimal time and lighting condition are critical considerations for
texture image data acquisition. Table 13 summaries mosaic results.
53
Chapter 6: Conclusions
TABLE 13
IMAGE MOSAIC RESULTS
Figure
Figure 23 (a)
Figure 23 (b)
Figure 24 (a)
Figure 24 (b)
Figure 25
No. of Frames
45
45
2
3
4
Pixel size
754 x 4744
720 x 4840
735 x 638
738 x 811
734 x 951
(a)
(b)
Figure 5.5: Moving trajectory. (a) 2D plot; (b) Plot in a high resolution
satellite image of UT campus. The van moves on a roadway along Knoxville
river, and then turns right into a parking lot.
54
Chapter 6: Conclusions
104 m
372 m
(a)
(b)
(c)
(d)
Figure 5.6: A large scale data (a) Large scale 3D model; (b) 1st round zoom and show
more details; (c) The scale of 1st round zoom operation; (d) Colorcoded surface patch.
55
Chapter 6: Conclusions
(a)
(b)
(c)
(d)
(e)
(f)
Figure 5.7: Large scale (a) Large scale 3D model; (b) Trajectory plot; (c) The scale of
117.1m
1st;
(d) plot trajectory in GIS image; (e) 2nd round zoom to show small details about
roughness of surface obtained from IVP; (f) Small crack patch obtained from IVP.
26.2 m
56
Chapter 6: Conclusions
(a) grass
(b) handicapped mark
(c) manhole
(d) Texture model
57
Chapter 6: Conclusions
(g) IVP surface patch in a L curve
(f) IVP cracking slit
(g) IVP crack slit
Figure 5.8: Varied 3D models built from Sick TOF laser and
IVP ranger.
Chapter 6: Conclusions
Figure 5.9: Mosaic result. The right result is obtained from GPS/IMU pose
estimation. The left one is obtained from matching features.
58
59
Chapter 6: Conclusions
(a)
(c)
Figure 5.10: Mosaic result. (a) Using 2 frames; (b) Using 3 frames.
Chapter 6: Conclusions
Figure 5.11: Mosaic result using 4 frames.
60
Chapter 6: Conclusions
6
61
CONCLUSIONS
The paper shows fusing laser, GPS, IMU data can achieving accurate and eyeappealing 3D surface models. These models together with image information can be
used for many applications, including terrain modeling, pavement inspection, tunnel
inspection and even urban architecture reconstruction. The integrated approach can be
great potential in fast digitizing and representing physical world with less human
efforts. That is our integrated system is capable of scanning both terrain and its
accompanying structures, at the desired level of resolution. For example, the micro
system scanning configuration is appropriate of inspection tasks of pavement surface.
The macro system is efficient for larger scale terrain profiling with little human
intervention kilometers’s worth of data is captured and represented in real time within
a few minutes. The system has been developed modularly that is the system can be
reconfigured according to the application with minimal amount
of effort and changes to the processing.
Latter efforts will go on system accuracy improvement and obtaining eye-appealing
3D models. Some of efforts will be:

A decent and robust algorithm for fast triangulating range points and
representing flat road surface;

Improving the navigation estimation algorithm in the ways of de-nosing,
fusing and interpolation;

An accurate algorithm of enhancing small-scale detail regions in the presence
of noise. Such details are as cracks, cutting, and rocks.
62
Bibliography
BIBLIOGRAPHY
Bibliography
63
[Belkasim, 1991] S. O. Belkasim, M. Shridhar, and M. Ahmadi, “Pattern Recognition
with Moment Invariants: a Comparative Study and New Results,” Pattern
Recognition, Vol. 24, No. 12, 1991, pp. 1117-1138.
[Chen, 1998] C. Y. Chen, “Image Stitching - Comparisons and New Techniques,”
1998.
[Cheng, 1990] H. D. Cheng, C. Tong, and Y. J. Lu, “VLSI curve detector,” Pattern
Recognition, Vol. 23, No.1/2, 1990, pp.35–50.
[Cheng, 1999a] H. D. Cheng, J. R. Chen, C. Glazier, and Y. G. Hu, “Novel Approach
to Pavement Cracking Detection Based on Fuzzy Set Theory”, J. of Computing in
Civil Eng., Vol. 13, No. 4, October 1999, pp. 270-280.
[Cheng, 2003] H. D. Cheng, X. J. Shi, and C. Glazier, “Real-Time Image
Thresholding Based on Sample Space Reduction and Interpolation Approach”, J. of
Computer in Civil Eng., Vol. 17, Issue 4, October, 2003, pp. 264-272.
[Cross, 1983] G. Cross and A. Jain, "Markov Random Field Texture Models," IEEE
Trans. PAMI, Vol. 5, No.1, 1983, pp.25-39.
[Feijoo, 1991] C. Feijoo, A. Asesio, and F. Perez, “New Practical Method for
Measurement Error Determination in Radar Systems under Real Traffic Conditions,”
IEEE Proc. –F, Vol. 138, No. 6, Dec. 1991, pp. 525 – 530.
[Feijoo, 1992] C. Feijoo, F. Perez, and A. Asesio, “Real Time Estimation of GPS
Measurement Errors from Vehicle Position Data,” Vehicle Navigation and
Information Systems, VNIS., The 3rd International Conference on September 2 - 4,
1992, pp. 369 – 374.
[Fischler, 1981] M. A. Fischler, and R. C. Bolles, “Random Sample Consensus: A
Paradigm for Model Fitting With Applications to Image Analysis and Automated
Cartography,” Communications of the ACM, Vol. 24, No. 6, 1981, pp. 381-395.
[Fu, 1981] K. S. Fu and J. K. Mui, “A Survey on Image Segmentation,’’ Pattern
Recognition, Vol.13, 1981, pp.3–16.
[Fukuhara, 1990] T. Rukuhara, K. Terada, M. Nagao, A. Kasahara, and S. Ichihashi,
“Automatic Pavement-Distress-Survey System,“ J. of Transp. Eng., Vol. 116, No.1,
January, 1990, pp. 280-286.
Bibliography
64
[Fundakowski, 1991] R. A. Fundakowski et al., “Video Image Processing for
Evaluating Pavement Surface Distress”, Final Report, National Cooperative Highway
Research Program1-27, September, Washington, D.C., 1991.
[Goldberg, 1989] D. E. Goldberg, “Genetic Algorithm in Searching, Optimization, and
Machine Learning,” Addison-Wesley, Reading, Mass, 1989.
[Gonzalez, 1992] R. C. Gonzalez and R. E. Woods, “Digital image processing”, 3rd
Ed., Addison-Wesley, Reading, Mass, 1992
[Grinstead, 2005] B. .Grinstead, A. Koschan, D. Page, A. Gribok, and M. A. Abidi,
“Vehicle-borne Scanning for Detailed 3D Terrain Model Generation,” SAE
International, 2005.
[Groeger, 2003] J. L. Groeger, P. Stephanos, P. Dorsey, and M. Chapman,
“Implementation of Automated Network-level Crack Detection Processes in
Maryland,” Transportation Research Record 1860, Transportation Research Board,
National Research Council, Washington, D.C., 2003, pp. 109-116.
[Guo, 2002] L. Guo, Q. Zhang and S. Han, “Position Estimate of Off-Road Vehicles
Using a Low-Cost GPS and IMU,” 2002 ASAE (The Society for Engineering in
Agriculture, Food and Biological System) Annual International Meeting, July, 2002.
[Hass, 1990] C. Haas, S. McNeil, "Criteria for Evaluating Pavement Imaging
Systems," Transportation Research Record, No. 1260, 1990.
[Howe, 1997] R. Howe, and G. G. Clemena, “Assessment of the Feasibility of
Developing and Implementing an Automated Pavement Distress Survey System
Incorporating Digital Image Processing,” Final Report for Virginia Department of
Transportation, Report No. VTRC 98-R1, Project No. 9128-010-940, Virginia
Transportation Research Council, Charlottesville, VA, November, 1887.
[Hu, 1962] M. K. Hu, “Visual Pattern Recognition by Moment Invariants,” IRE
Transactions on Information Service, Vol. 1, IT-8, February, 1962, pp. 179-187.
[Hu, 2000] R. Hu and M. M. Fahmy, “Texture Segmentation Based on a Hierarchical
Markov Random Texture Field,” Signal Processing, No. 26, Vol.3, March, 1992, pp.
285-305.
[Javidi, 2003] B. Javidi, “Pilot for Automated Detection and Classificationo of Road
Surface Degradation Features,” Report for Project JHR 03-293, Optical/Imaging
Processing Laboratory, The University of Connecticut, November, 2003.
Bibliography
65
[Javidi, 2004] B. Javidi, D. Kim and S. Kishk, “A Laser-Based 3D Data Acquisition
System for the Analysis of Pavement Distress and Roughness,” Report for Project
JHR 03-300, Optical/Imaging Processing Laboratory, The University of Connecticut,
December, 2004.
[Laurent, 1997] J. Laurent, M. Talbot, and M. Doucent, “Road Surface Inspection
using Laser Scanners Adapted for the High Precision Measurements of Large Flat
Surfaces”, International Conference on Recent Advances in 3-D Digital Imaging and
Modeling (3DIM '97), Ottawa, Canada, Dec., 1997.
[Lee, 2002] S. C. Lee and P. Bajcsy, “Multi-Instrument Analysis from Point and
Raster Data”, Tech. Report, No. alg04-02, National Center for Supercomputing
Applications, Champaign, IL, 2002.
[Leica, 2003] Leica AG. Heerburgg, “Outside World Interface (OWI) Interface
Control Document, Project GPS System 500,” No. TUH 3421, May 2003.
[Li, 1991] L. Li, P. Chan, A. Rao, and R. L. Lytton, “Flexible Pavement. Distress
Evaluation Using Image Analysis,” Proc., 2nd Int. Conf. on Applications of Advanced
Technologies in Transp. Engrg., 1991, pp. 473-477.
[Lu, 2002] J. J. Lu, X. Mei and M. Cunaratne, “Development of an Automatic
Detection System for Measuring Pavement Crack Depth on Florida Roadways,” Final
Research Report to Florida Dept. of Transp., Department of Civil and Environment
Engineering, University of South Florida, January, 2002.
[Luhr, 1999] D. R. Luhr, "A proposed methodology to Quantify and Verify
Automated Crack Survey measurements," the 78th Annual Transportation Research
Board Meeting, Washington, D.C., January, 1999.
[Meignen, 1997] D. Meignen, M. Bernadet,, and H. Briand, “One Application of
Neural Network for Detection of Defects Using Video Data Bases: Identification of
Road Distress,“ J. IEEE Computer Society, 1997, pp. 459-464.
[McGhee, 2004] K. H. Mcghee, “Automated Pavement Distress Collection
Techniques – A Synthesis of Highway Practice,” Report for National Cooperative
Highway Research Program (Synthesis 334), Transportation Research Board of the
National Academies, 2004.
[Monti, 1995] M. Monti, “Large-area laser scanner with holographic detector optics
for real-time recognition of cracks in road surfaces,” Optical Engineering, Vol. 34, No.
7, Brian J. Thompson, Ed., 1995, pp. 2017-2023.
Bibliography
66
[Nassar, 2003] S. Nassar, K. P. Schwarz, A. Noureldin and N. E. Sheeimy, “Modeling
Inertial Sensor Errors Using Autoregressive (AR) Models”, Proc. ION NTM-2003,
Anaheim, USA, January 22-24, 2003, pp.116-125.
[Ritchie, 1990] S. G. Ritchie, “Digital Imaging Concepts and Applications in
Pavement Management,” J. of Transp. Eng., Vol. 116, No.1, January, 1990, pp. 287298.
[Rumelhart, 1986] D. E. Rumelhart, G. E. Hinton, and R. J. Williams, “Learning
Representations by Back-Propagating Errors,” Nature, Vol. 323, No. 9, 1986, pp. 533536.
[Sheimy, 2004] N. El-Sheimy and N. Nassar, “Wavelet De-Nosing for IMU
Alignment,” IEEE Aerosp. Electron. Syst. Mag., Vol.19, No.10, October 2004, pp.32 39.
[Sick, 2002] Sick AG, “LMS 200 / LMS 211 / LMS 220 / LMS 221 / LMS 291 Laser
Measurement Systems Technical Description,” Division Auto Indent., Germany, 2002.
[Simon, 1998] D. Simon, “Kalman Filtering”, Innnovatia software, 1998.
www.Innovatia.com/software
[Skaloud, 1999] J. Skaloud, “Optimizing Georeferencing of Airborne Survey Systems
by INS/GPS,” Ph.D. Thesis, Dept. of Geometrics Engineering, Univ. of Calgary,
Calgary, Alberta, Canada, UCGE Reports No. 20126.
[Standard, 2001] Standard Practice for Quantifying Cracks in Asphalt Pavement
Surface, AASHTO Designation PP44-01, American Association of State Highway and
Transportation Officials, Washington, D.C., Apr. 2001.
[Szeliski, 1996] R. Szeliski, “Video Mosaics for Virtual Environment”, IEEE
Computer Graphics and Applications, vol. 16, No. 2, March, 1996.
[Tao, 1994] S. Tao, N. Kehtarnavaz, and R. Lytton, “Image-Based Expert-System
Approach to Distress Detection on CRC Pavement,” J. Transportation Engineering,
Vol. 120, No. 1, January/February 1994, pp. 52-64.
[Tao, 1998] C. V. Tao, “Mobile Mapping Technology for Road Network Data
Acquisition,” International Symposium of Urban Multimedia/3D Mapping, Tokyo,
Japan, June 8-9, 1998.
Shape
Histograms
Bibliography
67
[Teague, 1993]
M. R. Teague, “Image Analysis via the General Theory of
Moments,” Optical Society American Journal, Vol. 70, No. 8, August, 1993, Aviation
Administration, the University of Wisconsin-Madison, 2004, pp. 920-930.
[Tsai, 1987] R. Y. Tsai, "A versatile Camera Calibration Technique for HighAccuracy 3D Machine Vision Metrology Using Off-the-Shelf TV Cameras and
Lenses", IEEE Journal of Robotics and Automation, Vol. RA-3, No. 4, 1987, pp. 323344.
[Walker, 2004] D. Walker, “Pavement Surface Evaluation and Rating – Asphalt
Airfield Pavements,” Manual for the Federal
[Wang, 2003] K. C. P. Wang, “Transportation Research Circular: Automated Imaging
Technologies for Pavement Distress Survey”, Committee A2B06, Transportation
Research Board, National Research Council, Washington, D.D., draft, June, 2003.
[Ye, 2002] C. Ye and J. Borenstein, “Characterization of a 2-D Laser Scanner for
Mobile Robot Obstacle Negotiation,” Proceedings of the 2002 IEEE International
Conference on Robotics and Automation, Washington DC, USA, May, 2002, pp.
2512-2518.
[Zadeh, 1968] L. A. Zadeh, “Probability Measures of Fuzzy Events,” J. Mathematical
Anal. and Applications, Vol. 23, 1968, pp. 421–427.
[Zgonc, 1996] K. Zgonc K and J. D. Achenbach, “A neural network for crack sizing
trained by finite element calculations”, NDT&E Int., Vol. 29, 1996, pp. 147-55.
[Zhao, 2003] H. Zhao and R. Shibasaki, “Reconstructing a Textured CAD Model of
an Urban Environment Using Vehicle-Borne Laser Range Scanners and Line
Cameras”, Machine Vision and Applications, Vol. 14, No. 1, April, 2003, pp. 35 - 41.
Download