THE AUTONOMOUS MINI HELICOPTER: A POWERFUL PLATFORM FOR MOBILE MAPPING

advertisement
THE AUTONOMOUS MINI HELICOPTER:
A POWERFUL PLATFORM FOR MOBILE MAPPING
Henri Eisenbeiss
Institute of Geodesy and Photogrammetry, ETH Zurich, CH-8093, Zurich, Switzerland, +41 44 633 32 87
ehenri@geod.baug.ethz.ch
Commission Ⅰ, ICWG I/V
KEY WORDS:
Acquisition, Automation, Integration, Mobile, Mapping, Processing, UAV
ABSTRACT:
In this paper, the developments related to an autonomous airborne mobile mapping system for photogrammetric processing will be
presented. During the last years the author has been involved in several projects related to mobile mapping using an autonomously
flying model helicopter, a so-called mini UAV (Unmanned Aerial Vehicle). The overall motivation of using mini UAVs for mobile
mapping, the developed workflow for UAV-data processing, the current status of the work, and recent developments related to
specific applications are described. In a first step our mini UAV-system and our approach are explained and the need for specific
developments is highlighted. In the following two applications will be described: The Randa rockslide and the maize project,
demonstrating our developments for the automation of the photogrammetric workflow using the mini UAV system. In the Randa
project a unique flight planning tool for UAV-monitoring of mountainous areas was developed. The tool allows for the adaptation of
the flight in a way, that image data with 2-3 cm resolution of the cliff could be acquired. Moreover, due to this high image resolution,
a DSM with a higher point density compared to standard airborne and helicopter-based LIDAR-systems, could be generated. In the
second project the focus was on the automation of the image orientation process. Hence, two test areas (A and B) were defined.
While in A two commercial photogrammetric workstation were evaluated, in B the data set was processed using GPS/INS data as
initial values for the image orientation. Therefore, the manual input for the aerial triangulation was reduced to a minimum user
interaction. Finally, with the oriented image data two DSMs for the analysis of the outcrossing in maize were generated. Because of
these developments, the tools for the planning and the data acquisition as well as the workflow for the processing of UAV-images
were automated. However, the complete workflow still requires manual interaction, which will be discussed here in detail including
a proposal for future developments.
1.
(Bendea et al., 2007) and, in contrast to Microdrones (Nebiker et
al., 2007), more stable against environmental conditions like wind.
The developments of model helicopters and comparable
autonomous vehicles are primarily driven by the artificial
intelligence community (AAAI, 2008) and have been used mainly
in the past for military applications with increasing use in the
civilian sector.
INTRODUCTION
Throughout the last years, UAVs (Unmanned Aerial Vehicles)
have become more and more suitable platforms for data
acquisition in photogrammetry and mobile mapping. “UAVs
are to be understood as uninhabited and reusable motorized
aerial vehicles.” (van Blyenburg, 1999). These vehicles are
remotely controlled, semi-autonomous, autonomous, or have a
combination of these capabilities. The term UAV is used
commonly in the computer science, robotics and artificial
intelligence communities. Supplementary, in the literature also
synonyms like Remotely Piloted Vehicle (RPV), Remotely
Operated Aircraft (ROA) and Unmanned Vehicle Systems
(UVS) can be found. The definition of UAVs encompasses
fixed and rotary wings UAVs, lighter-than-air UAVs, lethal
aerial vehicles, aerial decoys, aerial targets, alternatively
piloted aircrafts and uninhabited combat aerial vehicles.
Sometimes, cruise missiles are also referred to as UAVs (van
Blyenburg, 1999).
In the past, model helicopters were already used in
photogrammetric applications (Eisenbeiss, 2004). However at that
time, the model helicopters were controlled manually via radio
link. Nowadays, these new technologies allow low cost
navigation systems to be integrated in model helicopters, enabling
them to fly autonomously. This kind of autonomous flying model
helicopter is called mini UAV system (Blyenburg, 1999;
Eisenbeiss 2004). These mini UAVs are highly manoeuvrable due
to the possibility of hovering, change of flight direction around
the center of rotation as well as the capability for turning the
mounted camera in horizontal and vertical direction. However,
due to the difficulty of keeping the ideal position and attitude, the
vibration of the helicopter and the manual planning of image
acquisition points, model helicopters have not been used
successfully in the past for measurements, precise modeling and
mapping of objects (Eisenbeiss, 2004). Latest developments
integrate GPS/INS (Global Positioning System / Inertial
Navigation System) together with a stabilization platform for the
camera. Because of the small size and the low payload, the
selection of the installed hardware is mostly limited to low cost
navigation systems with low precision. Nevertheless, the
However, a model helicopter was selected for our
investigations. Model helicopters are clearly defined by the
Unmanned Vehicle Systems (UVS) International Association as
mini, close short and medium range UAVs depending on their
size, endurance, range and flying altitude (UVS, 2008).
In contrast to standard airplanes, model helicopters are able to
operate closer to the object. In addition, these systems are
highly flexible in navigation compared to fixed wing UAVs
977
The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences. Vol. XXXVII. Part B1. Beijing 2008
parameters. Additionally, the interfaces have the functionality for
data transformation and formatting. Therefore, with our proposed
workflow several UAV applications can be handled. The main
modules are defined as follows: Project parameters (PP), Flight
planning (FP), Autonomous photogrammetric flight (APF),
Quality check of the data (QCD), UAV Block Triangulation
(UAV BT), DSM, orthophoto, 3D Model (DO3D). The individual
procedures of a module can be a process or a part of a process
from commercial or in-house developed software.
Figure Figure 1 show which module are or will be developed by
IGP and which focus more on the evaluation of existing
commercial or in-house developed software packages.
Additionally, figure 1 explains which part of the work will be
possible during or after the field work or both.
combination of GPS/INS sensors with image data for
navigation allows more precise and reliable results.
Furthermore, the integration of GPS/INS and image data in a
real time triangulation method will drastically reduce time and
cost required for post processing.
Mini UAVs have been used recently for civilian applications
like 3D city modeling (Wang et al., 2004), modeling of a
medieval castle in Sarnen (Pueschel et al., 2008), Horcher and
Visser (2004) proposed applications in forestry like BMP
Inspections, road maintenance of forest roads and trespas.
Further on, UAVs were used in applications related to
agriculture (Herwitz et al., 2004; Rovira-Más et al., 2005;
Eisenbeiss, 2007; Reidelstuerz et al., 2007; Rovira-Más et al.,
2008) and for the documentation of an archaeological site
(Lambers et al., 2007). In these projects the focus was on a fast
data processing, which reduces the accuracy of the results or in
contradiction the results had high accuracy, while the
processing cost still a lot of manual effort.
In the following, two developments related to real applications of
such a mini UAV-system, which lead to the automation of the
workflow, will be presented. Before describing the developments
in detail, our mini UAV-system and its ground control station will
be described. After all, the outcome of the developments and the
direct benefits using an UVS system in the two specific
applications are discussed. Finally, we will propose a new UAVplatform, which can lift a bigger payload.
Hence, our goal is the automation of the complete workflow in
a way that the manual input can be reduced to an acceptable
minimum. A further objective of our work is the high reliability,
precision and resolution of the photogrammetric products like
elevation models, orthophoto and textured 3D models.
Therefore, the aim is to develop procedures for the UAV data
acquisition and processing that can lead to a highly automated
workflow and accurate complete results (see Figure 1).
2. AUTONOMOUS MINI HELICOPTER
The mini UAV-system Copter 1b (see Figure 4 and Table 1) was
developed by the French company Survey-Copter. It is equipped
with an on-board navigation system (wePilot 1000) from
weControl. The wePilot includes a GPS (Global Positioning
System), an INS (Inertial Navigation System), which allows the
mini UAV to perform attitude stabilization, position control and
to navigate autonomously. The private company weControl
GmbH is a spin-off of ETH Zurich that specialized in the
development of miniature flight control systems for UAV-systems.
Mini UAV-system Copter 1b
Length
2m
Rotor diameter
1.8m
Maximum takeoff weight
15kg
Payload capacity
5kg
Flight endurance
Max. 45min
Altitude
1500m
Range
5km
Table 1: Specifications of the mini UAV Copter 1b
(SurveyCopter, 2008).
The navigation system features the following main characteristics:
an altitude stabilization and velocity control, position and RC
transmitter sticks interpreted as velocity commands, integrated
GPS/INS system, altimeter, magnetometer, payload intensive
flight controller, built-in data logger and telemetry capability,
programmable hardware for rapid customization and an embedded
computer system. Furthermore, the system consists of a ground
control station (a laptop with monitoring software (weGCS)), a
convertible gimbal for still video cameras like Canon EOS D10,
D20 and the Nikon D2Xs, communication links, power supply,
video link (incl. video camera) for visual control for monitoring
image overlap, and transport equipment.
Figure 1: Modular Workflow for processing of UAV data
showing the main functionalities.
All processing steps of the workflow are categorized in
modules, which communicate with each other via interfaces.
The interfaces are established in a way that for the individual
module the attributes can change, depending on the project
978
The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences. Vol. XXXVII. Part B1. Beijing 2008
The ground control station also includes a flight simulator,
which allows the simulation of the real flight to verify the
predefined flight path (see Figure 2).
3.
3.1
Therefore, a high-resolution digital surface model (DSM) with a
spatial resolution of 10–20cm had to be generated and oriented
images with a footprint of 2-5cm in object space had to be
acquired. Because of the non-nadir case and a difficult terrain, the
standard aerial flight planning procedures could not be applied.
RANDA
Furthermore, the sight was acquired by a manned helicopter in
November 2007, using the Helimap system, combining laser
scanning and oblique images (Vallet, 2007). The system is based
on a Riegl LMS-Q240i-60 laser scanner for point cloud
generation and a Hasselblad H1 camera with a Dos Imacon
Xpress 22Mpix back, which is normally used for texture mapping.
Project Aims
The focus of the geological project Randa is to characterize the
large-scale rockslide in Randa (Wallis, Swiss Alps; see Figure
3). This rockslide will be used as an example of the ultimate
evolutionary stages corresponding to the complete development
of the basal shear surface(s) and internal rock mass deformation.
In order to understand this failure stage in a better way, the
interrelationships between pre-existing tectonic fractures,
subsequent tensile fracturing and shearing of intact rock bridges
will be investigated. Furthermore, for the understanding of the
rock mass structure and kinematics during this evolutionary
stage, the basal shear planes must be identified, the internal
rock mass deformation has to be further studied, and the 3D
time-dependent displacement field has to be measured more in
detail using the UAV-image data and the generated DSM
(Randa, 2008).
Using the two comparable systems, the autonomous flying model
helicopter and the manned helicopter, at one site, allowed the
analysis of the performance of the systems. Moreover, it is
possible to compare and integrate the two data sets.
3.2 Flight planning
Before doing the flight, the main parameters of the autonomous
flight were defined. For the Randa rockslide we decided to use the
Nikon D2Xs, which has a CMOS-sensor with 4288x2848 pixels
using a 50mm lens (Nikon, AF NIKKOR 50mm 1:1.8D) and a
pixel size of 5.5µm. For the recognition of features with a length
of 10-20cm, an image scale of 1:4500 was selected which results
in a pixel footprint of approximately 3cm. The distance to the cliff
was defined to 230m, which is equal to the normally used flying
height above ground. Finally, the side and end lap were set to
75%. Using this high overlapping along and across strip it was
possible to avoid occlusions and gaps in the image data (see
Figure 6). The flight velocity was defined to 3m/s, while the
shutter speed was 1/1000s. Therefore, the influence of the image
motion was negligible.
After defining these parameters, the most recent elevation model
with the highest available point density was used. Therefore, the
LiDAR data provided by swisstopo was selected. Using this data
set and the available orthophoto (swissimage, swisstopo®)
allowed the definition of the area of interest. Hence, the area was
separated in three sections. The sections were selected using the
inclination and aspect values of the slope.
Figure 2: Screenshot of the ground control station software
showing the flight lines of the Randa data acquisition. In offline mode the software can be used for the simulation of the
flight. The orange shading shows the lower part of the Randa
rock cliff.
Figure 4: Our model helicopter during a test flight with the
inclined camera looking perpendicular to the flight direction.
For each particular section a plane was defined in a way that the
average 3D distance of the surface to the plane was reduced to a
minimum. While looking perpendicular to the surface, we
assumed to reduce the occlusion in the images. After the
definition of the plane equation, the normal of the plane was
Figure 3: The Randa rockslide failure (Wallis, Switzerland).
The lower part is situated 1400 a.s.l. The height difference from
the bottom to the top is more than 800m.
979
The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences. Vol. XXXVII. Part B1. Beijing 2008
calculated and with the given flying distance to the surface a
parallel plane was defined (see Figure 5). In the new plane the
area of interest was projected and the inclined flight lines for
the helicopter flight were generated. Finally, using this
information the image acquisition points were determined and
transferred to the navigation unit (wePilot1000) of the mini
UAV-system. Due to the inclination of the plane the camera
was mounted with a pitch angle of 15-30° (respectively for the
individual sections) on the helicopter (see Figure 4).
3.3
Field Work
Since the Randa rockslide covers a height from 1400m -2300m
a.s.l. the mini UAV-system was tested at the Flueelapass (~2400m
a.s.l., Graubuenden, Switzerland). For the flight at Flueelapass
weight dummies instead of the mounted camera were used on the
helicopter. To enable the evaluation of the performance at such a
height above mean sea level, the helicopter was started and
switched to the assisted flying mode. In this mode the current
parameters of the engine were checked visually and saved on
board of the helicopter. This specific test showed that the engine
of our mini UAV-system already achieved the limit for the
maximal turning moment of the engine. Therefore, to have a
buffer, we decided to do the first flight only at the lower part of
the rockslide (1400-1800m a.s.l., see Figure 3), while the upper
part was also covered by the Helimap flight.
Such precise planning was necessary due to the large distance
between operator and mini UAV, which limited the visibility of
the mini UAV. Furthermore, before flying along the predefined
flight path, the mini UAV was tested at an altitude of 2400m
a.s.l. in order to ensure that the engine of the helicopter would
not fail due to low air pressure. This test was important, since
the helicopter manufacturer set the maximum flying altitude to
1500m a.s.l. In the following, the results from the first flight in
Randa and an elevation model extracted from the acquired data
will be described.
Since the illumination conditions were acceptable only in the
morning, we decided to do the flight at Randa in the morning
hours. In the afternoon strong shadows made the data processing
more complicated, while a preprocessing of shadow areas had to
be accomplished. Furthermore, the GPS-satellite availability was
simulated in the preparation process. Thus, the existing elevation
model of the surrounding (DHM25, swisstopo®) was integrated
in a GPS sky plot software, which allowed the calculation of the
accessible number of GPS and the postulated GDOP (Geometric
Dilution Of Precision) value. Using this information, it was
possible to have an approximate a-posteriori value for the
accuracy of the GPS-position of the helicopter, which was crucial
in such an environment.
Given that at the bottom the site a large debris zone is situated
(see Figure 3), the accessibility of the site is quite complicated.
Therefore, the start and landing point was defined close to a road
at the bottom of the debris zone (see Figure 2). The field work
itself was done in few minutes, while the longest part of the flight
was the way to go from the starting point to the first image
acquisition point and the way back. At the end of the first flight,
the helicopter did a fast in-explainable turn along its own main
axis. Hence, we decided to stop the test and to evaluate the flight
data, having already the first high resolution images from the
rockslide.
Figure 5: Zoom-in to one section of the cliff showing the
LiDAR-DSM with the approximated and parallel plane from
the flight planning.
Figure 6: Left: Derived surface model from image matching, Middle: Zoom-in of an UAV-image, Right: Point cloud of Helicopterbased LiDAR projected on the derived surface model.
980
The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences. Vol. XXXVII. Part B1. Beijing 2008
3.4 Results
4.2 Field work
Using the acquired UAV-images, a DSM of the lower part of
the Randa rockslide with 15cm resolution was generated (see
Figure 6). The complete site was documented with the Helimap
system. For the Helimap system, the pixel size was 6cm to 8cm.
In comparison to the flight defined for the mini UAV-system,
the Helimap flight was controlled by a pilot, who followed the
flight lines with an accuracy of 30m. Therefore, for the whole
area the scale varies depending on the complexity of the terrain.
Furthermore, the Helimap system acquired the cliff with an
oblique field of view, which evokes the gaps through occlusion
in the data set. For the laser scan, the final point density was
approximately 3pt/m2 (see Figure 6).
After anthesis, digital pictures were captured with a still-video
camera (Canon D20) mounted on our mini UAV-system
(Eisenbeiss, 2007).
In the first step, a flight planning was performed for the
autonomous flight. The image resolution was defined to have 3
cm per pixel in object space. Furthermore, the maximum flying
height was set to 100m above ground, due to the nearness to
Zurich airport. With the selected camera the remaining parameters
were calculated (see Table 2).
After definition of the flight parameters, the area of interest was
roughly identified in the Swiss National Map 1: 25 000 provided
by swisstopo. With the defined flight parameters, a trajectory was
calculated by starting at one corner of the field and 3Dcoordinates of the flight trajectory were established. The
autonomous flight was done using our UAV system. Following
the flight trajectory in the stop mode for experiment A, the
camera captured the images automatically on the predefined
positions in the cruising flight mode for experiment B. In the stop
mode the helicopter flew autonomously to each image acquisition
point and the operator triggered the image via radio link, while
during the cruising mode the image acquisition was completely
autonomous.
The visual comparison of the extracted UAV-DSM and the
LiDAR-DSM shows clearly that the fine structure of the cliff
could be modeled out of the UAV-DSM, while the LiDARDSM shows big holes and less resolution (see Figure 6).
The analysis of the unexpected turning of the UAV-system
during the flight showed that the effect was caused by the fact
that the neighboring points of the flight plan were defined to
close to each other. Furthermore, the effect evoked while the
neighboring flight lines had a distance in horizontal and vertical
direction, which caused a miss-interpretation in the flight
control software. Hence, during the flight, one predefined data
acquisition point was skipped and the turning occurred during
the flight at Randa. Therefore, the flight control software was
improved to handle autonomous flights flown with this type of
configuration.
4.3
Due to the autonomous flight with stop points we could capture
all images for experiment A in 20 minutes. By using the cruising
flight modus we reduced the flight time for experiment B to 5
minutes.
4. MAIZE FIELD
4.1
Project Aims
For the image triangulation for experiment A 2-3 tie points were
measured manually, while for the experiment B the orientation
values calculated using the Kalman filter implemented in the
navigation unit on board the mini UAV were used. For 5-10
percent of the images the 2-3 manually measured points were not
enough, since during the image acquisition the light breeze moved
the leaves of the plants slightly. Therefore, the measurement of
more manual points was essential (Eisenbeiss, 2007). The
manually measured points (experiment A) and the approximate
positions of the camera (experiment B) served as initial values for
automated tie point extraction. For the generation of the tie points,
the standard procedure implemented in LPS (Leica
Photogrammetry Suite) was used. Therefore, the initial
approximations of each image location within the projects had to
be defined. Hence, for area B the function for exterior orientation
parameter input and for area A the function for manual tie point
measurement were selected. After the generation of tie points, the
control points were manually measured in the images.
The goal of this study was to analyze the influence of terrain
characteristics on pollen dispersal in maize. For that application,
the task was to generate 3D elevation models as well as to use
the oriented images for a stereoscopic inspection of maize
fields, and to combine them with high resolution crosspollination data. The elevation model and the stereoscopic
measures were used to determine the heights of pollen donor
and receptor plants with respect to their absolute orthometric
height. For these investigations, two test areas (2005: A and
2006: B) were defined.
Parameter
Value
Image scale
1:4000
Side / end lap
75% / 75 %
Flying height above ground
~ 80 m
Camera
Canon EOS 20D
Focal length (calibrated)
Pixel (Image format)
20.665, RMSE 1.5e-003 mm
8.25 megapixels (3520x2344)
Flying velocity
3 m/s
Data Processing
Since the maize fields had a highly repetitive structure in the
image, it was necessary to detect blunders before doing image
orientation. In LPS, a blunder can be detected only by manual
checking until the project has proceeded to the point where a
bundle adjustment is possible. Therefore the blunder detection
with ORIMA (Leica Geosystems) integrated in LPS and ISDM
(Image Station Digital Mensuration, Intergraph) was attempted.
These software packages allow a relative orientation for each
stereo pair, multi-image blocks and the detection of blunders.
Finally, with both software packages the blunders were detected
and the orientation of the UAV-images was performed with an
Table 2: Flight parameters for the image observation of the
maize field in 2005 and 2006
981
The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences. Vol. XXXVII. Part B1. Beijing 2008
comparable fixed wing systems due to the system characteristics.
Therefore, the used autonomous flying helicopter system can be
introduced in several applications like crop maps for biomass
measurement, as input for tractor-automated guidance (Kise et al.,
2005), in forestry, for monitoring and detection and prediction of
erosion using the extracted image data, orthophotos and elevation
model. Furthermore, the mini UAV-system can be used for 3D
city modeling in combination with an autonomous car (Lamon et
al., 2006).
sigma 0 a posteriori of approx. 0.5 pixel and the RMS error for
the control points was below 1dm for both test areas.
4.4 Results
After image orientation, a DSM with a resolution of 10 cm was
produced automatically with the in-house developed software
package SAT-PP (Satellite imagery Precise Processing). The
software was initially developed for the processing of satellite
images and later adapted such that it was capable to handle
still-video camera and aerial images. Due to the combination of
the multi image approach and the combination of different
matching methods (area, feature and edge matching) (Zhang,
2005) allowed the generation of a precise, reliable and
consistent 3D model of the maize field. Finally, an orthophoto
with a footprint of 3 cm was produced using the generated
DSM and integrated into a GIS for further analysis (see
Figure 7).
In comparison to the mobile stereovision system, proposed by
Rovira-Mas et al. (2005) and Rovira-Mas et al. (2008), our
method allows an autonomous documentation with the UAV
system by producing accurate and reliable elevation models and
orthoimages of the site. The produced data can be directly
integrated into several GIS systems using absolute coordinates.
Therefore, the data can be easily combined with any other
reference data. Furthermore, by using medium format still-video
cameras, the image and the derived DSM resolution allows
analysis at a larger scale as with stereovision systems using the
same flight height. Therefore, large sites can be documented
completely in a shorter time period with our method.
Orientation values were calculated by using the Kalman filter
implemented in the navigation unit on board the mini UAV as
initial exterior orientation parameters for the photogrammetric
processing for experiment B; we could reduce the manual effort
during the photogrammetric processing. However, the complete
workflow still requires manual measurements for the control
points and statistical analysis of the results. This lack of
automation has to be the emphasis for future work by
development of online-triangulation methods and the automated
measurements of control points using coded targets.
In future work, we will also focus on the combination of LiDAR
and image data. Since our model helicopter has a maximum
payload of 5kg, a more powerful UAV-system (Aeroscout Scout
B2-120; Aeroscout) will be used, which also features the flight
control system of weControl. The combination of LiDAR and
image data allows us to use the strength of both systems: the laser
scanner is good in areas with less texture, while image data is
advantageous for edge measurement and texture mapping.
Figure 7: Screenshot of the 3D-model of experiment B (2006)
using the Swissimage orthoimage (swisstopo®) draped over the
DHM25 (swisstopo®) combined with the orthoimage of the
maize field and the position of the cross-pollination data.
5.
CONCLUSIONS AND FUTURE WORK
ACKNOWLEDGEMENTS
In this study, a workflow for the processing of UAV data was
presented. The processing steps of the workflow are based on
modules, which allow the independent adaptation of the
individual steps.
The author thanks Kirsten Wolff and Martin Sauerbier for their
contribution to the paper and Manuel Kaufmann for his input to
the Randa project.
The autonomous flight allowed us to predefine the flight
trajectory for an optimal photogrammetric processing. The
flight trajectories for both projects were calculated
automatically by in-house developed software using existing
topographic maps, orthophotos and elevation models. For the
Randa project, the flight planning was modified to the nonnadir case. Two autonomous flying modes were applied in the
maize field project: stop modus and cruising modus. By using
the cruising mode the flight time could be reduced by 75
percent for the observation of one field. The performance of the
system during the cruising mode is comparable with fixed-wing
UAV systems (Horcher and Visser, 2004), while the stabilizer
system of the helicopter allowed a continuous and stabilized
flight. In addition, the helicopter system is more flexible than
REFERENCES
AAAI. 2008. American Association for Artificial Intelligence.
http://www.aaai.org (accessed May 8th 2008).
Bendea, H. F., Chiabrando, F., Tonolo, F. G., Marenchino, D.,
2007. Mapping of archaeological areas using a low-cost UAV the
Augusta Bagiennorum Test site. XXI International CIPA
Symposium, Athens, Greece.
Eisenbeiss, H., 2004. A mini unmanned aerial vehicle (UAV):
system overview and image acquisition. International Archives of
Photogrammetry. Remote Sensing and Spatial Information
Sciences. 36(5/W1). (on CD-ROM).
982
The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences. Vol. XXXVII. Part B1. Beijing 2008
Eisenbeiss, H., 2007. Applications of photogrammetric
processing using an autonomous model helicopter. Revue
Francaise de Photogrammetrie et de Teledetection. Symposium
ISPRS Commission Technique I "Des capteurs a l'Imagerie",
n°185 (2007-1), Saint-Mande Cedex, France.
UVS International, 2008. Unmanned Vehicle Systems
International http://www.uvs-international.org (accessed 08. May
2008).
Vallet, J., 2007. GPS-IMU and LiDAR integration to aerial
photogrammetry: Development and practical experiences with
Helimap
System®,
Voträge
Dreiländertagung
27.
Wissenschaftlich-Technische Jahrestagung der DGPF. 19-21 June
2007, Muttenz.
Herwitz, S.R., Johnson, L.F., Dunagan, S.E., Higgins, R.G.,
Sullivan, D.V., Zheng, J., Lobitz, B.M., Leung, J.G., Gallmeyer,
B., Aoyagi, M., Slye, R., E., Brass, J., 2004. Demonstration of
UAV-based imaging for agricultural surveillance and decision
support. Comput. Electron. Agr. 44:49-61.
van Blyenburg, P., 1999. UAVs: an Overview. Air & Space
Europe. Vol. I No 5/6. pp. 43-47.
Horcher, A., Visser, R.J.M., 2004. Unmanned aerial vehicles:
Applications for natural resource management and monitoring.
COFE (Council on Forest Engineering) Annual Meeting 2004,
Proceedings;
http://www.cnr.vt.edu/ifo/VT%20Andy%20COFE%202004%2
0Drone%20Paper1.pdf (accessed 08. May 2008).
Wang, Y., Yang, X., Stojic, A., Skelton, B., 2004. Toward higher
automation and flexibility In commercial digital photogrammetric
systems. IAPRS, Vol. XXXV, p. 838-840, Istanbul.
Zhang, L., 2005. Automatic Digital Surface Model (DSM)
Generation from Linear Array Images. Ph. D. Dissertation,
Institute of Geodesy and Photogrammetry, ETH Zurich,
Switzerland, Report No. 88.
Kise, M., Rovira-Más, F., Zhang, Q., 2005. A stereovisionbased crop row detection method for tractor-automated
guidance. Biosystems Eng. 90:357-367.
Lambers, K., Eisenbeiss, H., Sauerbier, M., Kupferschmidt, D.,
Gaisecker, T., Sotoodeh, S., Hanusch, T., 2007. Combining
photogrammetry and laser scanning for the recording and
modelling of the late intermediate period site of Pinchango Alto,
Palpa, Peru. J. Archaeol. Sci. 34:1702-1710.
Lamon, P., Stachniss, C., Triebel, R., Pfaff, P., Plagemann, C.,
Grisetti, G., Kolski, S., Burgard, W., Siegwart, R. 2006.
"Mapping with an Autonomous Car", Proc. of The Workshop
on Safe Navigation in Open and Dynamic Environments
(IROS).
Nebiker, S., Christen, M., Eugster, H., Flückiger, K. and Stierli,
C., 2007. Integration von mobilen Geosensoren in kollaborative
virtuelle Globen, Dreiländertagung der SGPBF, DGPF und
OVG: Von der Medizintechnik bis zur Planetenforschung Photogrammetrie und Fernerkundung für das 21. Jahrhundert.
DGPF Tagungsband Nr. 16, FHNW, Muttenz, pp. 189-198.
Pueschel, H., Sauerbier, M., Eisenbeiss, H., 2008. A 3D model
of Castle Landenberg (CH) from combined photogrammetric
processing of terrestrial and UAV-based images. XXI ISPRS
Congress , Beijing, China, 03.-11. July 2008, (In Press).
Randa rockslide, 2008. http://www.rockslide.ethz.ch/ (accessed
May 8th 2008).
Reidelstuerz, P., Link, J., Graeff, S., Claupein, W., 2007. UAV
(unmanned aerial vehicles) für Präzisionslandwirtschaft. 13.
Workshop Computer-Bildanalyse in der Landwirtschaft & 4.
Workshop Precision Farming 61:75-84.
Rovira-Más, F., Zhang, Q., Reid, J.F., 2005. Creation of threedimensional crop maps based on aerial stereoimages.
Biosystems Eng. 90:251-259.
Rovira-Más, F., Zhang, Q., Reid, J.F., 2008. Stereo vision
three-dimensional terrain maps for precision agriculture.
Comput. Electron. Agr. 60:133-143.
SurveyCopter,
2008.
http://perso.wanadoo.fr/surveycopter/eindex.htm (accessed May 8th, 2008).
983
The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences. Vol. XXXVII. Part B1. Beijing 2008
984
Download