Sensors and media INF 5300 Data fusion for image analysis Anne Solberg ()

advertisement
Sensors and media
INF 5300
Data fusion for image analysis
• Remote sensing:
– Electromagnetic signals, from visible to
thermal and infrared and microwave
energy
– Sensors: Optical, microwave, thermal,
laser altimeter
Anne Solberg (anne@ifi.uio.no)
Today:
• Medical image analysis:
•Background
– Ultrasound: acoustical signals
– Magnetic Resonnance Imaging
(MR)/CT
– PET/SPECT
– X-ray
•Multisensor data registration
•Multisensor image classification
•Multitemporal image analysis
• Sonar imaging:
•Multiscale image analysis
– Multifrequency echosounders
F1 20.3.06
INF 5300
1
F1 20.3.06
Multifrequency imaging
INF 5300
2
Multitemporal data
• Multispectral and hyperspectral sensors provide
information on different frequencies. Different targets
are sensitive to certain frequencies.
– Imaging at several frequencies>separating more detalied
classes
• By imaging on object at different times, more
information can be derived.
– Discriminate between objects based on their temporal
variation (e.g. vegetation classification)
• Challenges:
• This applies to other types of signals, too.
– Sensor calibration/weather changes
– Changes in the classes occurr
– Multifrequency echosounders used for fishery
• A new road is built, trees are cut
F1 20.3.06
INF 5300
3
F1 20.3.06
INF 5300
4
Multisensor data
Multisensor data registration
• Each sensor has it own characteristics
–
–
–
–
–
–
–
• Data registration is a pre-requisite for data fusion
• Data registration can be simple if the data are
georeferenced (map co-ordinates are known)
Noise level
Reliability
Ability to discriminate certain object types
Resolution
Availability/Coverage
Limitations due to e.g. weather/clouds
Imaging geometry and geometrical errors
F1 20.3.06
INF 5300
– Registration is then just resampling the images to a common
map projection
• An image matching step is often necessary to obtain
subpixel accuracy in the co-registration.
5
Image registration – an overview
F1 20.3.06
INF 5300
6
Registration step 1: feature extraction
• Purpose: extract regions, edges, contours etc. that
can be used to represent tie-points in the set of
images.
• Tie-points: characteristic points that can be identified
in all images.
• Area-based methods:
The main steps in a full image registration process are:
1. Feature extraction
2. Feature matching
3. Transformation selection
4. Image resampling
– Match the grey levels in the images directly using statistical
measures.
• Feature-based methods:
– Define a set of useful features (e.g. edges, lines,
intersections etc.) tailored to the application.
– Extract these features using either spatial-domain methods
or frequency-domain methods (Fourier- or wavelets)
F1 20.3.06
INF 5300
7
F1 20.3.06
INF 5300
8
Registration step 3:
Transformation selection
Registration step 2: feature matching
• Purpose: establish the correspondence between the
tie-points found during step 1.
• Area-based methods:
• Purpose: find the mapping function that describes
the relation between the positions of the tie-points in
the different images.
• Affine tranforms are often used.
– Match using correlation, Fourier-transform methods or
optical flow.
Affine transformasjoner (lineære):
x’=a0x+a1y+a2
y’=b0x+b1y+b2
Kvadratiske transformasjoner (kalles warping):
x’=a0x2+a1y2+a2xy+a3x+a4y+a5
y’=b0x2+b1y2+b2xy+b3x+b4y+b5
• Feature-based methods:
– Perform the matching in the Fourier-domain using the
equivalence between correlation and multiplication in the
Fourier-domain.
• Optical flow methods:
– Estimate the relative motion between the two images.
• This is commonly used in video analysis.
F1 20.3.06
INF 5300
9
F1 20.3.06
Registration step 4:
Image resampling
The data can be fused at different level:
• Signal-based fusion
– Fuse the signals prior to image formation
• Pixel-based fusion
– Merge the multisensor images on a pixel-by-pixel basis
• Feature-based fusion
– Merge features derived from the different signals or sensors.
• Decision-level fusion
• See INF2310 for details of affine transforms and
interpolation.
http://www.ifi.uio.no/~inf2310/v2006/forelesninger/20060228_geometriskeOperasjoner.pdf
INF 5300
10
Fusion at different levels
• Purpose: compute the new pixel co-ordinates in the
co-registered images and interpolate the pixel values.
• Transforming the co-ordinates will give non-integer
co-ordinates. Interpolation is used to combine pixel
values from neighboring pixels.
• Normally, either nearest-neighbor interpolation or
bilinear interpolation is used.
F1 20.3.06
INF 5300
11
– Fuse the information at a higher level by first classifying
each image independently, then combining the preliminary
classifications.
F1 20.3.06
INF 5300
12
Pixel-level fusion
Pixel-level fusion
• Advantages:
– Simple. No special classifier software needed.
– Correlation between sources utilized.
– Well suited for change detection.
• Limitations:
– Assumes that the data can be modelled using a common
probability density function. Source reliability cannot be
modelled.
– Not suited for heteorogenous data.
F1 20.3.06
INF 5300
13
F1 20.3.06
Feature-level fusion
INF 5300
14
Feature-level fusion
• Advantages:
– Simple. No special classifier software needed.
– Sensor-specific features gives advantage over pixel-based
fusion.
– Well suited for change detection.
• Limitations:
– Assumes that the data can be modelled using a common
probability density function.
– Source reliability cannot be modelled.
F1 20.3.06
INF 5300
15
F1 20.3.06
INF 5300
16
Decision-level fusion
Decision-level fusion
• Image X1,... XP from P sensors. C are the class labels
of the scene.
• Assign each pixel to the class that maximizes the
posterior probabilities
• The measurements from the different sensors are
often assumed to be conditionally independent:
Compute the class-conditional probabilities for each class,
then combine these probabilities.
F1 20.3.06
INF 5300
17
F1 20.3.06
INF 5300
18
Selecting the weights λi
Combination of single-sensor classifiers
• The weights λi are often selected based on heuristics
• The can be found based on grid search and crossvalidation
• Consensus theory deals with
methods for combing single-sensor
classification results.
• The output from each sensor
(probability estimates) are often
weighted according to the source’s
reliability.
• Linear opinion pool:
– Divide the training set in two, training and validation
– Select the weights that give best performance on the
validation set.
• Neural-networks can be trained to estimate weights.
• Logarithmic opinion pool:
F1 20.3.06
INF 5300
19
F1 20.3.06
INF 5300
20
Neural nets for
multisensor classifcation
Decision-level fusion
• Advantages:
– Suited for data with different probability densities.
– Source-specific reliabilities can be modelled.
• Limitations:
– Special software often needed.
F1 20.3.06
INF 5300
21
F1 20.3.06
Multitemporal image
classification - applications
– Different fusion levels can be used.
– Examples of pixel-level methods: image math (difference or ratio),
image regression, principal component analysis.
• Data normalization must be done prior to using these methods.
• Improved quality in discriminating between a set of classes.
– Cover limitations due to e.g. partly cloud coverage, snow or soil
moisture conctents.
– A fusion model that takes source reliability into account should be
used.
Direct multidate classification:
• Merge the data on a pixel-basis into one
measurement vector.
• Classify this using traditional classifiers.
• Advantage:
– Simple. Temporal feature correlation between image
measurements utilized.
• Discriminate between classes based on their temporal signature
development
– Example: Discriminate vegetation based on how the chlorophyll
content varies during the growth season.
INF 5300
22
Multitemporal classifiers 1
• Monitor/identify specific changes
F1 20.3.06
INF 5300
23
• Limitations:
– Not suited for data sets containing noisy images.
F1 20.3.06
INF 5300
24
Multitemporal classifiers 2
Multitemporal classifiers 3
Cascade classifiers:
• At time t1, classify image x1.
• At time t2, compute the conditional probability for
observing class ω given the image x1 and x2:
Markov chain/Markov random field classifiers:
• Model the probability of class changes from time t1
to time t2.
• Advantages:
P(ω| x1, x2)
– Spatial and temporal correlation of class labels utilized.
– Information about special class transitions can be modelled.
• Advantages:
– Temporal correlation of class labels considered.
– Information about special class transitions can be modelled.
• Limitations:
– Special software needed.
• Limitations:
– Special software needed.
F1 20.3.06
INF 5300
25
Multitemporal classifiers 4
26
• Most common approach: resample one of the images
to the resolution of the other images. Classify using
traditional classifier.
• If resampling all images to the image with the
coarsest resolution, fine details might be lost.
• Resamping all images to the finest resolution can be
done by copying coarse-resolution pixels (or
performing interpolation of neighboring pixels)
– Tilstander: vinterdvale, blomstring, fullt utviklet, høst
• Classification is done based on the likelihood of observing a
chain of class transitions (and when they occurr).
• Markov chains are used to model class transitions.
• Advantages:
– Can discriminate between classes not separable at a single point in
time.
– Can be used at different levels.
• Limitations:
– Feature level approaches can be sensitive to noise.
– A time series (>2 images) needed (sometimes difficult to get).
INF 5300
INF 5300
Multiscale image classification
Temporal trajectory classification:
• Use integrated models that incorporate the expected
development of the classes during a time period.
• Example: vegetation in the growing season:
F1 20.3.06
F1 20.3.06
27
– Interpolation should be used with care as it alters the grey
level distributions.
• More advanced classifiers for true multiscale data
exists. The need for such models depends on the
desired detail level in the results.
F1 20.3.06
INF 5300
28
Summary – multisensor image
classification
Multiscale image visualization
• Many satellite sensors give a combination of a highresolution panchromatic band (grey-level) with high
resolution (e.g. 1m) and a set of multispectral bands
with lower resolution (e.g. 4).
• Many methods for overlaying the multispectral image
on the panchromatic image with fine spatial
structures exists.
• Common methods are based on wavelets or Fouriertransform to model data at various spatial scales.
• In general, there is no consensus on which multisource or
multitemporal methods that is overall best. Choices should be
application specific.
• The fusion level should be considered carefully. Pixel-based
fusion can be suited for simple analysis, but decision-level gives
best control and allows weighting the influence of each sensor.
• To find the best classifier for a multitemporal data set, the
complexity of the class separation problem and the available
data set must be selected.
• Multiscale data can either be resampled to a common
resolution, a classifier with implicit modelling of the relationship
between the different scales can be used. The latter is
recommended for problems involving small or detailed
structures.
F1 20.3.06
F1 20.3.06
INF 5300
29
INF 5300
30
Download