图像与视频处理计算机学院

advertisement
图像与视频处理
Machine Learning Group
3.11 Video
Enhancement
and Restoration
计算机学院
University of Texas at Austin
1
图像与视频处理
Machine Learning Group
1. Introduction
• Video enhancement and restoration has always been
important, not only to improve the visual quality but
also to increase the performance of subsequent tasks
such as analysis and interpretation.
applications
• one encounters in astronomy, forensic sciences, and
medical imaging
• preserving motion pictures and videotapes recorded
over the last century : reusing old film and video
material
计算机学院
University of Texas at Austin
2
图像与视频处理
Machine Learning Group
1. Introduction
• difference between the enhancement and restoration:
the amount of data to be processed
• noise (Section 2): the most common artifact
encountered in the above mentioned applications.
• blotches (Section 3) and intensity flicker (Section 4)
: the removal of two other important impairments that
rely on temporal processing algorithms.
计算机学院
University of Texas at Austin
3
图像与视频处理
Machine Learning Group
1. Introduction
计算机学院
University of Texas at Austin
4
图像与视频处理
Machine Learning Group
2. Spatiotemporal Noise Filtering
• the ideal uncorrupted image sequence f(n, k).
• The recorded image sequence g(n, k) corrupted by
noise w(n, k) is then given by
g(n ,k )  f(n ,k )  w(n ,k )
where n = (n1, n2) refers to the spatial coordinates
and k to the frame number in the image sequence.
计算机学院
University of Texas at Austin
5
图像与视频处理
Machine Learning Group
2. Spatiotemporal Noise Filtering
• 2.1 Linear Filters
• 2.2 Order-Statistic Filters
• 2.3 Multiresolution Filters
计算机学院
University of Texas at Austin
6
图像与视频处理
Machine Learning Group
2.1 Linear Filters
• Temporally Averaging Filters
• Temporally Recursive Filters
计算机学院
University of Texas at Austin
7
图像与视频处理
Machine Learning Group
2.1 Linear Filters
• Temporally Averaging Filters
the restored image sequence is obtained by
fˆ(n ,k ) 
K
h(l )g(n ,k

l K

 l ) (1)
Here h(l) are the temporal filter coefficients used to
weight 2K+l consecutive frames.
计算机学院
University of Texas at Austin
8
图像与视频处理
Machine Learning Group
2.1 Linear Filters
• Temporally Averaging Filters
the filter coefficients can be optimized in a minimum
mean-squared error fashion,
h(l )  min E[(f(n ,k )  fˆ(n ,k ))2 ]
h(l )
yielding the well-known temporal Wiener filtering
solution:
计算机学院
University of Texas at Austin
9
图像与视频处理
Machine Learning Group
2.1 Linear Filters
• Temporally Averaging Filters
The motion artifacts can greatly be reduced by operating the
filter, along the picture elements (pixels) that lie on
the same motion trajectory .
计算机学院
University of Texas at Austin
10
图像与视频处理
Machine Learning Group
2.1 Linear Filters
• Temporally Averaging Filters
Equation (2) then becomes a motion-compensated temporal
filter.
fˆ(n ,k ) 
K
h(l )g(n

l K
1
 d x(n1 ,n2; k ,l ),n2  d y(n1 ,n2; k ,l ),k  l )

Here 𝑑 𝑛; 𝑘, 𝑙 = (𝑑𝑥 𝑛1 , 𝑛2 ; 𝑘, 𝑙 , 𝑑𝑦 (𝑛1 , 𝑛2 ; 𝑘, 𝑙)) is the
motion vector for spatial coordinate (𝑛1 , 𝑛2 )estimated between
the frames k andl .
计算机学院
University of Texas at Austin
11
图像与视频处理
Machine Learning Group
2.1 Linear Filters
• Temporally Averaging Filters
In order to avoid the explicit estimation of motion, which might be
problematic at high noise levels, two alternatives are available that
turn Eq. (1) into a motion-adaptive filter.
In the first place, in areas where motion is detected (but not
explicitly estimated) the averaging of frames should be kept to a
minimum.
Second, filter (1) can be operated along M a priori selected motion
directions at each spatial coordinate.
计算机学院
University of Texas at Austin
12
图像与视频处理
Machine Learning Group
2.1 Linear Filters
• Temporally Averaging Filters
Filter (1) can be extended with a spatial filtering part. The most
straightforward extension of Eq. (1) is the following 3-D
weighted averaging filter:
fˆ(n,k ) 
h(m ,l)g(n  m ,k

ml S
 l)
( , )
Here S is the spatiotemporal support or window of the 3-D filter
(see Fig. 3).
计算机学院
University of Texas at Austin
13
图像与视频处理
Machine Learning Group
2.1 Linear Filters
• Temporally Averaging Filters
Here S is the spatiotemporal support or window of the 3-D filter
(see Fig. 3).
计算机学院
University of Texas at Austin
14
图像与视频处理
Machine Learning Group
2.1 Linear Filters
• Temporally Averaging Filters
Disadvantages with the 3-D Wiener filter:
• The requirement that the 3-D autocorrelation function for the
original image sequence is known a priori.
• The 3-D wide-sense stationarity assumptions, which are
virtually never true because of moving objects and scene
changes.
计算机学院
University of Texas at Austin
15
图像与视频处理
Machine Learning Group
2.1 Linear Filters
• Temporally Averaging Filters
Simpler ways of choosing the 3-D filter coefficients are usually
preferred, one such choice for adaptive filter coefficients is the
following:
h(m ,l; n ,k ) 
c
1  max(a ,(g(n ,k )  g(n  m ,k  l ))2 )
计算机学院
University of Texas at Austin
16
图像与视频处理
Machine Learning Group
2.1 Linear Filters
• Temporally Recursive Filters
The general form of a recursive temporal filter is as
follows:
fˆ(n,k(2)
)  fˆb(n,k )  (n,k )[g(n,k )  fˆb(n,k )]
Here 𝑓𝑏 n, k is the prediction of the original kth frame on the
basis of previously filtered frames, and α n, k is the filter gain
for updating this prediction with the observed kth frame.
计算机学院
University of Texas at Austin
17
图像与视频处理
Machine Learning Group
2.1 Linear Filters
• Temporally Recursive Filters
A popular choice for the prediction 𝑓𝑏 n, k is the previously
restored frame, either in direct form
fˆ(n ,k )  fˆ(n ,k  1)
or in motion-compensated form:
fˆb(n,k )  fˆ(n  d(n; k ,k  1),k  1)
计算机学院
University of Texas at Austin
18
图像与视频处理
Machine Learning Group
2.1 Linear Filters
• Temporally Recursive Filters
A switching filter is obtained if the gain takes on the values a and
1, depending on the difference between the prediction 𝑓𝑏 n, k and
the actually observed signal value g(n, k):
计算机学院
University of Texas at Austin
19
图像与视频处理
Machine Learning Group
2.1 Linear Filters
• Temporally Recursive Filters
A finer adaptation is obtained if the prediction gain is optimized
to minimize the mean-squared restoration error , yielding
Here
is an estimate of the image sequence variance
in a local spatiotemporal neighborhood of (n, k).
计算机学院
University of Texas at Austin
20
图像与视频处理
Machine Learning Group
2.2 Order-Statistic Filters
• Order-statistic (OS) filters are nonlinear variants of
weighted averaging filters.
• The distinction is that in OS filters the observed noisy
data, usually taken from a small spatiotemporal
window, are ordered before being used.
计算机学院
University of Texas at Austin
21
图像与视频处理
Machine Learning Group
2.2 Order-Statistic Filters
The general structure of an OS restoration filter is as
follows:
fˆ(n ,k ) 
|S |
h r (n ,k )g r (n ,k )

r
( )
( )
1
𝑔𝑟 (𝑛, 𝑘) : the ordered intensities, or ranks, of the
corrupted image sequence;
|S| : the number of intensities in this window.
The objective is to choose appropriate filter
coefficientsℎ𝑟 (𝑛, 𝑘) for the ranks.
计算机学院
University of Texas at Austin
22
图像与视频处理
Machine Learning Group
2.2 Order-Statistic Filters
• The most simple order-statistic filter is a
straightforward temporal median, for instance taken
over three frames:
• the multistage median filter (MMF):
计算机学院
University of Texas at Austin
23
图像与视频处理
Machine Learning Group
2.2 Order-Statistic Filters
• an example of the spatiotemporal supports of the
multistage median filter
计算机学院
University of Texas at Austin
24
图像与视频处理
Machine Learning Group
2.2 Order-Statistic Filters
• If the coefficients are optimized in the mean-squared
error sense, the following general solution for the
restored image sequence is obtained :
计算机学院
University of Texas at Austin
25
图像与视频处理
Machine Learning Group
2.2 Order-Statistic Filters
• The overall filter structure thus obtained is shown in
Fig. 5.
计算机学院
University of Texas at Austin
26
图像与视频处理
Machine Learning Group
3. Blotch Detection and Removal
计算机学院
University of Texas at Austin
27
图像与视频处理
Machine Learning Group
3. Blotch Detection and Removal
• A model for blotch is the following:
• The overall blotch detection and removal scheme :
计算机学院
University of Texas at Austin
28
图像与视频处理
Machine Learning Group
3.1 Blotch Detection
three characteristic properties:
• blotches are temporally independent and therefore
hardly ever occur at the same spatial location in
successive frames.
• the intensity of a blotch is significantly different
from its neighboring uncorrupted intensities.
• blotches form coherent regions in a frame, as
opposed to, for instance, spatiotemporal shot noise.
计算机学院
University of Texas at Austin
29
图像与视频处理
Machine Learning Group
3.1 Blotch Detection
• pixel-based blotch detector : the spike-detector index
(SDI)
• A blotch pixel is detected if SDI(n,k) exceeds a
threshold:
计算机学院
University of Texas at Austin
30
图像与视频处理
Machine Learning Group
3.1 Blotch Detection
• order-statistic-based detector : the rank order
difference (ROD) detector
• A blotch pixel is detected if any of the rank order
differences exceeds a specific threshold Ti:
计算机学院
University of Texas at Austin
31
图像与视频处理
Machine Learning Group
3.1 Blotch Detection
计算机学院
University of Texas at Austin
32
图像与视频处理
Machine Learning Group
3.1 Blotch Detection
计算机学院
University of Texas at Austin
33
图像与视频处理
Machine Learning Group
3.1 Blotch Detection
postprocessing the blotch mask in two ways:
• removing small blotches
• completing partially detected blotches: hysteresis
thresholding
计算机学院
University of Texas at Austin
34
图像与视频处理
Machine Learning Group
3.1 Blotch Detection
计算机学院
University of Texas at Austin
35
图像与视频处理
Machine Learning Group
3.2 Motion Vector Repair and Interpolating
Corrupted Intensities
Two strategies in recovering motion vectors:
• take an average of surrounding motion vectors
• validate the corrected motion vectors using intensity
information directly neighboring the blotched area
计算机学院
University of Texas at Austin
36
图像与视频处理
Machine Learning Group
3.2 Motion Vector Repair and Interpolating
Corrupted Intensities
• In a multistage median interpolation filter, five
interpolated results are computed by using the
(motion-compensated) spatiotemporal neighborhoods.
计算机学院
University of Texas at Austin
37
图像与视频处理
Machine Learning Group
3.2 Motion Vector Repair and Interpolating
Corrupted Intensities
• Each of the five interpolated results is computed as the median
over the corresponding neighborhood Si:
• The final result is computed as the median over the five
intermediate results:
计算机学院
University of Texas at Austin
38
图像与视频处理
Machine Learning Group
3.2 Motion Vector Repair and Interpolating
Corrupted Intensities
计算机学院
University of Texas at Austin
39
图像与视频处理
Machine Learning Group
3.2 Motion Vector Repair and Interpolating
Corrupted Intensities
• Result of the blotch-corrected frame:
计算机学院
University of Texas at Austin
40
图像与视频处理
Machine Learning Group
4. Intensity Flicker Correction
• Intensity flicker is defined as unnatural temporal
fluctuations of frame intensities that do not originate
from the original scene.
计算机学院
University of Texas at Austin
41
图像与视频处理
Machine Learning Group
4. Intensity Flicker Correction
• A model describing the intensity flicker:
g( n, k )  ( n, k )f ( n, k )  ( n, k )  ( n, k ) .
Here , ( n, k ) and ( n, k ) are the multiplicative
and additive unknown flicker parameters, ( n, k )
is a independent noise term.
计算机学院
University of Texas at Austin
42
图像与视频处理
Machine Learning Group
4.1 Flicker Parameter Estimation
• If the flicker parameters were known, then one could
form an estimate of the original intensity from a
corrupted intensity by using the following
straightforward linear estimator:
fˆ( n, k )  h1( n, k )g( n, k )  h0( n, k ) .
计算机学院
University of Texas at Austin
43
图像与视频处理
Machine Learning Group
fˆ( n, k )  h1( n, k )g( n, k )  h0( n, k ) .
• In order to obtain estimates for the coefficients h1( n, k )
, the mean-squared error between f ( n, k ) and fˆ( n, k )
is minimized, yielding the following optimal solution:
2



( n, k )
1
w
h0( n, k )  
E[ g( n, k ) ]  ,
 ( n, k )  2

( n, k ) 
 g( n, k )

1
h1( n, k ) 
( n, k )
计算机学院
University of Texas at Austin
 g2( n, k )   w2( n, k )
 g2( n, k )
44
图像与视频处理
Machine Learning Group
4.1 Flicker Parameter Estimation
• If the observed image sequence does not contain any
noise, then:
( n, k )
h0( n, k )  
,
( n, k )
1
h1( n, k ) 
( n, k )
计算机学院
University of Texas at Austin
45
图像与视频处理
Machine Learning Group
4.1 Flicker Parameter Estimation
• In practice, the true values for the intensity-flicker
parameters α(n, k) and β(n, k) are unknown and have
to be estimated from the corrupted image sequence
itself.
• Since the flicker parameters are spatially smooth
functions, we assume that they are locally constant:
( n, k )   m( k ) 
 n  Sm,
( n, k )   m( k ) 
where Sm indicates a small frame region.
计算机学院
University of Texas at Austin
46
图像与视频处理
Machine Learning Group
4.1 Flicker Parameter Estimation
• By computing the averages and variances of both
sides of
g( n, k )  ( n, k )f ( n, k )  ( n, k )  ( n, k ) ,
one can obtain:
计算机学院
University of Texas at Austin
47
图像与视频处理
Machine Learning Group
4.2 Estimation on Sequences with
Motion
• We assume that the image sequence intensities do not
change significantly over time previously. Clearly,
this is an incorrect assumption if motion occurs.
• Because of the intensity flicker this assumption is
violated heavily. The only motion that can be
estimated with sufficient reliability is global motion
such as camera panning or zooming.
计算机学院
University of Texas at Austin
48
图像与视频处理
Machine Learning Group
4.2 Estimation on Sequences with
Motion
There are various approaches for detecting local motion
• the detection of large differences between the current
and previously (corrected) frame
• compare the estimated intensity-flicker parameters to
threshold values
计算机学院
University of Texas at Austin
49
图像与视频处理
Machine Learning Group
4.2 Estimation on Sequences with
Motion
• For frame regions S, where the flicker parameters
could not be estimated reliably from the observed
image sequence, the parameters are estimated on the
basis of the results in spatially neighboring regions.
• For the regions in which the flicker parameters could
be estimated, a smoothing post processing step has to
be applied to avoid sudden parameter changes that
lead to visible artifacts in the corrected image
sequence.
计算机学院
University of Texas at Austin
50
图像与视频处理
Machine Learning Group
4.2 Estimation on Sequences with
Motion
计算机学院
University of Texas at Austin
51
图像与视频处理
Machine Learning Group
Conclude
This chapter has described methods for enhancing and
restoring corrupted video and film sequences. Although
the focus has been on noise removal, blotch detection
and correction, and flicker removal, the approaches and
tools described in this chapter are of a more general
nature, and they can be used for developing
enhancement and restoration methods for other types of
degradation.
计算机学院
University of Texas at Austin
52
Download