Presentation - Duke Electrical and Computer Engineering

advertisement
From Learning Models of Natural Image Patches to
Whole Image Restoration
Daniel Zoran
Interdisciplinary Center for Neural Computation
Hebrew University of Jerusalem
Yair Weiss
School of Computer Science and Engineering
Hebrew University of Jerusalem
Presented by Eric Wang
Duke University
6/18/2012
Introduction
• Patch based learning on images has significant computational
advantages over learning dictionaries over the entire image.
• A primary concern of this paper is the affect the choice of
dictionary priors have on the performance of the model.
• This paper addresses 3 main questions
– (1) Do priors that give high likelihoods yield better patch restoration
performance?
– (2) Do priors that give high likelihoods yield better whole-image
restoration performance?
– (3) Can we learn better priors?
Motivation and Patch Restoration
• Answer to question (1): Priors with higher likelihoods will yield
improved per-patch denoising performance, as shown with
several popular priors
PSNR of restoration on patches vs. dictionary (prior) likelihoods. Trained on
50,000 8x8 patches of natural images of faces. Tested on unseen patches.
From Patches to Whole Image Restoration
• Good patch-based image restoration does not guarantee high
quality image restoration, many reconstruction methods can
generate significant artifacts.
Expected Patch Log-Likelihood
• Choosing random patches can provide a solution to minimize
artifacting, but has the issue that most random patches will
have low likelihood to a given dictionary.
• This paper presents an optimization algorithm that maximizes
expected patch log-likelihood (EPLL) with the constraint that
the reconstructed image be close to the original image.
• The EPLL under prior p is defined as
• Where
image x.
is a mask that extracts the ith random patch from
Cost Function
• Let
be the corruption model on the image where x
is a vectorized image, A defines the corruption model and y is
the vectorized noisy observation.
• The cost function to minimize is then
• Direct optimization of this cost function is intractable, so an
alternative method called half quadratic splitting is used,
where
are a set of per-patch auxiliary variables.
The Corruption Model
• The choice of the matrix A is determined by the application.
• For denoising, A is an identity matrix, and
precision
is the noise
• For deblurring, A is a convolution matrix with a known kernel
• For inpainting, A is a diagonal matrix with zeros for the
missing elements.
Optimization
• The EPLL optimization involves two steps: (1) solving for
given
• And (2) solving for
given , which is dependent on the
dictionary and involves solving a MAP estimate of the most
likely dictionary element for a particular patch.
•
is either set by hand or set to
estimated noise standard deviation in
where
.
is the
Image Restoration
• Answer to question (2): It is shown that priors with higher
patch likelihoods yield improved whole image restoration
Building a better dictionary via the GMM
• Gaussian zero mean data can usually be well represented by
the top-m eigenvectors of its covariance matrix.
• This paper proposes clustering the patches (pixels) via a GMM.
Patches sharing a cluster also share a dictionary.
Sample dictionaries for six mixture components
Building a better dictionary via the GMM
• Answer to question (3): The GMM prior outperforms other
priors in both patch and whole-image restoration.
PSNR values, 200 GMM components
Building a better dictionary via the GMM
• The GMM prior is also shown to outperform the ICA prior in
reconstruction with noise level
using EPLL.
Image Denoising
• 68 images from the Berkeley dataset, 8x8 patches,
comaprison is in PSNR
Image Deblurring
• 68 images from the Berkeley dataset with known blur kernels
and 1% white Gaussian noise (comparison is in PSNR)
Download