Digital Image Processing - Weizmann Institute of Science

advertisement
Digital Image Processing in Life Sciences
March 14th, 2012
Lecture number 1: Digital Image Fundamentals
Lecture’s outline
What Is Digital Image Processing?
(The Origins of Digital Image Processing)
Fundamental Steps in Digital Image Processing
Image Sampling and Quantization
Spatial and Gray-Level Resolution
Some Basic Relationships Between Pixels
Zooming and Shrinking Digital Images
Lookup tables
Color spaces
Terms to be conveyed:
Pixel
Gray level
Bit depth
Dynamic range
Connectivity types/neighborhood
Interpolation types
Look-up tables
Book:
Digital Image Processing, Rafael C. Gonzales and Richard E.Woods
Web resources:
www.microscopy.fsu.edu (very thorough and informative)
www.cambridgeincolour.com (beautiful examples, excellent tutorials)
Next topics:
2. Image enhancement in the spatial domain
3. Segmentation
4. Image enhancement in the frequency domain
5. Multi dimensional image processing
6-7. Guest lectures-TBD
What Is A Digital Image?
Image= “a two-dimensional function, f(x,y), where x and y are spatial
coordinates, and the amplitude of f at any pair of coordinates (x, y) is
called the intensity (gray level of the image) at that point. When x, y, and
the amplitude values of f are all finite, discrete quantities, we call the
image a digital image.” (Gonzalez and Woods).
These sets of numbers can be depicted in terms of frequencies
http://cvcl.mit.edu/hybrid_gallery/gallery.html
Digital Image Processing-Points to consider:
Why process?
Are both the input and output of a process images?
Where does image processing stop and image analysis start?
Are the processing results intended for human perception or for machine perception?
Character recognition and fingerprint comparisons vs intelligence photos…
We can define three types of computerized processes:
Low-, mid-, and high-level.
Low: image preprocessing, noise reduction, enhance contrast etc.
Mid: segmentation, sorting and classification.
High: assembly of all components into a meaningful coherent form
Digital image originsThe digital image dates back to…
the 1920’s and the Bartlane cable picture transmission system between NY and
London. The image took 3 hours to transmit, instead of more than one week.
They started with 5 tone levels and increased to 15 levels by 1929.
Taken from Gonzalez and Woods
Lecture’s outline
What Is Digital Image Processing?
(The Origins of Digital Image Processing)
Fundamental Steps in Digital Image Processing
Image Sampling and Quantization
Spatial and Gray-Level Resolution
Some Basic Relationships Between Pixels
Zooming and Shrinking Digital Images
Lookup tables
Color spaces
Essential steps when processing digital images:
Acquisition
Enhancement
Restoration
Outputs are
digital images
Color image restoration
Wavelets
Morphological processing
Segmentation
Representation
Recognition
Outputs are
attributes of
the image
Image acquisition
Acquire or receive an image for further processing.
This step has a major impact over the entire procedure of processing and
analysis.
Image Enhancement
Improving quality subjectively (e.g. by change of contrast)
Image Restoration
Improving quality objectively (e.g. by removing psf)
microscopy.fsu.edu
microscopy.fsu.edu
microscopy.fsu.edu
Morphological processing
Extracting components for the purpose of representing shapes
Segmentation
Deconstructing the image into its constituent objects. A crucial step for
successful recognition of the image contents.
Morphological processing
Extracting components for the purpose of representing shapes
Segmentation
Deconstructing the image into its constituent objects. A crucial step for
successful recognition of the image contents.
Representation
Feature selection-classification/grouping of objects
Lecture’s outline
What Is Digital Image Processing?
(The Origins of Digital Image Processing)
Fundamental Steps in Digital Image Processing
Image Sampling and Quantization
Spatial and Gray-Level Resolution
Some Basic Relationships Between Pixels
Zooming and Shrinking Digital Images
Lookup tables
Color spaces
Sampling and quantization
Keep in mind:
The sensor we used to create the image has a continuous output.
But, the transition from a continuum to a digital image requires two processes:
sampling and quantization.
Sampling is the process of digitizing the spatial coordinates.
Quantization is the process of digitizing the amplitude values at those spatial
coordinates.
The arrangement of the sensor used to create the image determines the sampling
method and its output.
Different limits determine the performance of the optical sensors and of the
mechanical sensors.
Sampling and quantization result in arrays of discrete quantities.
By convention, the coordinate (x,y)=(0,0) is located at the upper leftmost
corner of the image.
(Gonzales and Woods)
picture elements=image elements=pels=pixels
Sampling results in typical image sizes that can vary from 128 x 128 to 4096 x
4096 or any combination thereof.
An Image Formation Model
Let l(x0, y0) be the gray level (gl) value at (x0, y0) : l=f (x0, y0)
l is bounded by Lmin and Lmax and the boundary [Lmin, Lmax] is the gray scale.
This interval is usually shifted to [0, L-1] where 0 represents black gl values,
and L-1 represents white gl values.
Quantization results in discrete values of gray levels, typically an integer power of
2: L=2k .
If k=8, the result is 256 gray levels, from 0 to 255.
Dynamic range- the portion of the gray levels in the image out of the entire gray
scale of the image.
Think about high vs low dynamic range images: how does the dynamic range
affect the contrast of the image? Next lecture…
How many bits are required to save a digital image?
b=M x N x k (or M2k for images of equal dimensions).
Resolution
Size (kb)
Gray level (bit-depth)
8 (256)
12 (4096) 16 (65536)
128
16.384
24.576
32.768
256
65.536
98.304
131.072
512 262.144 393.216
524.288
1024 1048.576 1572.864 2097.152
2048 4194.304 6291.456 8388.608
8bit images- values are integers, unsigned
16bit images- values are integers, some softwares allow signed.
32bit images-floating-point, signed.
Lecture’s outline
What Is Digital Image Processing?
(The Origins of Digital Image Processing)
Fundamental Steps in Digital Image Processing
Image Sampling and Quantization
Spatial and Gray-Level Resolution
Some Basic Relationships Between Pixels
Zooming and Shrinking Digital Images
Lookup tables
Color spaces
Spatial and gray-level resolution
Spatial resolution is rather intuitive, and is determined by the quality and “density” of
the sampling.
Sampling theories (eg Nyquist-Shannon) state that sampling should be performed at
a rate that is at least twice the size of the smallest object/highest frequency.
Based on this, over-sampling and under-sampling (=spatial aliasing) can occur.
Gray level resolution is a term used to describe the binning of the signal rather than
the actual difference we managed to obtain when we quantized the signal. 8-bit and
16-bit images are the most common ones, but 10- and 12-bit images can also be
found.
Changing the resolution of the image without changing bit-depth
checker board patterns
512 x 512
256 x 256
128 x 128
64 x 64
Changing the bit-depth of the image without changing resolution
False contouring
8bit
4bit
3bit
2bit
1bit
Lecture’s outline
What Is Digital Image Processing?
(The Origins of Digital Image Processing)
Fundamental Steps in Digital Image Processing
Image Sampling and Quantization
Spatial and Gray-Level Resolution
Some Basic Relationships Between Pixels
Zooming and Shrinking Digital Images
Lookup tables
Color spaces
Neighbors of a pixel
(x-1, y-1) (x, y-1) (x+1, y-1)
(x-1, y)
(x,y)
(x+1, y)
(x-1, y+1) (x, y+1) (x+1, y+1)
(x+1, y), (x-1, y), (x, y+1), (x, y-1)= 4 neighbors of p, or N4(p)
(x+1, y+1), (x+1, y-1), (x-1, y+1), (x-1, y-1)= the four diagonal neighbors, or Nd(p).
N4(p) together with Nd(p) are N8(p).
Consider the case of image borders.
Adjacency/Connectivity, Regions, and Boundaries
Pixels are said to be connected if they are neighbors and if their gray levels
satisfy a specified criterion of similarity.
Consider this example of binary
pixels
V- the set of gray levels used to define adjacency. In this binary example, V={0}
to define adjacency of pixels with the value 0. In non-binary images, the values
of V can have a wider range.
The region R of an image- a subset of pixels which is a connected set, meaning that
there exists a path that connects the adjacent pixels.
The boundary (=border=contour) of R is the set of pixels in R that have one or more
neighbors that are not in R.
What happens when R is the entire image?
Do not confuse boundary with edge. The edge is formed by discontinuity of gray
levels at a certain point.
In binary images, edges and boundaries correspond.
Distances between pixels
Between (x,y) and (s,t):
Eucladian distance: given by Pythagoras
D4 distance (=city-block distance): D4(p, q) = |x – s| + |y – t|.
Pixel coordinates:
1,3
3,1
2
2,2
3,2
4,2
2,3
3,3
4,3
2,4
3,4
4,4
3,5
5,3
2
2
1
1
0
2
1
2
1
2
2
Diamond pattern
2
D8(p, q) =max( |x – s| , |y – t|) results in a square pattern around the center pixel.
2
2
2
2
1
1
2
1
0
2
1
1
2
2
2
2
1
1
1
2
2
2
2
2
2
Lecture’s outline
What Is Digital Image Processing?
(The Origins of Digital Image Processing)
Fundamental Steps in Digital Image Processing
Image Sampling and Quantization
Spatial and Gray-Level Resolution
Some Basic Relationships Between Pixels
Zooming and Shrinking Digital Images
Lookup tables
Color spaces
Zooming and shrinking digital images
Zoom: 1. Create new pixel locations
2. Assign gray level values to the locations
For increasing the size of an image an integer number of times, the
method of “pixel replication” is used.
For example, when changing a 512 x 512 image to 1024 x 1024, every
column and every row in the original image is duplicated.
At high magnification factors, checkerboard patterns appear.
Nearest neighbor interpolation
Bilinear interpolation (2 x 2)
Bicubic interpolation (4 x 4)
Examples of non-adaptive interpolation
Scaling up using different methods
Pixel replication
Bilinear
Bicubic
Lecture’s outline
What Is Digital Image Processing?
(The Origins of Digital Image Processing)
Fundamental Steps in Digital Image Processing
Image Sampling and Quantization
Spatial and Gray-Level Resolution
Some Basic Relationships Between Pixels
Zooming and Shrinking Digital Images
Lookup tables
Color spaces
Look up tables:
Save computational time (LUTs can be found early in history…)
Require a mapping or transformation function- an equation that converts the
brightness value of the input pixel to another value in the output pixel
Do not alter pixel values
Image transformations that involve look-up tables can be implemented by either
one of two mechanisms: at the input so that the original image data are
transformed, or at the output so that a transformed image is displayed but the
original image remains unmodified.
www.microscopy.fsu.edu
Lecture’s outline
What Is Digital Image Processing?
(The Origins of Digital Image Processing)
Fundamental Steps in Digital Image Processing
Image Sampling and Quantization
Spatial and Gray-Level Resolution
Some Basic Relationships Between Pixels
Zooming and Shrinking Digital Images
Lookup tables
Color spaces
There are ways to describe color images other than the RGB space
Color space=color gamut
RGB= 3 X 8-bit channels= 24bit= true color
The histograms of RGB images can be viewed either as separate channels or as
the weighted average of the channels.
Some representations of color images calculate a weighted average of green,
red and blue.
Hue-Saturation-Intensity (more intuitive, as we perceive the world):
Hue= color spectrum, Saturation= color purity, Intensity= brightness
More: Hue-Saturation-Lightness; Hue-Saturation-Brightness
End of Lecture 1
Thank you!
Download