Image Segmentation Techniques Supradeep Narayana (104902068) (For Partial fulfillment of the course ESE 558 requirements) Image Segmentation Introduction: Segmentation is a process of that divides the images into its regions or objects that have similar features or characteristics. Some examples of image segmentation are 1. In automated inspection of electronic assemblies, presence or absence of specific objects can be determined by analyzing images. 2. Analyzing aerial photos to classify terrain into forests, water bodies etc. 3. Analyzing MRI and X-ray images in medicine for classify the body organs. Some figures which show segmentation are Segmentation has no single standard procedure and it is very difficult in non-trivial images. The extent to which segmentation is carried out depends on the problem Specification. Segmentation algorithms are based on two properties of intensity valuesdiscontinuity and similarity. First category is to partition an image based on the abrupt changes in the intensity and the second method is to partition the image into regions that are similar according to a set of predefined criteria. In this report some of the methods for determining the discontinuity will be discussed and also other segmentation methods will be attempted. Three basic techniques for detecting the gray level discontinuities in a digital images points, lines and edges. The other segmentation technique is the thresholding. It is based on the fact that different types of functions can be classified by using a range functions applied to the intensity value of image pixels. The main assumption of this technique is that different objects will have distinct frequency distribution and can be discriminated on the basis of the mean and standard deviation of each distribution. Segmentation on the third property is region processing. In this method an attempt is made to partition or group regions according to common image properties. These image properties consist of Intensity values from the original image, texture that are unique to each type of region and spectral profiles that provide multidimensional image data. A very brief introduction to morphological segmentation will also be given. This method combines most of the positive attributes of the other image segmentation methods. Segmentation using discontinuities Several techniques for detecting the three basic gray level discontinuities in a digital image are points, lines and edges. The most common way to look for discontinuities is by spatial filtering methods. Point detection idea is to isolate a point which has gray level significantly different form its background. w1=w2=w3=w4=w6=w7=w8=w9 =-1, w5 = 8. Response is R = w1z1+w2z2……+w9z9, where z is the gray level of the pixel. Based on the response calculated from the above equation we can find out the points desired. Line detection is next level of complexity to point detection and the lines could be vertical, horizontal or at +/- 45 degree angle. Responses are calculated for each of the mask above and based on the value we can detect if the lines and their orientation. Edge detection The edge is a regarded as the boundary between two objects (two dissimilar regions) or perhaps a boundary between light and shadow falling on a single surface. To find the differences in pixel values between regions can be computed by considering gradients. The edges of an image hold much information in that image. The edges tell where objects are, their shape and size, and something about their texture. An edge is where the intensity of an image moves from a low value to a high value or vice versa. There are numerous applications for edge detection, which is often used for various special effects. Digital artists use it to create dazzling image outlines. The output of an edge detector can be added back to an original image to enhance the edges. Edge detection is often the first step in image segmentation. Image segmentation, a field of image analysis, is used to group pixels into regions to determine an image's composition. A common example of image segmentation is the "magic wand" tool in photo editing software. This tool allows the user to select a pixel in an image. The software then draws a border around the pixels of similar value. The user may select a pixel in a sky region and the magic wand would draw a border around the complete sky region in the image. The user may then edit the color of the sky without worrying about altering the color of the mountains or whatever else may be in the image. Edge detection is also used in image registration. Image registration aligns two images that may have been acquired at separate times or from different sensors. Figure e1 Different edge profiles. There is an infinite number of edge orientations, widths and shapes (Figure e1). Some edges are straight while others are curved with varying radii. There are many edge detection techniques to go with all these edges, each having its own strengths. Some edge detectors may work well in one application and perform poorly in others. Sometimes it takes experimentation to determine what the best edge detection technique for an application is. The simplest and quickest edge detectors determine the maximum value from a series of pixel subtractions. The homogeneity operator subtracts each 8 surrounding pixels from the center pixel of a 3 x 3 window as in Figure e2. The output of the operator is the maximum of the absolute value of each difference. new pixel = maximum{½ 1111½ , ½ 1113½ , ½ 1115½ , ½ 1116½ ,½ 1111½ , ½ 1116½ ,½ 1112½ ,½ 1111½ } = 5 Figure e2 How the homogeneity operator works. Similar to the homogeneity operator is the difference edge detector. It operates more quickly because it requires four subtractions per pixel as opposed to the eight needed by the homogeneity operator. The subtractions are upper left lower right, middle left middle right, lower left upper right, and top middle bottom middle (Figure e3). new pixel = maximum{½ 1111½ , ½ 1312½ , ½ 1516½ , ½ 1116½ } = 5 Figure e3 How the difference operator works. First order derivative for edge detection If we are looking for any horizontal edges it would seem sensible to calculate the difference between one pixel value and the next pixel value, either up or down from the first (called the crack difference), i.e. assuming top left origin Hc = y_difference(x, y) = value(x, y) – value(x, y+1) In effect this is equivalent to convolving the image with a 2 x 1 template Likewise Hr = X_difference(x, y) = value(x, y) – value(x – 1, y) uses the template –1 1 Hc and Hr are column and row detectors. Occasionally it is useful to plot both X_difference and Y_difference, combining them to create the gradient magnitude (i.e. the strength of the edge). Combining them by simply adding them could mean two edges canceling each other out (one positive, one negative), so it is better to sum absolute values (ignoring the sign) or sum the squares of them and then, possibly, take the square root of the result. It is also to divide the Y_difference by the X_difference and identify a gradient direction (the angle of the edge between the regions) The amplitude can be determine by computing the sum vector of Hc and Hr Sometimes for computational simplicity, the magnitude is computed as The edge orientation can be found by In real image, the lines are rarely so well defined, more often the change between regions is gradual and noisy. The following image represents a typical read edge. A large template is needed to average at the gradient over a number of pixels, rather than looking at two only Sobel edge detection The Sobel operator is more sensitive to diagonal edges than vertical and horizontal edges. The Sobel 3 x 3 templates are normally given as X-direction Y-direction Original image absA + absB Threshold at 12 Other first order operation The Roberts operator has a smaller effective area than the other mask, making it more susceptible to noise. The Prewit operator is more sensitive to vertical and horizontal edges than diagonal edges. The Frei-Chen mask In many applications, edge width is not a concern. In others, such as machine vision, it is a great concern. The gradient operators discussed above produce a large response across an area where an edge is present. This is especially true for slowly ramping edges. Ideally, an edge detector should indicate any edges at the center of an edge. This is referred to as localization. If an edge detector creates an image map with edges several pixels wide, it is difficult to locate the centers of the edges. It becomes necessary to employ a process called thinning to reduce the edge width to one pixel. Second order derivative edge detectors provide better edge localization. Example. In an image such as The basic Sobel vertical edge operator (as described above) will yield a value right across the image. For example if is used then the results is Implementing the same template on this "all eight image" would yield This is not unlike the differentiation operator to a straight line, e.g. if y = 3x-2. Once we have gradient, if the gradient is then differentiated and the result is zero, it shows that the original line was straight. Images often come with a gray level "trend" on them, i.e. one side of a regions is lighter than the other, but there is no "edge" to be discovered in the region, the shading is even, indicating a light source that is stronger at one end, or a gradual color change over the surface. Another advantage of second order derivative operators is that the edge contours detected are closed curves. This is very important in image segmentation. Also, there is no response to areas of smooth linear variations in intensity. The Laplacian is a good example of a second order derivative operator. It is distinguished from the other operators because it is omnidirectional. It will highlight edges in all directions. The Laplacian operator will produce sharper edges than most other techniques. These highlights include both positive and negative intensity slopes. The edge Laplacian of an image can be found by convolving with masks such as or The Laplacian set of operators is widely used. Since it effectively removes the general gradient of lighting or coloring from an image it only discovers and enhances much more discrete changes than, for example, the Sobel operator. It does not produce any information on direction which is seen as a function of gradual change. It enhances noise, though larger Laplacian operators and similar families of operators tend to ignore noise. Determining zero crossings The method of determining zero crossings with some desired threshold is to pass a 3 x 3 window across the image determining the maximum and minimum values within that window. If the difference between the maximum and minimum value exceed the predetermined threshold, an edge is present. Notice the larger number of edges with the smaller threshold. Also notice that the width of all the edges are one pixel wide. A second order derivative edge detector that is less susceptible to noise is the Laplacian of Gaussian (LoG). The LoG edge detector performs Gaussian smoothing before application of the Laplacian. Both operations can be performed by convolving with a mask of the form where x, y present row and column of an image, s is a value of dispersion that controls the effective spread. Due to its shape, the function is also called the Mexican hat filter. Figure e4 shows the cross section of the LoG edge operator with different values of s. The wider the function, the wider the edge that will be detected. A narrow function will detect sharp edges and more detail. Figure e4 Cross selection of LoG with various s. The greater the value of s, the wider the convolution mask necessary. The first zero crossing of the LoG function is at . The width of the positive center lobe is twice that. To have a convolution mask that contains the nonzero values of the LoG function requires a width three times the width of the positive center lobe (8.49s). Edge detection based on the Gaussian smoothing function reduces the noise in an image. That will reduce the number of false edges detected and also detects wider edges. Most edge detector masks are seldom greater than 7 x 7. Due to the shape of the LoG operator, it requires much larger mask sizes. The initial work in developing the LoG operator was done with a mask size of 35 x 35. Because of the large computation requirements of the LoG operator, the Difference of Gaussians (DoG) operator can be used as an approximation to the LoG. The DoG can be shown as The DoG operator is performed by convolving an image with a mask that is the result of subtracting two Gaussian masks with different a values. The ratio s 1/s 2 = 1.6 results in a good approximation of the LoG. Figure e5 compares a LoG function (s = 12.35) with a DoG function (s1 = 10, s2 = 16). Figure e5 LoG vs. DoG functions. One advantage of the DoG is the ability to specify the width of edges to detect by varying the values of s1 and s2. Here are a couple of sample masks. The 9 x 9 mask will detect wider edges than the 7x7 mask. For 7x7 mask, try For 9 x 9 mask, try Segmentation using thresholding. Thresholding is based on the assumption that the histogram is has two dominant modes, like for example light objects and an dark background. The method to extract the objects will be to select a threshold F(x,y)= T such that it separates the two modes. Depending on the kind of problem to be solved we could also have multilevel thresholding. Based on the region of thresholding we could have global thresholding and local thresholding. Where global thresholding is considering the function for the entire image and local thresholding involving only a certain region. In addition to the above mentioned techniques that if the thresholding function T depends on the spatial coordinates then it is known as the dynamic or adaptive thresholding. Let us consider a simple example to explain thresholding. Figure: for hypothetical frequency distribution of intensity values for fat , muscel and bone A hypothetical frequency distribution f(I) of intensity values I(x,y) for fat, muscle and bone, in a CT image. Low intensity values correspond to fat tissues, whereas high intensity values correspond to bone. Intermediate intensity values correspond to muscle tissue. F+ and F- refer to the false positives and false negatives; T+ and T- refer to the true positives and true negatives. Basic global thresholding technique: In this technique the entire image is scanned by pixel after pixel and hey is labeled as object or the background, depending on whether the gray level is greater or lesser than the thresholding function T. The success depends on how well the histogram is constructed. It is very successful in controlled environments, and finds its applications primarily in the industrial inspection area. The algorithm for global thresholding can be summarized in a few steps. Select an initial estimate for T. 2) Segment the image using T. This will produce two groups of pixels. G1 consisting of all pixels with gray level values >T and G2 consisting of pixels with values <=T. 3) Compute the average gray level values mean1 and mean2 for the pixels in regions G1 and G2. 4) Compute a new threshold value T=(1/2)(mean1 +mean2). 5) Repeat steps 2 through 4 until difference in T in successive iterations is smaller than a predefined parameter T0. Basic adaptive thresholding technique: Images having uneven illumination make it difficult to segment using the histogram. In this case we have to divide the image in many sub images and then come up with different threshold to segment each sub image. The key issues are how to divide the image into sub images and utilize a different threshold to segment each sub image. The major drawback to threshold-based approaches is that they often lack the sensitivity and specificity needed for accurate classification. Optimal thresholding technique: In the above two sections we described what global and adaptive thresholding mean. Below we illustrate how to obtain minimum segmentation error. Let us consider an image with 2 principle gray levels regions. Let z denote the gray level values. Values as random quantities and their histogram may be considered an estimate of probability P(z). Overall density function is the sum or mixture of two densities, one of them is for the light and other is for the dark region. The total probability density function is the P(z) = P1 p1(z)+P2 p2(z) ,Where P1 and P2 are the probabilities of the pixel (random). P1 +P2 =1. The overall error of probability is E(T) = P2 E1(T) + P1 E2(T), where E1 and E2 are the probability of occurrence of object or background pixels. We need to find the threshold value of the error E(T) , so by differentiating E w.r.t T we obtain P1p1(T) = P2p2(T). So we can use Gaussian probability density functions and obtain the value of T. T = (μ1+μ2)/2 + (σ^2/(μ1-μ2))*ln(P2/P1). Where μ and σ^2 are the mean and variance of the Gaussian function for the object of a class. The other method for finding the minimum error is finding the mean square error , to estimate the gray-level PDF of an image from image histogram. Ems = (1/n)* (∑ (p(zi) –h(zi))^2 ) for i= 1 to n. Where n is the number of points in the histogram. The important assumption is that either one of the objects or both are considered. Probability of classifying the objects and background is classified erroneously. Region based segmentation. We have seen two techniques so far. One dealing with the gray level value and other with the thresholds. In this section we will concentrate on regions of the image. Formulation of the regions: An entire image is divided into sub regions and they must be in accordance to some rules such as 1. Union of sub regions is the region 2. All are connected in some predefined sense. 3. No to be same, disjoint 4. Properties must be satisfied by the pixels in a segmented region P(Ri)=true if all pixels have same gray level. 5. Two sub regions should have different sense of predicate. Segmentation by region splitting and merging: The basic idea of splitting is, as the name implies, to break the image into many disjoint regions which are coherent within themselves. Take into consideration the entire image and then group the pixels in a region if they satisfy some kind of similarity constraint. This is like a divide and conquers method. Merging is a process used when after the split the adjacent regions merge if necessary. Algorithms of this nature are called split and merge algorithms. consider the example of the split and merge process. Fig: Image tree split –merge. Segmentation by region growing Region growing approach is the opposite of split and merges. 1. An initial set of small area are iteratively merged based on similarity of constraints. 2. Start by choosing an arbitrary pixel and compared with the neighboring pixel. 3. Region is grown from the seed pixel by adding in neighboring pixels that are similar, increasing the size of the region. 4 When the growth of one region stops we simply choose another seed pixel which does not yet belong to any region and start again. 5 This whole process is continued until all pixels belong to some region. 6 A bottom up method. Some of the undesirable effects of the region growing are . Current region dominates the growth process -- ambiguities around edges of adjacent regions may not be resolved correctly. Different choices of seeds may give different segmentation results. Problems can occur if the (arbitrarily chosen) seed point lies on an edge. However starting with a particular seed pixel and letting this region grow completely before trying other seeds biases the segmentation in favor of the regions which are segmented first. To counter the above problems, simultaneous region growing techniques have been developed. Similarities of neighbouring regions are taken into account in the growing process. No single region is allowed to completely dominate the proceedings. A number of regions are allowed to grow at the same time. o similar regions will gradually coalesce into expanding regions. Control of these methods may be quite complicated but efficient methods have been developed. Easy and efficient to implement on parallel computers. Segmentation by Morphological watersheds: This method combines the positive aspects of many of the methods discussed earlier. The basic idea to embody the objects in “watersheds” and the objects are segmented. Below only the basics of this method is illustrated without going into greater details. The concept of watersheds: It is the idea of visualizing an image in 3D. 2 spatila versus gray levels. So all points in such a topology are either 1. belonging to regional minimum. 2. all with certain to a single minimum. 3. equal to two points where more than one minimum A particular region is called watershed if it is a region minimum satisfying certain conditions. Watershed lines: Simple if we have a hole and water is poured at a constant rate. The level of water rises and fills the region uniformly. When the regions are about to merge with the remaining regions we build dams. Dams are boundaries. The idea is more clearly illustrated with the help of diagrams. The heights of the structures are proportional to the gray level intensity. Also the entire structure is enclosed by the height of the dam greatest of the dam height. In the last figure we can see that the water almost fills the dams out, until the highest level of the gray level in the images researched. The final dam corresponds to the watershed lines which are the desired segmentation result. The principle applications of the method are in the extraction of uniform objects from the background. Regions are characterized by small variations in gray levels, have small gradient values. So it is applied to the gradient than the image. Region with minimum correlated with the small value of gradient corresponding to the objects of interest. Use of Motion in segmentation: Motion of objects can be very important tool to exploit when the background detail is irrelevant. This technique is very common in sensing applications. Let us consider two image frames at time t1 and t2, f(x,y,t1) and f(x,y,t2) and compare them pixel to pixel. One method to compare is to take the difference of the pixels D12(x,y) = = 1 if | f(x,y,t1) – f(x,y,t2)| > T, 0 otherwise. Where T is a threshold value. This threshold is to signify that only when the there is a appreciable change in the gray level, the pixels are considered to be different. In dynamic image processing the D12 has value set to 1 when the pixels are different; to signify the objects are in motion. Image segmentation using edge flow techniques: A region-based method usually proceeds as follows: the image is partitioned into connected regions by grouping neighboring pixels of similar intensity levels. Adjacent regions are then merged under some criterion involving perhaps homogeneity or sharpness of region boundaries. Over stringent criteria create fragmentation; lenient ones overlook blurred boundaries and over-merge. Hybrid techniques using a mix of the methods above are also popular. A connectivity-preserving relaxation-based segmentation method, usually referred to as the active contour model, was proposed recently. The main idea is to start with some initial boundary shape represented in the form of spline curves, and iteratively modifies it by applying various shrink/expansion operations according to some energy function. Although the energy-minimizing model is not new, coupling it with the maintenance of an ``elastic'' contour model gives it an interesting new twist. As usual with such methods, getting trapped into a local minimum is a risk against which one must guard; this is no easy task. In [2], the authors create a combined method that integrates the edge flow vector field to the curve evolution framework. Theory and algorithm of Edge flow and curve evolution: Active contours and curve evolution methods usually define an initial contour C0 and deform it towards the object boundary. The problem is usually formulated using partial differential equations (PDE). Curve evolution methods can utilize edge information, regional properties or a combination of them. Edge-based active contours try to fit an initial closed contour to an edge function generated from the original image. The edges in this edge function are not connected, so they don't identify regions by themselves. An initial closed contour is slowly modified until it fits on the nearby edges. Let C( ):[0,1] →R2 be a parameterization of a 2-D closed curve. A fairly general curve evolution can be written as: 1 where κ is the curvature of the curve, N is the normal vector to the curve, , α β are constants, and S is an underlying velocity field whose direction and strength depend on the time and position but not on the curve front itself. This equation will evolve the curve in the normal direction. The first term is a constant speed parameter that expands or shrinks the curve, second term uses the curvature to make sure that the curve stays mooth at all times and the third term guides the curve according to an independent velocity field. In their independent and parallel works, Caselles et al.and Malladi et al. initialize a small curve inside one of the object regions and let the curve evolve until it reaches the object boundary. The evolution of the curve is controlled by the local gradient. This can be formulated by modifying (1) as: (2) where , F ε are constants, and g = 1/(1+ I ) . I is the Gaussian smoothed image. This is a pure geometric approach and the edge function, g, is the only connection to the image. Edge flow image segmentation [3] is a recently proposed method that is based on filtering and vector diffusion techniques. Its effectiveness has been demonstrated on a large class of images. It features multiscale capabilities and uses multiple image attributes such as intensity, texture or color. As a first step, a vector field is defined on the pixels of the image grid. At each pixel, the vector’s direction is oriented towards the closest image discontinuity at a predefined scale. The magnitude of the vectors depends on the strength and the distance of the discontinuity. After generating this vector field, a vector diffusion algorithm is applied to detect the edges. This step is followed by edge linking and region merging to achieve a partitioning of the image. Details can be found at [3]. Two key components shaping the curve evolution are the edge function g and the external force field F. The purpose of the edge function is to stop or slow down the evolving contour when it is close to an edge. So g is defined to be close to 0 on the edges and 1 on homogeneous areas. The external force vectors F ideally attract the active contour towards the boundaries. At each pixel, the force vectors point towards the closest object boundary on the image. In [2], the authors use the edgeflow vector as the external force Summary Image segmentation forms the basics of pattern recognition and scene analysis problems. The segmentation techniques are numerous in number but the choice of one technique over the other depends only on the application or requirements of the problem that is being considered. In this report we have considered illustrating a few techniques. But the numbers of techniques are so large they cannot be all addressed. Some of the demonstrations of the techniques discussed in the report one can visit [6]. References: 1. Rafael C. Gonzalez and Richard E. Woods,” Digital Image Processing”, 2Ed, 2002. 2. B. Sumengen, B. S. Manjunath, C. Kenney, "Image Segmentation using Curve Evolution and Flow Fields,” Proceedings of IEEE International Conference on Image Processing (ICIP), Rochester, NY, USA, September 2002. 3. W. Ma, B.S. Manjunath, “Edge Flow: a technique for boundary detection and image segmentation,” Trans. Image Proc., pp. 1375-88, Aug. 2000 4. Venugopal, “Image segmentation “written reports 2003. 5. Jin Wang,” Image segmentation Using curve evolution and flow fields”, written reports 2003. 6. http://vision.ece.ucsb.edu/segmentation