Neural Network Classifier for Segmentation and Classification of Skin Lesions from Digital Images Jain Elza Jose Anina John Department of Computer Science Department of Computer Science 4th Semester MTech Student Assistant Professor Caarmel Engineering College, Perunad Caarmel Engineering College, Perunad Abstract- Melanoma is a cancer that make melanin in the melanocytes. Melanoma tumors are brown or black. There is a need for an automated system to assess a patient’s risk of melanoma using images of their skin lesions captured using a standards digital camera. Locating the skin lesion in the digital image is a challenging one. The segmentation accuracy is also less. In the existing system, it segments the skin lesions and classify the skin as lesion and normal skin. In the proposed system, texture distictiveness lesion segmentation algorithm is used to segment the skin lesions. The TDLS algorithm consists of two steps. First, a set of sparse texture distributions of normal skin and lesion skin are learned. A TD metric is calculated to measure the dissimilarity of a texture distribution from all other texture distributions. Second, the TD metric is used to classify regions in the image as part of the skin class or lesion class. Finally a neural network classifier is used to classify the lesion as benign or malignant. The proposed framework has higher segmentation accuracy compared to all other tested algorithms. superpixels). The goal of segmentation is to simplify and/or change the representation of an image into something that is more meaningful and easier to analyze. Image segmentation is typically used to locate objects and boundaries (lines, curves, etc.) in images. Index Terms- Melanoma, neural network, segmentation, texture. I. INTRODUCTION Image processing [17] is a signal processing for which the input is an image, such as a photograph. The output may be either an image or a set of characteristics or parameters related to the image. Most image-processing techniques involve treating the image as a two-dimensional signal and applying standard signal-processing techniques to it. Image processing usually refers to digital image processing, but optical and analog image processing also are possible. This article is about general techniques that apply to all of them. The acquisition of images (producing the input image in the first place) is referred to as imaging. Image segmentation [5][16] is the process of partitioning a digital image into multiple segments (sets of pixels, also known as The result of image segmentation is a set of segments that collectively cover the entire image, or a set of contours extracted from the image (see edge detection). Each of the pixels in a region are similar with respect to some characteristic or computed property, such as color, intensity, or texture. Adjacent regions are significantly different with respect to the same characteristics. Melanoma [1][2] is a cancer that begins in the melanocytes. Most of these cells still make melanin, so melanoma tumors are often brown or black. Melanoma most often starts on the trunk (chest or back) in men and on the legs of women, but it can start in other places, too. Having dark skin lowers the risk of melanoma, but a person with dark skin can still get melanoma. Melanoma can almost always be cured in its early stages. But it is likely to spread to other parts of the body if it is not caught early. Due to the increase in incidence rates, early detection of melanoma is essential. To reduce the cost of screening melanoma an automated melanoma screening [2] algorithms have been proposed. A dermatoscopes [3] is a handheld device that optically magnifies, illuminates and enhances skin lesions, allowing the dermatologist to better view the lesion features. Use of the dermatoscope has been found to improve diagnosis, compared to the naked eye. Dermatoscope was used for screening melanoma by the dermatologists. The main reasons against using the dermatoscope include a lack of training or interest. Recent work includes screening the melanoma from digital images. It includes segmentation algorithms for locating the lesion border of the skin. This is important while classifying the lesion based on the features. The main features include the ABCD scale: asymmetry, border irregularity, color variegation, and diameter. The examples of digital images of melanoma are shown in fig. 1(a) and (c). (a) (b) Fig. 1. (a) and (b) are examples of digital images of melanoma. II. RELATED WORKS Existing illumination correction algorithms [6] adjust pixel intensities in an image based on an estimated illumination map. The goal of these algorithms is to remove any external illumination, so that the resulting image is independent of any illumination effects. Some of these effects that should be removed include shadows and bright areas caused by illumination variation. The motivation of this preprocessing step is to improve the performance of subsequent steps, including lesion segmentation and classiffication. Existing general illumination correction algorithms focus on correcting for illumination variation in standard digital images. These algorithms are general and can be applied to any image. Recent work by cavalcanti et al. [6] proposes a correction algorithm specific for skin lesion images. The algorithm fits pixel intensities from the four corners of the photograph to a parametric surface. The disadvantage with this algorithm is that only a very small subset of pixels is used to fit the parametric surface. This results in the estimated illumination map being over-or underestimated. The purpose of image segmentation algorithms is to find and outline distinct objects of importance in an image. For example, for images of skin lesions, the border of the skin lesion should be identified. Segmentation in general is a very well-researched area and many different algorithms have been proposed. Image segmentation [5][16] is perhaps the most studied area in computer vision, with numerous methods reported. A segmentation method is usually designed taking into consideration the properties of a particular class of images. The majority of algorithms only use features derived from pixel color to drive the segmentation. Segmentation is difficult due to the illumination variation. Thresholding algorithms are used for bright areas where there is a reflection of camera flash. In preprocessing, a color image is first transformed into an intensity image in such a way that the intensity at a pixel shows the color distance of that pixel with the color of the background. The color of the background is taken to be the median color of pixels in small windows in the four corners of the image. Preprocessing step is to correct shadows and bright spots caused by illumination variation. Most of the segmentations algorithms use color variation to identify the skin lesions. Textures are also used to identify the skin lesion and normal skin. Since normal skin and lesion skin has different textures. It includes, smoothness, roughness, bumps or ridges etc. Stoecker et al. [7] analyzed texture in skin images using basic statistical approaches, such as the gray-level cooccurrence matrix. They found that texture analysis could accurately find regions with a smooth texture and that texture analysis is applicable to segmentation and classification of dermatological images. Existing texture analysis extracts features and measurements of a texture, allowing textures from different regions to be compared. Texture analysis is useful for image segmentation because different parts of the same object will usually match in texture. Algorithms include using first level statistics, gray-level co-occurrence matrix or haralick statistics. Model-based algorithms use probability models, such as the autoregressive model or markov random field model [4][8l, to characterize textures. Structural algorithms deconstruct and characterize the texture as a number of texture elements. The algorithm proposed by xu et al. [9] learns a model of the normal skin texture using pixels in the four corners of the image, which is later used to find the lesion. Hwang and celebi [10] use gabor filters to extract texture features and use a g-means clustering approach for segmenting the lesion. Fig. 2. Architecture of the System In this paper, it proposes a texture distinctiveness lesion segmentation algorithm (TDLS) to locate skin lesion in the digital images. The TD metric measures the dissimilarity of texture distributions with all other distributions. Then classify it as normal skin or lesion skin. A probabilistic neural network classifier is used to classfy the skin as benign or malignant. In section III system design, the process of learning the sparse texture model and calculating a metric to measure td is described. Finally the lesion is separated. Then the neural network classifier is used to classify the lesion as benign or malignant. In section IV experimental results and discussion about future are explained. III. SYSTEM DESIGN Each row in the neighbourhood is concatenated sequentially. An unsupervised algorithm is used to learn the textures. The k-means clustering algorithm is used here. It is used to find the k clusters of texture data. It is also used to increase the robustness and to speed up the number of iterations. The mean and standard deviations are calculated to find the TD metric. TD means the texture distinctiveness metric. The TD metric is used to classify the image as normal skin and lesion skin. For the normal skin, the dissimilarity of the textures from other is very small. So the TD metric is also small for the normal skin. But for the lesion skin, the TD metric will be large. Because of its dissimilarity from other textures. The proposed system includes a texture distinctiveness lesion segmentation algorithm (TDLS) to learn the texture distributions and to calculate the TD metric. It is used to classify the image as normal skin and lesion skin. Then finally a neural network classifier is used to the lesion as benign or malignant. A. Texture distributions In texture disributions, the input image will transform the RGB image to XYZ image. It will convert the pixel values into XYZ. For each pixel in the image a local texture vector is obtained. The texture vector contains pixels in the neighbourhood of size n centered on the pixel of interest. Fig.3. Map of the texture distinctive metric. In (a) and (c), the original images are shown. In (b) and (d), maps of the texture distinctive metric In Fig. 3, it will show the map of texture distinctiveness metric. Fig. 3 (b) and (d) displays the textural distinctiveness metric for each pixel in the image. In both figures, the lesion is white, which means that it has highest textural distinctiveness metric. lesion as malignant. The resultant graph is shown in fig.4. This will shows the better analysis with the existing classification algorithms. B. Statistical region merging In this, first of all the image is oversegmented, that image is divided into number of regions. For that SRM algorithm is used. It includes mainly two steps that is a sorting step and a merging step. In the sorting step , a four connected graph is constructed. The horizontal and vertical pixels are sorted based on their similarity. In the merging step, it will merge the regions based on the pixel intensities. Here the normal skin and lesion skin are classified. Fig.4. Performance Analysis Graph C. Segmentation refinement In the segmentation refinement, the lesion is refined here. To refine the lesion border postprocessing are applied. It includes two steps: morphological dilation and region selection. The morphological dilation is used to fill the holes and smooth the border of the lesion. The region selection is used to select the lesion region and eliminate the unwanted region. The region which touches the edge of the photographs can be eliminated. Since the lesion will not touch the edge. D. Neural network classifier In this research, as a future enhancement probabilistic feed forward neural network classifier is used. In this, the lesion is segmented. The features such as the area of the lesion, orientation, major axis length, minor axis length, eccentricity, centroid, momentum, red average value are extracted. The training towards the benign and melanoma are tested. Finally, it is used to classify the lesion as benign or malignant. IV. EXPERIMENTAL RESULTS AND DISCUSSIONS The objective of this experiment is to measure the classification accuracy of the lesion that is classified as benign or malignant. This experiment is done with 50 benign images and 50 malignant images. The experiment shows about 98% accuracy for classifying the lesion as benign and 99% accuracy for classifying the lesion as malignant. In the existing systems it is only 80% for classifying the lesion as benign and 81% for classifying the V. CONCLUSION AND SCOPE OF THE WORK In summary, a novel lesion segmentation algorithm using the concept of learning is proposed. Texture distinctiveness lesion segmentation algorithm is used. It captures dissimilarity between the texture distributions. Then image is divided into smaller regions and classified as lesion or skin based on TD map. In future enhancement, it will work with neural network classifier. The classifier will extract features and finally classify the lesion as benign or malignant. Experience and training based learning is an important characteristic of neural networks that makes it ideal for diagnosis applications. From the segmented images neural network learns skin and lesion pixel values. The proposed framework achieves higher segmentation and classification accuracy. As a future work, it is expected to work with border accuracy of benign or malignant image. REFERENCES [1]. D. S. Rigel, r. J. Friedman, and a. W. Kopf, “the incidence of melanoma in the united states: issues as we approach the 21st century,” j. Amer. Acad. Dermatol., vol. 34, no. 5, pp. 839–847,1996. [2]. R. Amelard, j. Glaister, a. Wong, and d. A. Clausi, \melanoma decision support using lighting-corrected intuitive feature models," in computer vision techniques for the diagnosis of skin cancer, j. Scharcanski and m. E. Celebi, eds. Springer,accepted, 2013. [3]. Starck, m. Elad, and d. Donoho, “image decomposition via the combination of sparse representations and a variational approach,” ieee trans. Image process., vol. 14, no. 10, pp. 1570–1582, oct. 2005. j. M. Grichnik, a. A. Marghoob, h. S. Rabinovitz, and s. W. Menzies, “border detection in dermoscopy images using statistical region merging,” skin res.technol., vol. 14, no. 3, pp. 347–353, 2008. [4]. C. Serrano and b. Acha, “pattern analysis of dermoscopic images based on markov random fields,” pattern recog., vol. 42, no. 6, pp. 1052–1057, 2009. [5]. L. Xu, m. Jackowskia, a. Goshtasby, d. Roseman, s. Bines, c. Yu, a. Dhawan, and a. Huntley, “segmentation of skin cancer images,” image vis. Comput., vol. 17, pp. 65–74, 1999. [13]. P. G. Cavalcanti, j. Scharcanski, andc. B. O. Lopes, “shading attenuation in human skin color images,” in advances in visual computing, g. Bebis, r. Boyle, b. Parvin, d. Koracin, r. Chung, r. Hammoud, m. Hussain, t. Karhan, r. Crawfis, d. Thalmann, d. Kao, and l. Avila, eds., (ser. Lecture notes in computer science), vol. 6453 heidelberg, germany: springer, 2010, pp. 190–198. [6]. P. G. Cavalcanti and j. Scharcanski, \automated prescreening of pigmented skin lesions using standard cameras," computerized medical imaging and graphics, vol. 35, no. 6, pp. 481{491, sept 2011. [14]. P. G. Cavalcanti and j. Scharcanski, “automated prescreening of pigmented skin lesions using standard cameras,” comput.med. Imag. Graph.vol. 35, no. 6, pp. 481–491, sep. 2011. [7]. W. V. Stoecker, c.-s. Chiang, and r. H. Moss, “texture in skin images: comparison of three methods to determine smoothness,” comput. Med. Imag. Graph., vol. 16, no. 3, pp. 179–190, 1992. [8]. C. Serrano and b. Acha, “pattern analysis of dermoscopic images based on markov random fields,” pattern recog., vol. 42, no. 6, pp. 1052–1057, 2009. [15]. Jeffrey glaister, alexander wong “segmentation of skin lesions from digital images using joint statistical texture distinctiveness”ieee trans.bio.med,april 2014,vol.61,no. 4. [9]. L. Xu, m. Jackowskia, a. Goshtasby, d. Roseman, s. Bines, c. Yu, a. Dhawan, and a. Huntley, “segmentation of skin cancer images,” image vis. Comput., vol. 17, pp. 65–74, 1999. [10]. S. Hwang and m. E. Celebi, “texture segmentation of dermoscopy images using gabor filters and g-means clustering,” in proc. Int. Conf.image process., comput. Vision, pattern recog, jul. 2010, pp. 882–886. [11]. G. Peyre, \sparse modeling of textures," journal of mathematical imaging and vision, vol. 34, no. 1, pp. 17{31, 2009. [12]. M. E. Celebi, h. A. Kingravi, h. Iyatomi, y. A. Aslandogan,w. V. Stoecker, r. H. Moss, j. M. Malters, [16] C. Carson, s. Belongie, h. Greenspan, and j. Malik. Blobworld: image segmentation using expectationmaximization and its application to image querying. Ieee trans. Pattern anal. And machine intell., 24(8):1026–1038, 2002. [17]. Jensen, J.R. 1996. Introduction to Digital Image Processing: A Remote Sensing Perspective.Practice Hall, New Jersey.