IMAGE-BASED OBJECT DETECTION AND IDENTIFICATION SALIM MOHAMMED HUMAID ALWAILI A thesis submitted in fulfillment of the requirements for the award of the degree of Master of Science (Remote Sensing) Faculty of Geoinformation and Science Engineering Universiti Teknologi Malaysia May 2007 PSZ 19:16 (Pind. 1/97) UNIVERSITI TEKNOLOGI MALAYSIA BORANG PENGESAHAN STATUS TESISυ JUDUL: IMAGE-BASED OBJECT DETECTION AND IDENTIFICATION ____________________________________________________________________ ____________________________________________________________________ SESI PENGAJIAN: Saya II(2006/2007) SALIM MOHAMMED HUMAID ALWAILI (HURUF BESAR) mengaku membenarkan tesis (PSM/Sarjana/Doktor Falsafah)* ini disimpan di Perpustakaan Universiti Teknologi Malaysia dengan syarat-syarat kegunaan seperti berikut: 1. Tesis adalah hakmilik Universiti Teknologi Malaysia. 2. Perpustakaan Universiti Teknologi Malaysia dibenarkan membuat salinan untuk tujuan pengajian sahaja. 3. Perpustakaan dibenarkan membuat salinan tesis ini sebagai bahan pertukaran antara institusi pengajian tinggi. 4. ** Sila tandakan (4) 9 SULIT (Mengandungi maklumat yang berdarjah keselamatan atau kepentingan Malaysia seperti yang termaktub di alam AKTA RAHSIA RASMI 1972) TERHAD (Mengandungi maklumat TERHAD yang telah ditentukan oleh organisasi/badan di mana penyelidikan dijalankan) TIDAK TERHAD Disahkan oleh -------------------------------------------------------- ---------------------------------------------------------- (TANDATANGAN PENULIS) (TANDATANGAN PENYELIA) Alamat Tetap: PO Box 113, PC 113 NATIONAL SURVEY AUTHORITY SULTANATE OF OMAN Tarikh: 15TH OF MAY 2007 CATATAN: EN. SAMSUDIN BIN AHMAD Nama Penyelia Tarikh: 15TH OF MAY 2007 * ** Potong yang tidak berkenaan. Jika tesis ini SULIT atau TERHAD, sila lampirkan surat daripada pihak berkuasa/organisasi berkenaan dengan menyatakan sekali sebab dan tempoh tesis ini perlu υ Tesis dimaksudkan sebagai tesis bagi Ijazah Doktor Falsafah dan Serjana secara penyelidikan, atau disertasi bagi pengajian secara kerja kursus dan penyelidikan, atau Laporan Projek Sarjana Muda (PSM). “I hereby declare that I have read this thesis and in my opinion this thesis is sufficient in terms of scope and quality for the award of the degree of Masters of Science (Remote Sensing)” Signature : .......................................................... Name of Supervisor I : EN. SAMSUDIN BIN AHMAD Date : 15TH OF MAY 2007 ii I declare that this thesis entitled “Image-Based Object Detection and Identification” is the result of my own research except as cited in the references. The thesis has not been accepted for any degree and is not concurrently submitted in candidature of any other degree. Signature :............................................................ Name : Salim Mohammed Humaid Alwaili Date :15th of May 2007 iii To the spirit of my beloved father, To my mother, To my wife Ziyana, And to my kids Bader, Ahmed, and Mohammed iv ACKNOWLEDGEMENT I wish to express my sincere appreciation to the Head of Remote Sensing section Professor Dr. Mazlan Bin Hashim, for his encouragement, guidance, patience, and support. I will always be indebted to him for her persistence in ensuring that this thesis is up to the academic world standards. I am also very thankful to my supervisor En. Samsudin Bin Ahmad for his guidance. I am also thankful to the staff in the Faculty of Geoinformation and Science Engineering for their assistance during the past three semesters. Last, but not least, I am extremely grateful and indebted to my sponsors, the National Survey Authority. In particular, I would like to thank the Head of the NSA Brigadier Mohammed Said Al-Kharousi and the Director of Support Services. v ABSTRACT Object detection is a process of identifying and locating objects in image scene. It is a process of combining digital image processing and computer vision. There are two main approaches for object detection, namely contour-based and region-based detection for both approaches. Segmentation plays an important role in object-detection. Detection and extraction of the spatial and spectral responses of different targets of interest could only be carried out with different approaches. In this study, object-based detection is carried out to identify and recognize typical military targets (Plane1, Plane2, Tank) located on natural background using spectral and spatial approaches. This is useful in military intelligence or strategic planning where it assists in identifying the aforementioned targets. In the spectral-based object detection, the spectroradiometer readings of the targets are individually observed on different backgrounds (white platform used as calibration plane, sand and camouflage). These spectroradiometer observations are reduced and examined for forming the unique spectral signatures of the targets. Spatial-characteristics of the targets are also examined using object-based detection where multi-resolution segmentation were employed for this purpose. Results of the study obtained that all the three targets can be extracted with more than 20% differentiation, while the spectral characteristics of the targets are also good indications for object detection with overall accuracy of RMSE = ± 0.6. vi ABSTRAK Pengesanan objek daripada imej adalah satu proses di mana objek sasaran dikenalpasti dan ditentukan. Proses ini menggabungkan pemprosesan imej digital dengan ‘ computer vision’. Terdapat dua pendekatam utama dalam proses pengesanan objek iaitu Berdasarkan Kontur dan Berdasarkan Bahagian. Teknik Segmentasi merupakan teknik utama bagi kedua-dua pendekatan ini. Pengesanan penentuan elemen spatial dan spektral objek sasaran adalah fokus utama dalam kajian ini, di mana sasaran utama yang digunakan mempunyai kepentingan ketenteraan, iaitu Kapal Terbang Pejuang 1, Kapal Terbang Pejuang 2 dan Kereta Kebal. Bagi Pendekatan Spektral, cerapan alat spektroradiometer dijalankan bagi semua objek sasaran dengan pelbagai variasi latarbelakang. Terdiri daripada permukaan putih yang digunakan sebagai rujukan kalibrasi, latarbelakang pasir dan bahan penyamaran. Cerapan spektroradiometer ini kemudiannya dilaraskan sebelum pencarian ciri-ciri ‘spectral signatures’ sasaran dibuat. Penentuan ciri-ciri spatial objek dijalankan dengan menggunakan ‘multi-resolution segmentation’. Hasil daripada kajian ini mendapati kaedah spektral boleh digunakan untuk mengenalpasti sasaran yang diuji dengan ketepatan RMSE = ± 0.6. vii TABLE OF CONTENTS CHAPTER 1 2 TITLE PAGE TITLE i DECLEARATION ii DEDICATION iii AKNOWLEDGMENT iv ABSTRACT v ABSTRAK vi TABLE OF CONTENTS vii LISTS OF TABLES ix LIST OF FIGURES x LIST OF APPENDIX xi INTRODUCTION 1 1.1 The Importance of Object Detection 1 1.2 Problem Statement 4 1.3 Objectives 4 1.4 Scope of the Study 5 1.5 Significance of the Research 6 1.6 Chapter Organization 6 OBJECT DETECTION APPROACHES AND APPLICATIONS 7 2.1 Introduction 7 2.2 Approaches to Object Detection and Recognition 7 2.3 Image Segmentation 11 2.3.1 Contour Segmentation 11 2.3.2 Region Segmentation 12 2.3.3 Object Oriented Approach 14 Reflectance and Object Detection 14 2.4 viii 3 2.5 Automatic Object Detection 15 2.6 Summary 17 METHODOLOGY 18 3.1 Introduction 18 3.2 Materials 20 3.2.1 Target Sets 20 3.2.2 The Data Capture 20 3.3 4 5 3.2.2.1 Digital Camera Nikon Coolpix 7900 21 3.2.2.2 Spectro-radiometer 21 3.2.2.3 Images of Target Sets 21 3.2.2.4 Reflectance Measurement 24 Data Pre-processing 25 3.3.1 Radiometer Correction 25 3.3.2 Geometric Rectification 27 3.3.3 Image Segmentation 28 3.3.4 Multiresolution Segmentation 29 3.3.4.1 Classification of Segmented Image 30 3.3.4.2 Classification Based Segmentation 33 3.3.4.3 Feature Space Optimization 34 RESULTS AND ANALYSIS 35 4.1 Introduction 35 4.2 Object Detection Based on Reflectance 36 4.3 Spectral Assessment 41 4.4 Segmentation Results 43 4.5 Assessment of Object Detection Based on Spatial Entities 46 4.6 Discussion 48 CONCLUSIONS AND RECOMMENDATIONS 50 5.1 Conclusions 50 5.2 Recommendations 51 ix LIST OF TABLES TABLE No. 3.1 TITLE Descriptive statistics and the amount shift applied to the PAGE 26 red and blue bands to normalize radiometric effects 3.2 The result of rectification process 27 3.3 Signatures of the class of interest of the target set 32 4.1 Separation of spectral reflectance curve at RGB bands of 37 the targets for the white board set 4.2 Separation of spectral reflectance curve at RGB bands of 39 the targets for the sand background set 4.3 Separation of spectral reflectance curve at RGB bands of 41 the targets for the camouflage set 4.4 Assessment for the spectral reflectance curves for the 3 42 target sets for the RGB bands 4.5 Overall assessment for the spectral reflectance curves for 42 the 3 target sets for the RGB bands 4.6 Accuracy assessment for the segmentation process 45 4.7 Results of feature space optimization 45 4.8 A matrix for feature space optimization 46 4.9 Results of spatial assessment 48 x TABLE OF FIGURES FIGURE No. TITLE PAGE 2.1 Block diagram of object detection 10 2.2 Region-based segmentation diagram 13 2.3 Automatic Target Detection Process 16 3.1 Flow chart for the methodology carried out in the study 19 3.2 The target set 22 3.3 Configuration of spectro-radiometer with respect with 25 respect target platform for calibration purpose 3.4 The histograms for the RGB bands of the calibration 26 board 3.5 The image segmentation process in the eCognition 29 system adopted in the study 3.6 Image classification 32 4.1 Reflectance curve for the white board target set 36 4.2 Reflectance curve for the sand target set 38 4.3 Reflectance curve for the camouflage target set 40 4.4 Results of the object detection using image segmentation 43 4.5 Assessment of spatial entities based on generated 47 template xi LIST OF APPENDIXES APPENDIX No. APPENDIX I TITLE Description of Generic Shape Features PAGE 57 CHAPTER 1 INTRODUCTION 1.1 The Importance of Object Detection Object detection is a process of identifying and locating objects in a scene. It is a process of combining digital image processing and computer vision (Gonzalez et al, 2001). The main goal of an object detection process is to detect and identify objects in a scene. Selected parameters, which are commonly used for object detection such as the interpretation parameters aimed to recognizes and identify objects, will be described in the following sections. The commonly used parameters in object detection for remote sensing images are elaborated in the following section. Image-based object detection has been widely used in the remote sensing field. Skilled interpreter and computer-assisted interpretation are two essential factors in digital-recognition objects. In the manual-interpretation approach, seven interpretation elements are usually used to interpret images; namely size, shape, shadow, tone, color, texture and pattern. The size of an object is defined by the most commonly parameters such as length, area, perimeter, area and volume. The shape of an object is determined by its geometric characteristics such as linear, curvilinear, circular, elliptical, radial, square, etc. The shadow is a silhouette caused by sideilluminated objects, often. The image tone of an object refers to the gray scale of the image, typically ranging from bright to total dark. It is used in extracting the scene of the height or relief of the object of interest. The color tone is also used in the interpretation of color images. Texture is a group of small patterns often used in extracting groups of objects. Pattern is usually used in the spatial arrangement of 2 objects on the ground (systematic, unsystematic, etc.). Apart from the above seven interpretation elements, other elements are used in manual interpretation such as the associated relationships as the combination of the elements, geographic characteristics and configuration of the objects. A well-trained interpreter can interpret the image using the above elements easily as a dichotomous key (Jensen 2000). The use of digital interpretation approach is widely used in various applications. In this approach, the computer systems (including appropriate software) are used to extract object features using basic interpretation elements. The most common elements used for digital recognition are tone, color and texture. Tone and color of individuals in the scene are the two dominant interpretation elements used in the computer-based object detection based on statistical pattern recognition techniques (Hardin and Thomson, 1992). At present, both the manual and computerbased interpretation are used hand-in-hand, where each of these approaches compliment each other as fully computer-based interpretation is rather limited due to each of the dedicated systems being built for specific purposes. The image-based object detection has formed wide applications, which includes military and biomedical imaging. In military applications, recognizing and detecting of targets within specific areas has been the main focus. Such targets are crucial inputs into intelligence planning and monitoring. In the biomedical applications, object-detection has been carried out for assisting medical practitioners in diagnosis of diseases/sickness (Yuan-Hui Yu and Chin-Chen Chang, 2006). These include also recognition of fingerprints and eye pupil as biometric identification, apart from the study of human perception to vision and recognition ability. Image-based object detection and recognition is important, particularly for intelligence where it is widely used in the planning of military operations. It is also very important for assisting pilots, tanker operators, and shooters to detect and identify targets of interest. Using this technique the operator could identify the target of interest faster and more precisely. Computer-assisted object detection could be used for automatic detection of any object or target of interest. This is referred to as 3 automatic target recognition (ATR), which generally refers to the self-directed or aided target detection and recognition by computer processing of images from a variety of sensors such as forward looking infrared (FLIR), synthetic aperture radar (SAR), inverse synthetic aperture radar (ISAR), laser radar (LADAR), millimeter wave (MMW) radar, multispectral/ hyperspectral sensors, low-light television (LLTV), video, etc. ATR is an extremely important approach that is useful for targeting and surveillance missions of defense weapon systems operating from a variety of platforms (BHANU et al, 1997). Automatic target recognition aims to reduce the workload of human operators such as image analysts, pilots, and tankers, who are tasked with a large scope of activities ranging from assessing the battlefield/battlespace situation over large areas and volumes to targeting individual targets on land, sea, or air. It is also used to reacquire targets when used on unmanned lethal weapon systems such as missiles. The effectiveness and efficiency of locating and detecting targets is greatly increased with this application (Clark et al, 1997). In the context of medical imaging, object detection is widely used in detecting symptoms peculiar to disease or related to malfunction or strange organs. It uses the concept of detecting foreign bodies inside the human body. Yfantis et al. (2000) proposed an algorithm for detecting cancer using ultrasound image. The algorithm is based on calculating a number of attributes in the section of interest in the image and finding the distance in the multivariate space of the vector of attributes in the section from 2 predetermined centroids (cancerous and non-cancerous). If the distance from the non-cancerous centroid is smaller than the distance from the cancerous centroid the region is classified as non-cancerous, otherwise it is cancerous. Gender classification could be also done with the help of object detection using face detection and principal component analysis technique (Zehang et al, 2002). In this study, artificial steel objects on a sand background will be detected using the object-oriented approach and spectral reflectance characteristics. The image sets in this study are captured using digital optical camera in a controlled 4 laboratory environment where a specrtoradiometer device is used for acquiring reflectance measurements from the object targets. 1.2 Problem Statement Remotely sensed images are formed as the resultant of the interaction of electromagnetic radiation (EMR) with the earth surface. The magnitudes of these interactions are dependent on the wavelength of the EMR and the characteristics of the earth's surface. Extraction of any feature from these images can therefore be made based on the spectral and spatial entities of the EMR interactions. Two typical algorithms established for object detection in remote sensing were based on both these the spatial and spectral interactions. In this study, the region-based algorithm is used for detecting and identifying spatial entities (shape and area), while the spectral reflectance characteristics of the objects are used for recognizing the target of interest. 1.3 Objectives The objectives of this study are: (1) To detect and extract selected targets from remotely sensed images based on the spectral and spatial characteristics; (2) To perform accuracy assessment of the objects extracted in objective (1). 5 1.4 Scope of the Study The scopes of the study are as follows: (i) The images used are captured in the laboratory using a digital camera, of a set of known miniature military targets with sand background. The targets comprised two steel model air fighters (Plane1, Plane2) and one steel model tank. (ii) The images of the targets are captured within the visible spectrum of EMR at different sets: (a) The white board, captured as a control; (b) Targets on the white board (capturing reflectance for individual target); (c) Sand reflectance; (d) Camouflage reflectance; (e) Targets on sand (capturing reflectance for individual target); and (f) Reflectance for the targets covered with camouflage. (iii)The corresponding spectral responses of the targets with the background were also observed using spectroradiometer in the range of 400-700nm, equivalent to the wavelength of the camera used to capture the digital image. (iv) The spectral responses of the targets are reduced and analyzed for the absolute responses within the given wavelength range for determining the best differentiation among the targets. (v) The detection and extraction of objects using spectral entities is performed based on region growing and region merging algorithm where it has to satisfy the condition (Hsiao et al, 2005). 6 1.5 Significance of the Research Detection and extracting the spatial and spectral responses of different targets are usually done through different approaches. This research helps users to identify and recognize different kinds of targets (Plane1, Plane2, Tank), which are sited on natural background using spectral and spatial approaches. It is useful for military purposes and could be utilized in different applications. 1.6 Chapter Organization This report consists of five chapters. Chapter 1 is an introduction and background of the project. Chapter 2 reviews some previous similar works for animage-based object detection/recognition. This includes segmentation approaches, the importance of reflectance in object detection and automatic target detection. The methodology of the study is described in Chapter 3, where the materials and method used in this study is elaborated. The results of the study are presented, analyzed and discussed in Chapter 4. Finally, Chapter 5 concludes the findings of this study, apart from highlighting the future recommendations for improving this study. CHAPTER 2 OBJECT DETECTION APPROACHES AND APPLICATIONS 2.1 Introduction This chapter reviews some previous similar works on object detection. The review focuses on object-based approaches; namely contour-based, region-based, boundary-based and feature extraction and morphology operations. The image segmentation approaches are also reviewed, which cover contour segmentation and region segmentation. The relation between reflectance and object detection is also discussed in this chapter. 2.2 Approaches to Object Detection and Recognition Image-based object detection and recognition have been widely used for the identification of artificial targets in the natural environment. There are at least three main approaches for such detection, which are (i) contour-based (Caron et al, 2002), (ii) region-based approach (Ternovskiy et al., 2003; Rizon et al., 2006), (iii) boundary-based approach (Chen and Gao, 2005; Lau and Ozawa, 2004; Zouagui et al., 2003). There are some derived approaches such as the one that depends on feature extraction and morphological operation (Wang et al., 2005) and others depending on the statistical repartition (Vincent et al., 2000), which are used for different applications. 8 Contour-based object detection can be achieved using two common approaches: (i) high pass filtering, and (ii) matching technique. The first approach includes Canny filters and adaptive filtering, while the matching approach is based on matching the object contour with the geometrical models. Both these approaches work only when the shape of the object to be detected is known. It needs a contour template, which defines the object of interest (Caron et al., 2002). The main advantages of using the contour-based approach are: (a) it preserves most of the shape information about the object, which is necessary for identifying and recognizing objects; and (b) 3-D models can be made from 2-D models. The contour-based approach can be used to extract shapes and morphology for features such as the shape of any desired object. Most of the contour detection algorithms extract broken or incomplete boundaries due to poor imaging conditions. The second approach to object detection is based on region-based approach, which describes the regions within the image. It is used as shape-based detection. There are three steps in detecting objects using the region-based approach, namely: (i) Fitting method: uses a fixed number of rectangles to fit the shape of the object which represents the shape as vector of numeric parameters (ii) Eigenmodes of finite element model: the eigenmodes method uses the finite element model to represent the shape. Shape similarity computed as the modal deformation energy needed to align 2 objects (Sebastian and Kimia, 2001) (iii)Statistical calculation for the grey level: can be used by calculating mean and median of the grey level, the minimum and maximum gray level values and the number of pixels with values above and below the mean. The advantage of using region-based approach is the simplicity of identifying the topology and the texture of the objects (Chen and Gao, 2005). An example of region-based detection application is image classification using maximum likelihood analysis. Lastly, the boundary-based approach for object detection has been reported by Lau and Ozawa (2004), Chen and Gao (2005) and Zouagui et al., (2003). This 9 approach assumes that the pixel values change at the object's boundaries where they are then converted to those values to a closed region boundaries. The assumption that the adjacent pixels in the same region have similar visual features is also a basic primitive in the region-based approach. Boundary-based approach is the best for determining edge information, and this can be achieved in the following procedures: (i) by local filtering using Canny edge detector; (ii) Using snake model where an active contour model is used to minimize the energy; and (iii) Using a balloon model that is used to solve and generalized the problems encountered by the main method. It is a tool used for deducing the region in order to use it for region-based approach. The advantages of using boundary-based approach are: (i) the ease of getting information about the interior characteristics of the objects, and (ii) it does not require order boundary points to preserve the shape of objects. Apart from the three above-mentioned approaches, object detection can also be carried out using feature extraction and morphological operations (Wang et al., 2005). The main aim of feature extraction is to discriminate different regions by combining features within the image. Feature extraction is carried out in the sequence of procedures, such as: (i) initial segmentation, (ii) morphology processing; and (iii) object detection. The initial segmentation stage is essential to segment the image to regions before extracting features within images. This is performed using texture properties of the image. The image texture is described using either statistical or texture measurements based on the co-occurrence matrix. The result of this stage is a binary image where it is be used in the morphology processing stage (César et al., 2007) The morphology processing starts with the binary image that is obtained from the first stage. It consists of the following processes: (1) object selection; (2) label extraction; and (3) boundary decision. The goal of object selection is to remove the noise from the image using dilation and erosion operations. Label extraction is meant to identify the presence of homogenous regions. As a result, labeled image 10 can be obtained which can identify the objects numbers of the original image. The purpose of boundary decision is to find the contour for every object, allowing for easily locating the object in the original image. As such, object detection operation can be performed. An example of this approach is detecting artificial objects in a natural background (Wang et al., 2005). Figure 2.1 below illustrates object-detection using the morphological approach. Original image Initial segmentation Feature extraction Fuzzy c-mean clustering Mathematical morphology processing Selecting and labeling objects Finding location of objects in labeled image Locating the objects areas in original image Object detection Object detection Source: Wang et al, 2005 Figure 2.1: Block diagram of object detection 11 2.3 Image Segmentation Image segmentation is defined as the process of assigning pixels to regions having common properties. In other words, subdividing an image into constituent regions in order to locate and extract outlines of certain objects in the image. It is the first step in the object detection/recognition process. Meanwhile, the level to which the segmentation takes place depends on the application to be used for and the interesting objects to be detected (Zouagui et al., 2003; Lau and Ozawa, 2004). Image segmentation algorithms are generally based on two basic properties of the intensity values: (i) discontinuity and (ii) similarity. Using the first property, the approach is to subdivide an image based on rapid changes in the intensity values such as the edges of the image. The similarity property of the intensity is based on partitioning an image into regions that are similar according to the predefined criteria. Examples of the second category are region growing, region splitting and region splitting and merging. Image segmentation can be achieved using: (a) contour segmentation; (b) region segmentation; and (c) object-oriented approach (Gonzalez et al., 2002). 2.3.1 Contour Segmentation The contour-based segmentation approach is called the valley-following method or the snake method. It starts with edge detection followed by a linking process that seeks to make use of curvilinear continuity (Malik et al., 2001; Zagorchev et al., 2007). The information that could be taken from the boundaries of region segmentation, could define contours. Contours can be discrete or continuous. Discrete contours are defined by a sequence of points whereas continuous contours are defined by a parametric curve such as B-Splines and non-uniform rational BSplines. In the discrete representation of the contour segments, the adjacent contour points are connected to create a continuous boundary. As a result, the contour defined by a parametric curve represents a continuous boundary. This segmentation approach is widely used for medical analysis. Fear et al. (2000) use this approach for 12 cancer detection and cancer contour creation. Liang et al. (2006) also use this approach to analyze and interpret X-rays and MRI images. 2.3.2 Region Segmentation Region segmentation is commonly used to subdivide an image into its constituent regions or objects. The subdivision of the image into appropriate level depends on the problem or the dimension of the object to be detected. The subdivision stops once the objects of the interest have been isolated. It consists of processes in a sequence, such as: (i) region growing; (ii) region merging; and (iii) region splitting. Region growing is a procedure that groups pixels or sub regions into larger regions based on the predefined criteria. It has advantages such as the boundaries of the growing regions are perfectly thin and connected and the algorithm used is very stable (Gonzalez et al., 2002). Region merging segmentation simply merges regions that have similar characteristics. In other words, if there are two adjacent regions having similar properties, these two regions are merged. However, region splitting is a process that aims to divide regions that are heterogeneous. Further explanation on the other region segmentation is given in Gonzalez et al. (2002). Figure 2.2 illustrates region segmentation methods. Figure 2.2 (a) represents region growing method, while (b) illustrates region splitting method. Figures (c) and (d) show one of the splitting techniques. It has a representation called quadtree (which means a tree in which nodes have exactly four descendants). 13 (a) Region merging method (b) Region splitting method R R1 R2 R3 R41 (c) Quadtree R43 R44 source Gonzalez et al 2002 R1 R3 (d) Partitioning image R42 R4 R2 R41 R42 R43 R44 source Gonzalez et al 2002 Figure 2.2: (a) Region merging method, (b) Region splitting method , (c) Qadtree; and (d) Partitioning image 14 2.3.3 Object Oriented Approach The object-oriented approach depends mainly on image segments, and not on single pixels. The expected results of the object-oriented segmentation are real world objects with the proper shape and classification, which cannot be achieved by a pixel-based approach. This approach has the advantages of extracting additional information that can be derived based on image object, apart from giving meaningful statistics and textures. In the object-oriented approach, it is possible to take context and semantic information into account. The input and the output could be vectorbased. The final output could be carried out to analyze aerial images and to update to update GIS information (Benz et al, 2004). Quartel et al. (2006) use this approach to extract beach morphology using video images. Xu et al. (2004) also use this approach but for land cover mapping using Quick Bird images using eCognition software. 2.4 Reflectance and object detection Reflectance is an important property of an object by which one could differentiate object type and materials. There are some approaches that have been used to find relations between reflectance of objects and recognition. Different objects having different subjects result in different reflectance magnitudes. Accordingly, the characterization of earth surface features based on their reflectance is a central concept in remote sensing. Two approaches described below concerning the importance of reflectance in object detection are: (i) reflectance ratio computation, (ii) the shape and reflectance properties. Any object appearance depends on its shape, the optical properties of its surface (reflectance), the surrounding distribution of the light, the viewing position, roughness and material. A method for computing the ratio of the reflectance of a region to that of its background is one of the approaches to use reflectance for object detection within an image. This is in turn resulted in a physical region with each scene that is invariant to the shape of the object as well as the intensity and direction 15 of illumination. From this photometric invariant (reflectance ratio), valuable information could be used for recognition process. To do such a process the image is first segmented followed by computing the reflectance ratio for each individual segment with respect to their background. Shape and reflectance are two important properties for an object. Russell et al. (2007) use such properties to detect objects. The shape itself is not enough to recognize any objects as a result reflectance play important role in detecting and recognizing an object. Russell et al. (2007) use two different experiments in order to detect and recognize object, one using gray scale image and the other one using color image. From the two experiments used, good differentiation between the faces of the male and the female has been achieved. 2.5 Automatic Object Detection Automatic object detection (ATR) is the process of detecting and identifying objects in a scene. It has been widely used in the extracting appropriate information from a remote sensing data. There are many application of ATR such as image analysis and interpretation. It is used to assist operators in image analysis to reduce the workload. Advancement in computing contributed significantly toward ATR usage. Two main approaches commonly used in ATR are (i) image-based and (ii) model-based approaches (Bhanu et al, 1997). The image-based approach requires combining different views using small reference images. The model-based approach depends on boundary segmentation and planar surface extraction to describe the scene. Target detection is done using trained neural networks or generic algorithms (Vasile and Marino, 2005). Automatic target detection could also be achieved by the matching technique (Huttenlocher et al., 1997). This can be accomplished by matching target models with images as sets of oriented edge pixels. The matching method, however, is very tedious particularly for producing a meaningful match of high significance. It has a disadvantage of artifacts appearing if the image used a considerable inherent clutters. 16 In order to reduce such problem, the orientation and the location of pixels have to be carefully considered. Typical data-processing for any ATR approach uses the following sequence of steps: (i) detection of objects, (ii) discrimination of objects, (iii) classification of objects, (iv) recognition of objects; and (v) identification of objects. In the detection stage, the target is detected based on the presence of its signature captured in the image. While in the discrimination stage any artifacts or noise are removed. In the third stage, the targets are classified according to their signatures and how they are differentiated from each other and to what class they are classified. This stage need predefinition of the targets by image analyst. The fourth stage is recognition, which is meant to recognize targets by their class. The final stage is the identification stage, which is the ability to precisely define the targets (Paget et al., 2001). All the procedures for all ATR are shown Figure 2.3 below. Objects Detection Objects Discrimination Objects Classification Objects Recognition Objects Identification Source : Paget et al, (2001). Figure 2.3: Automatic Target Detection stages 17 2.6 Summary The approaches to image-based object detection and applications were reviewed. In this chapter, the object detection/recognition process has two main approaches, which are (i) contour-based, (ii) region-based. There are also hybrid approaches such as boundary-based approaches, feature extraction and morphological operations, which are derived from the above two main ones. The contour-based approach employs high pass filters and matching technique, whereas the region-based detection is achieved by the fitting method, eignmode of finite element model and statistical calculation of gray level. Object detection/recognition cannot be done without image segmentation, which is one of the crucial procedures in image analysis and interpretation. Accordingly, there are diverse segmentation approaches that can be used for object detection ranging from contour-based segmentation and region-based segmentation. Object detection, on the hand, can also be based on characteristics of the object of interest. Detection of objects can be done by computing reflectance ratio and by obtaining shape and reflectance properties of the object to be detected. Automatic target detection (ATR) was also reviewed in this chapter. CHAPTER 3 METHODOLOGY 3.1 Introduction This chapter highlights the methodology carried out for the study. There are two main parts of the methodology, namely: (1) materials used; and (2) the method employed in the study. Detailed description of all materials used is given in the following sections along with a step-by-step explanation on the methods used in the object detection in this study. The flowchart of the methodology employed for the entire study is shown in Figure 3.1. 19 Image capture Spatial-based • Digital camera (RGB) • Lab-based spectroradiometer object detection object detection Image processing • Geometric correction • Spectral-based Spectral reflectance Radiometric correction Reduction of spectral observation Image segmentation • Region-growing approach • Object oriented Analyze of SRC's for target detection Template Evaluation • shape Evaluation Results on spatialbased object detection Results on spectral - End based object detection End Figure 3.1: Flowchart of the methodology carried out in the study 20 3.2 Materials There are four types of materials used in this study, which are: (a) Plane 1, (b) Plane 2, (c) Tanker; and (d) Sand used as natural background. 3.2.1 Target Set Specially-made miniaturized target sets are configured for this study. Figure 3.2 (a) illustrates the target set with selected targets, three steel miniaturized models, namely Planes (Palne1, Plane2) and tank are placed on a rectangular board as background. The dimension of the board are measured and demarcated, to identify the entire test points and the position of the object/targets. Plane 1 represents F-15 fighter with the dimension 36cm main length and wings of 12cm length, whereas Plane2 represents F-14 fighter with the length of 15cm wings of 6cm each side. The tanker is Sherman Tank M4A3 model with a length of 17cm and width of 8cm. 3.2.2 The Data Capture Two types of data capture are carried out in this study. Firstly, to capture digital image using digital camera positioned on a goniometer to produce an image of the target set. This image is used to examine the extraction of the known targets based on optical-based object detection algorithm. Configuration of the digital image capture is illustrated in Figure 3.2 (a), where the digital camera is used instead by spectro-radiometer. The second data capture is carried out using spectro-radiometer of the target set on the wavelength range of 400 -700 nm for the RGB of the visible wavelength. The data being captured for the reflectance is stored as txt file and are later analyzed using statistical analysis software. The technical descriptions of the digital camera and the specroradiometer are given in the following sections. 21 3.2.2.1 Digital Camera Nikon Coolpix 7900 The digital camera used in the study is Nikon model Coolpix 7900. It has data capture capability of 7 megapixel (7MP resolution) with 3x optical Zoom offering focal length of 38-114mm. Three types of image sizes can be acquired, which include: (i) 7 million pixels at 3072 x 2304 scene, (ii) 5 million pixels at 2592 x 1944scene; and (iii) 3 million pixels at 2048 x 1536 scene. 3.2.2.2 Spectro-radiometer The spectro-radiometer is used for capturing the reflectance of the individual objects of the target sets with various background settings. The spectro-radiometer employed is ASD FieldSpec Pro JR, which is manufactured by Analytical Spectral Devices Inc. It has the ability to acquire the data for full spectrum range of 3502500nm with rapid data collected at 1/10th of the second per spectrum. It is a full workstation with field laptop and Gun to measure reflectance. This instrument is designed to collect solar reflectance, radiance and irradiance measurements. It can be used for different applications such as remote sensing, geology, ecology, forestry, plant physiology, and many others (ASD manual). 3.2.2.3 Images of Target Set Figure 3.2 (a) represents the configuration of the specto-radiometer in the laboratory set up to measure the reflectance of the target. The goniometer is used for placement of spectro-radiometer with target to appropriate illumination geometry for the light source and objects of interest. Figure 3.2 (b) shows the white board used for calibration, fully grided for assisting geometric rectification. It is divided into 5 divisions in x-axes and 10 divisions in y-axes, respectively. Each division represents 22 10cm intervals. Figure 3.2 (c) shows camouflage used to cover the targets. On the mean time, Figure 3.2 (d) demonstrates the setting of the object on the white board. However, Figure 3.2 (e) shows the setting of the targets on the natural environment (sand). The axes origin (x,y) is set to the left bottom corner the background board. The orientation of the targets and the position of the control points used for rectification are shown in Figure 3.2 (f). (a) L R G x O y R: Radiometer O: Object set L: source of light, G: goniometer (b) 4 3 2 1 2 3 4 5 6 7 8 9 Figure 3.2: the target set: (a) Configuration of target set placed on goniometer, (b) the white board background used for calibration…./continue 23 (c) (d) X Y (5,10) (e) Plane1 Plane Tank X (0,0) Y Figure 3.2: the target set: (c) the camouflage on the top of targets, (d) the targets sited on the white board background, (e) the targets sited on the sand background…./continue 24 (f) 4 3 7 Plane1 5 Tank Plane 1 6 2 X Y Figure 3.2: the target set: (f) the targets sited on the sand background with the control points used for rectification. 3.2.2.4 Reflectance Measurement The reflectance measurement of the target set is made using spectroradiometer in the controlled environment. The captured reflectance sets are: (i) the white board which is used for calibration, (ii) targets (Plane1, Plane2,Tank) on the white board, (iii) camouflage only, (iv) sand only, (v) targets (Plane1, Plane2,Tank) on the sand; and (vi) targets (Plane1, Plane2,Tank) covered with camouflage. The output of this process is spectral reflectance curves for each set mentioned and for each target in different environment. This is done to establish spectral reflectance curves in order to discriminate between the reflectance behaviors for different target sets (refer to Figure 3.2 (c) - (f)). Calibration for reflectance measurement is performed by capturing the reflectance of objects of interest from different angles and taking the average as the ideal one for calibration. Figure 3.3 is a diagram representing calibration method been taken. Reflectance is captured from different angles (α, α1, α2) with constant height of the sensor (radiometer) from the background, i.e. 2 m. Then a graphical representation for the 3 reflectance readings are averaged as the calibrated spectral reflectance. 25 α, α1, α2 are the angles for Radiometer observation set for the spectral reflectance of background α2 α 2m α1 + Demarcated grid for geometric reference + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + Target background Figure 3.3: configuration of spectro-radiometer with respect to target platform for calibration purpose 3.3 Data Pre-processing Image processing started with a data preprocessing stage where the image is radiometrically corrected and geometrically rectified. The preprocessed data is then subjected to segmentation approach adopted in the study is elaborated further in Section 3.3.3. 3.3.1 Radiometric Correction The radiometric correction is aimed to reduce radiometric errors or distortion. These are due to: (i) sensor sensitivity (the digital camera), (ii) the source of illumination, (iii) light fallout on object due to scattering, and (iv) absorption of radiation inside the laboratory. The histogram of the three bands of the digital data is shown in Figure 3.4 below, which shows the radiometric value of white background. The shift in the peaks of the red, green and blue bands is actually the amount of the "offset" values needed to radiometracally the data. A simple approach to normalize 26 these three bands for radiometric shift is carried out, such that the green band is set as the reference while the red and the blue are shifted by the amount of differentiation the center peak values with the green peak respectively. Table 3.1 below tabulates the descriptive statistics and the amount of shift applied to the red and blue bands to normalizing of radiometric effects. Once the "offset" for the red and blue bands is applied, the resultant image of white platform is totally "white" with one "superimposed" histogram for the RGB bands. Figure 3.4: The histograms for the RGB bands of the calibration board Table 3.1: descriptive statistics and the amount of shifted applied to the red and blue bands to normalizing of radiometric effects Band Min Max Mean Stdev Amount of radiometric shift to normalized radiometric effects 1 0 248 105.9 24.525952 105.9 -99.9 = 6 20.601978 106.5 -99.9 = 6.6 2 0 255 106.5 3 0 255 99.9 22.359181 Zero green) shift (reference 27 3.3.2 Geometric Rectification The Geometric rectification aims to assign absolute coordinates of the object to the digital image so accurate measurements of the image can be made. The image to map approach is used. The control points used are acquired from the target set for this study, which is the white board of (50cm x 100cm). This rectification process has two steps: (i) transformation of image into the map coordinate system, (ii) the resampling of the image brightness into the transformation coordinate system. The 1st degree polynomial equation (Equation 3.1) is used for transformation due to small size of the image used. The resampling method used the nearest neighbor strategy. It does not need a lot of computation and it preserves the original image from loosing original information. Table 3.2 shows the result of the rectification process. RMS for X is 0.14 pixels and RMS for Y is 0.01 pixels, with overall RMSXY of = (0.14) 2 + (0.01) 2 =0.14 pixels. X = a1 + a 2 x + a3 y Y = b1 + b2 x + b3 y } Equation 3.1 where, X, Y= are the absolute coordinates obtained from the template measurements, x,y = image coordinates of pixels, and a1 ,a2 , a3, b1 ,b2 , b3 are coefficients of transformation. Table 3.2: The result of rectification process ID X Y Xuncorrected Yuncorrected ΔX ΔY 1 2 1 441.97 1000.97 0.01 0 2 7 1 1983.98 1105.98 -0.02 0 3 9 4 194.06 2583.56 -0.02 0 4 1 4 116.50 25.50 0.02 0 28 5 6 3 1663.97 449.10 0.10 -.01 6 4 1 1059.03 1043.03 0.01 0 7 4 3 1047.02 407.02 -0.10 .01 RMSx 0.14 3.3.3 RMSy 0.01 RMSXY 0.14 Image Segmentation Image segmentation is the process of partitioning an image into segments or regions in order to analyze and interpret it. In this study, it is aimed to detect the shape of selected targets. The image segmentation of the eCognition system is used for this purpose and such process is achieved through the following procedures: (a) importing of the image; (b) multiresolution segmentation; (c) classification of the segmented image; (d) classification-based segmentation; and (e) feature space optimization. Figure 3.5 represents the flowchart segmentation process used in the study. 29 Load/import image/s and define working units Multiresolution Segmentation Classification process Classification Based Segmentation Feature Space Optimization Figure 3.5: The image segmentation process in the eCognition system adopted in the study 3.3.4 Multiresolution Segmentation Multiresolution segmentation is a method of generating image objects. It produces homogenous segments in any chosen image resolution. It is the bottom up region growing-based approach, where small pixels are merged into bigger ones. Multiresolution segmentation has the following parameters: (i) Image weight: used to define the layer weight for the selected image layers to define image importance, the higher the number the higher layer importance. (ii) Level: to determine which level the segmentation will be saved, and determines the hierarchy of the file. 30 (iii) Scale parameter: used to determine the heterogeneity of the output image. Increasing the value creates larger image object. The scale parameter refers to exclusively to the weights of the image layers (weight* scale parameter). Chosen the best scale parameters depend on the image and objects used. In this study, the scale parameter is 100. (iv) Compactness and smoothness: used to define how shape homogeneity is described. As the smoothness increases, compactness decreases. The total scale is 1. In this project, Compactness is 0.2 and the smoothness is 0.8. (v) Segmentation mode: there are 3 segmentation modes (a) normal, (b) sub- object line analysis; (c) spectral difference segmentation. Normal mode is the default mode where the scale parameters can be used to create sub-objects for calculation of line features. Whereas the spectral difference mode is to join neighboring objects according to color differences. The parameters to be used rely on the image used and object to be segmented. To get to the result, classification for the multiresolution segmentation must be done. 3.3.4.1 Classification of Segmented Image The classification process is used to assign objects to certain classes according to the class's description. The class description for the main classes in this study (Plane1, Plane2, Tank and the background) uses the manual interpretation parameters to interpret and identify the objects. eCognition software used two main approaches: (i) knowledge-based, and (ii) fuzzy approach. The knowledge approach depends on knowledge of the interpreters about the object to be classified, whereas the fuzzy approach depends on the class description using fuzzy sets of object features defined by membership function. To achieve good classification, the approach to be used should combine both approaches. Classification can be performed either using supervised or unsupervised approaches. For the supervised approach, training sets for each class are generated before the classification takes place. While the unsupervised classification starts 31 with grouping of spectral classes using clustering method, labeling of spectral classes formed are needed for the final output. In the eCognition, system the classification uses the following sequence of steps: (a) Open class hierarchy: in order to start supervised classification and collecting training samples. (b) Insert class: for every feature has its own class to distinguish between different classes. (c) Select samples: In the case of target set (plane 1) the camouflage color of dark green is one class whereas the grey color of targets is another class, the rockets have different class etc. (d) Start classification using nearest neighbor which is used by the software. (e) Apply standard nearest neighbor to all classes (f) Classify: where the system take care to apply the classification according to training sets. Figure 3.6 (a) below shows the selection of training sets for classification. Samples as demarcated by the polygon for seven classes the signature of these classes is given in Table 3.3. The classes are marked according to the signature of the segmented item as it appears in Figure 3.6 (b) (class 2 (a), (b), etc.), where the results of classification are represented. As it appears, editing for the resulted classes is necessary to get the object of interest with the appointed class. 32 1 Background 2 Tank 3 Parts from Plane1 4 Plane2 5 Parts from Plane1 6 Parts from Plane1 7 Parts from the background (a) (b) Figure 3.6: (a) selected training sample, (b) the resultant classification Table 3.3: signatures of the class of interest of the target set Class Mean Std Dev R G B R G B 1(background) 112.20 109.85 102.42 14.19 12.93 13.36 2 (a) 35.04 54.39 51.92 15.75 14.58 16.24 2 (b) 57.38 60.52 55.65 16.58 16.21 15.98 2 (c) 60.02 68.81 57.99 19.23 21.14 22.44 2 (d) 35.57 54.63 58.30 15.58 11.35 10.83 3 (a) 105.27 128.42 120.62 21.78 22.63 27.90 3 (b) 117.02 143.33 138.44 20.26 20.84 26.38 3(c) 108.60 129.81 117.18 18.71 20.30 21.83 3 (d) 134.28 142.73 150 16.53 24.58 29.45 33 4 (a) 105.26 128.42 120.62 21.78 22.63 27.90 5 (a) 62.16 85.15 76.08 28.27 29.53 36.64 5 (b) 123.08 147.08 154.50 21.55 25.02 33.55 5 (c) 90.55 115.97 128.97 46.97 51.07 52.26 6 (a) 84.25 102.63 102.55 32.02 35.54 46.47 7 (a) 53.20 57.14 55.27 20.27 20.04 21.47 7 (b) 93.97 93.39 87.97 13.21 13.09 12.95 7 (c) 52.81 55.34 58.42 13.21 12.97 12.69 7 (d) 142.96 140.06 134.25 10.64 10.87 10.70 7 (e) 153.28 171.63 187.13 21.69 25.30 28.50 7 (f) 53.20 57.14 55.48 20.27 20.04 21.47 7 (g) 92.29 90.90 86.10 11.61 11.20 10.35 In order to perform the next process of segmentation, a structure group has to be defined. It is aimed to merge classes together to perform classification-based segmentation. This stage has two main steps: (i) definition of structure group (ii) object generation. The definition of structure groups aims to create image objects based on the classification. In this project, all classes of the same objects are joined in this stage. In the case of Plane 1, a single class formed by combining other classes that identify Plane 1. 3.3.4.2 Classification Based Segmentation Classification based segmentation is another approach for object oriented segmentation which further segment the classified image. It differs from multiresolution segmentation in that the earlier method depends on the homogeneity whereas the latter segmentation uses knowledge base to generate object. This is done by editing structure groups, which was discussed earlier. 34 3.3.4.3 Feature Space Optimization Feature space optimization is a process to find a well-suited feature combination to separate classes in conjunction with a nearest neighbor classifier. It takes selected classes to be separated and results in a best separation with certain features. In order to do feature space optimization the criteria have to be initially selected to separate the objects. The features must be predefined before the generic shape features of the object of interest are identified. The following steps show feature space optimization processes in the eCognition: (1) Select the classes to calculate the optimal feature space: the objects of interest (2) Select the initial feature space to be reduced, the features to be optimized with (generic shape features) (3) Select the image object level: to what level shall be used (level2). (4) Enter the maximum dimension according to the features selected: how many features to be optimized e.g. 11 features. (5) Start the optimization (6) The result of this process shown in the optimized feature space which defines the best separation distance (the distance between closest samples of the chosen class) and the dimension used: what is best feature for optimizing between objects of interest (7) Separation reports of the optimized targets. In this study, the generic shape features used are based on Baat and Benz (2001) as the followings: (a) area, (b) length/width ratio, (c) length, (d) width, (e) border length, (f) shape index, (g) density, (h) asymmetry,(i) compactness, (j) elliptic fit, and (k) rectangular fit. Full description about those parameters is given in the Appendix 1. CHAPTER 4 RESULTS AND ANALYSIS 4.1 Introduction Object detection from the captured images was performed using the spectral and spatial approaches. There are three main image sets used for reflectance capture, which include the following: (a) white board set consists of: (i) white board (calibration), (ii) Plane 1 on the white board, (iii) Plane 2 on the white board, (iv) Tank on the white board (b) Sand set, consists of: (i) Sand only, (ii) Plane 1 on the sand, (iii) Plane 2 on the sand, (iv) Tank on the sand (c) Camouflage set consists of: (i) Camouflage, (ii) Plane 1 covered with camouflage, (iii) Plane 2 covered with camouflage, (iv) Tank covered with camouflage. Image segmentation is carried out adopting region-based and object-oriented approaches. The image set used was captured with the sand background. For other image sets such as the camouflage, it is difficult to detect objects because of them being covered. Hence, this image set is not worth preserving for further analysis. The results of image-based object detection using reflectance and segmentation processes are presented in this Chapter. 36 4.2 Object Detection Based on Reflectance The following graphs show the results of the reflectance capture for the different target sets. The first set includes the results of: (a) calibration of the white board; and (b) Comparison of spectral reflectance curve for the objects on the white board background is shown in Figure 4.1. Calibration G B 1.4 R Reflectance 1.2 1 R1 R2 R3 0.8 0.6 0.4 0.2 0 400 450 500 550 600 650 700 750 800 Wavelength (nm) (a) R1: the reflectance at α, R2: reflectance at α1, R3: reflectance at α 2 Comparison for the whiteboard targets 1.2 R G B Reflectance 1 Calibration 0.8 Plane1w 0.6 Plane2w 0.4 Tankw 0.2 0 400 450 500 550 600 650 700 750 800 Wavelength (nm) (b) Figure 4.1: reflectance curve for the target sets: (a) reflectance of the white background used for calibration of RGB bands; and (b) comparison of object inserted on the white background. 37 The spectral reflectance represented for all the target sets are on the visible band of the spectrum, equivalent to the image capture using Nikon Coolpix 7900 digital camera operating in red, green and blue bands. Blue band corresponds to 400-500nm, while Green band has the wavelength of 500-600nm and Red band equivalent to 600-750nm wavelength. Figure 4.1 (a) shows the calibration where there are three RGB curves, which correspond to the positions captured in order to determine the amount of shift to normalize each band respectively. Speckles occurred in all the three RGB bands at wavelength of less than 450nm. Minimal and acceptable spectral curves occur after the 450nm range, with smoother curve in wavelength larger than 600nm. Figure 4.1 (b) shows all the three targets placed on the white background are well separated in the wavelength larger than 500nm with optimum separation in the larger wavelengths. This separation is minimal and non-uniform in the wavelengths less than 500 nm. These fluctuations are mainly due to the existence of noises as mentioned above. Separation between Plane 1 and Plane 2 is 10% and separation of these two plane targets to the tank is more than 20% due to their differing materials, as shown in Table 4.1. Table 4.1: separation of spectral reflectance curve at (a) Blue band, (b) Green band, (c) Red band for the white board target set λ = 500 Plane1 Plane2 Tank λ = 600 Plane1 Plane2 Tank Plane1 φ 10% 20% Plane1 φ 10% 20% Plane2 10% φ 30% Plane2 10% φ 30% Tank 20% 30% φ Tank 20% 30% φ (a) (b) λ = 700 Plane1 Plane2 Tank Plane1 φ 10% 20% Plane2 10% φ 30% Tank 20% 30% φ (c) 38 B 0.6 G Sand R Reflectance 0.5 0.4 0.3 0.2 0.1 0 400 450 500 550 600 650 700 750 800 Wavelength (nm) (a) Comparison between the reflectance curves for targets B G R Reflectance 0.8 Sand 0.6 Plane1S 0.4 Plane2s TankS 0.2 0 400 450 500 550 600 650 700 750 800 Wavelength (nm) (e) Figure 4.2: (a) Spectral reflectance Sand, (b) Comparison of spectral reflectance curve for the objects on the sand. Figure 4.2 (a) shows the reflectance response of the sand background with RGB bands. Speckles occurred mostly at wavelengths less than 500nm, the blue range. The reflectance smoothly increases with the green and red band ranges starting from the wavelength 500nm onwards. Figure 4.2 (b) shows all the three targets placed on the sand background are well separated in wavelengths larger than 500nm with optimum separation in the larger wavelengths. Fluctuations also appear in this target set, which is mainly because of noise existence. This separation is minimal and non-uniform in wavelengths less than 500 nm. Separation between Plane 1 and Tank is 30%, while 39 Plane 2 has the same reflectance curve as the sand background. Table 4.2 shows that Plane 2 and the tank have a separation of 40% at the 500nm wavelength, 15% at 600nm wavelength and 0% at 700nm. It also shows that the separation between Plane 2 and Plane1 is 10% at 500nm, 20% at 600nm and 30% at 700nm. Table 4.2 also shows that there is no separation between Plane 2 and the background, which is due to the smaller size of Plane 2 and the effects of sand reflectance as the background with respects to other materials. Table 4.2: separation of spectral reflectance curve at (a) Blue band, (b) Green band, (c) Red band for the sand background set λ = 500 Plane1 Plane2 Tank λ = 600 Plane1 Plane2 Tank Plane1 φ 10% 30% Plane1 φ 10% 20% Plane2 10% φ 40% Plane2 10% φ 15% Tank 30% 40% φ Tank 20% 15% φ (a) (b) λ = 700 Plane1 Plane2 Tank Plane1 φ 10% 30% Plane2 30% φ 0% Tank 20% 0% φ (c) 40 Camouflage G B 0.5 R Reflectance 0.4 0.3 0.2 0.1 0 400 450 500 550 600 650 700 750 800 w a velength (nm) (a) Comparison of the Camouflaged curves 0.5 B G R Reflectance 0.4 Camouflage 0.3 Plane1C 0.2 Plane2C TankC 0.1 0 400 450 500 550 600 650 700 750 800 Wavelength (nm) (b) Figure 4.3: (a) spectral reflectance curve for Camouflage material, (b) Comparison of spectral reflectance curve of the objects covered with camouflage with camouflage Figure 4.3 (a) shows the reflectance response of the camouflage material with RGB bands. Speckles occurred mostly at wavelengths less than 500nm - the blue range - while speckles also continued with other bands. These noises appear more clearly in this target set due to camouflage material used. Figure 4.3 (b) shows all the three targets covered by camouflage on the sand background. They are well separated in wavelengths larger than 500nm with best separation in the larger wavelengths. The fluctuation also appears in this target set, which is mainly because of noise existence and the covering camouflage of material. Separation between Plane 1 and Plane 2 is 10% and separation of these two planes 41 targets to the tank is more than 10%, as shown in Table 4.3. This closer separation between the targets in this set is because of the effect of the camouflage on the reflectance at RGB bands. Table 4.3: separation of spectral reflectance curve at (a) Blue band, (b) Green band, (c) Red band for the camouflage set λ = 500 Plane1 Plane2 Tank λ = 600 Plane1 Plane2 Tank Plane1 φ 10% 10% Plane1 φ 10% 10% Plane2 10% φ 10% Plane2 10% φ 10% Tank 10% 10% φ Tank 10% 10% φ (a) (b) λ = 700 Plane1 Plane2 Tank Plane1 φ 10% 10% Plane2 10% φ 10% Tank 10% 10% φ (c) 4.3 Spectral Assessment Apart from the segmentation analysis mentioned in the next section, the absolute reflectance curves for the targets are also analyzed. Table 4.4 tabulates the reflectance capture for the three targets used. Table 4.5 represents the assessment for all RGB bands. Among those readings, Plane 1 covered with camouflage has the lowest reflectance where it has a mean reflectance of 0.0276 and the highest reflectance is for the calibration (white board with mean reflectance of 1.0047). From the standard deviation results for the reflectance curves, it appears all the curves are acceptable. 42 Table 4.4: Assessment for the spectral reflectance curves for the three target sets for RGB bands R Target Range G B Mean Std Dev Range Mean Std Dev Range Mean Std Dev White board set Plane1 0.35 0.1860 0.0629 0.07 0.2175 0.0109 0.03 0.210 0.0049 Plane2 0.41 0.3006 0.0770 0.06 0.3469 0.0131 .03 0.323 0.0064 Tank 0.29 0.0626 0.0468 0.07 0.0449 0.0116 0.03 0.035 0.0048 Plane1 0.35 0.1860 0.0629 0.07 0.2175 0.0109 0.03 0.210 0.0050 Plane2 0.35 .01254 0.0604 0.25 0.2610 0.0710 0.15 0.455 0.0409 Tank 0.64 0.4863 0.0722 0.03 0.4889 0.0086 0.04 0.484 0.0067 Sand set Camouflage set Plane1 0.14 0.0171 0.0270 0.04 0.0185 0.0104 0.05 0.027 0.0081 Plane2 0.35 0.1090 0.0540 0.09 .1418 0.0197 0.13 0.198 0.0297 Tank 0.19 0.0492 0.0430 0.07 .0508 0.0133 0.06 0.070 0.0114 Table 4.5: Overall assessment for the spectral reflectance curves for the three target sets for the RGB bands Object Minimum Maximum Mean Std. Deviation Calibration 0.98 1.05 1.0047 0.00932 Plane1w 0.17 0.25 0.2133 0.00796 Plane2w 0.31 0.38 0.3342 0.01467 Tankw 0.01 0.08 0.0380 0.00915 Sand 0.12 0.55 0.4048 0.12356 Plane1s 0.17 0.25 0.2133 0.00796 Plane2s 0.14 0.55 0.4043 0.11680 Tanks 0.46 0.51 0.4870 0.00780 Camouflage 0.11 0.38 0.2361 0.06867 Plane1c 0.00 0.06 0.0276 0.01265 Plane2c 0.08 0.33 0.1990 0.06292 Tankc 0.00 0.11 0.0694 0.02165 43 4.4 Segmentation Results There are four main outputs from the segmentation process, namely: (a) multiresolution segmentation, (b) classification of segments, (c) segmentation of defined segments; and (d) final classification of segmented image. Figure 4.4 shows the image of the outputs. The most component of the segmentation results is the final output of the process, as shown in Figure 4.4 (d). The ultimate shape of the tested targets against the background is shown in Figure 4.4 (d), which is the shape expected from the segmentation process. (a) (b) Figure 4.4: results of the object detection using image segmentation (a) the result of multiresolution segmentation, (b) the resulted image of the classification 44 (c) Legend Plane1 Plane2 Tank (d) Background Figure 4.4: results of the object detection using image segmentation (a) the result of multiresolution segmentation, (c) the resulted image of the classification based segmentation, (d) the image after classification (cont’d) Table 4.6 shows the details of final classification of the targets. It is clearly shown that all the three targets tested are segmented as one object, respectively. This indicates effectiveness of the adopted process. In Table 4.6, objects refer to the number of segments or regions detected within each of the tested targets. Plane 1, Plan 2 and the Tanker are identified as unique object, which is expected. The signature statistics of these three targets are also uniquely depicted by zero standard deviation. In case of the background, variations of 118 segments are normal as sand is used. In natural environment, these variations could be greater because of the 45 atmospheric effects. The main signature features of the targets with the sand (simulating the natural settings), the Tanker, the Plane 2 differs to sand by 10% and 30% respectively with variation of ± 20% from sand background. The Tanker is not likely to be successfully detected and identified. Adding the worst case of 10% obscuring by the atmospheric effects, the Tanker is confirmed as not being a good candidate for this detection approach. Taking the same natural setting and possible effects, Plane 2 is marginally detected, but with the weakest magnitude. Table 4.6: Accuracy assessment for the segmentation process Class Objects Mean StdDev Minimum Maximum Plane1 1 0 1 1 1 Background 118 0.6979 0.213 0.004926354 1 Plane2 1 0.2356 0 0.2356 0.2356 Tanker 1 0.7964 0 0.7964 0.7964 Feature space optimization is used to discriminate between different features/ objects. The generic shape features have been used after classification-based segmentation. The parameters have been mentioned earlier in Chapter 3. This process is done using the sets presented in Table 4.7. Table 4.7: Results of feature space optimization Optimization Set Detected and identified All objects together (Plane1, 11 (area, length, width, elliptic fit, rectangular fit, Plane2, Tank, Background) border length, asymmetry, density, shape index, length/width, compactness) Plane1 and the background 3 (area, border length, width) Plane2 and the background 10 (area, length, width, elliptic fit, rectangular fit, border length, asymmetry, density, shape index, length/width) Tank and the background 1 (area) Plane1 and plane2 3 (area, border length and width) Plane 1 and the tank 3 (area, border length and width). 46 Plane2 and the tank 5 (area, rectangular fit, elliptic fit compactness and asymmetry) Plane1, plane2 and the tank 5 (area, rectangular fit, elliptic fit compactness, asymmetry ) As it appears from the above table, different parameters are used for discriminating between the shapes of the targets. These parameters are the best ones for differentiation purposes. To discriminate between the three (Plane 1, Plane 2, Tank) targets and the background, 11 features can be used to differentiate between them. Among the 11 features, the area feature is the only feature, which is used in all sets. On the other hand, feature space optimization using generic shape feature could be represented by the matrix shown in Table 4.8. Table 4.8: A matrix for Feature Space Optimization Object Plane1 Plane2 Tank Background Plane1 0 3 3 3 Plane2 3 0 5 10 Tank 3 5 0 1 10 1 0 Background 3 4.5 Assessment of Object Detection Based on Spatial Entities The spatial assessments in this study focus on the analysis of shape of the tested targets (Plane1, Plane2, Tank). This is carried out by comparing the dimensions of the targets using a template (created in ground truth) with the segmented shape derived. Figure 4.5 shows the template used and the parameters for the assessment. Figure 4.5 (a) illustrates the template used to compare Plane1 and Plane2, which are identified by the main axis, minor axis1 for the right wing and minor axis2 for left wing. Tank is identified by major axis and a minor axis. From the comparison shown in Table 4.9 below, it appears that almost all objects fit the 47 template. There is a minute difference for the axis of Plane 1, which is because of the segmentation process and measurement analysis. d j a b l k c e m Plane1 f g h i n Tanker Plane 2 (a) Shape determinants: Plane1 Major axis is dbe, minor axis are ab, bc, Plane2 major axis is jgi, minor axis fg, gh, Tanker major axis is kl, and minor axis is mn. Major axis Major axis Major axis (b) Figure 4.5: Assessment of spatial entities based on generated template (a) the linear graphs typical to objects tested, (b) template overlaid the original image. 48 Major axis Major axis Major axis (c) Figure 4.5: Assessment of spatial entities based on generated template (c) the template overlaid on ultimate output of the segmentation approach (continued) Table 4.9: Results of spatial assessments Object Plane1 Georectified ΔMajor Δminor Δminor axis axis1 axis2 RMS Major Minor Minor Major Minor Minor axis axis1 axis2 axis axis1 axis2 36 12 (ab) 12 36.6 11.7(a'b') 11.7(b'c') 0.6 0.3 0.3 1.161895 (bc) (b'd'e') 6 (fg) 6 (gh) 15(j'g'i') 6 (f'g') 6 (g'h') 0 0 0 0 8 (mn) - 16.9 (k'l') 7.8 (m'n') 0.1 0.2 (bde) Plane2 Segmented image 15 (jgi) Tank 17 0.223607 (kl) 4.6 Discussion Spectral and spatial object detection approaches were used in this study. It concentrated mainly on region-merging and object oriented based segmentations (Gonzalez et al., 2002 ; Xu et al., 2004). As it appears in Section 4.4, the image is segmented to certain regions then these regions are classified using the supervised classification approach. Those segments are then merged to form objects using 49 classification-based segmentation, which uses the object-oriented approach to detect the shape of the three targets. Accordingly, the spectral approach was done using reflectance-based approach (Russell et al., 2007), where the objects could be detected from the spectral reflectance responses at the RGB band of the spectrum. For the three target sets - the white background, sand background and the camouflage sets the separations between the three targets are clear for all visible bands. CHAPTER 5 CONCLUSIONS AND RECOMMENDATION 5.1 Conclusions Following are the conclusions drawn from the study carried out: (a) The selected targets labeled using the object-based detection and identification using both spectral and spatial entities show good outputs. In the case of spectral approach, the segmentation of the three selected targets is very distinct in the visible range with optimum separation of 10% between Plane 1 and Plane 2 and 20% between the two planes and the Tanker. Even in the natural background such as sand, Plane 1 and the Tanker targets are recognizable at reasonably good separation of 30% at all visible band tested, while Plane 2 and other targets are separated by an average separation of 10%. However, the three targets are separated by 10% for all spectral RGB bands. (b) Spectral and spatial assessments have been carried out for the three targets derived from the segmentation process using the object-based approach. In the case of the targets sited on the white background, all targets have minimal standard deviations of less than 0.007 for all the RGB band of the spectrum. Even the natural background set of sand spectrally have good signature of less than 0.05 standard deviations for all the visible bands. The camouflage set shows good results for all targets from their standard deviation values of less than 0.03. The output objects have very good correlation with the original target set. The RMS values for Plane 1 is 1.1, Plane 2 is zero and 51 Tanker is 0.2. This is an indication of proper application of the spatial approach in this study. 5.2 Recommendations The recommendation and the outlooks for this study consist of: (i) Martial and environment. This recommendation is to use material and environment that simulate the real world. In the real world, the aircrafts are sited on airports where the background is made from asphalt. The environment used for this project is applicable for the tank case. It could be more beneficial for spectral capture using real material, i.e. using the models from the same materials used in reality. This could be done by contacting the Plane or the military equipment factories. If the reflectance measurements have to be done indoor, it is recommended to measure reflectance in a proper way, such as measuring in a dark room where there is no interfaces from other sources of light such as from the window or doors. The floor should be of a similar color and material to the background used in the study, in order to have good results without interference from other subjects. If the objects materials are from materials used in the real life, the reflectance should be better. Furthermore, the settings of the experiment should closely mimic the real world environment. (ii) Satellite image Satellite images will help significantly if used in this research. They will give experience of using satellite image whether they are RADAR images or highresolution optical images in object detection. (iii)Thermal Camera Another recommendation is to use a thermal camera to measure the temperature of the objects. Using the output of such a process could be combines with the results of the object-oriented approach to detect the shape of the object. This will incorporate the thermal characteristics of the objects into the analysis. 52 LIST OF REFERENCES Abe, Y. and Hagiwara, M.,(1997), Hierarchical Object Recognition from a 2D Image using a Genetic Algorithm, IEEE, International Conference on Computational Cybernetics and Simulations, 3: (2549 – 2554) Baatz, M., and Benz, U., (2001), Definciens Imaging, eCognition object oriented image analysis manual Benz, U. C., Hofmann, P., Willhauck, G., Lingenfelder, I., and Heynen, M., (2004), Multi-resolution object-oriented fuzzy analysis of remote sensing data for GIS-ready information, ISPRS Journal of Photogrammetry & Remote Sensing, 58: (239– 258) Bhanu, B. and Peng J., ( 2000) Adaptive Integrated Image Segmentation and Object Recognition, IEEE, Image Processing, 6 (1): 427-441 Bhanu, B., Dudgeon, D.E., Zelnio, E.G., Rosenfeld, A., Casasent, D., Reed, I.S. (1997), Guest Editorial Introduction to Special Target Detection and Recognition, IEEE, Image Processing, 6 (1): 1-6 Canny, J.F., (1986), A computational approach to edge detection, IEEE Transaction Pattern Analysis, 8 (6): 679–698 Caron, Y., Vincent, N., Markis, P.,(2002), Artificial object detection in natural environments, RFAI Team Publication, First Annual Meeting on Health, Science and Technology, Ecole Doctorale SST, Tours (France), 1 : (1-10) 53 César, A.B., Castañón, Jane, S, F., Sandra, F., Arthur, G., Luciano, d.F., Costab,(2007), Biological shape characterization for automatic image recognition anddiagnosis of protozoan parasites of the genus Eimeria, Science Direct, Pattern Recognition,40 :( 1899 – 1910) Chen, H., and Gao, Q., (2005), Efficient Image Region and Shape Detection by Perceptual Contour Grouping, IEEE,International Conference on Mechatronics & Automation, 2: (793- 798) Clark, F.O. and Daniel, P. H., (1997), Automatic Target Recognition by Matching Oriented Edge Pixels, IEEE Transactions on Image Processing, 6 (1):103 Cohen, L.D., (1991), On active contour models and balloons, Computer Vision, Graphics, Image Process, Image Understanding 53 (2): 211–218 Conzalez, R. C., (2002), Digital Image Processing, Richard E. Woods, 2nd edition, New Jersey Cubron, C. A. and Peddle, D.R, (2006) A low cost filed & laboratory gonimeter system for estimating hyperspectral bidirectional reflectance, Can. J Remote Sensing , 32 (3): 244-253 Fear, E.C., Stuchly, M.A., (2000), Microwave detection of breast cancer, IEEE, Transactionson Microwave Theory and Techniques, 48(11): 1854 - 1863 Gevers, T., and Smeulders, A.W. M., (1998), Image Indexing using Composite Color and Shape Invariant Features, IEEE, 16th International Conference on Computer Vision, PP: 576-581, Bombay, India. Gool, L. V., Moons, T., Ungureanu, D., (1996), Geometric/Photometric Invariants for Planar Intensity Patterns, 4th European Conference Computer Vision, 1: (642651) 54 Hardin, P.J., and Thomson, C.N., (1992), Fast Nearest Neighbor Classification Method For Multispectral Imagery, Professional Geographer, 44(2):191-201 Hsiao, Y.T., Chuang, C.L., Jiann J.A., Chien, C.C., (2005), A contour based image segmentation algorithm using morphological edge detection, IEEE International Conference on Systems, Man and Cybernetics, 3: (2962 – 2967) Huttenlocher, D. P., and Olson, C. F., (1997), Automatic Target Detection by Matching Oriented Edge Pixels, IEEE Transaction, Image Processing, 6 (1) :1103113 Jensen, J.R, (1996) , Introductory Digital Image Processing, A Remote Sensing Perspective, Prentice Hall, 2nd edition, New Jersey Jensen, J.R, (2000), Remote sensing of the Environment, an Earth Resource Perspective, Prentice Hall, New Jersey Kass, M., Witkin A., Terzopoulos, D., (1988), Snake: active contour models, Int. J. Computer Vision, 1 (4): 312–331 Lau, P. Y. and Ozawa, S. (2004), A region-based approach combining marker- controlled active contour model and morphological operator for Image Segmentation, Process 26th Annual International Conference of the IEEE EMBS San Francisco, CA, USA, 1 (3) :1652-1655 Lazzerini, B., and Marcelloni, F., (2001), A Fuzzy Approach to 2-D Shape Recognition, IEEE Transaction of fuzzy systems, 9 (1): 5-16 Liang, J., McInerney, T., Terzopoulos, D., (2006) , United Snakes , Medical Image Analysis, 10 : (215–233) Malik, J., Belongie, S., Leung, T., and Shi, J., (2001),Contour and Texture Analysis for Image Segmentation, International Journal of Computer Vision, 43(1): 7–27, Kluwer Academic Publishers 55 Nayar, S.K. and Bolle, R.M. (1996), Reflectance based object recognition, International Journal of Computer Vision, 17 (2) :219-240 Paget, R., Homer, J., and Crisp, D., (2001), (Automatic) target detection in synthetic aperture radar imagery via terrain recognition, IEEE, International Conference on Image Processing, 3: (54-57) Quartel, S., Addink, E. A., (2006), Object-oriented extraction of beach morphology from video images, International Journal of Applied Earth Observation and Geoinformation, 8 : (256–269) Ralesc, A. L., and Shananhan, J. G., (1999), Pereptual Organisation for inferring Object Boundaries in an Image, Pattern Recognition, 32 : ( 1923-1933) Rizon, M., Yazid,H., Saad, P., Md Shakaff, A. Y., and Saad , A. R., (2006), Object Detection using Geometric Invariant Moment, American Journal of Applied Sciences 2 (6): 1876-1878 Russell, R., Biederman, I., Nederhouser, M., Sinha, P., (2007), The utility of surface reflectance for the recognition of upright and inverted face, Vision Research, 47: (157–165) Sarkar, S. and Boyer, K.,(1993), Perceptual Organization in Computer Vision: A review and a proposal for classificatory structure, IEEE Transaction on Systems, Man and Cybernetics, 23 (2) : 382~398 Sebastian, T. B., and Kimia, B. B., (2001), Curves vs skeletons in object recognition, International Conference on Image Processing, Thessaloniki, Greece, IEEE, 3 : ( 22-25) Ternovskiy, I., Nakazawa, D., Campbell, S., and Suri, R., E., (2003), Biologically Inspired Algorithms For Object Recognition, IEEE International Conference on Integration of Knowledge Intensive Multi-Agent Systems, PP: 364-367 56 Vasile, A. N. and Marino, R. M. (2005), Pose-Independent Automatic Target Detection and Recognition Using 3D Laser Radar Imagery, Lincoln Laboratory Journal, 15 (1): 61-78 Vincent, N., Makris, P., Brodier, J., (2000), Compressed Image Quality and Zipf’s Law, 5th International Conference on Signal Processing , Beijing (China), 2 : (10771084 Wang, L., Li, X., and Fang K. , (2005), Object Detection Based on Feature Extraction and Morphological Operations, IEEE, International Conference on Neural Networks and Brain, 2: ( 1001 – 1003) Xu, W., Wu, B., Huang, J., Zhang, ,Y., Tian, Y., (2004), A Segmentation and Classification Approach of Land Cover Mapping Using Quick Bird Image, International Geoscience and Remote Sensing Symposium, IEEE, 5 : ( 3368-3370) Yfantis, E. A., Lazarakis, T., Bebis, G., and Gallitano, G. M., (2000), An Algorithm for Cancer Recognition and Ultrasound Images, International Journal on Artificial Intelligence Tools, 9 (2) : 265-276 Yuan-Hui, Y., and Chin-Chen, C.,(2006), A new edge detection approach based on image context analysis Image and Vision Computing, Image and Vision Computing, 24 (10) : 1090-1102 Zagorchev, L., Goshtasby A., and Satter, M., (2007), R-snakes, Image and Vision Computing , 25 (6) : 945-959 Zehang, S., Bebis, G., Xiaojing, Y., Yuan, X., and Louis, S.J., (2002), Genetic Feature Subset Selection for Gender Classification: A Comparison Study, 6th IEEE Workshop on Applications of Computer Vision, PP : 165 – 170 Zouagui, T., Benoit-Cattin, H., and Odet, C..(2003), Image segmentation functional model, Pattern Recognition, 37: (1785 – 1795) 57 APPENDIX I: DISCRIPTION OF GENERIC SHAPE FEATURES (i) Area: In nongeoreferenced data the area of a single pixel is 1. Consequently, the area of an image object is the number of pixels forming it. If the image data is georeferenced, the area of an image object is the true area covered by one pixel times the number of pixels forming the image object. (ii) Length/width: There are two ways to compute the length/width ratio of an image object. (a) The ratio length/width is identical to the ratio of the eigenvalues of the covariance matrix with the larger eigenvalue being the numerator of the fraction. γ = eig1 ( S ) l = , eig1 ( S ) > eig 2 ( S ) w eig 2 ( S ) (2) (b) The ratio length/width can also be approximated using the bounding box. Feature value range: (0;1) γ = l a 2 + ((1 − f ).b) 2 = w A (3) (iii)Length: The length can also be computed using the length-to-width ratio derived from a bounding box approximation. It is approximated as follows: l= A.γ (4) 58 Another possibility which works better for curved image objects is to calculate the length of an image object based on its sub-objects. Feature value range: (0 ; depending on shape of image object) (iv) Width: Also the width of an image object is approximated using the length-to width ratio. In eCognition the width is approximated as follows: w= A γ (5) Again, for curved image objects the use of sub-objects for the calculation is the superior method. Feature value range: (0 ; depending on shape of image object) (v) Border length: The border length e of an image object is defined as the sum of edges of the image object that are shared with other image objects or are situated on the edge of the entire scene. In nongeoreferenced data the length of a pixel edge is 1. Feature value range: (4 ; depending on shape of image object) (vi) Shape index: Mathematically the shape index is the border length e of the image object divided by four times the square root of its area A. Use the shape index s to describe the smoothness of the image object borders. The 59 more fractal an image object appears, the higher its shape index. Feature value range: (1 ; depending on shape of image object) s= (vii) e (7) 4. A Density: The density d can be expressed by the area covered by the image object divided by its radius. eCognition uses the following implementation, where n is the number of pixels forming the image object and the radius is approximated using the covariance matrix: d= n (8) 1 + var( X ) + Var (Y ) Use the density to describe the compactness of an image object. The ideal compact form on a pixel raster is the square. The more the form of an image object is like a square, the higher its density. Feature value range: (0; depending on shape of image object) (viii) Asymmetry: The lengthier an image object, the more asymmetric it is. For an image object, an ellipse is approximated which can be expressed by the ratio of the lengths of minor and major axes of this ellipse. The feature value increases with the asymmetry. Feature value range: (0 ; 1) K = 1− n m (10) (ix) Compactness: In eCognition the compactness c, used as a feature, is calculated by the product of the length m and the width n of the 60 corresponding Object and divided by the number of its inner pixels a. Feature value range: (0; inf ) c= n.m a (11) (x) Elliptic Fit: As a first step in the calculation of the elliptic fit is the creation of an ellipse with the same area as the considered object. In the calculation of the ellipse also the proportion of the length to the width of the Object is regarded. After this step the area of the object outside the ellipse is compared with the area inside the ellipse that is not filled out with the object. While 0 means no fit, 1 stands for a complete fitting object. Feature value range: (0; 1) (xi) Rectangular fit: A first step in the calculation of the rectangular fit is the creation of a rectangle with the same area as the considered object. In the calculation of the rectangle also the proportion of the length to the width of the object in regarded. After this step the area of the object outside the rectangle is compared with the area inside the rectangle, which is not filled out with the object. While 0 means no fit, 1 stands for a complete fitting object. Feature value range: (0; 1)