International Journal of Engineering Trends and Technology (IJETT) – Volume 4 Issue 9- Sep 2013 Background Subtraction with Feature Extraction Based on FPGA Sistla V Sudheer Kumar*, B.N. Srinivasa Rao**, E. Govinda*** *Final M.Tech Student, Dept of Electronics and Communication Engineering, Avanthi Institute of Engineering & Technology, Narsipatnam, Andhra Pradesh. **Assistant Professor, Dept of Electronics and Communication Engineering, Avanthi Institute of Engineering & Technology, Narsipatnam, Andhra Pradesh ***Head of Department of Electronics and Communication Engineering, Avanthi Institute of Engineering & Technology, Narsipatnam, Andhra Pradesh Abstract:- Motion estimation is the hard to estimate the depth of the visual features. In our work we implemented L-K algorithm with large scale estimations over FPGA. In our architecture multi scale estimation is best process for 32 frames per second and also our work increases accuracy and estimation. And also the resources are incremented using this approach. The idea of background subtraction is to subtract or difference the current image from a reference background model. This paper proposes a new method to detect moving object based on background subtraction. First owe establish a reliable background updating model based on statistical and use a dynamic optimization threshold method to obtain a more complete moving object. Here we have written the core processor Microblaze is designed in VHDL (VHSIC hardware description language) is implemented using XILINX ISE 8.1 implementation apts the algorithm is written in system C Language and tested in SPARTAN-3 FPGA kit by interfacing a test circuit with the PC using the RS232 cable. This is tested and the results are seen to be satisfactory. Area taken and the speed of the algorithm are also evaluated. I. Introduction A field-programmable gate array (FPGA) is an integrated circuit designed to be configured by a customer or a designer after manufacturing. The configuration is generally specified using a hardware description language (HDL) and it is same to that used for an application-specific integrated circuit (ASIC) (circuit diagrams were previously used to specify the configuration and they were for ASICs this is increasingly rare). The traditional FPGAs have large resources of logic gates and RAM blocks to implement complex digital computation operations. The FPGA designs employ very fast IOs and bidirectional data buses it becomes a challenge to verify correct timing of valid data within setup ISSN: 2231-5381 time. Floor planning enables resources allocation within FPGA to meet these time constraints. The traditional FPGAs can be used to implement any logical function that an ASIC could perform. There is an ability to update the functionality after shipping, and partial re-configuration of a portion of the design and the low non-recurring engineering costs relative to an ASIC design (not with standing the generally higher unit cost) and offers the advantages for many applications. A field-programmable gate array contain programmable logic components is known as logic blocks and a hierarchy of reconfigurable interconnects that allow the blocks to be "wired together"—somewhat like many (changeable) logic gates that can be inter-wired in (many) different configurations and the logical blocks can be configured to perform complex combinational functions or simple logic gates like AND and XOR. Most of the FPGAs logic blocks also include memory elements and which may be simple flip-flops or more complete blocks of memory. Some FPGAs have analog features in addition to digital functions. Very common analog feature is programmable slew rate and drive strength on each output pin and that is allowing the engineer to set slow rates on lightly loaded pins that would otherwise ring unacceptably and to set stronger faster rates on heavily loaded pins on high-speed channels that would otherwise run too slow. There is another relatively common analog feature is differential comparators on input pins designed to be connected to differential signalling channels. A few "mixed signal FPGAs" have integrated peripheral analog-todigital converters (ADCs) and digital-to-analog converters (DACs) with analog signal conditioning blocks allowing them to operate as a system. The devices blur the line between an FPGA and it carries digital ones and zeros on its internal programmable interconnect fabric, and field-programmable analog array (FPAA), which carries analog values on its internal programmable interconnect fabric. http://www.ijettjournal.org Page 3814 International Journal of Engineering Trends and Technology (IJETT) – Volume 4 Issue 9- Sep 2013 In computer vision, image segmentation is the process of partitioning a digital image into multiple segments and the sets of pixels also known as super pixels. The segmentation goal is to simplify and/or change the representation of an image into something that is more meaningful and easier to analyzing and the segmentation of image is typically used to locate objects and boundaries (lines, curves, etc.) in images. Image segmentation is the process of assigning a label to every pixel in an image such that pixels with the same label share certain visual characteristics and the result of image segmentation is a set of segments that collectively cover the entire image or a set of contours extracted from the image. Every pixel in a region is similar with respect to some characteristic or computed property and such as colour and intensity or texture. The regions those are significantly different with respect to the same characteristic(s). When applied to a stack of images and the typical in medical imaging and resulting contours after image segmentation can be used to create 3D reconstructions with the help of interpolation algorithms like marching cubes. Mathematical morphology (MM) is a theory and technique for the analysis and processing of geometrical structures are based on set theory such as lattice theory and random functions. It is generally applied to digital images and it can be employed as well on graphs and many other spatial structures. Topology and geometric space concepts such as size and geodesic distance were introduced by MM on both continuous. Mathematical morphology is also the foundation of morphological image processing and it consists of a set of operators that transform images according to the above characterizations. Its originally developed for binary images and was later extended to gray scale functions and images. Generalization to complete lattices is widely accepted today as MM's theoretical foundation. II. RELATED WORK A) Optical flow or optic flow: The pattern of apparent motion of objects surfaces and edges in a visual scene caused by the relative motion between an observer (an eye or a camera) and the scene and the optical flow was introduced by the American psychologist to describe the visual stimulus provided to animals moving through the world. The importance of optic flow for affordance perception and the ability to discern possibilities for action within the environment. Followers and his ecological approach to psychology have further demonstrated the role of the optical flow stimulus for: the perception of movement by the observer in the world; perception of the shape distance and movement of ISSN: 2231-5381 objects in the world and the control of loco-motion. New term optical flow has been co-opted by roboticists to incorporate related techniques from image processing and control of navigation and such as motion detection and also object segmentation, time-to-contact information and the expansion calculations, luminance, and motion stereo disparity measurement. The motion estimation of sequences of ordered images allow as either instantaneous image velocities or discrete image displacements. The optical flow methods try to calculate the motion between two image frames which are taken at times t and t + δt at every position. Based on the Taylor series approximation of the image signal and using partial derivatives with respect to the spatial and temporal coordinates optical flow is estimated. For a 2D+t dimensional case (3D or n-D cases are similar) a voxel at location (x,y,t) with intensity I(x,y,t) will moved by δx, δy and δt between the two image frames and the following image constraint equation can be given as, I(x,y,t) = I(x + δx,y + δy,t + δt) Assuming the movement to be small and the constraint at I(x,y,t) using Taylor series can be developed to get Where Vx,Vy are the velocity of x and y components or optical flow of I(x,y,t) and , and are the derivatives of the image frames at (x,y,t) in the corresponding directions. IxVx + IyVy =− It or This is an equation of two unknowns and cannot be solved as such and problem is said as http://www.ijettjournal.org Page 3815 International Journal of Engineering Trends and Technology (IJETT) – Volume 4 Issue 9- Sep 2013 aperture problem of the optical flow algorithms. To find the accurate optical flow another set of equations is needed, given by some additional constraint. For estimating the actual flow all optical flow methods introduce additional conditions. The first task is to find a representation which in effect de-correlates the image pixels to design an efficient compression code and represent the image directly in terms of the pixel values. An image pyramid is the representation of an image at different resolutions. The image pyramids are mainly used to generate a number of homogeneous parameters that represent the response of a bank of filters at different scales and possibly different orientations. The Gaussian pyramid is computed as follows and the original image is convolved with a Gaussian kernel and described above the resulting image is a low pass filtered version of the original image and the frequency limit can be controlled using the parameter . The Laplacian is then computed as the difference between the original image and the low pass filtered image and the process is continued to obtain a set of band-pass filtered images (since each is the difference between two levels of the Gaussian pyramid). The method Laplacian pyramid is a set of band pass filters. B) Lucas-Kanade Algorithm: The main concern of optical flow estimation is the pixel displacements between successive frames. Here each pixel have the velocity vector (u,v) and assumed that the intensity of the pixel in the(x,y) position in the image at time t , and the pixel(x+u,y+v) in the image of time(t+1) does not change. This only have good accuracy for slow-moving objects between frames (it depends on the distance from the object to the camera and on the 3-D object velocity and on the camera frame-rate) By assumption in Lucas-Kanade method the displacement of the image contents between two nearby instants (frames) is small and approximately constant within a neighbourhood of the point p under consideration. Optical flow equation can be assumed to hold for all pixels within a window centered at p. Namely, the local image flow (velocity) vector (Vx,Vy)must satisfy Ix(q1)Vx + Iy(q1)Vy =− It(q1) Ix(q2)Vx + Iy(q2)Vy =− It(q2) Ix(qn)Vx + Iy(qn)Vy =− It(qn) ISSN: 2231-5381 Where are the pixels inside the window and Ix(qi),Iy(qi),It(qi) are the partial derivatives of the image I with respect to positions are x, y and time t, evaluated at the point qi and at the current time and the equations can be written in matrix form Av = b, where and This system has more equations than unknowns and thus it is usually over-determined. The Lucas-Kanade method using the least squares principle and obtains a compromise solution. Namely, it solves the 2×2 system ATAv = ATb or v = (ATA) − 1ATb Where AT is the transpose of matrix A. That is, it computes with the sums running from i=1 to n. The matrix ATA is often called the structure tensor of the image at the point p. The same importance is given to all n pixels qi in the window as the plain least squares solution. In practice it is usually better to give more weight to the pixels that are closer to the central pixel p. One uses the weighted version of the least squares equation, ATWAv = ATWb Or v=(ATWA) −1ATWb Where W is an n×n diagonal matrix containing the weights Wii = wi to be assigned to the equation of pixel qi. The weight wi is usually set to a Gaussian function of the distance between qi and p. Due to the disadvantage of L&K algorithm in the accuracy of the estimations of large displacements, high performance approach is consider with the multi-scale implementation that computes the optical flow velocity components for each different resolution with good accuracy. Digital images are everywhere: consumer, business, industrial, medical, military, scientific. However, the original raw data coming from the acquisition device is often of poor quality because of distortions such as noise, blurring, quantization, geometrical, aberrations, etc. To obtain an image of acceptable quality some processing is necessary to suppress these degradations. Such processing is called signal restoration/reconstruction. Restoration is even more important if we consider that, before they reach the end-user, the majority of images undergoes some further processing (re sampling, equalization, compression, enhancement) that cannot be directly performed on distorted data. The Scaling circuit reads old partial optical flow values and scales them with a bilinear interpolation (the new values are multiplied by 2 to adapt the optical flow values to the next scale). • The Warping circuit reads the pyramid images and displaces them using the expanded motion. • The Median filtering stage removes the outliers homogenizing the partial and the final results. This stage consists in a median bi-dimensional filter and http://www.ijettjournal.org Page 3816 International Journal of Engineering Trends and Technology (IJETT) – Volume 4 Issue 9- Sep 2013 can be a cascade of them. This filter also contributes by incrementing the density of the final results removing the non-confident values and filling the holes with the filter results (in our case, the filter computes on a 3*3 neighbourhood). Frames First Motion estimation and scaling, warping Motion Estimation Gaussian Filtering The Merging stage allows the sum of the previous optical flow estimation and the current one in order to compute the final estimation. The interaction with memory is a very critical problem and needs a dedicated circuit and a specific memory mapping. The parallel access to the RAM blocks is allowed using the multiple available banks and a sequential operation strategy. Pyramid: A pyramid is built by a smoothing and sub-sampling circuit (see Fig. 4). Each pyramid scale is obtained sequentially (mainly, due to the limitations of the sequential access to the external memory). Input and output images are directly read/stored into an external RAM memory. Warping: It consists in a bilinear interpolation of the input images with the increment values that we have stored from the optical flow in the previous scale in a LUT and computation of each warped pixel requires the reading of the pair from their correspondent matrices as well as the pixel P. Then the numerical part of the pair is used for retrieving from memory the four pixels of the original image. After that warped pixel is calculated with the fractional part performing a weighted bilinear interpolation. Merging: This module computes the addition of the previous optical flow estimation and the current and the resultant is stored for the next iteration. The non-valid values are propagated from the coarsest scales to the finest ones. These nonconfident values are obtained at each scale applying ISSN: 2231-5381 the threshold mentioned before as the eigenvalues product. At the last scale, the finest one, we make the logical “and” operation between its non-valid values and the propagated ones for the final estimation. The propagation for the other scales is implemented using an “or” logical operation; this difference in the computation is performed to weight more the nonvalid values of the finest scale because they are the more exact ones in terms of non-confidence. The main problem at this module is the synchronization between the current and the stored results. C) Understanding the Dilation and Erosion Morphology is a broad set of image processing operations that process images based on shapes. The morphological operations apply a structuring element to an input image and the creating an output image of the same size. The value of each pixel in the output image is based on a comparison of the corresponding pixel in the input image with its neighbours. The size and shape of the neighbourhood, you can construct a morphological operation that is sensitive to specific shapes in the input image. The most basic morphological operations are dilation and erosion. It adds pixels to the boundaries of objects in an image and while erosion removes pixels on object boundaries. The pixels added or removed from the objects in an image depends on the size and shape of the structuring element used to process the image. Morphological dilation and erosion includes separations the state of any given pixel in the output image is determined by applying a rule to the corresponding pixel and its neighbours in the input image. Rule used to process the pixels defines the operation as dilation or erosion. The table shows the rules for both dilation and erosion. a)Dilation Rule: The value of the output pixel is the maximum value of all the pixels in the input pixel's neighbourhood. In a binary image and if any of the pixels is set to the value 1 and the output pixel is set to 1. b) Erosion Rule: The value of the output pixel is the minimum value of all the pixels in the input pixel's neighbourhood. In a binary image and if any of the pixels is set to 0 and output pixel is set to 0. The following figure illustrates the dilation of a binary image. How the structuring element defines the neighbourhood of the pixel of interest and that which is circled. The dilation function applies the appropriate rule to the pixels in the neighbourhood and assigns a value to the corresponding pixel in the output image. The morphological dilation function sets the value of the output pixel to 1 because one of http://www.ijettjournal.org Page 3817 International Journal of Engineering Trends and Technology (IJETT) – Volume 4 Issue 9- Sep 2013 the elements in the neighbourhood defined by the structuring element is on. Morphological dilation of binary image The above figure illustrates this processing for a gray scale image. The figure shows the processing of a particular pixel in the input image and how the function applies the rule to the input pixel's neighbourhood and uses the highest value of all the pixels in the neighbourhood as the value of the corresponding pixel in the output image. D) Processing Pixels at Image Borders (Padding Behaviour): Morphological functions position the origin of the structuring element, its centre element, over the pixel of interest in the input image. For pixels at the edge of an image and the parts of the neighbourhood defined by the structuring element can extend past the border of the image. To process border pixels and morphological functions assign a value to these undefined pixels and if the functions had padded the image with additional rows and columns. These padding pixels a value varies for dilation and erosion operations. The table describes the padding rules for dilation and erosion for both binary and gray scale images. E) Rules for padding images Dilation: Pixels beyond the image border are assigned the minimum value afforded by the data type. In binary images these pixels are assumed to be set to 0. For gray scale images, the minimum value for unit 8 images. Erosion: Pixels beyond the image border are assigned the maximum value afforded by the data type. For binary images and these pixels are assumed to be set to 1. For gray scale images, the maximum value for uint8 images. By using the minimum value for dilation operations and the maximum value for erosion operations and the toolbox avoid border effects where regions near the borders of the output image do not appear to be homogeneous with the rest of the image. Consider an example if erosion padded with a minimum value and the eroding an image would result in a black border around the edge of the output image. ISSN: 2231-5381 Opening and closing are two important operators from mathematical morphology. Both derived from the fundamental operations of erosion and dilation. Those operators they are normally applied to binary images and although there are also gray level versions. The basic effect of an opening is somewhat like erosion in that it tends to remove some of the foreground (bright) pixels from the edges of regions of foreground pixels. It is less destructive than erosion in general. As with other morphological operators and the exact operation is determined by a structuring element. Such effect of the operator is to preserve foreground regions that have a similar shape to this structuring element and that can completely contain the structuring element and eliminating all other regions of foreground pixels. Closing is an important operator from the field of mathematical morphology. Like its dual operator opening and it can be derived from the fundamental operations of erosion and dilation. Those operators like it are normally applied to binary images and there are gray level versions. Closing some ways are similar to dilation in that it tends to enlarge the boundaries of foreground (bright) regions in an image (and shrink background colour holes in such regions) and it is less destructive of the original boundary shape and with other morphological operators and the exact operation is determined by a structuring element and it effects of the operator is to preserve background regions that have a similar shape to this structuring element or that can completely contain the structuring element, while eliminating all other regions of background pixels. Fig: Opening and Closing III. Kit for Design Flow The MicroBlaze is a soft processor core designed for Xilinx FPGAs from Xilinx. As a soft-core processor and Micro-Blaze is implemented entirely in the general-purpose memory and logic fabric of http://www.ijettjournal.org Page 3818 International Journal of Engineering Trends and Technology (IJETT) – Volume 4 Issue 9- Sep 2013 Xilinx FPGAs. In terms of its instruction-set architecture and MicroBlaze is very similar to the RISC-based DLX architecture described in a popular computer architecture book by Patterson and Hennessy. Considering few exceptions and the MicroBlaze can issue a new instruction every cycle and maintaining single-cycle throughput under most circumstances. The MicroBlaze has a versatile interconnect system to support a variety of embedded applications. Micro Blaze’s primary I/O bus and Core Connect PLB bus, is a traditional system-memory mapped transaction bus with master/slave capability. A latest version of the MicroBlaze and supported in both Spartan-6 and Virtex-6 implementations and as well as the 7-Series that supports the AXI specification. Majority of vendor-supplied and thirdparty IP interface to PLB directly (or through a PLB to OPB Bus Bridge.) For access to local-memory (FPGA BRAM), MicroBlaze uses a dedicated LMB bus and reduces loading on the other buses. It is defined coprocessors are supported through a dedicated FIFO-style connection called FSL (Fast Simplex Link). Co-processors interface can accelerate computationally intensive algorithms by offloading parts or the entirety of the computation to a user-designed hardware module. Many aspects of the MicroBlaze can be user configured: cache size, pipeline depth (3-stage or 5stage) and embedded peripherals and memory management unit and bus-interfaces can be customized. The area-optimized version of MicroBlaze and which uses a 3-stage pipeline and sacrifices clock-frequency for reduced logic-area and the performance-optimized version expands the execution-pipeline to 5-stages, allowing top speeds of 210 MHz (*on Virtex-5 FPGA family.) and also the key processor instructions which are rarely used but more expensive to implement in hardware can be selectively added/removed (i.e. multiply, divide, and floating-point ops.) and the customization enables a developer to make the appropriate design tradeoffs for a specific set of host hardware and application software requirements. With the memory management unit and the MicroBlaze is capable of hosting operating systems requiring hardware-based paging and protection and those are the Linux kernel. Otherwise it is limited to operating systems with a simplified protection and virtual memory-model: e.g. Free RTOS or Linux without MMU support. Micro Blaze’s overall throughput is substantially less than a comparable hardened CPU-core (such as the PowerPC440 in the Virtex-5.) ISSN: 2231-5381 Xilinx Platform Studio (XPS) is a key component of the ISE Embedded Edition Design Suite and the helping the hardware designer to easily build and that connect and configure embedded processor-based systems; from simple state machines to full-blown 32-bit RISC microprocessor systems. XPS employs graphical design views and sophisticated correct-bydesign wizards to guide developers through the steps necessary to create custom processor systems within minutes and true potential of XPS emerges with its ability to configure and integrate plug and play IP cores from the Xilinx Embedded IP catalog and with custom or 3rd party Verilog and VHDL designs. The highly-custom processors can be designed according to project-specific needs include; peripheral and IO requirement and the real-time responsiveness general purpose processing power and the floating point performance on-chip or off-chip memory minimal power consumption and much more. Firmware and software developers benefit from XPS integration with Xilinx SDK which allows the automatic generation of critical system software such as boot loaders, bare metal BSP and Linux BSPs. The effeciency ensures that OS porting and applications development can begin without delay caused by firmware development. A) Configuring and customizing ZynqTM-7000 All Programmable SoCs using XPS: The Zynq-7000 AP SoC delivers the pinnacle of programmable SoCs functionality through dual ARM Cortex A9 dual microprocessors and then hardened peripheral set with functions such as Ethernet etc and coupled with Xilinx programmable logic where custom soft peripherals and logic and the devices and accelerators can be initiated and XPS accelerates every aspect of design creation for Zynq devices through easy-to-use graphical wizards including clock domain setup, interrupts, DMAs, external connections for the hardened peripherals and the interface connections for the soft peripherals in programmable logic. That designers can immediately begin their custom design without fear of defining incompatible interfaces or connections. B) Customizing Xilinx MicroBlaze TM using XPS: With extraordinary scalability and customization potential and ranging from an 8-bit state machine all the way up to complex and that C-like 32 bit RISC designs, Xilinx MicroBlaze meets a diverse set of project-specific processing requirements. The engineers can create hundreds of different MicroBlaze designs by using XPS to integrate prevalidated processor-internal IP such as pipelines and interrupt controllers; and processor-peripheral IP; such as memory controllers and much more that are available through the Xilinx Embedded IP catalog. http://www.ijettjournal.org Page 3819 International Journal of Engineering Trends and Technology (IJETT) – Volume 4 Issue 9- Sep 2013 This processor configuration capability is combined with the ability to integrate 3rd party RTL and custom-IP blocks and the engineers can truly produce unique the custom designs that meet their precise requirements. C) Adding new Plug and Play Peripherals and Devices with XPS: XPS supports drag-and-drop integration of IP cores from the Xilinx Embedded IP catalog and within custom processor designs. Examples of such IP cores include peripherals and accelerators such as AXI bridges and GPIO, BRAM and external memory controllers, Serial Peripheral and Quad SPI Interfaces, Analog to Digital converters and Timers, UARTs Interrupt controllers and much more. D) Integrating Custom or 3rd Party Peripherals with XPS: Although many kinds of systems can be created from the peripherals available within the Xilinx catalog and it is often necessary to create and import custom peripherals for new functionality and Xilinx Create and Import Peripheral wizard allows hardware designers to create AXI (version 4) peripherals in Verilog or VHDL, or both (for a mixed-language design) and then to import them into an XPS projects for connection to any AXI4-Lite, AXI4 (Burstenabled) and AXI4-Stream interface and wizard also enables you to integrate your PLB (version4.6) or FSL peripherals PLB-based designs. Upon import into XPS and custom peripheral is managed just like any off-the-shelf module available from the Xilinx Embedded IP catalog. ISSN: 2231-5381 E) Connecting Peripherals with XPS : XPS makes it easy to connect each of the IO pins and internal programmable logic end-points to their desired endpoint. XPS easily manages this for connections that link off-chip to the PC board via a physical pin or to another device within the programmable logic XPS guarantees proper routing signal and voltage-rail correctness. For the Zynq-7000 AP SoC device family and XPS also manages configuration of the built in IO multiplexer which routes the processing system devices to their appropriate output pins. There are two options available for debugging the application created using EDK namely: Xilinx Microprocessor Debug (XMD) for debugging the application software using a Microprocessor Debug Module (MDM) in the embedded processor system and Software Debugger that invokes the software debugger corresponding to the compiler being used for the processor. The Software Development Kit Xilinx Platform Studio Software Development Kit (SDK) is an integrated development environment and complimentary to XPS, that is used for C/C++ embedded software application creation and verification. Built on the Eclipse open source framework. Soft Development Kit (SDK) is a suite of tools that enables you to design a software application for selected Soft IP Cores in the Xilinx Embedded Development Kit (EDK).The software application can be written in a "C or C++" then the complete embedded processor system for user application will be completed and else debug & download the bit file into FPGA. It behaves like processor implemented on it in a Xilinx Field Programmable Gate Array (FPGA) device. http://www.ijettjournal.org Page 3820 International Journal of Engineering Trends and Technology (IJETT) – Volume 4 Issue 9- Sep 2013 HDL or ISE Add Embedded Source Other Sources -RTL -Core Generator -System Generator Implementation into Bit stream 1. Synthesis 2. Translate 3. MAP 4. PAR 5. Timing 6. Bit stream Generation XPS XPS Launches Automatically SDK Design Entry 1. Create design in base system builder 2. Modify design in system assembly view Create/identify a workspace Export to SDK (.xml files Create a new project/Board Support package Application Net list Generation with platgen Download to FPGA .bi t .b Debug Export to SDK (.xml, .bit, .bmm files) Board IV.CONCLUSION In this proposed work we introduced L-K algorithm with multi-scale estimation. It efficiently works on large scale images. It reduces cost complexity and time complexity during processing in the real systems. It shares hardware resources. It adapts visual features as depth and phase. In the scaling phase it reads optical flow and reduces the computation complexity of visual features and The proposed method is inherently parallel, since computations for each pixel of each sequence frame can be done concurrently with no need for communications. This can help in lowering execution times for high-resolution sequences. Moreover, the approach is suitable to be adopted in a layered framework, where, operating at region-level and it improve detection results allowing to more efficiently tackle problem and to distinguish morphological Image by the morphological operator. This is a very desirable operative mode, considering that a very actual visual segmentation with high accuracy is achieved. REFERENCES: ISSN: 2231-5381 [1] K. Pauwels, N. Kruger, M. Lappe, F. Worgotter, and M. M. V. Hulle, “A cortical architecture on parallel hardware for motion processing in real-time,” J. Vision, vol. 10, no. 18, pp. 1–21, 2010. [2] A. Kokaram, “On missing data treatment for degraded video and film archives: A survey and a new bayesian approach,” IEEE Trans. Image Process., vol. 13, no. 3, pp. 397–415, 2004. [3] A. Wali and A. M. Alimi, “Event detection from video surveillance data based on optical flow histogram and high-level feature extraction,” in Proc. 20th Int. Workshop Database Expert Syst. Appl., 2009, pp.221–225. [4] J. Bergen, P. Anandan, K. Hanna, and R. Hingorani, “Hierarchical model-based motion estimation,” in Computer Vision ECCV’92, ser. Lecture Notes in Computer Science, G. Sandini, Ed. New York: Springer-Verlag, 1992, vol. 588, pp. 237– 252. [5] M. Anguita, J. Diaz, E. Ros, and F. J. FernandezBaldomero, “Optimization strategies for highperformance computing of optical-flow in general-purpose processors,” IEEE Trans. Circuits Syst. for Video Technol., vol. 19, no. 10, pp. 1475– 1488, Oct. 2009. http://www.ijettjournal.org Page 3821 International Journal of Engineering Trends and Technology (IJETT) – Volume 4 Issue 9- Sep 2013 [6] A. Wali and A. M. Alimi, “Event detection from video surveillance data based on optical flow histogram and high-level feature extraction,” in Proc. 20th Int. Workshop Database Expert Syst. Appl., 2009, pp. 221–225. [7] A. Elgammal, D. Hanvood, and L. S. Davis, “Nonparametric model for background subtraction,” in Proc. ECCV, 2000, pp. 751–767 [8] K. Kim, T. H. Chalidabhongse, D. Harwood, and L. S. Davis, “Real-time foreground-background segmentation using codebook Model,” Real-Time Imag., vol. 11, pp. 172–185, 2005. [9] C. Stauffer and W. E. L. Grimson, “Adaptive background mixture models for real-time tracking,” in Proc. IEEE Conf. Computer Vision and Pattern Recognition, 1999, pp. 246–252. [10] R. Cucchiara, M. Piccardi, and A. Prati, “Detecting moving objects, ghosts, and shadows in video streams,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 25, no. 10, pp. 1–6, Oct. 2003. [11] M. Tomasi, M.Vanegas, F. Barranco, J. Diaz, and E. Ros, “High-performance optical-flow architecture based on a multiscale, multi-orientation phase-based model,” IEEE Trans. Circuits Syst. for Video Technol., vol. 20, no. 12, pp. 1797–1807, Dec. 2010. [12] Xilinx, San Jose, CA, “FPGA and CPLD solutions from Xilinx, Inc.,” 2011. [Online]. Available: http://www.xilinx.com/ [13]G.L.Foresti,”A Real Time System for Video Surveillance of Unattended Outdoor Environments”. [14] C. Stauffer and W. E. L. Grimson, “Adaptive background mixture models for real-time tracking,” in Proc. IEEE Conf. Computer Vision and Pattern Recognition, 1999, pp. 246–252. [15] R. Cucchiara, M. Piccardi, and A. Prati, “Detecting moving objects, ghosts, and shadows in video streams,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 25, no. 10, pp. 1–6, Oct. 2003 ISSN: 2231-5381 BIOGRAPHIES: Sistla V Sudheer Kumar completed B.Tech Electronics and Communication Engineering in Avanthi Institute of Engineering & Technology, Visakhapatnam currently pursuing M.Tech in from Avanthi Institute of Engineering & Technology, Visakhapatnam. Andhra Pradesh. Interesting research areas VLSI custom design. B.N. Srinivasa Rao received his B.Tech degree in Electronics and Communication Engineering from JNT University, Hyderabad, India and M.Tech in VLSI System Design from JNT University, Hyderabad, India. He is currently working as an Assistant Professor in Avanthi Institute of Engineering and Technology, Visakhapatnam, Andhra Pradesh, India. He has 5 years teaching and 9 years industrial experience. He has 10 publications in various International conferences. His area of interest VLSI Semi and full custom design. He guided many projects for B.Tech and M.Tech students. http://www.ijettjournal.org Page 3822 International Journal of Engineering Trends and Technology (IJETT) – Volume 4 Issue 9- Sep 2013 E. GOVINDA received his B.Tech degree in Electronics and Communication Engineering from V.R. Siddhartha Engineering College and M.Tech in DSCE from JNT University, Hyderabad, India. He is currently working as a Head of department of Electronics and Communication Engineering in Avanthi Institute of Engineering and Technology, Visakhapatnam, Andhra Pradesh, India. He has 10 years teaching experience. He has 15 publications in various International conferences. He guided many projects for B.Tech and M.Tech students. ISSN: 2231-5381 http://www.ijettjournal.org Page 3823