Performance Analysis of Different Motion Detection Techniques. Himani K.Borse Bharati Patil G.H. Raisoni College of Engg and Mgmt.,Wagholi,Pume G.H. Raisoni College of Engg and Mgmt., Wagholi, Pume Himaniborse1992@gmail.c om Bharti.patil@email.com ABSTRACT 2.2 Frame Separation: The first step of video analysis is background subtraction. In this, we implemented the four methods and one of the method in FPGA. Detection of foreground or moving object from background is important task of any motion detection techniques. We discussed the four methods as Background subtraction; Frame difference, neural map and Background subtraction with dirichlet process Gaussian mixture model. From the parameters obtained from this methods, we shows how Background subtraction with dirichlet process Gaussian mixture model is better than others methods. first step in the background subtraction algorithm is the Frame processing, the purpose of this step is to make the improved video frames by removing noise and undesirable object’s in the frame in order to increase the quantity of information gained from the frame and the sensitivity of the algorithm. Input video Frame Separation Preprocessing Extraction of object Different motion detection techniques Postprocessing Keywords Background subtraction, Dirichlet processes, video analysis, Frame difference, neural map, 1 INTRODUCTION Human body motion analysis is an essential technology which modem bio-mechanics along with computer vision and has normally used in intelligent control, human computer interaction, motion analysis and virtual reality and other fields. In which the moving human body recognition is the most significant part of the human body analysis, the purpose is to identify moving human body from background image in video sequence, and for the follow-up action such as the target classification its effective recognition plays a very major role. Three are various application of human activity in various fields, the most significant of which is surveillance. Other applications include advanced intelligent user interfaces, character animation for games and movies avatars for teleconferencing, biomechanical analysis of actions for sports and medicine the human body tracking and behavior understanding. 2 BLOCK DESCRIPTION: DIAGRAM AND Target object Figure1: Block diagram of the complete system 2.3 Pre-processing: Preprocessing is a procedure of collection of simple image processing tasks that alters the raw input video into a format. This can be done by succeeding steps. Preprocessing of the video is essential to improve the detection of moving objects. For example, by spatial and temporal smoothing, snow as moving leaves on a tree, should be eliminated by morphological processing of the frames later the identification of the moving objects 2.4 Motion detection by different techniques: There are different techniques are implemented for motion detection such as background subtraction, frame difference method, neural map and background subtraction with Gaussian mixture model. 2.1 Introduction to Digital Video: Digital video means to the capturing, manipulation, and storing of moving images that can be displayed on computer screens. It involves that the moving images be digitally held using the computer. The word digital signifies to a system built on discrete or Discontinuous events, as opposite to a continuous, analog event. Video can be in any format AVI, MPEG, MOV. 2.1.5. Post-Processing: The output of foreground detection has noise. Typically, it affects by various noise factors. To overcome this problem of noise, it involves additional pixel level processing. For elimination of noise from foreground pixel map low pass filter and morphological operations, erosion and dilation are used. Aim in applying these operations is eliminating noisy foreground pixels that do not resemble to actual foreground regions, and to eliminate the noisy background pixels close and inside object regions that are truly foreground pixels. 2.1.6. Extraction of Moving Human Body: Certain accurate edge regions will be got, after median filtering and morphological operations, but the region belongs to the moving human body could not be determined. 2.1.7. Target object: Accurate moving or foreground object is detected form background. 3 MOTION DETECTION DIFFERENT TECHNIQUES: BY 3.1 Background subtraction method: A basic method of detecting moving object is background subtraction, where two frames are compared one is reference or fixed frame and another is current frame. Pixels in the current frame that differ significantly from the background are detected to be moving objects. Further processing of the “foreground” pixels are done for object localization and tracking. Subsequently background subtraction is frequently the major step in various computer vision applications; it’s significant that the extracted foreground pixels perfectly match to the motion objects of interest. There are numerous problems that a background subtraction algorithm must resolve appropriately. Consider a video sequence from a static camera. As it is an outside environment, a background subtraction algorithm must adjust to different stages of illumination at dissimilar periods of the day and operates on adverse weather condition like fog or snow that changes the background. Varying shadow, cast by moving objects, should be eliminated so that reliable features can be collected from the objects in successive processing. . The basis of the method is that of identifying the moving objects from the difference of the current frame and a reference frame, normally called the “background copy”, or “background replica”. As the background image must be an illustration of the scene with no motion objects and should be kept frequently updated so as to adjust to the changing luminance conditions and geometry settings. Simple scheme of background subtraction is to subtract the current image from a reference image that represents the background scene. Normally, the simple steps of the algorithm are as follows: Background modeling considers a reference image signifying the background. Threshold selection defines suitable threshold values used in the subtraction procedure to get a desired detection rate. Subtraction operation or pixel classification categorizes the kind of a given pixel, i.e., the difference is greater than threshold is considered as the pixel is the part of foreground or it is a moving object. Otherwise it is background object If Difference > threshold foreground object Else background object. 3.2 Frame difference method: The aim of motion detection is to identify motion of objects originate in the two given images. Also, finding objects motion can corresponds to objects identification. Therefore, the key objective of this method is to distinguish pixels belonging to the same object. This method gives the whole movement information and identifies the moving entity from the background well. In the frame subtraction method the existence of moving objects is determined by computing the difference between two successive images. Detection of moving object from a sequence of frames received from a static camera is broadly accomplished by frame difference method. The frame difference method is the common technique of motion recognition. It implements pixel-based difference to find the moving object. The frame difference method is the common method of motion detection. This method takes on pixel-based difference to detect the moving object. Firstly, the first frame is taken through the static camera and afterward that sequence of frames is captured at fixed intervals. Secondly, the absolute difference between the consecutive frame and the difference image is calculated and stored in the system Thirdly, this difference image is changed into gray image and then converted into binary image. Lastly, for noise removal morphological filtering is done. Transformation of absolute difference image to Gray Image: There are holes present in moving entity region, and outline of moving object remains not closed. The absolute differential image is converted to grey image to simplify further-operations. RGB to Gray: Y=0.299R + 0.587G + 0.114B Difference of Two Successive Frames: Ik is the value of the kth frame present in image sequences. Ik+1 is the value of the (k+1)th frame present in image sequences. The absolute difference image is defined as follows: 𝐼𝑑(𝐾,𝐾+1) = | 𝐼𝑘+1 – 𝐼𝑘 | Object detection: After difference calculation, object is detected by setting threshold value. | 𝐼𝑘+1 – 𝐼𝑘 | > threshold If difference is greater than threshold then foreground object else background object. 3.3 Neural map method: This method based on selforganization over artificial neural networks, broadly useful in human image processing systems and further commonly in cognitive science. The proposed method can handle scenes containing moving backgrounds, slow illumination deviations and camouflage, has no bootstrapping limits, can contain into the background model shadows cast by moving things, then accomplishes robust detection for various kinds of videos taken with stationary cameras. The key difficulty of the background subtraction approach to motion object recognition is its highly sensitivity to dynamic scene variations due to lighting and extraneous events. While they are commonly identified, they leave behind “holes” where the recently showing background imagery varies from the well-known background model (ghosts). Whereas the background model ultimately adjusts to these “holes,” they create false alarms for a small period of time. Therefore, it is extremely required to construct an method to motion detection built on a background model that automatically adjusts to variations in a self-organizing style and without a priori knowledge. This method accepts a biologically motivated problem-solving method depends on visual attention mechanisms. The purpose is to find the objects that keep the user attention in accordance with a group of predefined features, containing gray level, motion and shape features. Each node computes a function of the weighted linear combination of received inputs, where weights look like the neural network learning. Doing so, each node could be signified by a weight vector, achieved from collecting the weights correspond to incoming links. In the following, the group of weight vectors will be referred as a model. An arriving pattern is mapped to the node whose model is “most similar” (similar to a predefined metric) to the pattern, and weight vectors in a neighbourhood of such node are updated. depicted in Fig. 4.1., the upper left pixel takes weight vectors stowed into the 3*3 components of the higher left portion of neuronal map 3.3.3 Subtraction and Update of the Background Model: After initialization, temporally subsequent samples are providing to the network. Each entering pixel 𝑃𝑡 of the the 𝐼𝑡 sequence frame is matched to the current pixel model to define if there occurs a weight vector that greatest matches it. If a greatest matching weight vector is establish, it means that pixel fits to the background and it is used as the pixel encoding approximation, and the greatest matching weight vector, together with its neighbourhood, is reinforced. Otherwise, if no allowable matching weight vector exists, we identified as belonging to a moving object (foreground). Euclidean distance of vectors gives the distance between two pixels in the HSV color hexcone, 𝑃𝑖 =(ℎ𝑖 , 𝑠𝑖 , 𝑣𝑖 ) and 𝑃𝑗 =(ℎ𝑗 , 𝑠𝑗 , 𝑣𝑗 ) as d(𝑃𝑖 , 𝑃𝑗 ) = ∥ (𝑣𝑖 𝑠𝑖 cos(ℎ𝑖 ), 𝑣𝑖 𝑠𝑖 sin(ℎ𝑖 )𝑣𝑖 ) − (𝑣𝑗 𝑠𝑗 cos(ℎ𝑗 ), 𝑣𝑗 𝑠𝑗 sin(ℎ𝑗 ), 𝑣𝑗 ) ∥22 Hence, the network acts as a reasonable neural network that implements a winner take- all function with an related mechanism, that alters the local synaptic plasticity of the neurons, letting learning to be limited spatially to the local neighbourhood of the maximum active neurons. Steps for neural map method: 1. 2. 3. 4. 5. Take background image and apply mapping 3*3 Assign constant for thresholding depending on environmental condition and camera Take current image and apply 3*3 mapping Apply distance calculation means find distance between background images and current image(Euclidian distance formula) Apply thresholding and make binary image 3.3.1 Applying mapping: For each color pixel, we assumed a neuronal map comprising of weight vectors. Every arriving sample is mapped to the weight vector that is nearest allowing to an appropriate distance measure, and the weight vectors in its neighbourhood are updated.. The representation of HSV values as vectors in the HSV color hexcone used in such distance measure avoids problems connected with the periodicity of hue and with the instability of hue for small values of saturation Weight vector 𝐶𝑚 , for some m, provides the greatest match for the entering pixel 𝑃𝑡 if its distance from is smallest in the model C of 𝑃𝑡 and is no larger than a fixed threshold. 𝑑(𝐶𝑚 , 𝑃𝑡 ) = lim 𝑑(𝐶𝑖 , 𝑃𝑡 ) ≤ 𝜀 𝑖=1,…,𝑛2 The threshold permits distinguishing between foreground and background pixels, and is selected as Fig 3..3.1(Left) simple image and (Right) Neural map structure 3.3.2 Initial Background Model: First image of our sequence is really a good initial estimation of the background, and consequently, for every pixel, the corresponding weight vectors are initializing with the pixel value. In order to signify each weight vector, we select the HSV color space, depend on the hue, saturation and value properties of every color. As 𝜀= { 𝜀1 , 𝜀2 , 𝑖𝑓 0 ≤ 𝑡 ≤ 𝐾 𝑖𝑓 𝑡≥𝐾 With 𝜀1 and 𝜀2 minor constants. Specifically, 𝜀1 must be greater than 𝜀2 , as greater values for 𝜀1 allow, within the first K sequence frames, to get a (possibly rough) background model containing numerous observed pixel intensity deviations whereas lower values for permit to 𝜀2 achieve a more correct background model in the online phase. 3.3.4 Updating in the Neighbourhood: In Fig. 4.1, Input Video(avi) if the greatest match for present pixel is the weight vector, then the weight vectors that are updated according to are weight vectors that belong in portion to the model of current image pixel. Color conversion 3.3.5 Shadow Detection: The simple idea is that a cast shadow blackens the background, whereas a moving entity can darken it or not dependent on its color. Cap model Lighting changes compensation Update mixture model Calculate frame probability 3.4 Dirichlet Process Gaussian Mixture Model (DP-GMM): This method is a non-parametric Bayesian method that spontaneously estimates the Number of mixture components is automatically to model the pixels background color distribution, e.g. Single mode pixel generate at the trunk and in the sky, when the tree is waving forward and backward in front of sky creates two modes pixels in the area where braches wave, i.e. pixel transition among leaf and sky regularly. If it requires more modes to denote multiple leaf color this will take place automatically, and, of excessive significance for term surveillance, this model will update with time. However, two issues of standard DP-GMM model: 1) Update techniques of the existing model cannot cope with the scene changes common in real-world applications; 2) if we used this model for continues video then more computation and memory is required. This model usages a Gaussian mixture model (GMM) for a Per-pixel density estimation and followed by connected component of regulation. Its mixture model has two components foreground and background. It classifies values based on their mixture components, which is allocated to the foreground or the background. larger components belongs to background and remaining belongs to foreground . 3.4.1 Block diagram: Video this is in AVI format is taken because processing take place on AVI format. Then this video is converted into frames and then color conversion is take place that convert the color (RGB) images into gray form. And reduces lighting effect occurs at the time of capturing of video.The proposed method normally splits into two parts—a per-pixel background model and a regularisation step 3.4.1.1 Per-Pixel Background Model: every single pixel has its multi-model density estimate, used to model P(x/bg) where x is the pixel color channels vector. It can be observed as the Dirichlet distribution prolonged to an infinite number of components, which permits it to obtain the true number of mixtures essential to indicate the data. Dirichlet process, first using the stick breaking construction then secondly using the Chinese restaurant process (CRP). Gives clean description of concept is provided by stick breaking, whereas the Chinese restaurant process integrates out inappropriate variables and offers the formulation we actually solve. Dirichlet process: In probability theory, are family of stochastic method whose realization is probability distribution. It is frequently used in Bayesian interference to define the preceding knowledge nearly the distribution of random variables, that is, how possibly it is that the random variable are distributed allowing to one or another specific distribution. Regulization Object detected Thresholding Figure 3.4.1 : Block diagram of the approach The Dirichlet process is presented by a base distribution H and a positive real number α called the concentration parameter. The base distribution is the probable the process value, that is, the dirichlet process does distribution “around” the base distribution in the way that a normal distribution draws real numbers around it. If the base dtribution is constant the distributions drawn from the Dirichlet process remain surely discrete. The concentration parameter identifies how robust this discretization: is in the limit of α→0, the realization are all concentrated on a particular value, whereas in the limit of α→∞ the realization becomes between the two extreme the realization are discrete distribution with fewer and fewer concentration as increases. The Dirichlet process that may also be understood as the infinite-dimensional generalization of the Dirichlet distribution. In the related manner as the dirichlet distribution is the infinite conjugate prior., In the same manner as the Dirichlet distribution is the categorical distribution conjugate prior, the dirichlet process is the infinite conjugate prior, nonparametric discrete distribution, A particularly important application of the dirichlet process is the prior probability distribution in infinite mixture model. Assume that the production of values 𝑋1 ,𝑋2 ,… can be described with the following algorithm. Input: H (Probability distribution called Base distribution), α (Positive real number called Concentration parameter) 1. 2. Draw 𝑋1 from the distribution H. For: n > 0 𝛼 1. With probability draw X from H. 𝛼+𝑛−1 𝑛 𝑥 2. With probability set 𝑥𝑛 = 𝑥, where 𝑛𝑥 is 𝛼+𝑛−1 the number of previous observations 𝑋𝑗 , 𝑗 < 𝑛 such that 𝑋𝑗 = 𝑋. The 𝑋1 ,𝑋2 ,… the observations are dependent, then we have to think through the previous results when producing value. They are still, replaceable. This fact may be displayed by computing joint probability distribution of the observations and seeing that the resultant formula only based on which value X occur among the observations and how many duplications they each have. Procedure of the above algorithm: 1. 1. Draw a distribution P from DP(H,𝛼) 2. Draw observations independently 𝑋1 , 𝑋2 ,… f rom P. The Chinese restaurant process: The "Chinese restaurant process" name is stated from the following analogy: imagine an infinitely big restaurant having an infinite tables, and capable to serve an infinite dishes. The restaurant in question works with a slightly unusual seating policy whereby new dinners are seated either at presently working table with probability proportional to the number of guests at present seated there, or at an unfilled table by means of probability proportional to a constant. Guests who sit at engaged table essential order the identical dish as those presently seated, whereas guests assigned a new table are served a different dish at random.The dishes distribution after J guests are served is sample drawn as described below. Suppose that J J samples, , {θj } j = 1 samples must before been got according to Chinese restaurant process, the (J + 1)th sample should be drawn from θ(J+1)~ 𝟏 (H(S)+J) 𝑱 (H+ ∑𝒋=𝟏 𝜹𝜽𝒋 ) Where 𝛿𝜃 is atomic distribution . Understanding this, two properties are clear: 1. Even if S is uncountable set, there is finite (i.e. non zero) probability that two samples will have nearly the similar value. A dirichlet process samples are discrete. 2. The dirichlet process shows a self-reinforcing property the further every so often a identified value has been sampled in the previous, the best probable it is to be sampled again 2. Where 𝛽𝑘 are independent random variables with the beta distribution Beta(1,𝛼). The correspondence to ‘stick-breaking’ can be realized through seeing as 𝛽𝑘 the length of a part of a stick. We begin with a unit-length stick and every step we halt a portion of the remaining stick according to 𝛽′ 𝑘 and assign this broken-off piece to 𝛽𝑘 . The formula can be understood by observing that subsequently the first k − 1 values have their portions allocated, the length ′ of the rest of the stick is∏𝑘−1 𝑖=1 (1 − 𝛽 𝑖 )and this ′ portion is broken according to 𝛽 𝑘 and becomes assigned to 𝛽𝑘 . The smaller α is, the fewer of the stick will be left consequent values (on average), and resulting further concentrated distributions. In Stick breaking, stick is continuously break infinite times and divides the samples into different Chinese restaurant sets. Integrating out the draw from the DP indications to better convergence, but more significantly replaces the infinite set of stick with the computationally controllable finite set of tables. 3.4.1.2 Probabilistic Regularisation: Per-pixel background model does not take information from the adjacent or neighboring pixel so causes it susceptible to noise and camouflage. Additionally, Gibbs sampling introduces certain amount of noise i.e. dithering effect at the boundary between foreground and background. This issues is resolved by using markow random field, with node of each pixel connected to four way neighborhoods. It is a binary labeling problem where every single pixel corresponds to the either foreground or to background. Threshold value is set to 45, the pixel above threshold is considered as foreground and below background. Cap model is used to update the model Instead of repeating the same calculation. Whatever output is obtained that is updated so when next same output is generated then the previous outputs are taken that are stored so reduces same calculations. Stick breaking: A third approach to the Dirichlet process is so-called stick-breaking process view, a dirichlet process are distributions over a set S. As the distribution drawn is discrete with probability 1.In the sticking-breaking procedure opinion, we clearly use the discreteness and provide the probability mass function of this random) discrete distribution as: 𝒇(𝜽) = ∑ 𝜷𝒌 . 𝜹𝜽𝒌 (𝜽) 𝒌=𝟏 Where 𝛿𝜃𝑘 is indicator function which estimates to zero all over, excepts for. δθk (θk ) = 1. Then this distribution is itself, its mass function is parameterized through two of random variable: the locations {𝜃𝑘 } ∞ 𝑘=1 and the corresponding probabilities{𝛽𝑘 }∞ 𝑘=1 . In the current deprived of proof what these random variables are. The locations 𝜃𝑘 are identically and independent distributed according to H, base distribution of the dirichlet process. The probabilities 𝛽𝑘 are specified by a procedure approximating the breaking of a unit-length stick: ′ 𝛽𝑘 = 𝛽′ 𝑘 .∏𝑘−1 𝑖=1 (1 − 𝛽𝑖 ) 3. HARDWARE: This method is implemented in FPGA. 3.1 FPGA: FPGAs contain an array of programmable logic blocks, and reconfigurable interconnects the logic blocks together, like various logic gates that can be present inter wired in different configuration. Logic blocks are configured to achieve difficult combinational functions or simple logic gates such as the AND or XOR. FPGA receives the current and reference frame data through serial communication bus UART. and stored data in RAM and Perform the and generates the output. This output is given to matlab through UART3. 3.2 RAM: A typical RAM cell has only four connections: Data in (the D pin on the D flip-flop), data out (the Q pin on the D flip-flop), Write Enable (often abbreviated WE; The C pin on the D flip-flop), and Output Enable (the Enable pin which we added).For our board, ram is 16 Mb. Ram stored current frame, reference frame, intermediately generated result (processing) and output. 3.3 UART: The universal asynchronous receiver/transmitter (UART) receives bytes of data and sends the individual bits in sequential manner.at the destination, a second UART re-assembles the bits into whole bytes Each UART holds a shift register which convert serial data into parallel form and vice versa. Transmission of single bit of bye in single wire is less costly than transmission of multiple data in parallel form in multiple wires. The MSE, also denote to the error signal ei,= xi − yi, which is the difference of the original and contaminated signals. 5.1.2 PSNR: It is the ratio of the maximum possible signal power to the corrupting noise power that distracts the fidelity of the representation. PSNR is usually stated with the help of the logarithmic decibel scale. The PSNR (in dB) is defined as: 4. OVERALL SYSTEM FLOWCHART: 𝑀𝐴𝑋 2 PSNR=10log10 ( 𝑀𝑆𝐸 ) =20.log10 ( Input video 𝑀𝐴𝑋1 √𝑀𝑆𝐸 ) =20.log10 (𝑀𝐴𝑋1 )-20.log10 (𝑀𝑆𝐸) (Matlab) Frames to binary image (Matlab) Here, MAXI is the extreme possible pixel image value. When the pixel is denoted with the help of bits per sample, this value is 255. Mostly, when denoted using linear PCM with B bits per sample, MAXI is 2B−1. 8 bit binary image(FPGA) 5.1.3 Entropy: Entropy is a statistical degree of Processsing uncertainty that can be used to describe the texture of the input image. Entropy is an index to evaluate the how much information (quantity) contained in an image. Entropy is defined as E=-∑𝐿−1 𝐼=0 𝑝𝑖 log 2 𝑝𝑖 (FPGA) Generates output (FPGA) Output 8 bit stream (Matlab) Display output (Matlab GUI) Figure 4.1 Overall system flowchart Where L is the total grey levels, 𝑝 = {𝑝0, 𝑝1, … . . 𝑝𝐿−1 } is the probability distribution of each level 5.1.4 Correlation: Normalized cross correlation are used to find out likenesses between current and reference image is given by the following equation NCC= 𝑛 ∑𝑚 𝑖=1 ∑𝑗=1(𝐴𝑖𝑗∗ 𝐵𝑖𝑗 ) 𝑛 2 ∑𝑚 𝑖=1 ∑𝑗=1(𝐴𝑖𝑗 ) 5.2 Results: Output of different motion detection a technique. 5.2.1 Backgound Subtraction: between current and reference frame Ddifference 5. PERFORMANCE AND RESULTS: 5.1 Output Parameters: 5.1.1 MSE: MSE as a measure signal fidelity. The signal fidelity compare the quantitative score of two signal so we can described how to signals are similar and level of distortion (noise) between them. Typically, it is considered that one signal is an original signal, while the other signal is distorted or contaminated with errors. Figure 5.2.1: Output of background subtraction method 5.2.2 Frame difference: difference between two consecutive frames Assume that x = {xi|i = 1, 2, · · · , N} and y = {yi|i = 1, 2, · · · , N} are two finite-length, discrete signals (e.g., visual images), where N is the number of signal samples (pixels, if the signals are images) and xi and yi are the i th samples in x and y, respectively. The MSE is defined as: 1 2 MSE(x,y)= ∑𝑁 𝑖=1(𝑥𝑖− 𝑦𝑖 ) 𝑁 Figure 5.2.2: Output of Frame Difference method 5.2.3 Neural Map method: pixel mapped into 3*3 mapping of current and reference frame. Object detection Video surveillance Object tracking Traffic monitoring 9. FUTURE SCOPE: Figure 5.2.3: Output of Neural Map method 5.2.4 Background Subtraction with Dirichlet process Gaussian mixture model (DP-GMM) Combining information from adjacent pixel in regulization does not fully achieve the information available. A more challenging method of spatial information transmission would be desirable-a conditional Dirichlet process might offer this. Rapid complex lighting variations are not controlled by this method i.e. it fails to handle certain indoor lighting variations. Still, a more sophisticated typical of the foreground and a clear model of left object could further improve our method. REFERENCES: [1] Weiming Hu, Tieniu Tan,”A Survey on Visual Surveillance of Object Motion and Behaviors” ieee transaction on systems, and cybernetics-Part C:application and revievs,vol . 34, no. 3,august 2004 [2] Qi Zang and Reinhard Klette,”Object Classification and Tracking in Video Surveillance”,unpublished Figure 5.2.4: Output of DP-GMM method 6. COMPARISION OF DIFFERENT METHODS PARAMETERS Method /parameter Backgroud subtraction Frame Difference Neural Map DPGMM [3] Kinjal A Joshi, Darshak G. Thakore,”A Survey on Moving Object Detection and Tracking in Video Surveillance System”,International Journal of Soft Computing and Engineering (IJSCE) ISSN: 2231-2307, Volume-2, Issue-3, July 2012 [4] Zhen Tang, Zhenjiang Miao ,”Fast Background Subtraction and Shadow Elimination Using Improved Gaussian Mixture Model”, HAVE 2007 - IEEE International Workshop on Haptic Audio Visual Environments and their Applications ,Ottawa - Canada, 12-14 October 2007 [5] Nishu Singla ,”Motion Detection Based on Frame Difference Method” International Journal of Information & Computation Technology. ISSN 0974-2239 Volume 4, Number 15 (2014), pp. 1559-1565 MSE 0.0115 0.0487 0.0245 0.0102 PSNR 67.5392 61.2563 64.243 68.673 [6] M. Julius Hossain, M. Ali Akber Dewan, and Oksam Chae ,”Edge Segment-Based Automatic Video Surveillance”, Journal on Advances in Signal Processing Volume 2008, Entropy 0.2950 0.0917 0.3766 0.2870 Correlation 0.8832 0.2783 0.8041 0.8812 [7] Tom S.F. Haines and Tao Xiang ,”Background Subtraction with Dirichlet Process Mixture Models” ieee transaction on pattern analysis and machine intelligence,vol. 36,no 4,April 2014 7. CONCLUSION: Dirichlet method is suitable for background modeling, and computationally scalable. This method handles the dynamic background. Infinite no of mixture components are used so whatever object is detected is more accurate than the other methods. And also handles the scene changes. It handles camouflage and shadow effect. Other methods require more computation and give less efficiency. From the above table, we can conclude that the output or human body detection done by Background subtraction with Dirichlet Process Gaussian Mixture model method is more accurate than the another methods 8. APPLICATIONS: [8] Lucia Maddalena and Alfredo Petrosino,,” A SelfOrganizing Approach to Background Subtraction for Visual Surveillance Applications”, ieee transaction on image processing vol. 17, No. 7, July 2008 [9]Babak Shahbaba and Radford Neal ,“Nonlinear Models Using Dirichlet Process Mixtures”,Journal of Machine Learning Research 10 (2009) 1829-1850 [10]Larissa ValmyAnd Jean Vaillant,”Bayesian Inference on a Cox Process Associated with a Dirichlet Process”, International Journal of Computer Applications (0975 8887) Volume 95 - No. 18, June 2014 [11]Ibrahim Saygin Topkaya, Hakan Erdogan and Fatih Porikli” Counting People by Clustering Person Detector Outputs”, 2014 11th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS)