JOURNAL OF INFORMATION, KNOWLEDGE AND RESEARCH IN ELECTRONICS AND COMMUNICATION ENGINEERING DIGITAL REVOLUSION -MOTION PICTURE COMPRESSION TECHNIQUES 1Charmy Patel, 2Dr.G.R.Kulkarni 1Research scholar Department of Electronics & Communication Engineering, Singhania university,junjuhnu,jaipur-Rajsthan,India Gujarat, India 2Principal of Shri C.U.Shah College of engineering and Technology Vadhvan city Gujarat, India charmychangela@gmail.com ABSTRACT: Digital television broadcasting is present and future of television. With the advent of digital technology, digital audio and video compression and other advanced signal processing, it is possible to transmit and receive broadband data over co-axial cable and through satellite and terrestrial. Digital television is a telecommunication system for broadcasting and receiving moving picture and sound by means of digital signals, in contrast to analogue signals in analogue (traditional) Television. It uses digital modulation data, which is digitally compressed and requires decoding by a specially designed television set or a standard receiver with a set-top box. For all delivery mechanism a digital set-top box is needed at the customer’s premises. It receives the digital transmission and decodes video, audio programs to display on normal Television Receiver. Digital television is practically and economically not feasible without motion picture compression technique. The main goal of any conditional access system is to control viewer’s access to digital television pay services and secure the service providers revenue streams. Keywords – digital video, spectral coding, spatial coding, MPEG, predictive picture I INTRODUCTION TO DIGITAL TELEVISION Digital Television is completely new way of broadcasting and is the future of television. It is a medium that requires new thinking and new revenuegenerating business models. Digital Television is the successor to analog Television and eventually all broadcasting will be done in digital format. Around the globe, Satellite, Cable and Terrestrial operators are moving to a digital environment. The digital age will improve the consumer viewing experience through cinema-quality pictures, CD –quality sound, hundreds of new channels. Digital television also opens up a new world of opportunities for companies who want to develop content and applications for the new paradigm. This includes the creative communities within the Television and film industry, internet content providers and software development houses. Finally, the new medium will allow viewers from the comfort of their homes to use a simple remote control to electronically purchase goods and service offered by various content providers. In existing Television system the input signal at the camera and the output to the picture tube are analog. If digital technique is to be used for Television system the composite video signal must be digitized i.e. a binary sequence of 0’s and 1’s must be used. Digital television uses the same language as computers- a long stream of binary digits, each of which is either 0 or 1. With digital television the signal is compressed and only the updated data transmitted. As a result, it is possible to squeeze 6 or 8 channels into frequency range that was previously occupied by only one analog Television channel. Digitization of the composite video signal involves high speed analog to digital converters and very high bit rates for transmission. The problem of high speed conversion and high bit rate for transmission has been solved with the emergence of Broadband- Integrated Services Digital Network (B-ISDN). In 1982 CCIR has unanimously adopted the ‘Recommendation 601’ for encoding parameter of digital Television for Studios. An arrangement has been made on a digital code compatible with both 525/60 and 625/50 standards. II TECHNICAL BACKGROUND OF TELEVISION The analog video format is based on the PAL/NTSC composite video signals, the standards used in Europe and the USA respectively. Black and white video consisted of a single luminance Y signal, which ISSN: 0975 – 6779| NOV 11 TO OCT 12 | VOLUME – 02, ISSUE - 01 Page 270 JOURNAL OF INFORMATION, KNOWLEDGE AND RESEARCH IN ELECTRONICS AND COMMUNICATION ENGINEERING indicated the brightness of pixels on the screen. This signal had a bandwidth of about MHz. When colour television was introduced, the video developers needed to maintain backwards compatibility, with the Black and White system. For this reason, the Y signal was maintained and the colour information was added in via the chrominance (C) signal. The C signal was positioned within the same spectrum as the Y signal at about 4.3MHz (PAL). The C signal is further composed of two colour difference signals, U and V, which are Quadrature Amplitude Modulated to form the C signal. This gives the U V Y format for the colour composite video signal. At the television receiver the U V Y format is converted in to RGB format, which informs the CRT what proportion of Red, Green and Blue light to shine on each pixel on the screen. The U V Y format is therefore a compressed way of transmitting the sequence of RGB frames, which make up the video signal. For digital transmission, the video must be digitized first using and A/D converter. The video is converted in to samples with, for example, 8bits per sample. The data is now simply a bit stream, which can be manipulated by computer. II.1 video signal after digitization Chrominance (Y): 13.5MHz Chrominance (Cr): 6.75MHz Chrominance (Cr): 6.75MHz Quantization : 8bit/sample Total bits per second: (13.5 + 6.75 +6.75)*8= 216M bits Total BW required : 108 MHz II.2 Digital Television with compression MPEG (Moving Picture Expert Group) The acronym MPEG stands for moving picture expert group, which workd to generate the spccifications under ISO, the International Organization for Standardization and IEC, the International Electrotechnical Commission. What is commonly referred to as ‘MPEG Video” actually consists at the present time of two finalized standards, MPEg-1 and MPEG-2 with a third standard, MPEG-4, was finalized in 1998 for very low bit rate audio-visual coding. The MPEG-1 and MPEG-2 standards are similar in basic concepts. They both are based on motion compensated block based transform coding techniques, while MPEG-4 deviates from these more traditional approaches in its usage of software image construct descriptors, for target bit-rate in the very low range. Because MEPG-1 and MPEG-2 are finalized standards and are both presently being utilized in a large number of applications this report concentrates on compression techniques relating only to these two standards. Note that there is now reference to MPEG-3.This is so because it was originally anticipated that this standard would refer to HDTV application, but it was found that minor extensions to MPEG-2 standard would suffice for this higher bit-rate, higher resolution application, so work on a separate MPEG3 standard was abandoned. To understand the concept of motion picture compression technique we can compare it with orange juice analogy as shown in fig. 1 Fig:1 comparision of compresion with orange juice analogy Filtering the video signal to remove the noice etc. otherwise it may corrupt the video signals.like washing of oranges. Removing unwanted information from the video signals like removing peel, seed, skin from oranges. By digitization process converting analog video into digital video like crusing process converting orange from orange juice. Asignificant amount of informationcontent in the video signal is redundant. An expensive encoder is used to remove the information from digirized video signal without destroying its quality like water content in orange juice is redundant. An expensive plant is used to remove the redundant water content from the orange juice without destroying its quality. Output of the encoder is digital video whose no of bits are much less than the bits of the digitize video fed to the encoder like the output of the plant is pulp whose physical quantity is less than the orange juice quantity fed to the plant. The bits of compressed digital video are convered into small packets. Hearder of each packet contains necessary information such as name, number of bits, and direction for decoding to convert it into almost like original video signal like the orange pulp is packed in small tins. Necessary information ISSN: 0975 – 6779| NOV 11 TO OCT 12 | VOLUME – 02, ISSUE - 01 Page 271 JOURNAL OF INFORMATION, KNOWLEDGE AND RESEARCH IN ELECTRONICS AND COMMUNICATION ENGINEERING such as name of company, details of ingredients, weight, directions for handling and converting pulp into almost like fresh orange juice. Packets of bits are multiplexed to form group of pictures. These are now ready for storage in hard disc or transmission by modulating on carrier, like small sealed orange tins are packed into big cartons for storage/transportation. These tins in the carton are transported by some carriers. The decoder on receiving the packets converts the compressed bits into almost like original video signal as per the information given in the header, like consumer followes the instructions written on each tin to prepare almost like fresh orange juice from the pulp. For this he is required to add correct quantity of water and ice etc. The decoding process of converting compressed bits in to video signal is much simpler than encoding process. Therefore decoders are cheaper and less complicated than encoder, like the process of making orange juice from the pulp is very simple as compared to the process of converting orange juice into pulp. Encoder is like a black box. The manufacturer does not specify, How coding is done. It is their trade secret .only bit stream and decoding format is described, like manufacturer does not specify how exactly pulp is prepared. It is trade secret. They only specify its ingredients and instruction about how to prepare orange juice from the pulp. II.3 types of redundancy in video signal Spectrum redundancy Spatial redundancy Temporal redundancy Entropy redundancy Psycho-visual redundancy II.4 coding techniques for compression Spectral coding Spatial coding Temporal coding II.4.1 Spectral coding The RGB signals form video cameras are highly correlated and take on large bandwidth To decrease the amount of video sample data based on human perception, the RGB color space is converted to Y Cr Cb color space. The y has the full bandwidth as it is very sensitive to human perception. The Cr and Cb components have a narrower bandwidth because these are less discernible by human eye. The chrominance components are usually decimated by two, both horizontally and vertically resulting in a reduced number of samples. II.4.2 Spatial coding Video signal exists in four dimensions i. The magnitude of the sample ii. The horizontal spatial axis iii. The vertical spatial axis iv. The time axis Since spatial coding is confined only to individual frame therefore it works in three dimensions only is on the horizontal and vertical spatial axis and on the sample values. In a picture, where there is a high spatial frequency content due to detailed areas of the picture, there is a relative small amount of energy at such frequency. Also picture may contain sizeable areas in which the same or similar picture values exist. This give rise to low spatial frequency contents e.g. areas of sky or grassland. The average brightness of the picture results in a substantial zero frequency (DC) components. Simply omitting the high frequency components is unacceptable as this causes an obvious softening of the picture. Human eye’s sensitivity to noise in high spatial frequencies is less. This is used for the coding gain. Coding gain is obtained b taking advantage of the fact that the amplitude of the spatial components falls with frequency. if the spatial spectrum is divided into frequency bands the high frequency bands can be described by fewer bits not only because their amplitudes are smaller but also because more noise can be tolerated. Fig: 2 block diagram of spatial coding process II.4.3Temporal coding Temporal coding or inter coding takes advantage of the similarities between successive pictures in real material. Instead of sending information for each picture separately, inter coders will send the difference between the previous picture and the current picture in a form of differential coding. A picture store is required at the coder to allow comparison to be made between successive pictures and a similar store is required at the decoder to make the previous picture available. The difference data may be treated as a picture it self and subjected to some form of transform based compression. ISSN: 0975 – 6779| NOV 11 TO OCT 12 | VOLUME – 02, ISSUE - 01 Page 272 JOURNAL OF INFORMATION, KNOWLEDGE AND RESEARCH IN ELECTRONICS AND COMMUNICATION ENGINEERING Motion compensation is a process which effectively measures motion of objects from one picture to the next so that it can allow for that motion when looking for redundancy between pictures. correction. A coding error in this picture does not propagate and the quantization step size is controlled to be relatively larger. B picture requires one quarter the data of an I picture. Temporal coding is achieved by creating three types of pictures. I(Intra) Picture P(predictive) Picture B(Bi-directionally predictive) Picture I (Intra) Picture This type of picture is intra frame coded that need no additional information for decoding They require a lot of data compared to other pictures types, and therefore they are not transmitted more frequently than necessary. They consist primarily of transform co-efficient and have no vectors. The I pictures are inserted periodically for the purpose of blocking propagation of errors caused by inter frame coding and for the purpose of realizing random access in broadcasting or storage media environments. The I picture affects the total image quality substantially. P (predictive) Picture This type of pictures are forward predicted from an earlier picture which could be an I picture or P picture. Picture data consists of vectors describing where, in the previous picture each macro block should be taken form and transform coefficients which describe the correction in difference data which must be added to that macroblock. The picture propagates coding errors to the following P and B pictures so it need to be quantized more finally than B pictures. B (Bi-directionally predictive) Picture B pictures are bi-directionally predicted from earlier or later I or P pictures. B data consist of vectors describing where in earlier or later pictures data should be taken from. It also contains the transform coefficients that provide the II.5 MPEG video compression Algorithms MPEG-2 video compression algorithm is composed of ME/MC. DCT scaler quantization. RLC and Huffman coding functions. It is a hybrid algorithm that combines the intra frame and inter frame coding schemes. The MPEG-2 video syntax has a hierarchical structure comprising the sequence layer, groups of picture (GOP) layer, picture layer, slice layer inacro block layer, and block layer. Among these the GDP layer is designed for random access and recovery from transmission errors. The slice layer is for resynchronization at the decoder in the case of transmission errors. When a slice is lost at the decoder due to transmission errors in the channel. The next slice can be received after resynchronization at the start of the slice. When the decoder is turned on during the transmission of the bit stream or when a channel change occurs in the broadcasting environment. The pictures are reconstructed and presented from the next I picture. II.5.1 MPEG-1 It was finalized in 1991, and was originally optimized to work at video resolutions of 352 x 240 pixels at 30 frames/sec for NTSC and 352 x 288 pixels at 25 frames/sec for PAL, commonly referred to as source input format video. It is often mistakenly thought that the MPEG-1 resolutions is limited to the above seize, but it in fact may go as high as 4095 x 4095 at 60 frames/sec.. The bit rate is optimized for applications of around 1.5Mb/sec, but again can be used at higher rate if required. MPEG-1 is defined for progressive frames only, and has no direct provision for interlaced video applications, such as in broadcast television application. II.5.2 MPEG-2 It was finalized in 1994, and addressed issues directly related to digital television broadcasting such as the efficient coding of field- interlaced video and scalability. Also the target Bit rate was raised between 4 and 9 Mb/sec, resulting in potentially very high quality video. ISSN: 0975 – 6779| NOV 11 TO OCT 12 | VOLUME – 02, ISSUE - 01 Page 273 JOURNAL OF INFORMATION, KNOWLEDGE AND RESEARCH IN ELECTRONICS AND COMMUNICATION ENGINEERING MPEG-2 consists of profiles and level. The profile defines the bit stream scalability and the colour space resolution, while the level defines the image resolution and maximum bit-rate per profile. Probably the most common descriptor in use currently is Main Profile, Main Level (MP@ML)which refers to 720 x 480 resolution video at 30 frames/sec, at bit-rate up to 15Mb/sec for NTSC video. Another example is HDTV resolution of 1920 x 1080 pixels at 30 frame/sec at a bit rate up to 80 Mb/sec. This is an example of the Main Profile, High level (MP@HL) descriptor. MPEG compression dramatically decreases the amount storage space required to record motion picture sequences by eliminating redundant and non essential image information from the stored data. Less total bits means that motion pictures can be transferred more rapidly, so rapidly in fact, that slow, but inexpensive, communication line and storage devices have entirely new moving picture applications. The human eye is forgiving of approximation and outright elimination of the finest details in images. This is fortunate because there is proportionally a lot more fine detail than coarse detail in images. MPEG compression approximates the intensity of fine detail with just a few shades (and progressively more shades for coarser detail) saving many bits over full representation. The eye also sees less colour change per inch then it does brightness variation. MPEG compression exploits this reduction by elimination non-essential colour detail that the human eye tends to gloss over anyway. II.4.4 Precaution with compression techniques The transmission systems using compressed data must incorporate more powerful error correction strategies and avoid compression techniques which are notoriously sensitive. If not necessary compression should not be used. If compression is to used, the degree of compression should be as small as possible. In other words highest practical bit rate should be used. Low bit rate coders should only be used for the final delivery of post produced signal to the end user. Compression quality should only be assessed subjectively. Compression quality varies widely with source material. One should be careful about demonstration which may use selected material to achieve low bit rates. While assessing the performance of a codec one should not hesitate in criticizing artifacts. Eyes and ears are good to assess the quality of the output of the codec. Use still frames to distinguish special artifacts from temporal artifacts, while assessing the performance. Digital television has several advantages over traditional TV, the most significant being use of a smaller channel bandwidth. This frees up space for more digital channels, other non-television services such as pay-multimedia services and the generation of revenue from the sales of frequency spectrum taken by analog TV. Using Motion Picture compression technique like MPEG, the difference picture is produced by subtracting every pixel in one picture from the same pixel in the next picture. MPEG-1 and MPEG-2 standards are similar in basic concepts. They both are based on motion compensated block based transform coding techniques, while MPEG-4 deviates from these more traditional approaches in its usage of software image construct descriptors, for target bit-rate in the very low range IV REFERENCES [1]. Digital Audio Visual Council http://www.davic.org/ [2]. Basic Television & Video System By Grob and Herndon, McGraw hill publication [3]. Real-Time Video Compression techniques and algorithms by Raymond Westwater Borko Furht, [4]. Kluwer Academic Publishers [5]. Phil tudor “Progressive transmission of interlaced picture ”BBC research and development [6]. J o Drewary “Interlaced and MPEG can motion compensation help?” British broadcasting communication, U .K. [7]. KEESMAN, G. et al., 1996. Trans coding of MPEG bit streams. Image Communication vol. 8, September, pp. 481-500. TUDOR, P. and WERNER, O., 1997. Real-time trans coding of MPEG-2 video bit streams. International Broadcasting Convention, Amsterdam, September, pp. 226-301. [8]. THOMAS, G. and DANCER, S., 1995. Improved motion estimation for MPEG coding within the RACE 'COUGAR' project. IBC, Amsterdam, September, pp. 238-243. [9]. KNEE, M., BRIGHTWELL, P., TUDOR, P. and WELLS, N., 1997. MPEG-2 Main Profile coding – towards the quality limits. International Broadcasting Convention, Amsterdam, September. [10]. STOREY, R., ALVES, A. et al., 1997. The ATLANTIC News Studio: reference model and field trial. Paper published at ECMAST, Milan, 21-23 May. [11]. SARGINSON, P., 1997. Dynamic multiplexing of MPEG-2 bit streams. International Broadcasting Convention, Amsterdam, September, pp. 566-571. III CONCLUSIONS ISSN: 0975 – 6779| NOV 11 TO OCT 12 | VOLUME – 02, ISSUE - 01 Page 274