7. O. Hadar, and A. Stern, “Image compression improvement by pre

advertisement
MTF as a quality measure for compressed images transmitted
over computer networks
Ofer Hadara, Adrian Sternb, Merav Huberc and Revital Huberc
a
Communication System Engineering, Department, Ben-Gurion University
of the Negev, Beer-Sheva, 84105 Israel
b
Electrical and Computer Engineering Department, Ben-Gurion University
of the Negev, Beer-Sheva, 84105, Israel
c
CCIT-Center for Communications and Information Technologies,
Computer, Network Laboratory, Faculty of Electrical Engineering,
Technion, Haifa 32000, Israel
ABSTRACT
One result of the recent advances in different components of imaging systems technology is that, these systems
have become more resolution-limited and less noise-limited. The most useful tool utilized in characterization of resolutionlimited systems is the Modulation Transfer Function (MTF). The goal of this work is to use the MTF as an image quality
measure of image compression implemented by the JPEG (Joint Photographic Expert Group) algorithm and transmitted
MPEG (Motion Picture Expert Group) compressed video stream through a lossy packet network. Although we realize that
the MTF is not an ideal parameter with which to measure image quality after compression and transmission due to the nonlinearity shift invariant process, we examine the conditions under which it can be used as an approximated criterion for
image quality. The advantage in using the MTF of the compression algorithm is that it can be easily combined with the
overall MTF of the imaging system.
Key words : Image compression, JPEG, video compression, MPEG, MTF, PSNR, packet loss models.
1. INTRODUCTION
Applications that require the transmission of digital captured images from a remote computer or server through
computer network has become common in various fields today, such as Video on Demand (VoD), video conferencing,
distance learning and military imagery. Such applications involve the incorporation of both digital transmission and imaging
systems. The combination of these two systems demands suitable tools for characterizing the image quality at the observer
site. The purpose of this work is to study the effect of the transmission subsystem on the image quality according to a
system-level analysis approach. The amount of video information is in general substantial and therefore it must be
compressed before it can be sent through the network. As a result the bandwidth, requirements are reduced and the
utilization of network resources increased.
Any communication network that involves transmission of digital images is confronted with two basic problems.
The first is that all lossy compression techniques distort and delay the signal. Degradations result from quantization and
subdivision to blocks, which are irreversible processes during encoding. The usual coding artifacts appear as small stains
around edges (mosquito noise), blurring of textured area and visible block boundaries in almost uniform areas. 1 In this paper
we consider two compression algorithms: the Joint Picture Experts Group (JPEG) designed for still images compression and
the Motion Picture Experts Group (MPEG) designed for video compression. The second problem is that information may be
altered or lost during transmission, due to buffer overflow and channel noise. For compressed bit streams, this loss can lead
to visual distortion at the decoder site. There are two kinds of transmission errors: random bit errors and erasure errors.
Random bit errors are caused by imperfect physical channels and result in bit inversion, bit insertion and bit deletion. The
used coding would determine the impact of bit errors: deep compressed signals would be affected more by bit errors.
Erasure errors can be caused by packet loss, burst errors in storage media due to physical defects, or short time system
failures.2 In section 4, two types of error models are applied to the communication channel simulation: the burst model and
the Independent Identical Distribution (IID) model. These two models represent the most common errors that may occur in

Correspondence: e-Mail : hadar@bgumail.bgu.ac.il; http://www.cse.bgu.ac.il/~hadar/; Tel: +972-7-6477233;
Fax:+972-7-6472883
a real communication network. The burst errors typically occur due to network congestion and the IID error model
represents losses that may occur due to random noise in the channel.
In this work, the transmission system is considered as a component of an overall imaging system that takes the
image from the object plane to the observer. In system engineering approach it is important to determine final image quality
or to design the system for a given required image quality. This requires quantitative criteria with which to define image
quality. An engineering tool very appropriate for system design and analysis is the concept of the Modulation Transfer
Function (MTF), which relates output and input to imaging system. Network transfer functions are an essential tool in
electrical engineering. Similarly, Optical Transfer Functions (OTFs) are essential in image systems. 3 Imaging system is
fabricated from many components, each of which can be characterized by its MTF. A block diagram for a typical imaging
communication system is presented in Fig. 1. In order to be able to evaluate the overall system performance, each
component should be described in the same manner. If each component in the system is described by its MTF then the
overall system MTF is the product of the component MTFs. The objective of our work is to quantify the degradation
obtained from compression and transmission over congested network and noisy channel.
Eye
Display
Environment
CCD
Compression
MTFsys = MTFenvironment · MTFoptics · MTFelectronics
MTFeye
Transmission
·
Decompression
MTFcompression+transmission= ? · MTFdisplay ·
Fig. 1: Block diagram of imaging system incorporating compression and transmission .
Today, most advanced imaging systems are resolution-limited rather than noise-limited, the common measure of
degradation in resolution-limited systems is the MTF. A system level analysis of imaging systems includes compression and
transmission requires the use of similar tools to characterize each component of the system. Compression algorithm
performances are in general evaluated with measures appropriated to noise-limited systems, such as the Mean Square Error
(MSE), and the Peak Signal to Noise Ratio (PSNR). These kinds of measures were appropriate for quantifying the image
quality while systems were more affected by noise rather than by resolution.
The paper is organized as follows. In Section 2 a short review on common image quality measures and their
limitations are presented. Section 3 presents the influence of the JPEG still image compression algorithm. Section 4 presents
the system simulation for transmission compressed digital video over noisy channels and Section 5 describes the results of
image quality from the network simulation. Section 6 concludes the paper.
2. IMAGE QUALITY MEASURES AND THEIR LIMITATIONS
As shown in Fig. 1, the compression algorithm can be understood as a component of the imaging system. Its
contribution to image degradation could be described by the component MTF. This is the goal of the sensor designer/
analyst, however the MTF approach presents a number of problems. First the MTF of the component describes the spatial
modulation throughput of the components, where the output frequency is some scaled value between 0 and 1 of the input
frequency. The mathematical conditions for applying the Fourier transform to define the MTF include linearity and
isoplanatism. In general optical imaging systems are not linear,4 however, they are approximated as Linear Shift Invariant
(LSI) in order to use the powerful mathematical tools developed for LSI systems. Nonlinear components of an imaging
system, such as imaging sensors, recording films, abberated optics, and so forth can be approximated as linear in a certain
range, and the MTF is applicable in that specific range. For shift variant systems, such as imaging through turbid media, the
average frequency response is calculated for isoplanatic patches. 5 Of course, the average performance of such an imaging
system does not in general coincide with the actual performance. However, the average measure gives us an indication
about the global performance of the system. Compression algorithms and the transmission process (Fig. 1) are not typically
linear, which means that not all the spatial frequencies at the output image originate from the same spatial frequencies at the
input image. The JPEG and the MPEG algorithms that are considered in this paper are not linear due to the use of the
irreversible quantization matrix in the compression process. These algorithms are space variant (not isoplanatic) because the
Discrete Cosine Transform (DCT) is applied on blocks. As a result, the MTF depends on the specific image contents to be
compressed.6,7 When the same compression quality factor is applied on different images, different MTFs and compression
ratios may be obtained. Therefore, the typical MTF cannot precisely describe the spatial frequency response of the
compression algorithm. However, the average MTF can be used by considering the image complexity within a spatial
frequency range.
The measures most commonly used for quantifying the quality of compressed images are the Mean Square Error
(MSE) and the Peak Signal to Noise Ratio (PSNR).8 They present the difference between the compressed image and the
original image. As an example, for an image of size N x M pixels the MSE and the PSNR are given as,
N M
1
MSE 
*   [Original _ image(n, m)  Degraded _ image(n, m)]2
M * N n 1 m 1
(1)
2
PSNR  10 log 10 (
P
eMSE
)
where P is the pixel's maximum value. The PSNR is a more appropriate measure because it is normalized to the intensity
scale of the image. Never the less, is not an ideal measure for the following reasons: 1) it is more suitable for comparison
purposes, and does not provide significant information about the absolute image; 2) unlike the MTF, it does not include
information about the transform behavior at different spatial frequencies. 3) The human visibility factor is not taken into
account in PSNR calculations. 4) The PSNR is a global measure and therefore does not give any local information about the
output image quality. It may occur that two images have the same PSNR value, but that the information loss is different
(Fig. 2).
Fig. 2 : Two kinds of distortion with same PSNR [Ref. 9 ]
Cox and Driggers6 proposed an Information Transfer Function (ITF) that can be used in a similar manner as the
MTF. This measure describes the correlation between the input image and the output image as a function of the spatial
frequency compression ratio and the image complexity. They suggest two types of ITF’s. The first is defined as the ratio of
the crosscorrelation between the original image and the compressed image and the autocorrelation of the original image
calculated at different spatial frequencies. The crosscorrelation and the autocorrelation functions are derived at zero spatial
displacement, so there is a hidden assumption of stationarity. The ITF defined this way is related to the MTF; actually it can
be easily shown that for stationary images the ITF is identical to the MTF. The second ITF measure is defined as 1-MSE,
which is calculated separately for each spatial frequency.
3. THE MTF OF JPEG COMPRESSED IMAGES
In this section, we investigate the behavior of the compression MTF with different test images and several compression
ratios, for which images were compressed and reconstructed with a JPEG compression algorithm. The JPEG standard 8 is a
collaborative effort of the CCITT (International Telegraph and Telephone Consultative Committee) and the ISO
(b)
(a)
Fig. 3 (a) “sun” and (b) “trees” test images.
(International Standards Organization), and actually comprises a variety of methods. We applied the implementation most
widely- used to separate the image into 8X8 blocks, perform two-dimensional (2D) DCT on each block, and quantize the
DCT coefficients with respect to a quantization matrix specified by the “q” quality factor. The 2D DCT coefficients are then
converted to a 1D matrix using a zig-zag scan. This 1D matrix is coded using predefined Huffman coding tables. The
remaining 1D data stream includes many zero coefficients that represent high spatial frequencies, which are coded by a run
length code. The decompression or image reconstruction procedure reverses the compression steps.
We used five test images: a point target, a line target, an edge target, a “sun” target and a “trees” image. The point
target is a white pixel in the center of a 512X512 image with a black background. The line target is a similarly-sized image
with a horizontal white line in the middle. The edge target is a half-white (512x256 pixels) half-black image. The “sun”
target and the “trees” image are shown in Fig. 3. The “sun” is a circular sinusoidal image with an angular frequency of 1/20
cyc./rad.
The decompressed point and line test images can be considered as a Point Spread Function (PSF) and a Line
Spread Function (LSF), respectively. The Fourier Transforms (FT) of these images yield the Optical Transfer Function
(OTF), and the absolute value of the OTF gives the MTF. The MTF obtained this way is accurate only for LSI transform.
However, since the JPEG algorithm is not LSI, the MTF is only an approximation of the transform frequency response. The
MTF calculated this way depends on the position of the point or the line. Shifting the point and line in the test images to
different locations yield different MTFs as demonstrated in Fig. 4 showing eight MTFs (dashed lines) appropriate to eight
lines in the center of the image. These MTFs represent all possible MTFs of horizontal test images. The solid line in Fig. 4
shows the average MTF in the horizontal direction for the line test image. As mentioned earlier the use of the average MTF
is common for other space variant systems as mentioned before. The average MTF does not specify local quality of the
decompressed image, but gives an indication about the global decompressed image quality.
Fig. 4 Dashed lines - Estimated MTFs for line test images with the
line placed at different locations; solid line – average MTF.
The LSF is estimated from the decompressed edge target by deviating the image in the edge direction.3 The
average 1D MTF is estimated by taking the absolute value of the LSF Fourier transform and averaging the pixels in the
direction normal to the edge. The edge test image is used to estimate the non-linearity effect of the compression on the LSF
because the edge target can be viewed as combination of line targets.
We designed the “sun” target [Fig, 3(a)] because it is isotropic and has spatial frequencies in a continuous range.
The intensity (image irradiance values) on each arc between two lines changes sinusoidally with a spatial frequency
depending on the distance from the origin:
f 
where

20r
(2)
f  is the spatial frequency in polar direction and r is the circle radius (distance from the origin). Our “sun” target is
related to commonly - used targets such as the “sector star target” and the “RCA” target, but also differs from them in that
its intensity is sinusoidal on a given arc rather than having a square–wave form. For each arc, we calculate the Modulation
Contrast3,4 (MC) defined as (Imax -Imin)/ (Imax+Imin), where Imax and Imin are the maximum and minimum intensity values on the
arc respectively. By dividing the MC of the decompressed image to the MC of the original image and by using Eq. 2 we
find the Modulation Contrast Function (MCF) for the spatial frequency related to the radius of the arc. For LSI transforms
of sinusoidal targets (such as the “sun” target) the MCF is identical to the MTF. We calculate the average MCF for each
spatial frequency by averaging the MCFs obtained for all possible arches with the same radius, which is related to the
spatial frequency according Eq. 2. Fig. 4 shows this average MCF for three quantization factors (quality factors) q=50, 10
and 0 yielding compression ratios (=original image size/compressed image size) of 9, 18 and 32 respectively.
Fig. 5: Average MTF derived from the “sun” target from 3 different quantization matrixes
As expected, it can be seen that the MTF is lower for deeper compression (higher compression ratios). However, the MTF
for all three cases does not decrease significantly, and the lowest value (up to spatial frequency of 0.35 cyc./pixel) is 0.75.
This observation indicates that the JPEG algorithm does not seriously affect the image contrast. We can be observe this
visually in Fig. 6 (b) which shows the “sun” image after compression with the lowest quality factor (highest compression
ratio) and decompression. In each ray the minimum and maximum intensity values (black and white) determining the MC
remain approximately similar to that of the original image. Even in the central area of the image containing high spatial
frequency components, the rays are resolvable approximately as they are in the original image [Fig. 6(a)]. Thus, for JPEG
compression, the average MTF is related to the average ability to resolve objects. However, the MTF does not indicate the
distortion that is apparent even at low frequencies (large distances from the center). This distortion is mainly due to
blocking effects8,7 and artifacts resulting from Gibb’s phenomenon. 8 From the perspective of the Fourier domain, the JPEG
algorithm only slightly affects the fundamental spatial frequency of a periodical image. Therefore, the contrast appears
almost unaffected. However, since JPEG is not an LSI transform, it adds spatial frequency components, which cause the
visible distortions. These distortions can be quantified by using the PSNR measure or the 1-MSE measure defined above.6
Unfortunately, for non-LSI transforms, the MSE based measures are image dependent and, therefore, can be characterized
only for a specific group of images.
(a)
(b)
Fig. 6 (a) original “sun” image, (b) “sun” image after compression with quality factor q=0. Note that
the rays are almost resolvable as in (a).
The “trees” image [Fig. 3(b)] was chosen to analyze the JPEG MTF for a non-artificial image. The MTF is calculated via
the power spectrum density (PSD) of the images 10:
MTF ( f ) 
S IO ( f )
S OO ( f )
(3)
where SIO(f) is the cross PSD of the decompressed and original image and SOO(f) is the original image PSD. The PSD is
estimated using Welsh method with overlapping blocks and a Hanning filter. 10 For LSI transforms of the original images,
S IO ( f )
. For JPEG transform Eq. 3, is considered as an approximation of the
S OO ( f )
the MTF can be described precisely by
mean MTF.
In Fig. 7 the average MTF calculated from the “trees” image is shown together with the average MTF for the
other four tested images. It can be seen that the MTF depends on target complexity. The more complex the image, the
higher is the MTF. In order to characterize the image complexity, we divide the image into N 16X16 blocks and calculate
the standard deviation  of each block. The complexity is given by:
1/ 2
1 N

complexity     i2 
 N i 1 
.
(4)
A similar approach is used to measure the clutter. 11 The size of the block dimension for the clutter measure is taken as twice
the target size. The lowest complexity using (4) is obtained for the point target and equals 1 and the highest is obtained for
the test image and equal 33.
Fig. 7 : MTF derivation for JPEG image compression from different objects.
4. MTF FOR TRANSMITTING-MPEG COMPRESSED IMAGES
In this section an imaging system with a compression and transmitting sub-block is considered (Fig. 1). The
compression-transmission scenario is illustrated in Fig. 8 where the server transmits the video stream through the channel
toward the client. The original analog signal is compressed by the MPEG algorithm. Then it is transmitted through a noisy
channel where packets may be lost. The received signal by the client is decompressed and transformed to an analog signal
by the MPEG player.
The MPEG standard was developed by the ISO/IEC JTC1/SC29 WG11, working group within the International
Standards Organization, for compressing motion pictures and multimedia, in contrast to JPEG, which is used for still
images. There are three types of pictures that are considered in MPEG: Intra pictures (I-frames), predicted pictures (PFrames) and bi-directional (interpolated) pictures (B-Frames). The I-Frame is the first frame of each GOP. The intra
pictures (I-Frames), provide reference points for random access and are subjected to moderate compression. From the
compression point of view, I-pictures are equivalent to images and can be DCT encoded using a JPEG-like algorithm. In P
and B pictures, the motion-compensated prediction errors are DCT coded. Only forward prediction is used in the P-pictures,
which are always encoded relative to the preceding I- or P- pictures. The prediction of the B- pictures can be forward,
backward, or bi-directional relative to other I- or P-pictures. In addition to these three, there is fourth type of picture called
D-pictures, containing only the DC component of each block. They are mainly useful in browsing at very low bit-rates. The
number of I, P, and B frames in a GOP is application dependent. For example, it depends on the access time requirement
and the bit-rate requirement.12
Server
Client
Analog video
Analog video
Encoder
Decoder
Lossy
packet
network
Transmitter
Original video stream
Receiver
Received video stream
Fig. 8 A video stream transmitted through the network
In order to measure the MTF from an MPEG video stream we created a special sequence of images that is suitable
for deriving the MTF in all directions. The sequence includes images of a spreading and shrinking white disk over a black
background. The radius of the circle changes at a rate of one pixel in each frame where there are 25 frames per second. The
isotropic symmetry of the images allows us to derive the MTF at three representative angles horizontal (0o), vertical (90o),
and diagonal (45o). After creating the sequence from a collection of separate frames it was compressed by the MPEG
encoder. Periodically the compressed video stream was transmitted through the simulated lossy channel several times in
order to acquire sufficient data, reaching a total stream of 1200 frames. The stream was divided into blocks of equal size,
each block containing 64 bytes. A model-based data loss generator was used to simulate packet network losses, which
included the burst and the IID models.
In the IID model, we draw a random number, uniformly distributed between 0 and 1. If the drawn number is higher
than a certain Packet Loss Ratio (PLR) value, the packet is considered as correctly accepted. Otherwise, the packet is
considered as lost. The PLR value is a pre-determined parameter in the simulation; we use three values of PLR: 0.01, 0.02
and 0.04. In the burst model, we draw two random numbers. The first number determines whether there will be a packet
loss, and the second determines the number of packets that will get lost all together. The burst size is randomly selected
between 1 and 10. The total error probability is the same as in the IID model. The losses occur randomly, such that any
packet may be lost, independent of the importance of the information type. The degradation of quality depends on the
picture type of the lost video data because of the predictions used for MPEG. For example, if loss occurs in a reference
picture (I or P picture), the lost macroblocks will affect the predicted macroblocks in subsequent frames. This is known as
temporal propagation.1
The analysis of the image quality after decoding is applied separately to each image from the sequence according
to the same method described in the previous section for JPEG compression. In this work, we assume that there is no use of
an algorithm for concealment of missing packets. Therefore, any lost packet may reduce the quality of the video stream.
Video quality is evaluated by the MTF and the PSNR tools presented earlier. These measures are calculated at the output
decoded video stream (Fig. 1), where the compression and transmission subsystems are considered as one unit. The results
of the MTF for PLR=0.01 are presented in Figs 9 and 10 for the IID and the burst loss models respectively.
Fig. 9 Average MTF curves, for IID loss model and PLR=0.01, for three angles: (a) 0 (b) 90 (c) 45.
Fig. 10 Average MTF curves, for burst loss model and PLR=0.01, for three angles: (a) 0 (b) 90 (c) 45.
Each figure includes three separate graphs, for the three angles: 0˚, 90˚, and 45˚. The MTF curve represents the
average MTF of the whole sequence. It is measured from the edge response along the three angles by using the Fourier
transform of the derivative of the edge response. The MTF is a measure of the edge sharpness. Therefore, only frames
which exhibit a blurring effect and not other kinds of degradation, were considered in the MTF derivation. As a result, we
applied an algorithm to select the appropriate frames from the whole sequence. Only MTF vectors whose zero spatial
frequency component was equal to one (MTF(f=0)=1) were taken into account.
The percentage of frames participating in the average MTF calculation was lower than 25 %, for all the sequences
tested. For example, in the IID loss model, at 0˚, and PLR=0.01, 16.5% of the frames were included, while for the burst loss
model 20.416%. The results at 0˚ for PLR=0.02 were: IID loss model 12.87% and burst 18.88%, while for PLR=0.04, IID
was 10.583% and burst 13.083%. These results show that most of the degradation resulting from compression combined
with packet loss cannot be measured by MTF but by other quality measures. From the simulations it appears that the
number of frames participating in the average MTF derivation decreases as the packet loss probability increases, and it is
always higher for the burst loss model. Similar results were obtained also for the other two angles of 45˚ and 90˚. As the
PLR decreases, the MTF measure represents more degraded frames, and the obtained average MTF is more reliable. In
general it would seen that the MTF is a more accurate measure for low packet loss probability.
By comparing the MTF acquired for the burst model to that acquired from the IID model, it can be observed that
smoother curves are obtained in the burst model. The reason is that more frames participate on deriving the average MTF in
the burst model than in the IID model. For all the simulations, the reduction as function of the spatial frequency of the
modulation contrast is not significant. The maximum reduction with the MTF is 35% at the highest normalized spatial
frequency (0.5 cycles/pixel) for 90˚, for burst loss model with PLR=0.04. In general, all the MTF curves behave similarly
and it is quite difficult to distinguish between the burst model and the IID model.
By comparing the MTF at the three angles, the greatest degradation occurred at the horizontal (0˚) direction. This
can be observed from the low number of frames participating in the average MTF. This can also be observed by viewing the
video streams with the MPEG decoder. The video quality in the horizontal direction suffers from degradation effects which
are more severe than smearing, such as block misplacing, block missing, and temporal propagation of errors. The
explanation for this result is that the number of pixels in the 0˚ direction is 80 compared to the other two directions, where
the number of pixels is 60. Thus, the probability of packet loss in the horizontal direction is higher.
For an additional comparison between the two loss models, we use the PSNR measurement calculated according to
Eq.(1). Here, the PSNR quantifies only the effects of the lossy network on the video streams. In Figs 11 and 12, the PSNR
histograms are presented for PLR equal to 0.01 and 0.04, respectively. Each figure contains two graphs: (a) the burst loss
model, and (b) the IID model. The histograms represent the total number of frames that have PSNR values in a certain range
(bin). The isolated right bin represents the number of undistorted received frames, which have theoretically infinite PSNR
(zero MSE) compared to the respectively transmitted frames.
(a)
(b)
(b)
(a) 11 PSNR histograms for PLR=0.01. (a) Burst loss model (b) IID loss model.
Fig.
Fig. 12 PSNR histograms for PLR 0.04. (a) Burst loss model (b) IID loss model
Table 1 summaries the results of the average PSNR, the variance PSNR, and the number of identical frames.
Table 1 : Summarized statistics of the PSNR
Packet Loss Probability :
0.01
0.02
0.04
81.0545
67.2887
68.6210
IID: average
1.2538e+03
400.7862
282.8412
Variance
90
24
19
Number of identical frames
128.1374
103.8463
92.1583
Burst: average
3.5010e+03
2.9371e+03
2.2896e+03
Variance
480
243
173
Number of identical frames
Fig. 12 PSNR histograms for PLR 0.01. (a) Burst loss model (b) IID loss model.
These results show that the average PSNR calculated from a video stream degraded by burst losses is much higher
than one degraded by IID losses. This can be explained by the fact that a larger part in a single frame is destroyed by burst,
unlike IID, where many small parts of different frames are degraded. Note that the packets after an important header is lost
are useless. Therefore, in the case of the burst losses where packets are lost in groups, packets that may be lost after header
packet will not actually effect the video stream. On the other hand, for the IID model, losses are randomly distributed.
Therefore each packet has the potential to affect the received video stream. It is evident from the results of Table 1 and Figs
11 and 12 that the number of undistorted frames is much larger in the burst model than in the IID model. The difference is
very noticeable, despite the use of the same loss-probability in both cases. As expected, as the PLR increases the average
PSNR decreases, and the number of undistorted frames decreases.
5. CONCLUSIONS
In this work we evaluated the average MTF obtained from a compressed image transmitted through a lossy
computer network. First, we evaluated the distortion of JPEG compressed images in terms of the MTF. We showed that the
average MTF represents the reduction with contrast as a function of spatial frequency, which is not the main cause for the
image distortion. This was an expected result, since nonlinear and space variant transformations cannot be precisely
evaluated by using frequency response measures. In the second part of the paper we analyzed the influence of packet losses
and video compression on the quality of the received video stream. We demonstrated that the burst loss model has a less
severe impact on the video stream as compared to the IID loss model. We found that the average MTF is a more accurate
measure for low packet loss ratios. As the number of lost packets increases there are additional distortions that cannot be
evaluated by the MTF, indicating more measures have to be developed. As we mentioned, since packet loss and
compression are not linear space and time invariant transformations, spatial frequency response measures can only partially
represent the distortion process.
6. ACKNOWLEDGMENTS
We would like to thank Doron Fuchs, and Hanan Dagan for their help in coding the simulation software, and to the
Computer network laboratory Faculty of Electrical Engineering in the Technion.
7. REFERENCES
1.
2.
3.
4.
5.
6.
O. Verscheure, X. Garcia, G. Karlsson and J.P. Hubaux, “User-Oriented QoS in Packet Video Delivery”, in
Proc. IEEE network, November/December 1998.
Y. Wang and Q.F. Zhu, “Error Control and Concealment for Video Communication: A Review”, in Proc.
IEEE, vol. 86, No. 5, May 1998.
N. S. Kopeika, A System Engineering Approach to Imaging, chap. 9, pp. 517-524, SPIE Optical Engineering
Press, Bellingham WA, 1998.
H. C. Holst, Electro-Optical Imaging System Performance, SPIE optical engineering press, Washington, 1995.
K. W. Goodman, Statistical Optics, John Wiley, New York, 1985.
P.G. Cox and R.G. Driggers, “Information Transfer Function for Image Compression Algorithm”, in Proc. of
SPIE, vol. 3063, pp. 23-30, 1998.
O. Hadar, and A. Stern, “Image compression improvement by pre-filtering”, in Proc. of SPIE, Vol. 3460, pp.
895-905, Applications of Digital Image Processing XXI, A.G. Tescher; Ed. San-Diego, CA, USA, 1998.
8.
J.C. Russ, The Image Processing Handbook, 2nd ed. IEEE Press, Boca Raton, Florida, 1994.
9.
http://cj.msu.edu/~jones/research/clemen41/psnr.html
10.
R. Shiavi, Introduction to Applied Statistical Signal Analysis, Irwin and Asken Associates, Inc., Boston , 1991.
11.
O. Hadar, A. Mendelblat, R. Sabat, N. S. Kopeika and S. R. Rotman, "Influence of sensor motion on target
acquisition," Infrared Physics and Technology, vol. 38, pp.373-382, June 1997.
12.
S. V. Raghavan, and S. K. Tripathi, Networked Multimedia Systems: Concepts, Architecture, and design.
Prentice Hall, 1998.
7.
1
2
Y. Wang and Q.F. Zhu - Error Control and Concealment for Video Communication: A Review.
- proceedings of the IEEE, vol. 86, No. 5, May 1998.
3N. S. Kopeika, A System Engineering Approach to Imaging, chap. 9, pp. 517-524, SPIE Optical
Engineering Press, Bellingham WA,1998.
Holst H. C., Electro-optical imaging system performance, SPIE optical engineering press,
Washington, (1995).
4
Goodman K. W. (1985) Statistical optics. John Wiley, New York.
5
P.G. Cox and R.G. Driggers - Information Transfer Function for Image Compression Algorithm. Proc.
6
of SPIE vol. 3063, p. 23-30.
O. Hadar, and A. Stern, “Image compression improvement by pre-filtering,” in
7
Proc. SPIE Vol. 3460, p. 895-905, Applications of Digital Image Processing XXI, Andrew G. Tescher;
Ed,
8Russ J. C. (1994) The image processing handbook. 2nd ed. IEEE Press, Boca Raton, Florida.
http://cj.msu.edu/~jones/research/clemen41/psnr.html
9
10. R. Shiavi, Introduction to Applied Statistical Signal Analysis, Irwin and Asken Associates, Inc. ,Boston, 1991.
11. ‫המאמר של עופר ב‬infrared phisics
S. V. Raghavan, and S. K. Tripathi, Networked Multimedia Systems : Concepts, Architecture, and
design. Prentice Hall, 1998.
12
Download