COMPUTER VISION, GRAPHICS, AND IMAGE PROCESSING &236-251 (1984) Image Enhancement Using the Median and the Interquartile Distance IRWIN SCOLLAR AND BERND WEIDNER Image Processing Laboratory Rheinisches Landesmuwum, Bonn, West Germany AND THOMAS S. HUANG Coordinated Science Laboratory, University of Illinois, Urbana-Champaign, Illinois 61801* Received October 30,1982; revised December 3,1982 Enhancement and noise cleaning algorithms based on local medians and interquartile distances are more effective than those using local means and standard deviations for the removal of spikelike noise, preserve edge sharpness better, and introduce fewer artifacts around high contrast edges present in the original data. They are usually not as fast as the mean-standard deviation equivalents, but all are suitable for large data sets to be treated in small machines in production quantities. 1. INTRODUCTION Many effective image enhancement algorithms make use of local statistics in a window of moderate size [l}. Such techniques are useful for noise cleaning, crispening, and adaptive contrast and brightness correction. In most cases the local statistics used are the mean and standard deviation of the pixels in a square or rectangular window. It is the goal of the present paper to show that, for certain classes of images, better results can be obtained if we replace the local mean and standard deviation by the local median and interquartile distance, respectively. Specifically, this replacement is desirable if the image contains high contrast edges or noise with heavy tailed probability density functions (pdf), especially spike- or pointlike noise, or its two dimensional analog which we will call thread noise because the noise looks like fuzz or threads in the image. It will be seen that the use of the median and the interquartile distance is more effective in removing such noise, in preserving edge sharpness, and in avoiding the introduction of artifacts on high contrast edges (light or dark banding). The images for which our enhancement algorithms were developed have inherently high resolution and large numbers of pixels. They are oblique aerial photographs of rural areas which show the outlines of buried archaeological sites as discolorations in growing crops or soils; metallographic X-ray pictures of corroded objects from archaeological excavations prior to restoration, infrared and ultraviolet images of paintings and other art objects; and thermal infrared scans made from the air to map buried ancient field systems. The users of such imagery demand hard copy of the highest quality without visible scan lines and with full photographic gray scale. To achieve this, most pictures are scanned with between two and four thousand lines at equivalent pixel resolution, resulting in data which are typically 4 *On leave at the Image Processing Laboratory, Rheinkches Landesmuseum, Bonn, under the auspices of the A. von Humboldt Stiftung, BOM, West Germany. 236 0734189X/84 $3.00 c~pyri%t 0 1984 by Academic Press, Inc. All rights Of reproduction in any form reserved. ENHANCEMENT 237 USING THE MEDIAN to 16 Mbytes per image. Production quantities must be processed at our installation. In such images, high spatial frequency but low contrast information is of great interest, but high contrast edges of objects of no interest are widely present. These edges may be modem field boundaries, roads, or structures, while the objects of interest have low contrast and are near the edges. Most of our pictures were obtained under unique conditions in the past and they cannot be repeated. Defects in such images are typically due to dirt and stratches on negatives because of improper storage or lack of care during processing, under- or overexposure in spatially varying amounts, very low or very high brightness with the information of interest restricted to a small part of the range, positionally dependent. Most pictures are on black and white aerial or X-ray film, some are direct recordings on magnetic tape from nonfilm sensors, some on 35mm film of all possible types. We have not processed the color pictures directly because we lack suitable hardware, but black and white color separation negatives from color slides have been treated. Apart from the above, noise in our images is also due to film grain or scanner photomultiplier noise, and these are usually Gaussian. Most types are usually present in an image at once. Signals are usually fine linelike structures with complex geometries or irregularly shaped blobs, all identifiable as man-made when seen by a trained observer. The aim of enhancement is to help him make the appropriate pattern recognition decisions by reinforcing the ability of the human eye to detect faint geometric structures in noisy backgrounds. Because of the very high resolution demanded and given the quantities of pictures to be processed, algorithms must be fast enough to keep up with picture input-output. Given the financial limitations, a large fast machine may not be used. The algorithms discussed in this paper possess the required properties and have been used successfully for a number of years. 2. PROPERTIES OF THE MEDIAN AND THE INTERQUARTILE DISTANCE Let us first establish our notation. f(i, i) = brightness at point (i, j) in the image. Local mean (window size (2k + 1) x (2k + 1)): Local standard deviation: l/2 a(i, j) = 1 . (2k + 1)’ Let pij(f) be the sample cumulative distribution {f(i function of - u, j - u)lu = -k ,..., 0 ,..., k; u = -k ,..., 0 ,..., k}. Local median: M(i, j) = pFc, such that Pii = l/2 238 SCOLLAR, Local interquartile WEIDNER, AND HUANG distance: Q(i, j) = Y - A where Pii = 3/4 and Pii = l/4. Note that if p is not unique, we take M(i, j) as the midpoint of the range of ~1;similarly for h and V. Note also that for the sake of simplicity we have used a square window, but generally the window may be of any shape: rectangular, cross, ring, etc. It is well known that the local median has the following properties [2]: (1) It tends to preserve sharp edges. In particular, the edges in an image are much better preservedin M(i, j) than in m(i, j) (seeSection8). (2) It tends to eliminate spike noise. More generally, the situation is as follows: let us consider the simple case where the signal part of the image is constant and zero, and the noise is zero mean, independent from sample to sample, and identically distributed with some probability density function. Then E[+, j)] =f(i, j) = E[M(kAl and the variances ui and I& can be used as measures of the noise cleaning power of the running mean and median filters. The smaller the remaining variance is, the more the noise filter has removed. It is well known that for Gaussian noise for a window of fixed size, of course, while for Laplacian noise (double-sided exponential) it has been observed empirically that uz > &. Generally, it has been observed empirically that when the tails of the probability density function of the noise become heavier, the running median filter becomes more effective relative to the running mean filter. A particular example is the case when the probability density function of the noise is of the form where A, a, and (Y are positive constants; then it can be readily shown that as (Y decreases (i.e., the tails of the pdf become heavier), (uz - I&) increases. A similar comparison exists between the variabilities of the local standard deviation and the local quartile distances. As the tails of the noise pdf become heavier, (u,’ - yu4) increases (where y is a constant inserted to make E[dyQ] equal to the standard deviation of the noise pdf. As shown in Section 8 of this paper, the interquartile distance is used as a good approximation to the absolute deviation from the mean, a measure of dispersion which is the order statistics analog of the standard deviation. The interquartile distance is much easier to compute than the absolute deviation, and for the data in question they seldom differ by more than a few percent, judging from empirical observation. A more detailed analysis of the advantages of order-based measures with regard to spike noise is given in Section 8. ENHANCEMENT 3. COMPUTATIONAL USING THE MEDIAN 239 REQUIREMENTS Assume that we are calculating m(i, j), a(& j), M(i, j), Q(i, j) for the entire image. Let the window size be A x B. It is well known that the local mean m(i, j) can be calculated using four additions per output point in a recursive manner independently of the window size [3, 41. One takes advantage of the separability of the two dimensional mean and does the calculation in one direction at a time; this is done recursively which requires only two additions per output point. The local standard deviation a(& j) can be calculated similarly. The recursive operation also needs four additions per output point, but we also need to subtract m(i, j) from each input point and square the result. After the recursion we need to take the square root, so that the total number of operations required per output point are five additions/subtractions, one multiplication for squaring, and one square root, independently of window size. To calculate the local median using the Huang-Yang-Tang histogram update algorithm [5] requires approximately 2A comparisons and 2A additions per output point. If we calculate the interquartile distance Q(i, j) at the same time we need 4A additional comparisons. If we calculate the separable median [6], we must compute it in one dimension horizontally, then vertically on the result. The computational requirement is then nearly the same as for the mean: four additions and four comparisons per output point independent of window size. However, the separable median is similar but not identical to the two dimensional median. For the same window size, the separable median removes less noise but preserves sharp comers better. Storage requirements are much greater, as we shall see in Section 4. Algorithms based on means and standard deviations are the fastest known for local enhancement and filtering, sometimes called box filtering [4]. Median methods may soon equal these since several companies are constructing VLSI chips for video-rate two dimensional median filtering. The initial work involves only small 3 x 3 or 5 x 5 windows, and interquartile distances have not been mentioned. When completed, with larger windows and the interquartile distances, this effort will bring down throughput times to video rates. A multipass algorithm for local enhancement using the median in place of the mean and the median absolute deviation (MAD) as a dispersion measure is suggested in [7]. To calculate the MAD it is necessary to produce a median filtered image, subtract it from the original, and then median filter the absolute differences. This requires three passes through the data plus two calculations of a median per point. A reduction in calculation can be obtained by using only the absolute deviation from the median as the dispersion measure instead of the MAD. Since the algorithm of [5] lends itself to simultaneous calculation of the interquartile distance as well as the median in one pass through the data, there seems to be no reason to use a multipass method in a nonhardware implementation. For very high resolution images (greater than 1500 lines, 1500 pixels per line) where the eye cannot see the scan structure at normal viewing distances, algorithms based on these techniques are the only practical ones for serial computers of modest speed and storage capability. The results cannot usually be distinguished by laymen from those obtained by more complex algorithms which can be implemented only on large machines when throughput time is important. Mean-standard deviation methods can usually be used if more than 400 pixels are available for the moving window. For 400 pixels or more defined to 256 levels, the median approaches the mean very 240 SCOLLAR, WEIDNER, AND HUANG closely for nearly all typical pdf’s encountered except at very sharp broad edges. Hence it is seldom of interest to use median based methods for windows larger than 19 X 19 unless the window is sampled. We have -obtained much faster computing times with larger windows, rather than the reverse, once the window contains more than 400 values, by sampling at suitable intervals. 4. STORAGE REQUIREMENTS In the case of the separable median, the computer must have enough directly addressable memory to store the entire input image or a very large part of it. In the normal two dimensional median technique using the Huang-Yang-Tang algorithm, a portion of the image in the vertical direction equal in size to the window height plus one additional line must be stored in a circular buffer at least. In the separable mean-standard deviation approach, only the upper, center, and lower picture lines of the strip in which the window moves need be stored, together with two accumulator arrays, one for the column sums of the moving window and one for the sums of squares. The latter usually needs twice the word length of the first. In all cases only one output line need be stored. Use of asynchronous input/output techniques which in our installation materially increase the speed of throughput require additional storage capabilities for the input and output data lines. Given the modest storage and word lengths of many micros and lower range minis, there is usually no way to implement the separable median for high resolution images directly. It can be done in two passes from the disks but the extra I/O time usually wipes out any advantage. Only the property of comer preservation may be a reason for its occasional use. We have programmed all algorithms described in this paper to work with at least 2000 line/pixel images and in some cases with 4000 line/pixel images in only 64K bytes of storage, including buffers for asynchronous input-output. 5. ALGORITHMS FOR NOISE CLEANING Let f( i, j) be the input image and g( i, j) the output image. We first list a number of simple algorithms for noise cleaning using local mean and standard deviation. Examples are shown in Fig. 1. Algorithm Nl, simple averaging: g(i, i) = m(i, i). Algorithm N2, averagingwith g(i, i) = 44 fixed threshold [8, 91: i) = f (i, i) if If(i, j) - m(i, j)l > t, if lf(i, j) - m(i, j)l 5 t where t is a threshold supplied interactively. Algorithm N3, averaging with variable threshold depending on contrast: g(i, j) = m(i, j) =f(i, j) where B is a constant. if If(i, j) - m(i, j)l > Mi, j>, if If (k j) - m(i, Al 5 Ba(i, i) ENHANCEMENT USING THE MEDIAN FIG. 1. Noise cleaning using local averaging (algorithms Nl-3); threshold lo%, N3: B = 1.5. 241 2000 line images, 7 x 7 window. N2: Algorithm Nl is used to smooth random Gaussian noise. It blurs edges. Algorithm N2 can be used with a window much larger than the noise features to smooth spike or thread noise. If t is well chosen, it may not blur edges as much as Nl. Algorithm N3 is a sign&ant improvement on N2 since the usual outlier criterion of 95% confidence limits in elementary statistics is satisfied when /I = 2. In N2, if t is chosen too large, then noise cleaning will not be good; if too small, edges will be blurred. N3 adapts to local contrast in the data, and sharp edges will hardly be blurred at all, providing that the window is large enough. More complicated algorithms have been proposed which can preserve edge sharpness, but they do not have the computational advantage of Nl, N2, or N3 [lo]. We propose here an alternative approach to high contrast edge preservation with small windows in the presence of spike or thread noise. In algorithms Nl-3 replace the local mean m(i, j) by the local median M(i, j) and the local standard deviation 242 SCOLLAR, FIG. 2. WEIDNER, AND HUANG Noise cleaning using local median (algorithms N’l-3). Parameters as in Fig. 1. u by the local interquartile distance Q(i, j) as an approximation to the local absolute difference. We call the three new algorithms Nl’, N2’, and N3’. Results using this substitution are shown in Fig. 2 for the data of Fig. 1. Algorithm N3’ is superior to algorithm N3 in removing spike noise and preserving edge sharpness since smaller windows can be used, the median and quartile differences being less alkted than the mean and standard deviation. Empirically we have observed that two- or threefold size reductions are possible. Algorithm Nl’ is superior to Nl in preserving edge sharpness, but the relative abilities of Nl and Nl’ depend on the noise pdf. Both Nl’ and N3’ will remove spike noise, but edge sharpness is better preserved by N3’, while Nl’ will remove more Gaussian noise. If large windows are used, comers will be better preserved with N3 than with N3’ using the two dimensional median. Performance is roughly the same using the separable median. For large windows N3 is the preferred algorithm; for small windows N3’. Artifacts with large windowed N3 runs may sometimes be objectionable on very large edges. ENHANCEMENT USING 6. ALGORITHMS THE FOR MEDIAN 243 CRISPENING By crispening we mean an apparent sharpening of the image without change of local brightness or contrast [ll]. A very effective algorithm which we have used most extensively because of its high speed is a variant on box filtering with roll-off control and brightness normalization. Algorithm C g(i, j) = w{ f(i, j) - am(i, j)) where (Y is a constant, 0 < a < 1, and w is a brightness normalization constant which depends on the window size and the value of (Y. The frequency response of this linear spatially invariant filter depends on the constant (Yand the window size used in computing the local mean m(i, j). The larger the value of (Ythe greater the crispening effect. The larger the window size, the more the effect is obtained over broader edges. However, for very high contrast edges already present in the picture, there is some overshoot which results in light or dark banding which may be objectionable when large window sizes are used. This is shown in Fig. 3. A large window size should be used to obtain stable means in noisy areas, but then the effect is most noticeable. To overcome this we recommend replacing the local mean by the local median to get our algorithm C’. For the same value of a and the same window size, algorithm C’ will crispen the image slightly less but this is hardly observable (Fig. 3). It is possible to combine crispening and noise cleaning in one pass. Algorithm C’ can be combined with N3’, C with N3, or mixtures of the two types used. For example, see the following. Algorithm C’N3’ g(i, i) = M(i, j) = w{f(i, j) - aM(i,i)} if V(i, j) - M(i, j)l > PQ(i,j), iflf(i, j) - M(i,j)l < PQ(i,i). It has been observed that crispening followed by noise cleaning in a two-pass algorithm is superior to the above or to the reverse order, because spike noise is first exaggerated by C’ and moved to the outer tail of the pdf where it can be removed more easily by N3’. It is useful to use C instead of C’ in the first pass to save time. Because of the variant data structures required, it is not useful to program this combination for a single pass. In general, it is desirable to use different window sizes for the crispening and cleaning parts of the treatment. Usually a smaller window for the crispening is desirable. This increases the storage requirements for CN3 since extra image lines have to be buffered and multiple accumtdators are needed. This is not the case for C’N3’ where a large circular buffer for the big window is available anyway. Despite the slower computational speed, the I/O saving is considerable. Multiple window versions of these algorithms can be combined to give bandpass and Gaussian filtering and are computationally effective, especially with the possible variants of 244 SCOLLAR, WEIDNER, AND HUANG FIG. 3. Crispening using local averaging and local median; 2000 line images, 7 X 7 window. algorithm C. There is some advantage, as pointed out in [12], in the sequential combination CC’ with values of a! equal to 1 and a larger window size for C than for C’ as a noise insensitive edge extractor. We also wish to note that a simple modification of C or C’ can be used for reduction of linear motion blur if it is uniform over the image and more or less at right angles or parallel to the scan direction. It is merely necessary to displace the central pixel in the window by a small amount depending on the direction of blur. Usually it is advantageous to use a rectangular window with the long axis parallel to .the blur direction. Blur which is not -parallel to the scan direction ought to be reducible with displacement at some angle from the true center. Large amounts of blur cannot be dealt with by this technique. Blur which is not uniform over the image might be dealt with by varying the amount and direction of offset over the image, but I we have not tried this. ..Here there will be a complex problem in determining the needed displacement and varying window sizes. ENHANCEMENT 7. ALGORITHMS FOR LOCAL USING THE MEDIAN CONTRAST 245 AND BRIGHTNESS ENHANCEMENT These algorithms are needed for exposure and development correction, especially in spatially variant cases which cannot be treated by simple histogram transformation of the entire image. One approach to enhance local contrast is to normalize the local standard deviation to some constant, a kind of automatic gain control [13]. Another possibility is raising the local standard deviation by multiplying the mean-pixel difference by a constant and replacing part of the output pixel [14]. The most satisfactory scheme which we have used judging by user acceptance has been that due to Wallis [15], as follows. Algorithm W g(i, j) = Au(if; + ud { f(i, j) - m(it d} + amd +tl - a>m(i? j) where md = a desired mean, ed = a desired standard deviation, A is a gain or contrast expansion constant (usually 4 < A < 20), and a is a brightness forcing constant, usually 0 < a < 0.3. This algorithm can be improved by limiting either the maximum observed standard deviation a(i, j) or the value of the pixels which contribute to it as suggested in [l]. The first is computationally preferable. Performance on high contrast edges is helped by this modification, reducing the edge artifacts somewhat; see Fig. 4. Experimentally, it has been found that good results which do not add noise to the image require windows of the order of 5 to 10% of the side length. Noise effects which have been noted [l] are due to the use of windows which are too small, giving consequent instability in the standard deviation in the presence of heavy-tailed pdf’s. However, if the image contains high contrast edges, then a large window will introduce dark or light bands along the edges; see Fig. 4. This artifact can be considerably reduced (Fig. 4) if we replace m and u by M and Q giving algorithm W ‘. The local contrast and brightness correction of W ’ are indistinguishable from W for the same window sire, but since smaller windows can be used, banding effects are less. Noise effects are the same since such large windows must be used for W. Very small windows produce a bit more noise in W than in W’, but on very high resolution images it is sometimes difficult to see the difference except in highly uniform gray areas. We have obtained a considerable speedup of W by using a regularly sampled window for those larger than 20 x 20 pixels. Presumably, this approach could also be applied to W’ but it does not appear to be worthwhile, since there is little difference in performance for the two algorithms in the noise sense when windows are so large. Only the difference in edge banding justifies the extra computational cost of W’. 8. ERROR NORMS AND ENHANCEMENT OPERATORS The analogies between means, standard deviations, medians, and quartile differences used throughout this paper are difficult to understand with standard filtering theory. We propose that they can best be understood in terms of error norms and their associated dispersion measures. For this we must introduce a few theoretical concepts. 246 SCOLLAR, WEIDNER, AND HUANG FIG. 4. Local contrast and brightness enhancement (algorithms W and W’) W window 50 x 50, av. = 127 (on 255 scale), (I = 50, a = 0.1, A = 20; W’ window 11 X 11. Same parameters for median and quartile difference as in W. A set is called a metric space if, for every pair of elements, a real distance is defined with satisfies the following conditions: 4x9 Y) > 09 while d(x, y) = 0 ifandonlyifx =y; d(x, y) = &Y, x); d(x, z) < d(x9 Y) + d(Y, z) An important triangle inequality. class of distance functions is generated by norms. A norm 1111on a ENMCEmT 247 USING THE MEDIAN linear space is a real function which satisfies the following axioms: PII h 03 II4 = lal while llxll = 0 ifandonlyifx . llxll for every number a = 0 lb + YII 5 Ilxll + IIYII 4x9 Y) = lb - YIl* In particular the L, or Minkowsky L,(a, norms are defined as b) = ( jblx(t)l’dr)? a Now, for our statistical measures, let p be a pdf defined on the interval I:= [-&in, -&ax ] and P be the corresponding cumulative distribution function or probability function. The expected value of a function f is then detined as E(f):= if dP. Let the dispersion of p around y E I with regard to the L, norm be d,(y):= {ax lsr<ao. - Yl’)I”’ The value d,:= $nd,( y) is called the dispersion of p with regard to the L, norm or simply the r dispersion of p. If there is a unique c, E I which minimizes d,(y) such that it will be called the r center of p. Let us now consider a few special cases: r = 2 (the Euclidean Norm); E(x - y)’ = min, * 9 = E(x), the mean value of p and d, = (E((x - E(x))~))~‘~, the standard deviation of p. This norm underlies all classical statistics and linear transform theory. r + co (Max Norm); in this case d,(y) converges to d,(y) = ~e~lx - YI 248 SCOLLAR, WEIDNER, AND HUANG independentlyof p suchthat the half range,and r = 1 (Manhattanor City Block Norm); E( lx - ~1) = min * /x~h~(y-x)dP+~x=-~(x-y)dP=o* [,dP=[-dP=i IlUll since@‘=~. The valueci with is just the medianand d, = E( Ix - ci I) is the absolutedeviationfrom the median. Becausethesemeasuresare dehnedin terms of the L, norm, attemptsto study their propertiesusingclassicaltransformtheory which dependson the L, norm will fail. For d, anotherexpressioncan be derived: d,= j”(q&in x)dP+[;-(x-q)dP= jX-xdPf,xdP. Cl OM Now let pl= = 2P9 x 2 Cl 0, x > md Cl pr = 0, = 2P, x 5 Cl, x > Cl be the left and right half distributions of p, respectively;then dl = l/2{ Epr(x) i.e., the differenceof the mean valuesof the two half distributions,which can be written, usingthe notation c,(p) for the r centerof p, E,,(x)}, 4 = :(cz(~r) - c,(~l))- It now seemsnatural to define,in an analogousmanner, Q:= t(c,< pr) - c~( Al)) with ci( pr) and ci( pl) as the mediansof the half-distributionshavingthe obvious properties s(p”dp I c$dJ’ = cp 5 jxm” dp = 4 Jxmim cl(Pr) 1 1 and are calledthe quartilesof p. ENHANCEMENT USING THE 249 MEDIAN 2Q is then called the interquartile distance. Let q1 = cr( pl), q2 = ct( p), q3 = cl( pr). To see what effect spikes in the data have, we may interpret p as a frequency distribution of samples and consider what happens if a small fraction h of samples is replaced by new samples with a fixed value z. Let p’ be the distribution of samples replaced with /dP’ = h, the fraction replaced. We also assume that the samples replaced are taken randomly, such that, approximately, and that z B cl. p’ = hp We then have c; = (1 - h)c, + hz. For the median c1 1 -= 2 1'; d(P - p') = /" d(P - P') + /“d(P Xmin = 9 XtN” +(I - h)l’;dP - P’) Cl = 9 +(l - h)P(c, 5 x 5 c;) (‘1 =y+(l-h)(P(x+;) -P(X~c;)=~~. 1 1 Analogous relations hold for the quartiles, so that for i = 1,2,3, P(X 5 q() = $A = $1 + h) for small h and Q’ = $( P-‘(a(1 + h)) - P-‘(a(1 + h)). All these relations are independent of z and give corrections only on the order of h, whereas the change for c2 is of the order hz. For spike, thread, and comparable “singularity” noise in pictures h, the fraction of spikes seen in a window is small ( < 0.1) whereas lz - ci 1 is usually quite big. In that case, as has been shown, the median and interquartile distance give much more noise tolerant estimates of local brightness and contrast than do the mean and standard deviation. IDefIning an edge as two adjacent areas with Gaussian distributions having different means and perhaps different standard deviations, let us examine what happens in a window when it passes over this edge, starting well out in the first area. When the window just touches the edge, the pixels belonging to the second distribution appear as singularities in the sense defined above (spikes or threads). 250 SCOLLAR, WEIDNER, AND HUANG Therefore the value of the median does not change immediately. It does change when the number of new pixels from the second distribution becomes a significant fraction of the total number in the window. This is why the normal type of median filtering blurs the edge less than does filtering with the mean. Given the advantages of the L, norm based measures, one would like to think that a theory of filtering could be built up using this norm by analogy with linear Wiener filtering which is based on an L, norm. L, norm filtering is based on the fact that the L, norm is derivable from a scalar product such that the notion of orthogonality is well defined. Unfortunately this is not the case with the L, norm. Classical filtering (Fourier, Hadamard, etc.) is efficient because of the analytical tractability of the L, norm approximations or estimates of a function. L, norm approximations to a function can only be obtained with linear programming techniques. This is not efficient, although filtering approaches using them are possible. As far as we know, using the median, quartile difference approach is the only efficient way to take advantage of the properties of the L, norm without using linear programming or something still worse. 9. CONCLUSIONS We have shown that, for certain types of images, the use of the local median and the interquartile distance is more effective than the use of the local mean and standard deviation at the expense of computation time unless the complete image can be stored. In this case separable medians and quartile differences restore the time advantage or even gain something since no squares or square roots are needed. Specifically, algorithms using M and Q have the following general advantages: (1) They are more effective in removing spike noise or noise with a heavy tailed pdf. (2) They preserve edge sharpness better. (3) They introduce fewer artifacts around high contrast edges already present in the original data. On the negative side they are generally slower to compute in a small general purpose machine. When VLSI chips for the median are available, video-rate operations will be possible. Combined algorithms with one pass through the data are presently a worthwhile improvement in serial machines, but the order of operations is important and different effects can be achieved with multipass algorithms depending on what is done first. The suite of algorithms is especially suitable for production quantities of high resolution images without excessive amounts of noise in an application environment, without the use of high-cost machines. REFERENCES 1. J. S. Lee, Digital image enhancement and noise filtering by use of local statistics, IEEE Truns. Pattern Anal. Much. Intell. PAMI-2, 1980. 165-168; Refined filtering of image noise using local statistics, Computer Graphics Image Processing 15, 1981. 380-389. 2. T. S. Huang. ed.. Two-Dimensional Digitul Signal Processing II: Transforms and Mediun Filters, Chaps. 1, 5. 6. Springer-Verlag, New York/Heidelberg. 3. J. M. Soha et al., Computer techniques for geological applications, Proceedings, Caltech/JPL Conference on Image Processing, Nov. 3-5, 1976, JPL SP 43-30, Sect. 4-1. 4. M. J. McDonnell, Box filtering techniques, Computer Graphics Image Processing 17, 1981. 65-70. ENHANCEMENT 5. USING THE MEDIAN 251 T. S. Huang, G. J. Yang, and G. Y. Tang, A fast two-dimensional median filtering algorithm, IEEE Trans. Acoust. Speech Signal Process. ASSP-21, 1979, 13-18. 6. P. M. Narendra, A separable median filter for image noise smoothing. Proceedings, IEEE Conference on Pattern Recognition and Image Processing, 1978, pp. 137-141. 7. R. T. Gray, D. C. McCaughy, and B. R. Hunt, Median masking technique for the enhancement of digital images. Proc. SPIE Annu. Tech. Symp. 24l7, 1979,142-145. 8. R. E. Graham, Snow removal-A noise-stripping process for picture signals, IRE Trans. If. Theov IT-& 1966,129-144. 9. J. M. S. Prewitt, Object enhancement and extraction, in Picture Processing and Psychopictorics (B. S. Lipkin and A. Rosenfeld, Eds.), Academic Press, New York, 1970. 10. M. Nagao and T. Matsuyama, Edge preserving smoothing, Computer Graphics Image Processing 9, 1979, 394-407. 11. W. K. Pratt, Digital Image Processing, pp. 322-323, Wiley-Interscience, New York, 1978. 12. B. J. Schachter and A. Rosenfeld, Some new methods of detecting step edges in digital pictures, Commun. Assoc. Comput. Mach. 21, 1978, 172-176. 13. A. H. Watkins, EROS Data Center image processing technology, Proceedings, Caltech/JPL Conference on Image Processing, Nov. 3-5, 1976, Sect. 16-3. 14. J. S. Lee, Digital image processing by use of local statistics, Proceedings, IEEE Conference on Pattern Recognition and Image Processing, 1978, p. 56, Eq. (4). 15. R. Wallis, An approach to the space-variant restoration and enhancement of images, in Image Science Mathematics (C. 0. Wilde and E. Barrett, Eds.) (Proceedings, Symposium on Current Mathematical Problems in Image Science, Monterey, Cal., Nov. 1976) Western Periodicals, North Hollywood, 1977, pp. 10-12.