ieee transactions on signal processing

advertisement
IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 50, NO. 7, JULY 2002
1583
Adaptive Null-Forming Scheme in
Digital Hearing Aids
Fa-Long Luo, Senior Member, IEEE, Jun Yang, Senior Member, IEEE, Chaslav Pavlovic, and
Arye Nehorai, Fellow, IEEE
Abstract—We propose an effective adaptive null-forming
scheme for two nearby microphones in endfire orientation that
are used in digital hearing aids and in many other hearing devices.
This adaptive null-forming scheme is mainly based on an adaptive
combination of two fixed polar patterns that act to make the null
of the combined polar pattern of the system output always be
toward the direction of the noise. The adaptive combination of
these two fixed polar patterns is accomplished by simply updating
an adaptive gain following the output of the first polar pattern
unit. The value of this gain is updated by minimizing the power
of the system output, and related adaptive algorithms to update
this gain are also given in this paper. We have implemented this
proposed system on the basis of a programmable DSP chip and
performed various tests. Theoretical analyses and testing results
demonstrated the effectiveness of the proposed system and the
accuracy of its implementation.
Fig. 1. Common direction processing system with two omnidirectional
microphones in endfire orientation.
Index Terms—Adaptive signal processing, array signal processing, beamforming, hearing aids, microphones, noise reduction,
speech enhancement.
I. INTRODUCTION
H
EARING devices with directionality make use of the direction difference between the target signal and the noise.
There are two types of directional devices: one with fixed directionality (or, say, with the fixed polar pattern) [1]–[4] and
the other with adaptive directionality that can track the varying
or moving noise sources [5]–[8]. Many techniques have been
used to achieve directionality in both fixed mode and adaptive
mode [3], [7], [9], [10]. However, most of these techniques can
be neither immediately employed nor implemented in instruments such as hearing aids because of the limit of hardware
size, the number and distance of microphones, computational
speed, mismatch of microphones, power supply, and other practical factors relating to these hearing instruments [7], [11], [12].
Manuscript received September 18, 2000; revised March 13, 2002. The work
of A. Nehorai was supported by the Air Force Office of Scientific Research
under Grants F49620-99-1-0067 and F49620-00-1-0083, the National Science
Foundation under Grant CCR-0105334, and the Office of Naval Research under
Grant N00014-01-1-0681. The associate editor coordinating the review of this
paper and approving it for publication was Dr. Brian Sadler.
F.-L. Luo was with the R&D Department, GN ReSound Corporation, Redwood City, CA 94063 USA. He is now with Quicksilver Technology, Inc., San
Jose, CA 95119 USA.
J. Yang was with the R&D Department, GN ReSound Corporation, Redwood
City, CA 94063 USA. He is now with ForteMedia, Inc., Cupertino, CA 95014
USA.
C. Pavlovic was with the R&D Department, GN ReSound Corporation, Redwood City, CA 94063 USA. He is now with Sound ID, Palo Alto, CA 94303
USA
A. Nehorai is with the Electrical and Computer Engineering Department, University of Illinois at Chicago, Chicago, IL 60607 USA.
Publisher Item Identifier S 1053-587X(02)05634-9.
Fig. 2. Three typical polar patterns obtained by the system in Fig. 1 with
using different delay values. From left to right, the patterns are bidirectional,
hypercardioid, and cardioid.
For example, in common hearing aids such as behind-the-ear
aids, there can be only two microphones, and the distance between these two microphones is only about 10 mm [13].
The most common technique in use in hearing aids is a directional microphone or a dual-omnimicrophone system with some
fixed polar patterns, as shown in Fig. 1. The directional system
in Fig. 1 can provide different polar patterns by selecting different values of delay . By way of example, Fig. 2 shows three
,
polar patterns with the value of delay being 0,
, respectively, where is the distance between the two
and
microphones, and is the sound speed. The direction directly
in front of the hearing-aid wearer is represented as 0 , whereas
180 represents the direction directly behind the wearer. The
thick line stands for the gain as a function of direction of the
sound arrival where the gain from any given direction is represented by the distance from the center of the circle. These
three polar patterns are called bidirectional pattern (with null
at 90 and 270 ), hypercardioid pattern (with null at 110 and
250 ), and cardioid pattern (with null at 180 ), respectively. Obviously, the cardioid system attenuates sound the most from directly behind the wearer, whereas the bidirectional system attenuates sound the most from directly to the left and to the right of
the wearer. In different listening environments, users select one
of these three polar patterns using control buttons to achieve the
best noise reduction performance, given the specific listening
1053-587X/02$17.00 © 2002 IEEE
1584
environment. However, for time-varying and moving-noise environments, this fixed directional system delivers degraded performance and therefore, the system with adaptive directionality
is highly desirable.
For a system with two nearby microphones in endfire orientation, the direct way to achieve adaptive directionality is to adaptively change the delay of the system in Fig. 1 so that its value
is equal to the transmission delay value of the noise between the
two microphones. From a performance and complexity point of
view, the key problem in this method is in effectively estimating
the delay value of the noise when noise and the target speech are
both present. Another problem related to this method is how to
implement this delay unit in real time since this delay is usually
a fractional sample delay. For example, if the sampling rate is
16 000 Hz, one sample interval is 62.5 s, but the largest transmission delay value of the noise between the two microphones
and about 29.1 s with
mm. In the scheme of
is
Fig. 1 with fixed mode, this fractional-sample delay is usually
implemented during the A/D stage by processing that is similar
to an up-sampling-shift method [13]. In the adaptive mode, the
up-sampling-shift method is no longer suitable, and the adaptive implementation of a fractional-sample delay is instead accomplished by adaptively updating the coefficients of a specific
filter (ideally, an all-pass filter with linear phase) [14]. These
problems prevent this idea from being of practical value in a
hardware implementation.
On the basis of these problems, this paper proposes a more
practical and effective adaptive directionality system for hearing
devices with two nearby endfire orientation microphones. This
adaptive directionality system is based mainly on an adaptive
combination of two fixed polar patterns that are arranged to
make the null of the combined polar pattern of the system output
always be toward the direction of the noise. The null of one of
these two fixed polar patterns is at 0 (straight ahead of the subject) and the other’s null is at 180 . Both polar patterns are cardioid. The first fixed polar pattern can be implemented by deand
laying the front microphone signal with the value being
subtracting it from the rear microphone signal. Likewise, the
second polar pattern is implemented by delaying the rear miand subtracting it from
crophone signal with the value being
the front microphone signal. The adaptive combination of these
two fixed polar patterns is accomplished by adding an adaptive
gain following the output of the first polar pattern. Different gain
values will provide the combined polar pattern with nulls at different degrees. The value of this gain is updated by minimizing
the power of the system output. Related adaptive algorithms to
update this gain are given in this paper. The hardware implementation of this proposed system has been completed and tested.
The test results demonstrated the accuracy and effectiveness of
the proposed scheme.
It should be noted that for systems with two nearby microphones, there have been also some very effective algorithms
such as the scheme proposed by Vanden Berghe and Wouters
[11] and the scheme proposed by Elko and Pong [15]. In comparison with the two-stage scheme in [11], our proposed system
and algorithm are very simple because ours requires only an
adaptive gain (one-tap adaptive filter).
As a matter of fact, the structure of Fig. 3 in this paper is the
same as that of Fig. 2 in [15], and this paper presents some sim-
IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 50, NO. 7, JULY 2002
Fig. 3.
Schematic of the proposed adaptive null-forming system.
ilar discussions concerning the algorithms, implementations,
and tests, as in [15], although we did our work independently.
In comparison with [15], the major contributions of this paper
could be summarized as follows.
1) The unique mapping relationship between the time delay
and the adaptive gain for forming a null is established in
our paper.
2) We prove that this unique relationship between the time
delay and the adaptive gain is independent of frequency
for the related configurations.
3) The entire shape of the polar pattern resulting from
changing the adaptive gain has been proved to be exactly
the same as the entire shape of the polar pattern resulting
from changing the time delay.
4) The optimum gain value and related adaptive algorithms
are derived in the case where the target signal and noise
are both present and the corresponding system has been
proved to provide the maximal signal-to-noise ratio
(SNR).
The rest of this paper is organized as follows. Section II
presents the proposed scheme and makes related theoretical
analyses. Hardware implementation of this proposed scheme
is dealt with in Section III. Section IV will give various test
results demonstrating the effectiveness and accuracy of the
proposed scheme. In Section V, we will give some conclusions.
II. PROPOSED ADAPTIVE NULL-FORMING SYSTEM
The proposed scheme is shown in Fig. 3, where the received
signals at the front microphone and the rear microphone are
and
, respectively.
delay unit in two channels;
output of the system;
adaptive gain;
output of the adaptive gain processing unit.
From the scheme of Fig. 1, it can be easily seen that the polar
is cardioid with the null at 180 ; likewise, the
pattern of
is cardioid but the null is at 0 . The polar
polar pattern of
is a combination of
pattern of the whole system output
and
and determined by the gain
. The relationwith the gain
ship of the null of the system output
is as follows:
(1)
is the angle of
where is the frequency of the signal, and
the null along the line between the two microphones. To prove
LUO et al.: ADAPTIVE NULL-FORMING SCHEME IN DIGITAL HEARING AIDS
1585
TABLE I
RELATIONSHIP BETWEEN THE NULL AND THE GAIN WITH DIFFERENT FREQUENCIES
this relationship, we use the front microphone as a reference
,
, and
as
channel and have
The null at the polar pattern means that the output power
at this direction, that is, we have
(2)
where is the angle of the source along the line between the
two microphones. It should be noted that all right-side terms
, which is
of (2) should be multiplied by a variable
the signal received in the front microphone. However,
could be considered as 1 for the sake of simplicity using the front
microphone as a reference channel. From (2), we can obtain the
output power of the proposed system as
(3)
Furthermore, using the related trigonometric identities yields
Rearranging the above equation yields (1) and concludes the
proof of (1).
, we will consider only the
Because
in the following, realizing that all related
interval
. It can been
conclusions apply also to the interval
, then
seen from (1) that for all frequencies, if
; if
, then
. This result
can also be obtained directly from Fig. 3 because
means
, the polar pattern of
is the same as that
, and because
means that
and
of
will be identical and will make
be zero when the signal
comes from 90 . Except for these two special cases, (1) shows
that the relationship between the null and the gain depends on
the frequency of the signal. However, with the approximation
for the frequency range of interests as used for obtaining the polar pattern of the scheme in Fig. 1, (1) can be approximated as
(4)
This result means that the relationship between the null and the
gain will be independent of the frequency. In effect, the difference between (1) and (4) is very small, especially when
is in the range from 90 to 180 . The lower the frequency, the
smaller this difference will be. To illustrate this, Table I shows a
mm and at frequencies 500 Hz, 1000
set of results with
1586
IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 50, NO. 7, JULY 2002
patterns, respectively, as shown in Fig. 2. In effect, the corresponding null can also be obtained from (4) or Table I.
These results suggest that the polar-pattern null of the system
output
for all frequencies can be moved by simply updating
and keeping the shape of the polar pattern
the gain value
the same as that of the scheme in Fig. 1 with the corresponding
delay value determined by (8).
Next, we show how to adaptively update the gain value so that
the corresponding null can always be toward the noise instead
of toward the target signal in the case where the target signal
and noise are both present. In this case, we have
(9)
(10)
(11)
Fig. 4. Comparison of the result (solid line) of (4) with the result (dashed line)
of (1) at 6000 Hz frequency.
Hz, 2000 Hz, 4000 Hz, 6000 Hz, and 8000 Hz, respectively. For
further illustrations, Fig. 4 compares the results of (1) and (4) at
6 kHz frequency.
From (4), we also have
for
(5)
This shows that the relationship between the null and the gain
, we
is a monotonic function. In other words, given a gain
can get the unique null for all frequency ranges of interest.
Previously, we have discussed the relationship of the polarwith the gain
. Now,
pattern null of the system output
we will investigate the shape of the whole polar pattern. Before
this, we first present the relationship of the output power
of the fixed system in Fig. 1 against the angle using the approxas follows [1].
imation
(12)
(13)
(14)
and
are the desired signal part and the noise
where
part in the front microphone, respectively, and is the delay
of the noise transmission from the front microphone to the rear
microphone. In the above, we also assume that the desired signal
comes straight ahead, that is, its angle is at 0 . By substituting
(9) and (10) into (12), we get
(15)
contains only the noise part. This property shows
that is,
is equivalent to minimizing
that minimizing the power of
because
the power of the noise part contained in the output
the target signal and the noise are not correlated. The optimal
that minimizes the power of
:
gain
(16)
(6)
Similarly, the output power
approximated from (3) as
of the proposed system can be
can be obtained by
(17)
(7)
A comparison of (6) with (7) shows that these two systems will
provide exactly the same polar pattern with mapping relationship
(8)
, 0.4903, and 0, reFor example, if we choose
,
spectively, then the corresponding delay equals 0,
, respectively, according to (8), and these two systems
and
will provide the bidirectional, hypercardioid, and cardioid polar
,
, and
are the power of
, the cross-corwhere
and
, and the power of
, respectively
relation of
[16]. Moreover, the objective function (16) is a quadratic func. On the basis of all
tion with unique minimization point
the above properties, we can conclude that the null of the system
, which is determined by minimizing the system output
power (16) or determined by (17), will always be toward the
direction of the noise when the target signal and the noise are
both present.
Now, the problem becomes how to adaptively update the opti, given the samples of
and
[as obmization gain
tained by (11) and (12), respectively] rather than the cross-corand the power
. We discuss this problem in
relation
Section III together with other issues of the hardware implementation of the system.
LUO et al.: ADAPTIVE NULL-FORMING SCHEME IN DIGITAL HEARING AIDS
1587
III. IMPLEMENTATION OF THE PROPOSED ADAPTIVE
NULL-FORMING SYSTEM
Having the sample of
and
, we can obtain the optimized gain using any available adaptive algorithms, such as the
LMS, NLMS, LS, or RLS algorithm, because (16) is a typical
quadratic optimization problem with only one coefficient. The
fact that only one coefficient need be calculated makes related
adaptive algorithms very simple and makes the real-time hardware implementation of the proposed system possible.
The LMS version for getting the adaptive gain can be written
as
(18)
where is a step parameter that is positive constant less than
, and is the power of the input
. For better performance and faster convergence speed, can also be time varying
(as used by the normalized LMS algorithm [16]). Based on this,
an algorithm with varying-step can be obtained:
(19)
is the estiwhere is a positive constant less than 2, and
.
mated power of
Equations (18) and (19) are suitable for the sample-by-sample
adaptive mode. However, in our hardware implementation of
this proposed scheme, we use the frame-by-frame adaptive
mode because of other processing in hearing aids is also based
on the frame-by-frame mode. With this mode, the following
steps are used to calculate the adaptive gain [16]. First, we
and
and the
estimate the cross-correlation between
at the th frame by
power of
(20)
(21)
is the number of all samples in a frame
respectively, where
and
of
and equals 56 in our implementation. Second,
and
, and then, the
(17) are replaced with the estimated
estimated adaptive gain is obtained by (17).
We have implemented this proposed scheme in one programmable (assembly code) DSP chip together with other
necessary processing such as A/D and D/A. Other practical
factors considered in our hardware implementation include the
following.
1) In this proposed scheme, the two microphones are assumed to have the same frequency response. However,
this is not the case since in real-world devices, these two
microphones are always mismatched. In order to overcome this problem, we first measure the frequency responses of the two microphones, and we then design and
add a matching filter at the rear channel on the basis
of the measured mismatch. From our measurements, a
first-order IIR filter can compensate well for the misand
have the desired carmatch and can make
dioid polar pattern with null at 180 and 0 , respectively.
Fig. 5 presents an example related to matching filter de-
Fig. 5. Upper curve is the measured mismatch, the bottom curve is the
frequency response of the designed matching filter, and the middle curve is the
frequency response difference of the two channels after matching.
sign where the upper curve is the measured mismatch between the front and rear microphones, the bottom curve
is the frequency response of the designed matching filter,
and the middle curve is the frequency response difference
of the two channels after matching.
2) In order to get a better estimate and make the frame-byframe processing smoother, we estimate the cross-correand
and the power of
by
lation between
(22)
(23)
and are two adjustable parameters such that
,
, and
. Obviously, if
and
, (22) and (23) becomes (20) and (21),
respectively.
would be at de3) The null of the total system output
gree less than 90 (that is, in the front hemisphere) when
is larger than 1.
the absolute value of the gain
This has the potential of canceling the target signal in the
front hemisphere. To avoid this problem, we limit the dynamic range of the gain to the range 1 to 1. This limit is
also convenient for the hardware implementation because
in hardware implementation, all digital values are in the
range from 1 to 1.
4) In the hardware implementation of this proposed system,
the total program memory usage is 129 words, the processing time is 0.32 ms, the propagation delay is 0.04 ms,
and the current is 38 A, which shows that the power cost
of this scheme is very small.
where
IV. TESTING RESULTS
With the DSP code to implement this algorithm being completed and being pulled to GN ReSound behind-the-ear (BTE)
devices, we made extensive tests to verify the algorithm and its
implementation. This kind of device used GN ReSound’s own
DSP chip and related platform (operation systems and assembly
languages). In these tests, a Knowles EM 4346 microphone and
a Knowles EM 3356 microphone were put in the endfire config-
1588
IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 50, NO. 7, JULY 2002
Fig. 6. (Left) Measured polar pattern of x(n) and (right) measured polar pattern of y (n).
Fig. 7.
(Left) Measured polar patterns of the system output z (n) with the gain value being
uration to the upper front and upper rear part of a BTE device,
respectively. The distance of these two microphones was about
1.2 cm. This BTE device was connected with a 2-cc coupler. The
sampling rate in this processing was 16 000 Hz. The distance
between the device and the speaker was about 1.0 m. The input
of the speaker was a frequency-sweep signal with frequencies
from 200 Hz to 8000 Hz and with 70 dB SPL. The room noise
was about 50 dB SPL. A Brüel & Kjær Type 2012 Time-Frequency Analyzer was used to measure and analyze related input
and output signals. In order to get the polar patterns, we changed
the angle between the device and speaker at each 10 by turning
the device, which is easier and more accurate than by moving
the speaker. With this setup, 36 sets of data with all frequencies
from 200 Hz to 8000 Hz, corresponding to 36 positions (angles)
of the device, can be obtained. These 36 sets of data were used
to obtain the polar patterns in combination with our polar-pattern generation software. The main testing can be summarized
as follows.
and
were measured to
1) The polar patterns of
verify that they were the desired fixed polar patterns, that
was cardioid with the
is, that the polar pattern of
was carnull at 180 and that the polar pattern of
dioid with the null at 0 . Fig. 6 shows the measured polar
and
. It should be noted that the polar
pattern of
was measured by reversing the position of
pattern of
00.2174 and (right) 00.4903, respectively.
the two microphones, and hence, its measured polar pattern should be the same as the measured one of
.
2) With a given specific value of the adaptive gain, the polar
was measured to see
pattern of the system output
if the null of the polar pattern and the given value of the
gain met the unique relationship given in (4) and Table I.
Two examples are shown in Fig. 7 with values of the gain
being 0.2174 and 0.4903, respectively.
3) A source (the target signal) was set straight ahead (0 ),
and another source (the noise) was set at specific degree
(toward which the null of the proposed adaptive system
should be) to see if the gain value obtained by the adaptive algorithm [(17), (22), and (23)] met the relationship
(4) and Table I. As a matter of fact, when we measured
each polar pattern, we moved the second source from 0
to 360 successively, as mentioned above. With this measurement method, the measured polar pattern obtained by
the proposed adaptive null-forming system is shown in
Fig. 8. This pattern results because the null of the system
should always be toward the degree of this moving source.
Note that we limit the gain value to the range from 1 to
zero.
4) A speaker with pure-tone sound (the target signal) was
set at 0 and a wide band noise with adjustable frequency
bandwidth and sound pressure level was set at different
LUO et al.: ADAPTIVE NULL-FORMING SCHEME IN DIGITAL HEARING AIDS
1589
a gain calculation is involved in the adaptive algorithm, the
computational complexity is low, which makes the proposed
scheme realizable in hardware implementation. Theoretical
analyses and various tests demonstrated the effectiveness of
the proposed system. As a result, this proposed scheme can
serve as a new and practical tool for noise reduction and signal
enhancement in many hearing devices.
ACKNOWLEDGMENT
Fig. 8. Measured polar pattern of the system output corresponding to testing 3.
angels. The angles of the noise source were changed,
(omni-directionality),
and then, the outputs of
(fixed cardioid directionality), and
(adaptive directionality) were measured to further see the performance
improvement of the proposed adaptive system over the
fixed system of Fig. 1. Moreover, in order to see how this
scheme function under real-world conditions, we have
also devised the hearing in noise test (HINT). The HINT
was utilized to assess speech understanding of sentences
at threshold level under varying noise conditions. The
HINT was an adaptive procedure whereby the masker
(uninterrupted speech-shaped noise) remained at 65 dBA
and the speech signal (sentences) level varied. The sentences were delivered at 0 azimuth, and the noise was
delivered at
1) 90 and 270 ;
2) 180 ;
3) 90, 180, and 270 ;
simultaneously. Subjects were tested in three different
conditions:
1) omni-directional system;
2) fixed directionality system in Fig. 1;
3) proposed adaptive directionality system.
All these tests showed that there was more mean benefit
over the omni-directional system for the proposed adaptive directionality system than for the fixed directional
system in Fig. 1.
V. CONCLUSIONS
In this paper, we propose an effective adaptive null-forming
scheme for two nearby microphones in endfire orientation
and deal with its implementation and tests. The null can be
adaptively moved by simply changing a gain value instead of
changing a fractional-sample delay value. We presented and
proved the unique mapping relationship between the null and
the gain value, and we discussed related adaptive algorithms
for updating the gain. The null corresponding to the gain
obtained by the given adaptive algorithm can be guaranteed
to be always toward the noise source in the case where the
target signal and the noise are both present. Because only
The authors are grateful to the anonymous reviewers and the
associate editor for their very useful suggestions and valuable
comments. They also thank N. Michael, S. Petrovic, C. Struck,
and other co-workers at the R&D Department, GN ReSound
Corporation, for their contributions to the implementation and
testing of this proposed system. A U.S. patent application based
on this proposed system is pending.
REFERENCES
[1] M. Valente, “Use of microphone technology to improve user performance in noise,” Trends Amplificat., vol. 4, no. 3, pp. 112–135, 1999.
[2] W. Soede, A. J. Berkhout, and F. A. Bilsen, “Development of a directional hearing instrument based on array technology,” J. Acoust. Soc.
Amer., vol. 94, no. 2, pp. 785–798, 1993.
[3] J. M. Kates and M. R. Weiss, “A comparison of hearing-aid array-processing techniques,” J. Acoust. Soc. Amer., vol. 99, no. 5, pp. 3138–3148,
1996.
[4] J. G. Desloge, W. M. Rabinowitz, and P. M. Zurek, “Microphone-array
hearing aids with binaural output-part I: Fixed-processing systems,”
IEEE Trans. Speech Audio Processing, vol. 5, pp. 529–542, Nov. 1997.
[5] D. P. Welker, J. E. Greenberg, J. G. Desloge, and P. M. Zurek, “Microphone-array hearing aids with binaural output-part II: A two-microphone adaptive system,” IEEE Trans. Speech Audio Processing, vol. 5,
pp. 543–551, Nov. 1997.
[6] J. E. Greenberg, “Modified LMS algorithm for speech processing with
an adaptive noise canceller,” IEEE Trans. Speech Audio Processing, vol.
6, pp. 338–351, July 1998.
[7] J. E. Greenberg and P. M. Zurek, “Evaluation of an adaptive beamforming method for hearing aids,” J. Acoust. Soc. Amer., vol. 91, no.
3, pp. 1662–1676, 1992.
[8] L. J. Griffiths and C. W. Jim, “An alternative approach to linearly constrained adaptive beamforming,” IEEE Trans. Antennas Propagat., vol.
AP-30, pp. 27–34, Jan. 1982.
[9] M. W. Hoffman, T. D. Trine, K. M. Buckley, and D. J. Van Tasell, “Robust adaptive microphone array processing for hearing aids: Realistic
speech enhancement,” J. Acoust. Soc. Amer., vol. 96, no. 2, pp. 759–770,
1994.
[10] H. Saruwatari, S. Kajita, K. Takeda, and F. Itakura, “Speech enhancement using nonlinear microphone array with complementary
beamforming,” in Proc. Int. Conf. Acoust., Speech, Signal Process.,
1999, pp. 69–72.
[11] J. Vanden Berghe and J. Wouters, “An adaptive noise canceller for
hearing aids using two nearby microphones,” J. Acoust. Soc. Amer.,
vol. 103, no. 6, pp. 3621–3626, 1998.
[12] F.-L. Luo, J. Yang, C. Pavlovic, and A. Nehorai, “An FFT Based
Algorithm for Adaptive Directionality of Dual Microphones,” Dept.
Elect. Eng. Comput. Sci., Univ. Illinois at Chicago, Chicago, IL,
UIC-EECS-00-8, 2000.
[13] B. W. Edwards, Z. Hou, C. J. Struck, and P. Dharan, “Signal processing
algorithms for a new, software based, digital hearing device,” Hearing
J., vol. 51, no. 9, pp. 44–52, 1998.
[14] T. I. Laakso, V. Vaelimaeki, M. Karjalainen, and U. K. Laine, “Splitting
tools for fractional delay filter design,” IEEE Signal Processing Mag.,
vol. 13, pp. 30–60, Jan. 1996.
[15] G. W. Elko and A.-T. Nguyen Pong, “A simple adaptive first-order differential microphone,” in Proc. IEEE Workshop Appl. Signal Process.
Audio Acoust., 1995.
[16] S. Haykin, Adaptive Filter Theory. Englewood Cliffs, NJ: PrenticeHall, 1996.
1590
Fa-Long Luo (SM’95) received the B.S., M.S.
and Ph.D. degrees, all with honors, in electronics
engineering from Xidian University, Xi’an, China,
in 1983, 1989, and 1992, respectively.
From 1983 to 1986, he was an engineer with
Changfeng Electronic Systems Corporation. He
was with Tsinghua University, Beijing, China, as a
Research Fellow from December 1991 to December
1993. From May 1994 to October 1998, he was with
University of Erlangen-Nuremberg, Nuremberg,
Germany, as a Principal Research Scientist, where
he was first supported by the Alexander von Humboldt Foundation of Germany
and then by the German Research Foundation. From October 1998 and
February 1999, he was a Senior Project Leader of Cybernetics InfoTech,
Inc. From March 1999 to May 2001, he was with GN ReSound Corporation,
Redwood City, CA, as a Senior Research Scientist and Project Manager. In
June 2001, joined Quicksilver Technology as a Senior Member of Technical
Staff. He has seven U.S. patents pending. He has authored two books: Applied
Neural Networks for Signal Processing (Cambridge, UK: Cambridge Univ.
Press, 1997, 1998, and 1999) and Neural Networks and Signal Processing
(Beijing, China: National Electronic Industrial Publishing House of China,
1993). He has also written more than 80 articles in journals and conferences.
As a principal investigator, he has proposed and conducted more than 20
research and industry projects in signal and data processing with various
applications and their implementation. His inventions on the spectral contrast
enhancement and adaptive microphone array system have been successfully
used and implemented in hearing aid products, and significant performance
improvements have been achieved.
Dr. Luo received the National Young Investigator Award of China, which is
biennially granted to up to 100 individuals who have made internationally or nationally recognized extraordinary contributions in science and technology and
is nominated from young scientists, engineers, educators, and technical executives over all China, in 1994. He received the National Outstanding Science
and Technology Book Award of China in 1995. He has been a technical committee member (Neural Networks Technical Committee) of IEEE Neural Networks Council since 1998. From June 1997 to December 2000, he was a technical committee member (Neural Networks for Signal Processing) of the IEEE
Signal Processing Society. As a Guest Editor, he edited a special issue of Signal
Processing on neural networks (vol. 64, no. 3, 1998). He is the Executive Guest
Editor of a special issue of Speech Communication on speech processing for
hearing aids. He was an Associate Editor of an IEEE Control Systems Society
Conference. As a reviewer, session chair, and technical committee member, he
has served on many international journals and conferences. He is an editorial
board member of International Journal of Information Fusion and IEEE Communication Surveys and Tutorials.
Jun Yang (SM’99) received the B.S. degree from
Huazhong (Central China) University of Science
and Technology in 1985, the M.S. degree from
Northwestern Polytechnic University in 1988, and
the Ph.D. degree from Xidian University, Xi’an,
China, in 1991, all in electronic engineering.
From November 1991 to October 1993, she was
with Tsinghua University, Beijing, China, where
she was promoted to Associate Professor in 1993.
She was with Oldenburg University of Germany and
Viennatone Corporation of Austria as a Research
Fellow and Principal Engineer, respectively, from October 1993 to November
1998. In December 1998, she joined GN ReSound Corporation, Redwood
City, CA, as a Senior Research Engineer and then became the Chief Scientist
of Spatializer Audio Laboratories, Inc. Since March 2001, she has been a
Senior Audio Engineer with VM Labs, Inc. and a Principal DSP Engineer
with ForteMedia, Inc., respectively. She has five U.S. patents (pending,
co-inventor) and has authored 40 journal articles and conference papers on
signal processing with its various applications such as audio, communications,
and hearing aids. The hearing devices with some of her inventions are well
sold in the world market. As a principal investigator, she has also conducted
several research projects on communications, speech coding and processing,
and an audio interface for blind computer users. Most of her projects, because
of the importance and excellence, were granted by government agencies or
foundations of Germany and China, respectively. Her works and contributions
have received wide attention and were reported by Science and Technology
Daily (the national science and technology newspaper of China) and published
IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 50, NO. 7, JULY 2002
by the State Science and Technology Committee of China. She has been a reviewer for Speech Communication
Dr. Yang has also received several university-level awards for her excellent
work in science and technology. She has been a reviewer for the IEEE
TRANSACTIONS ON SPEECH AND AUDIO PROCESSING. She is also a Member of
Audio Engineering Society and was a Senior Member of the Chinese Institute
of Electronics.
Chaslav Pavlovic received the M.S. and B.S. degrees
in electrical engineering and the Ph.D. degree in audiology.
He is currently the Executive Vice President for
Research and Development at Sound ID. Previously,
for eight years, he was with GN ReSound (formerly ReSound Corporation), a hearing healthcare
company, where he was the Senior Vice President
for Research and Development. During his tenure,
ReSound (subsequently GN ReSound) established
itself as a technology leader in hearing products and
grew to become the second largest hearing instrument company in the world.
He was a Full Professor of speech technology at Aix en Provence, France, and
an Associate Professor of Audiology at the University of Iowa, Iowa City, He
has held other academic appointments. He has more than 100 publications on
hearing aids, speech intelligibility, and speech quality.
Dr. Pavlovic was the Chair of the 1997 ASA work group that developed ANSI
S3.5: the Speech Intelligibility Index Standard. He has also been the Coordinator
of the European Audiological Tests and Station (EURAUD) project; Chair of the
American National Standards Institute S3-79 Writing Group (Calculation of the
Articulation Index); USA representative to the International Standards Organization ISO/TC 43/SC1; Coordinator of the Overall Quality Assessment Subgroup European Consortium for Speech Assessment Methods (SAM, Project
Esprit); Coordinator of participating French laboratories on projects TIDE and
OSCAR (pattern extraction hearing aids); Member of the American National
Standards Institute S12-8 Writing Group (rating noise with respect to speech
interference); Member of the Editorial Board of Acoustics; Staff Editor of the
Journal D’Acoustique; Board of Directors of the Journal D’Acoustique; and
Member of the Technical Committee on Speech Communication of the Acoustical Society of America.
Arye Nehorai (S’80–M’83–SM’90–F’94) received
the B.Sc. and M.Sc. degrees in electrical engineering
from the Technion—Israel Institute of Technology,
Haifa, in 1976 and 1979, respectively, and the Ph.D.
degree in electrical engineering from Stanford University, Stanford, CA, in 1983.
After graduation, he worked as a Research Engineer for Systems Control Technology, Inc., Palo Alto,
CA. From 1985 to 1995, he was with the Department of Electrical Engineering, Yale University, New
Haven, CT, where he became an Associate Professor
in 1989. In 1995, he joined the Department of Electrical Engineering and Computer Science, The University of Illinois at Chicago (UIC), as a Full Professor.
From 2000 to 2001, he was Chair of the Department’s Electrical and Computer
Engineering (ECE) Division, which is now a new department. He holds a joint
professorship with the ECE and Bioengineering Departments at UIC. His research interests are in signal processing, communications, and biomedicine. He
is on the Editorial Board of Signal Processing and was an Associate Editor for
Circuits, Systems, and Signal Processing.
Dr. Nehorai is Editor-in-Chief of the IEEE TRANSACTIONS ON SIGNAL
PROCESSING. He is also a Member of the Publications Board of the IEEE Signal
Processing Society. He has previously been an Associate Editor of the IEEE
TRANSACTIONS ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, the IEEE
SIGNAL PROCESSING LETTERS, the IEEE TRANSACTIONS ON ANTENNAS AND
PROPAGATION, and the IEEE JOURNAL OF OCEANIC ENGINEERING. He served
as Chairman of the Connecticut IEEE Signal Processing Chapter from 1986
to 1995 and is currently the Chair and a Founding Member of the IEEE Signal
Processing Society’s Technical Committee on Sensor Array and Multichannel
(SAM) Processing. He was the co-General Chair of the First IEEE SAM Signal
Processing Workshop, held in 2000, and will again serve in this position in
2002. He was corecipient, with P. Stoica, of the 1989 IEEE Signal Processing
Society’s Senior Award for Best Paper. He received the Faculty Research
Award from the UIC College of Engineering in 1999. In 2001, he was named
University Scholar of the University of Illinois. He has been a Fellow of the
Royal Statistical Society since 1996.
Download