Acoustic talker direction finders have potential applications to camera pointing for teleconferencing and to microphone array beam steering for audio communication and voice processing systems. This paper describes a laboratory setup and computer interface that was developed for testing talker direction finder algorithms in the Bell Labs Varechoic chamber, a room with computer-controlled absorbing panels. The procedure exploits the full capability of the facility by automatically stepping through a sequence of room panel configurations, outputting a digital speech signal, running the processor, and collecting the data. The advantage of this technique is that it allows for testing under a multitude of different acoustic conditions in the same physical location, thereby enabling a general characterization of the algorithm under evaluation. As an example of the technique, we have implemented the Fischell–Coker talker direction finder algorithm using real-time C-code running on an SGI workstation, which is the same machine that is used to orchestrate the automatic testing procedure.

1.
H. F.
Silverman
and
S. E.
Kirtman
, “
A two-stage algorithm for determining talker location from linear microphone array data
,”
Comput. Speech Lang.
6
,
129
152
(
1992
).
2.
M. S.
Brandstein
,
J. E.
Adcock
, and
H. F.
Silverman
, “
A closed-form location estimator for use with room environment microphone arrays
,”
IEEE Trans. Speech Audio Process.
5
,
45
50
(
1997
).
3.
D. V.
Rabinkin
,
R. J.
Renomeron
,
A.
Dahl
,
J. C.
French
, and
J. L.
Flanagan
, “
A DSP implementation of source location using microphone arrays
,”
Proc. SPIE
2846
,
88
99
(
1996
).
4.
B.
Champagne
,
S.
Bedard
, and
A.
Stephenne
, “
Performance of time-delay estimation in the presence of room reverberation
,”
IEEE Trans. Speech Audio Process.
4
,
148
152
(
1996
).
5.
A. Stephenne and B. Champagne, “Cepstral prefiltering for time delay estimation in reverberant environments,” in Proceedings of the IEEE International Conference on Acoustics, Speech, Signal Processing (IEEE, New York, 1995), pp. 3055–3058.
6.
M. Bodden, “Modeling human sound-source localization and the cocktail-party-effect,” Acta Acoustica 1, 43–55 (1993).
7.
B.
Kollmeier
and
R.
Koch
, “
Speech enhancement based on physiological and psychoacoustical models of modulation perception and binaural interaction
,”
J. Acoust. Soc. Am.
95
,
1593
1602
(
1994
).
8.
W.
Lindemann
, “
Extension of a binaural cross-correlation model by contralateral inhibition. I. Simulation of lateralization for stationary signals
,”
J. Acoust. Soc. Am.
80
,
1608
1622
(
1986
).
9.
W.
Gaik
, “
Combined evaluation of interaural time and intensity differences: Psychoacoustic results and computer modeling
,”
J. Acoust. Soc. Am.
94
,
98
110
(
1993
).
10.
W.
Lindemann
, “
Extension of a binaural cross-correlation model by contralateral inhibition. II. The law of the first wave front
,”
J. Acoust. Soc. Am.
80
,
1623
1630
(
1986
).
11.
A. G.
Dabak
and
D. H.
Johnson
, “
Function-based modeling of binaural processing: Level and time cues
,”
J. Acoust. Soc. Am.
94
,
2604
2616
(
1993
).
12.
D. R. Fischell and C. H. Coker, “A speech direction finder,” in Proceedings of the IEEE International Conference on Acoustics, Speech, Signal Processing (IEEE, New York, 1984), pp. 19.8.1–19.8.4.
13.
W. C. Ward, G. W. Elko, R. A. Kubli, and W. C. McDougald, “The new varechoic chamber at AT&T Bell Labs,” in Proceedings of the Wallance Clement Sabine Centennial Symposium (Acoustical Society of America, Woodbury, NY, 1994), pp. 343–346.
This content is only available via PDF.
You do not currently have access to this content.