The paper reports on the ability of people to rapidly adapt in localizing virtual sound sources in both azimuth and elevation when listening to sounds synthesized using non-individualized head-related transfer functions (HRTFs). Participants were placed within an audio-kinesthetic Virtual Auditory Environment (VAE) platform that allows association of the physical position of a virtual sound source with an alternate set of acoustic spectral cues through the use of a tracked physical ball manipulated by the subject. This set-up offers a natural perception-action coupling, which is not limited to the visual field of view. The experiment consisted of three sessions: an initial localization test to evaluate participants’ performance, an adaptation session, and a subsequent localization test. A reference control group was included using individual measured HRTFs. Results show significant improvement in localization performance. Relative to the control group, participants using non-individual HRTFs reduced localization errors in elevation by 10° with three sessions of 12 min. No significant improvement was found for azimuthal errors or for single session adaptation.
Skip Nav Destination
April 12 2012
Rapid head-related transfer function adaptation using a virtual auditory environment
Gaëtan Parseihian, Brian F. G. Katz; Rapid head-related transfer function adaptation using a virtual auditory environment. J. Acoust. Soc. Am. 1 April 2012; 131 (4): 2948–2957. https://doi.org/10.1121/1.3687448
Download citation file: