Noise fields encountered in real-life scenarios can often be approximated as spherical or cylindrical noise fields. The characteristics of the noise field can be described by a spatial coherence function. For simulation purposes, researchers in the signal processing community often require sensor signals that exhibit a specific spatial coherence function. In addition, they often require a specific type of noise such as temporally correlated noise, babble speech that comprises a mixture of mutually independent speech fragments, or factory noise. Existing algorithms are unable to generate sensor signals such as babble speech and factory noise observed in an arbitrary noise field. In this paper an efficient algorithm is developed that generates multisensor signals under a predefined spatial coherence constraint. The benefit of the developed algorithm is twofold. Firstly, there are no restrictions on the spatial coherence function. Secondly, to generate sensor signals the algorithm requires only mutually independent noise signals. The performance evaluation shows that the developed algorithm is able to generate a more accurate spatial coherence between the generated sensor signals compared to the so-called image method that is frequently used in the signal processing community.
Skip Nav Destination
November 01 2008
Generating nonstationary multisensor signals under a spatial coherence constraint
Emanuël A. P. Habets;
Emanuël A. P. Habets, Israel Cohen, Sharon Gannot; Generating nonstationary multisensor signals under a spatial coherence constraint. J. Acoust. Soc. Am. 1 November 2008; 124 (5): 2911–2917. https://doi.org/10.1121/1.2987429
Download citation file: