Subspace algorithms based on higher-order cumulants were developed to achieve high-resolution separation in non-Gaussian processes. However, singular value decomposition (SVD) of a huge matrix is an unavoidable step of these algorithms. The memory space and running time required by the decomposition are super-linear with respect to the size of the matrix, which is prohibitive in terms of practical applications. Thus, in this paper, a fast raypath separation algorithm based on low-rank matrix approximation is proposed in a shallow-water waveguide. The experimental results illustrate that the proposed algorithm dramatically reduces the consumption of time and space, with arbitrarily small error, compared to conventional higher-order cumulant-based algorithms.
1. Introduction
In shallow water, acoustic waves propagate in multi-paths due to refraction and reflection at the ocean surface and the continental shelf. Multi-path propagation has been widely used in many applications, such as active sonar, ocean acoustic tomography,1 and source localization, because more raypaths provide more information. However, the successful use of the information is based on the premise of first eliminating the raypaths' interferences with each other produced by the multi-path propagation. Among the existing algorithms to separate raypaths, subspace-based algorithms are noticeable for their large resolution improvement. In particular, the multiple signal classification (MUSIC) algorithm,2 as a representative subspace-based method, was applied to separate raypaths under the assumption of an uncorrelated narrow-band signal. However, MUSIC fails to separate raypaths when they are fully correlated or coherent. To overcome this drawback, Jiang et al.3,4 developed a high-resolution method called the smoothing-multiple signal classification active large band (MUSICAL) algorithm, which greatly improves the separation performance, especially when two fully correlated raypaths arrive close together. Although the smoothing-MUSICAL algorithm was developed under the assumption of white Gaussian noise, colored noise exists in real ocean environments. Recently, Jiang et al.5 presented a higher-order wideband active algorithm to separate raypaths interrupted by colored noise. Naturally, it takes advantage of fourth-order cumulants containing non-Gaussian information in non-Gaussian processes. The experimental results obtained from tank data at small scale5 illustrate that the higher-order algorithm can obtain more accurate results and suppress noise; however, it has large computational cost (both time and memory space) for computing the singular value decomposition of a huge cumulants matrix. This cost is prohibitive from a practical perspective. To overcome this drawback, we propose in this study a fast higher-order separation algorithm, which has the advantages of low computational cost and high resolution. In addition, its performance will be tested using synthetic data, experimental data obtained in a tank and ocean data. The most important contribution is that the proposed algorithm provides a strategy to find the signal subspace efficiently, which can be extended to all the subspace-based algorithms. Based on the strategy, their computational cost can be largely reduced especially when the cumulants matrix has large dimensions.
This paper is organized as follows. Our fast raypath separation algorithm is introduced in Sec. 2. In Secs. 3 and 4 we discuss the performance and the error of the proposed method. In Sec. 5, its performance test using synthetic data, experimental data obtained in a tank, and ocean data is conducted, respectively. We draw conclusions in Sec. 6.
2. A fast higher-order separation algorithm
We consider the case where P raypaths arrive on a vertical array composed of M sensors. The received signal of the mth sensor in the frequency domain is modeled by
where the term is the deterministic amplitude of the emitted signal at frequency ν. The amplitude of each raypath, ap, is considered random, where Tp represents the arrival time of the raypath on the reference sensor and is the delay between the reference sensor and the sensor. Additionally, , where θp is the arrival direction of raypath p on the array, d is the distance between two adjacent sensors, c is the propagation velocity of the acoustic signal, and is the additive noise received at the sensor.
We use F frequency bins of the wideband signal. Hence, Eq. (1) is written in matrix form as follows:
where with ; that is, is a long-vector of size M × F, which is arranged according to all the frequency bins on all the sensors. is similarly the concatenation of the observed noise vectors at each frequency. Specifically, with . is a vector of dimension P. with , where is the transfer function between the source and the mth sensor, and + indicates transposition.
We first apply spatial-frequency smoothing4,5 as a preprocessing step to eliminate the interferences among raypaths. That is, for a spatial smoothing factor Ks, an antenna composed of M sensors is divided into sub-antennas. The th sub-antenna is composed of the sensors in the range of . At the same time, a similar frequency-smoothing process is also applied, where Kf denotes the frequency-smoothing factor. Finally, using the recurrences produced by the spatial-frequency smoothing, we can estimate the wideband trispectral matrix with the following formula:
where represents the Kronecker product, indicates conjugation, and H indicates the conjugate transpose. Based on the preprocessing, the rank of is equal to (it is necessary to select K greater than P).
The traditional subspace-based algorithms use methods such as the singular value decomposition (SVD) to span a signal subspace, however, they require memory and time which are super-linear in the size of the trispectral matrix (or spectral matrix). It is prohibitive for a practical application, especially when the sizes are very large. Inspired by Frieze et al.,6–8 we introduce a randomized strategy to span the signal subspace at low computation cost.
The strategy is based on the following three principles.
Principle 1: We can approximate the product of two matrices by an algorithm called the basic matrix multiplication.7 That is, when we input two matrices, and , a probability distribution and a number , two output matrices and are returned, such that , where is a matrix whose columns are c randomly chosen columns of with suitably rescale.
Principle 2: An appropriate choice of the probability distribution is crucial steps of the proposed algorithm. Although a uniform distribution could be always used, the optimal probabilities are proved and defined as the following equation:7 Suppose , such that , and are such that , and such that for some positive constant : .
Principle 3: According to linear algebra, for symmetric matrices the left and right singular vector are the same. The singular values of are the nonnegative square roots of the eigenvalues of and of ; in addition, the left singular vectors of are eigenvectors of and the right singular vectors are eigenvectors of .
The general process of the proposed algorithm is first described: (1) based on principle 3, finding the eigenvectors of the matrix is equivalent to find the eigenvector of the matrix ; (2) based on principles 1 and 2, the matrix can be approximated by with the optimal probabilities, where consists of those c columns of (after appropriate rescaling); it has been proven that by choosing appropriate sampling probabilities and enough columns the approximation error can be made arbitrarily small; (3) based on principle 3, finding the eigenvectors of the matrix can be transformed to find the eigenvectors of the matrix , then from approximations to the top singular values and corresponding singular vectors of may be computed, which are the approximations of the eigenvectors of the matrix . Thus, the fast algorithm are finally operated in four specific steps.
To sample c columns of () randomly according to a judicious probability q, we first compute for all columns (where is the ith column of matrix and indicates Frobenius norm). Subsequently, we get a random number r, where . If , we retain the ith column. The sampling step will be repeated another c – 1 times and finally we will get c columns of .
According to the rule provided by the first step, we sample c columns where , rescale each column by the factor , and arrange all c columns to construct the approximation matrix of , which is denoted as . We would like to approximate the eigenvectors of square matrix . Normally, we can obtain these vectors directly through performing a SVD on matrix , but its size is still the same as that of . Thus, we first perform a SVD on , whose size is quite smaller, to find its eigenvectors, which are also the right singular vectors of ; then, they can be transformed into the left singular vectors of , which are also the eigenvectors of the matrix . Therefore in the next steps first we carry out the SVD on matrix to obtain right singular vectors of ; Second we compute P2 left singular vectors of , which successively approximate P2 eigenvectors of square matrix
SVD is performed on matrix , and the P2 eigenvector , corresponding to the P2 largest eigenvalues , is identified.
Eigenvectors are calculated using the equation .
Based on the eigenvectors computed previously, the signal subspace is projected as follows:
where is composed of .
Finally, the direction of arrival (DOA) and the arrival time of each raypath are obtained by maximizing the following estimator:
where the wideband steering vector is the Kronecker product of ,
where is the concatenation of the classical steering vectors , that is,
3. Error analysis of the low-rank matrix approximation
The error analysis of the low-rank matrix approximation is performed in two steps: (1) define the bound of the approximation error and (2) choose c and define the sampling probabilities.
The low-rank approximation incurs two error components: first, the error between and (assuming is the most optimal P2-rank approximation to matrix ); second, incurs an error depending on or, equivalently, . It has been proven7 that the error appears entirely within the following two bounds with respect to the Frobenius norm and spectral norm, respectively,
where indicates the Frobenius norm and indicates the spectral norm. The optimal P2-rank approximation matrix is , where is composed of the top P2 left singular vectors of .
To control the approximation error, the method used to select c and to define the sampling probabilities is important. According to the theory of Drineas et al.,7 we can determine the sampling probabilities as follows. We assume is constructed by sampling c columns of with probabilities to make for some positive . Set . ϵ represents an additional error coefficient and δ denotes the failure probability.
- If a Frobenius norm bound is desired: set ; then, with probability of at least ,(10)
- If a spectral norm bound is desired: set ; then, with probability of at least ,(11)
It is necessary note that the approximation error can be made arbitrarily small by choosing enough columns.
4. Analysis of the implementation and running time of the low-rank approximation
In the proposed algorithm, we can sample one column in one pass over the matrix without previously knowing its size. With the algorithm running in parallel, we can obtain more samples in a single pass in O(c) additional space and time. Given the index to be sampled, we need additional space and time to construct matrix . Computing requires additional space and additional time, and we consider the SVD of to require additional space and additional time. Then, computing requires P2 matrix-vector multiplications for additional space and time. Overall, the proposed fast algorithm requires additional space and additional time, which have linear relationships with the number of rows in the high-order cumulants square matrix.
5. Performance test
In this section, we test the performance of the proposed algorithm with many sets of simulation data, a set of real data obtained in a tank9 (the experiment is based on the principle that if we amplify the frequency of the signals by a factor while reducing the spatial distance by the same factor, the actual physical phenomena occurring in nature can be reproduced on a smaller scale) and finally a set of ocean data.10 (The experiment was performed in July 2005 north of Elba Island, Italy.) The conventional higher-order algorithm in Ref. 5 is provided for comparison. These tests are performed on a server with 6 Intel Xeon CPUs(X7542) with 48 cores @2.67 GHz and 256.0 GB RAM on the Redhat Linux 5.6 operating system. All the tests are performed between a point source and a vertical array, which are composed of M equally spaced sensors. We show the parameters including the number of sensors (M), the source location (zs), the range of the array covering the ocean from the position at m to the position at , the interval between two adjacent sensors d and the distance between the source and the reference sensor D in each configuration in Table 1. We give the central frequency (νc), the bandwidth of the emitted signals (νw), the number of the frequencies ( from νb to νe) and the taps of the broadband time-domain signal. (N) used in the tests in Table 2.
The configuration parameters of simulation, small-scale experiment and the at-sea experiment.
. | M . | . | . | d (m) . | D (m) . |
---|---|---|---|---|---|
Simulation | 6 | 25 | 0.75 | 750 | |
Small-scale experiment | 11 | 1 | |||
Ocean data | 32 | 96 | 80–112 | 0.75 |
. | M . | . | . | d (m) . | D (m) . |
---|---|---|---|---|---|
Simulation | 6 | 25 | 0.75 | 750 | |
Small-scale experiment | 11 | 1 | |||
Ocean data | 32 | 96 | 80–112 | 0.75 |
The test parameters of simulation, small-scale experiment and the at-sea experiment.
. | νc (Hz) . | νw (Hz) . | (Hz) . | N . | δ . | ϵ . |
---|---|---|---|---|---|---|
Simulation | 39() | 128 | 0.01 | 2 | ||
Small-scale experiment | 1.2 | 1 | 17(0.48276 ) | 145 | 0.01 | 4 |
Ocean data | 7() | 60 | 0.01 | 2 |
. | νc (Hz) . | νw (Hz) . | (Hz) . | N . | δ . | ϵ . |
---|---|---|---|---|---|---|
Simulation | 39() | 128 | 0.01 | 2 | ||
Small-scale experiment | 1.2 | 1 | 17(0.48276 ) | 145 | 0.01 | 4 |
Ocean data | 7() | 60 | 0.01 | 2 |
The proposed fast algorithm is applied to the simulation data, a set of real data obtained in a tank and finally a set of ocean data and the separation results are shown in Fig. 1 and Fig. 2, respectively. The expected positions are marked with black crosses. Figure 1(A) illustrates the separation result with a set of simulation data. We can see that the five raypaths are all successfully separated with few errors compared with their expected values [(0.5, 0), (0.9, −0.09), (1.2, 0.15), (1.9, −0.254), and (2.5, 0.235)]. Moreover, the efficiency of the proposed algorithm is tested with many sets of simulation data for various numbers of snapshots. Ten realizations are used for each number of snapshots. Figure 1(B) shows that the time reduction ratio increases sharply with increasing number of snapshots. Figure 1(C) shows the separation result with a set of real data obtained in a tank. The proposed algorithm successfully separates the seven raypaths with few errors. Their expected values are (0.6789, 0), (0.6799, 3.1), (0.68, −3.2), (0.683, 6.3), (0.683, −6.3), (0.688, 9.3), and (0.6882, −9.4). Figures 2(A) and 2(B) show the separation results for the at-sea experimental data with the beamforming algorithm and our proposed algorithm, respectively. The proposed fast algorithm are more efficient for noise suppression and obtain the smaller spots representing the positions of the raypaths compared to the beamforming algorithm. Figure 2(C) displays the three raypaths in Figs. 2(A) and 2(B), which propagate between the centers of source and receive arrays. The centers of source and receive arrays both located at 96 m in the water. Due to the sound speed variation described in the Ref. 10, the three raypaths arrived at close time. There are a surface-reflected ray and two refracted rays. Each refracted ray has a turning point near the depth of maximum sound-speed variability. Figures 2(a) and 2(b) illustrate that the proposed fast algorithm is robust to the fluctuation of environmental factors, for instance the water column sound speed variation we have described. The performance of the 4-smoothing-MUSICAL algorithm and the proposed fast algorithm are almost the same, which are not included in the paper due to space limitations. Table 3 compares the time and space consumed by the 4-smoothing-MUSICAL algorithm and the proposed fast algorithm, illustrating that with the proposed algorithm, the cost of the time in decomposition step has been reduced by approximately 99.62% (simulation), 99.7% (small-scale experiment), and 99.9% (ocean data), respectively. The peak memory has been reduced by approximately 95.50% (simulation), 99.2% (small-scale experiment), and 99.75% (ocean data), respectively.
(Color online) Separation results with the proposed fast algorithm. Black crosses indicate the expected positions of the raypaths. (A) Separation results with the simulation data. (B) The time reduction ratio with the simulation data (the proposed algorithm vs 4-smoothing-MUSICAL). (C) Separation results with the small-scale data.
(Color online) Separation results with the proposed fast algorithm. Black crosses indicate the expected positions of the raypaths. (A) Separation results with the simulation data. (B) The time reduction ratio with the simulation data (the proposed algorithm vs 4-smoothing-MUSICAL). (C) Separation results with the small-scale data.
(Color online) Results comparison for the data obtained from at-sea experiment. Black crosses indicate the expected positions of the raypaths. (A) Separation results with the beamforming algorithm. (B) Separation results with the proposed fast algorithm. (C) Raypaths between the source and the reference sensor for the three rays.
(Color online) Results comparison for the data obtained from at-sea experiment. Black crosses indicate the expected positions of the raypaths. (A) Separation results with the beamforming algorithm. (B) Separation results with the proposed fast algorithm. (C) Raypaths between the source and the reference sensor for the three rays.
The comparison of time and space consumption.
. | . | Conventional EVD . | Linear time EVD . | Reduction ratio . |
---|---|---|---|---|
Simulation | CPU time (s) | 5195.208 | 19.642 | 99.62% |
Peak memory (Kb) | 15 167 030 | 682 820 | 95.5% | |
Small-scale experiment | CPU time (s) | 19 042 | 58 | 99.7% |
Peak memory (Kb) | 38 797 300 | 291 500 | 99.2% | |
Ocean data | CPU time (s) | 55 080 | 55 | 99.9% |
Peak memory (Gb) | 95.3 | 0.235 | 99.75% |
. | . | Conventional EVD . | Linear time EVD . | Reduction ratio . |
---|---|---|---|---|
Simulation | CPU time (s) | 5195.208 | 19.642 | 99.62% |
Peak memory (Kb) | 15 167 030 | 682 820 | 95.5% | |
Small-scale experiment | CPU time (s) | 19 042 | 58 | 99.7% |
Peak memory (Kb) | 38 797 300 | 291 500 | 99.2% | |
Ocean data | CPU time (s) | 55 080 | 55 | 99.9% |
Peak memory (Gb) | 95.3 | 0.235 | 99.75% |
6. Conclusion
In this paper we proposed a fast raypath separation method that achieves large reductions in time and space consumption compared with the conventional high-order cumulant-based algorithm. In the future, the method will be extended to improve the efficiency of the multi-dimensional subspace-based algorithms.
Acknowledgments
This paper was supported by the National Natural Science Foundation of China (Grant Nos. 61401085 and 31400842), the State Key Laboratory of Acoustics, Chinese Academy of Sciences (Grant No. SKLA201604), and the Scientific Research Foundation for the Returned Overseas Chinese Scholars. The Focused Acoustic Forecasting experiment (FAF05) experiment was performed in collaborative experiments with the NATO Underwater Research Centre (NURC), La Spezia, Italy, with Mark Stevenson as Chief Scientist. Scientists who contributed to these experiments include Tuncay Akal, W. A. Kuperman, W. H. Hodgkiss, H. C. Song, B. D. Cornuelle, Piero Boni, Piero Guerrini, other NURC staff, and the officers and crew of the RV Alliance.