A compressive sensing based compressed neural network for sound source localization
Summary (2 min read)
INTRODUCTION
- In the sound source localization techniques, location of the source has to be estimated automatically by calculating the direction of the received signal [1] .
- Feature extraction is the process of selection of the useful data for estimation of DOA.
- The important key insight is the use of the instantaneous crosspower spectrum at each pair of sensors.
- After this step the authors have compressed the neural network that is designed with these feature vectors.
- The next section presents a review of techniques for sound source localization.
II. SOUND SOURCE LOCALIZATION
- The assumption of far field sources remains true while the distance between source and reference microphone is larger than [2] fig.
- And D is the microphone array length.
- So, the time delay of the received signal between the reference microphone and the − ℎ microphone would be [15] : EQUATION.
- Therefore, is the amount of time that the signal traverses the distance between any two neighboring microphones, Fig. 1 EQUATION where, r is the distance between source and the first microphone [15] .
III. FEATURE SELECTION
- The aim of this section is to compute the feature vectors from the array data and use the MLP (Multi Layer Perceptron) approximation property to map the feature vectors to the corresponding DOA, as shown in Fig. 3 [6] .
- The authors summarized their algorithm for computing a real-valued feature vector of length (2( − 1) + 1) , for dominant frequencies and M sensors below: Preprocessing algorithm for computing a real-valued feature vector: 1. Calculate the -point FFT of the signal at each sensor.
- In conclusion, their purpose is to design a neural network with least number of hidden neurons (or weights) that has the minimum increase in error given by‖ − ‖.
- This problem is equivalent to finding which most of its rows are zeros.
- Comparing these equations with (7) the authors can conclude that these minimization problems can be written as CS problems.
VI. RESULTS AND DISCUSSION
- As mentioned before, assuming that the received speech signals are modeled with 10 dominant frequencies, the authors have trained a two layer Perceptron neural network with 128 neurons in hidden layer and trained it with feature vectors that are obtained with CS from the cross-power spectrum of the received microphone signals.
- After computing network weights the authors tried to compress network with their algorithms.
- With these outputs the authors can infer that CS algorithms are faster than other algorithms and have smaller error in compare with other algorithms.
- This means that, According to the number of Measurement vectors, the algorithm that uses single-measurement vector (SMV) is faster than another algorithm that uses multiple-measurement vector (MMV) but its achieve error is not smaller.
VII. CONCLUSION
- Particularly, using the pursuit and greedy methods in CS, a compressing methods for NNs has been presented.
- The key difference between their algorithm and previous techniques is that the authors focus on the remaining elements of neural networks; their method has a quick convergence.
- The simulation results, demonstrates that their algorithm is an effective alternative to traditional methods in terms of accuracy and computational complexity.
- Results revealed this fact that the proposed algorithm could decrease the computational complexity while the performance is increased.
Did you find this useful? Give us your feedback
Citations
32 citations
15 citations
Cites background or methods or result from "A compressive sensing based compres..."
...In next step, to evaluate the proposed sound source localization system, we have compared its performance with those for two of the previously-reported CS-based target localization algorithms, namely DTL [8] and CSNN [9]....
[...]
...The Compressive Sensing-based Neural Network (CSNN) method [9] employs a neural network for the calculation of spectral feature vectors in each microphone....
[...]
...In [9], authors have tried to reduce computational complexity by employing a feature extraction process that selects useful data for estimation of DOA....
[...]
...Comparison between the localization performance of the proposed system, CSNN [9] and DTL algorithm [8] in the case of two sound sources and two microphones....
[...]
9 citations
Cites methods from "A compressive sensing based compres..."
...Localization performance of (a) the proposed scheme, (b) DTL algorithm in [10], and (c) CSNN algorithm in [23], for three patients and six receiver nodes....
[...]
...We compare the performance of the CS-2FFT-based scheme with that of two CS-based target localization algorithms, namely DTL [10] and CS-based neural network (CSNN) [23]....
[...]
...EML algorithms in [10], [23], and [24] in the case of three patients and six receiver nodes....
[...]
...pared to other classical positioning algorithms such as the EML, DTL, and CSNN approaches in [10] and [23]....
[...]
5 citations
4 citations
References
40 citations
Additional excerpts
...sidual error iven signal u ples [14]....
[...]
38 citations
"A compressive sensing based compres..." refers methods in this paper
...All of the traditional algorithms, such as Optimal Brain Damage (OBD)[16], Optimal Brain Surgeon (OBS)[17], and Magnitude-based pruning (MAG)[ 18], Skeletonization (SKEL)[6], non-contributing units (NC)[7] and Extended Fourier Amplitude Sensitiv ity Test (EFAST)[13], are available in SNNS (CSS1 is name of algorithm that uses SMV for sparse representation and CSS2 is another technique that uses MMV for sparserepresentatio n)....
[...]
14 citations
"A compressive sensing based compres..." refers background in this paper
...If the distances between source and microphones are not far enough, then time delay of the received signal between the reference microphone and the − h microphone would be [15] fig....
[...]
...So, the time delay of the received signal between the reference microphone and the − h microphone would be [15]: = ( − 1) = ( − 1) (1)...
[...]
...2: = ( ) (( ) ) (2) where, r is the distance between source and the first (reference) microphone [15]....
[...]
10 citations
"A compressive sensing based compres..." refers methods in this paper
...Before we formulate the problem of network pruning as a compressive sampling problem we introduce some definitions [ 11 , 10]:...
[...]
4 citations
"A compressive sensing based compres..." refers background in this paper
...ethods for c gories: 1) we D) [10], O nitude-based ing, e....
[...]
...: Sk s (NC) [10] a (EFAST) [2]...
[...]