Multiple and single snapshot compressive beamforming
March 08, 2015 Β· Declared Dead Β· π Journal of the Acoustical Society of America
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Peter Gerstoft, Angeliki Xenaki, Christoph F. MecklenbrΓ€uker
arXiv ID
1503.02339
Category
math.ST
Cross-listed
cs.IT
Citations
165
Venue
Journal of the Acoustical Society of America
Last Checked
1 month ago
Abstract
For a sound field observed on a sensor array, compressive sensing (CS) reconstructs the direction-of-arrival (DOA) of multiple sources using a sparsity constraint. The DOA estimation is posed as an underdetermined problem by expressing the acoustic pressure at each sensor as a phase-lagged superposition of source amplitudes at all hypothetical DOAs. Regularizing with an $\ell_1$-norm constraint renders the problem solvable with convex optimization, and promoting sparsity gives high-resolution DOA maps. Here, the sparse source distribution is derived using maximum a posteriori (MAP) estimates for both single and multiple snapshots. CS does not require inversion of the data covariance matrix and thus works well even for a single snapshot where it gives higher resolution than conventional beamforming. For multiple snapshots, CS outperforms conventional high-resolution methods, even with coherent arrivals and at low signal-to-noise ratio. The superior resolution of CS is demonstrated with vertical array data from the SWellEx96 experiment for coherent multi-paths.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
π Similar Papers
In the same crypt β math.ST
R.I.P.
π»
Ghosted
R.I.P.
π»
Ghosted
An introduction to Topological Data Analysis: fundamental and practical aspects for data scientists
R.I.P.
π»
Ghosted
Minimax Optimal Procedures for Locally Private Estimation
R.I.P.
π»
Ghosted
Optimal Best Arm Identification with Fixed Confidence
R.I.P.
π»
Ghosted
Fast low-rank estimation by projected gradient descent: General statistical and algorithmic guarantees
R.I.P.
π»
Ghosted
User-friendly guarantees for the Langevin Monte Carlo with inaccurate gradient
Died the same way β π» Ghosted
R.I.P.
π»
Ghosted
Language Models are Few-Shot Learners
R.I.P.
π»
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
π»
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
π»
Ghosted