A heuristic search algorithm for the multiple measurement vectors problem

In this paper, we address the multiple measurement vectors problem, which is now a hot topic in the compressed sensing theory and its various applications. We propose a novel heuristic search algorithm called HSAMMV to solve the problem, which is modeled as a combinatorial optimization. HSAMMV is proposed in the framework of simulated annealing algorithm. The main innovation is to take advantage of some greedy pursuit algorithms for designing the initial solution and the generating mechanism of HSAMMV. Compared with some state-of-the-art algorithms, the numerical simulation results illustrate that HSAMMV has strong global search ability and quite good recovery performance.

[1]  Philip Schniter,et al.  Efficient High-Dimensional Inference in the Multiple Measurement Vector Problem , 2011, IEEE Transactions on Signal Processing.

[2]  Bruce E. Hajek,et al.  Cooling Schedules for Optimal Annealing , 1988, Math. Oper. Res..

[3]  Bhaskar D. Rao,et al.  Sparse solutions to linear inverse problems with multiple measurement vectors , 2005, IEEE Transactions on Signal Processing.

[4]  Xinpeng Du,et al.  A reduced l 2-l 1 model with an alternating minimisation algorithm for support recovery of multiple measurement vectors , 2013, IET Signal Process..

[5]  Riccardo Poli,et al.  Particle swarm optimization , 1995, Swarm Intelligence.

[6]  Xinpeng Du,et al.  A Swarm Intelligence Algorithm for Joint Sparse Recovery , 2013, IEEE Signal Processing Letters.

[7]  Xinpeng Du,et al.  Three stochastic measurement schemes for direction-of-arrival estimation using compressed sensing method , 2014, Multidimens. Syst. Signal Process..

[8]  Shanhe Wang,et al.  A hybrid simulated annealing thresholding algorithm for compressed sensing , 2013 .

[9]  Dmitry M. Malioutov,et al.  A sparse signal reconstruction perspective for source localization with sensor arrays , 2005, IEEE Transactions on Signal Processing.

[10]  Jie Chen,et al.  Theoretical Results on Sparse Representations of Multiple-Measurement Vectors , 2006, IEEE Transactions on Signal Processing.

[11]  Yoram Bresler,et al.  Subspace Methods for Joint Sparse Recovery , 2010, IEEE Transactions on Information Theory.

[12]  Kathryn A. Dowsland,et al.  Simulated Annealing , 1989, Encyclopedia of GIS.

[13]  H. Rauhut,et al.  Atoms of All Channels, Unite! Average Case Analysis of Multi-Channel Sparse Recovery Using Greedy Algorithms , 2008 .

[14]  Donald Geman,et al.  Stochastic Relaxation, Gibbs Distributions, and the Bayesian Restoration of Images , 1984, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[15]  Zongben Xu,et al.  $L_{1/2}$ Regularization: A Thresholding Representation Theory and a Fast Solver , 2012, IEEE Transactions on Neural Networks and Learning Systems.

[16]  Stochastic Relaxation , 2014, Computer Vision, A Reference Guide.

[17]  Jong Chul Ye,et al.  Compressive MUSIC: Revisiting the Link Between Compressive Sensing and Array Signal Processing , 2012, IEEE Transactions on Information Theory.

[18]  Bruce L. Golden,et al.  Solving the traveling salesman problem with annealing-based heuristics: a computational study , 2002, IEEE Trans. Syst. Man Cybern. Part A.

[19]  I. Osman,et al.  Simulated annealing for permutation flow-shop scheduling , 1989 .

[20]  Riccardo Poli,et al.  Particle Swarm Optimisation , 2011 .

[21]  Olgica Milenkovic,et al.  Subspace Pursuit for Compressive Sensing Signal Reconstruction , 2008, IEEE Transactions on Information Theory.

[22]  Yuantao Gu,et al.  Retrieval of sparse solutions of multiple-measurement vectors via zero-point attracting projection , 2012, Signal Process..

[23]  Robert H. Halstead,et al.  Matrix Computations , 2011, Encyclopedia of Parallel Computing.

[24]  Yonina C. Eldar,et al.  Rank Awareness in Joint Sparse Recovery , 2010, IEEE Transactions on Information Theory.

[25]  Peter Rossmanith,et al.  Simulated Annealing , 2008, Taschenbuch der Algorithmen.

[26]  David L Donoho,et al.  Compressed sensing , 2006, IEEE Transactions on Information Theory.

[27]  Bruce E. Rosen,et al.  Genetic Algorithms and Very Fast Simulated Reannealing: A comparison , 1992 .

[28]  C. D. Gelatt,et al.  Optimization by Simulated Annealing , 1983, Science.

[29]  Zhen Li,et al.  Signal recovery from multiple measurement vectors via tunable random projection and boost , 2012, Signal Process..

[30]  Peter Salamon,et al.  Ensemble approach to simulated annealing , 1991 .

[31]  D. Janaki Ram,et al.  Parallel Simulated Annealing Algorithms , 1996, J. Parallel Distributed Comput..

[32]  T. Blumensath,et al.  Iterative Thresholding for Sparse Approximations , 2008 .

[33]  Michael Elad,et al.  Optimally sparse representation in general (nonorthogonal) dictionaries via ℓ1 minimization , 2003, Proceedings of the National Academy of Sciences of the United States of America.

[34]  Bhaskar D. Rao,et al.  Sparse Signal Recovery With Temporally Correlated Source Vectors Using Sparse Bayesian Learning , 2011, IEEE Journal of Selected Topics in Signal Processing.

[35]  Richard W. Eglese,et al.  Simulated annealing: A tool for operational research , 1990 .

[36]  Deanna Needell,et al.  CoSaMP: Iterative signal recovery from incomplete and inaccurate samples , 2008, ArXiv.

[37]  Emmanuel J. Candès,et al.  Decoding by linear programming , 2005, IEEE Transactions on Information Theory.

[38]  Stéphane Canu,et al.  $\ell_{p}-\ell_{q}$ Penalty for Sparse Linear and Sparse Multiple Kernel Multitask Learning , 2011, IEEE Transactions on Neural Networks.

[39]  Yonina C. Eldar,et al.  Reduce and Boost: Recovering Arbitrary Sets of Jointly Sparse Vectors , 2008, IEEE Transactions on Signal Processing.