Eigenvalues and condition numbers of random matrices
SIAM Journal on Matrix Analysis and Applications
Sparse Approximate Solutions to Linear Systems
SIAM Journal on Computing
Optimal Solutions for Sparse Principal Component Analysis
The Journal of Machine Learning Research
Subspace pursuit for compressive sensing signal reconstruction
IEEE Transactions on Information Theory
Generalized Power Method for Sparse Principal Component Analysis
The Journal of Machine Learning Research
On verifiable sufficient conditions for sparse signal recovery via ℓ 1 minimization
Mathematical Programming: Series A and B - Special Issue on "Optimization and Machine learning"; Alexandre d’Aspremont • Francis Bach • Inderjit S. Dhillon • Bin Yu
Decoding by linear programming
IEEE Transactions on Information Theory
IEEE Transactions on Information Theory
IEEE Transactions on Information Theory
Hi-index | 0.00 |
The restricted isometry constant (RIC) of a matrix $A$ measures how close to an isometry is the action of $A$ on vectors with few nonzero entries, measured in the $\ell^2$ norm. Specifically, the upper and lower RICs of a matrix $A$ of size $n\times N$ are the maximum and the minimum deviation from unity (one) of the largest and smallest, respectively, square of singular values of all ${N\choose k}$ matrices formed by taking $k$ columns from $A$. Calculation of the RIC is intractable for most matrices due to its combinatorial nature; however, many random matrices typically have bounded RIC in some range of problem sizes $(k,n,N)$. We provide the best known bound on the RIC for Gaussian matrices, which is also the smallest known bound on the RIC for any large rectangular matrix. Our results are built on the prior bounds of Blanchard, Cartis, and Tanner [SIAM Rev., to appear], with improvements achieved by grouping submatrices that share a substantial number of columns.