Skip to main content
Log in

Randomized Algorithms for Low-Rank Matrix Factorizations: Sharp Performance Bounds

  • Published:
Algorithmica Aims and scope Submit manuscript

Abstract

The development of randomized algorithms for numerical linear algebra, e.g. for computing approximate QR and SVD factorizations, has recently become an intense area of research. This paper studies one of the most frequently discussed algorithms in the literature for dimensionality reduction—specifically for approximating an input matrix with a low-rank element. We introduce a novel and rather intuitive analysis of the algorithm in [6], which allows us to derive sharp estimates and give new insights about its performance. This analysis yields theoretical guarantees about the approximation error and at the same time, ultimate limits of performance (lower bounds) showing that our upper bounds are tight. Numerical experiments complement our study and show the tightness of our predictions compared with empirical observations.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2

Similar content being viewed by others

Notes

  1. To accommodate this, previous works also provide bounds in terms of the singular values of \(A\) past \(\sigma _{k+1}\).

  2. This follows from Hölder’s inequality \(\mathbb {E}|XY| \le (\mathbb {E}|X|^{3/2})^{2/3} (\mathbb {E}|Y|^3)^{1/3}\) with \(X = g^{2/3},\,Y = g^{4/3}\).

References

  1. Chen, Z., Dongarrar, J.J.: Condition numbers of gaussian random matrices. SIAM J. Matrix Anal. Appl. 27, 603–620 (2005)

    Article  MathSciNet  Google Scholar 

  2. Geman, S.: A limit theorem for the norm of random matrices. Ann. Probab. 8, 252–261 (1980)

    Article  MATH  MathSciNet  Google Scholar 

  3. Halko, N., Martinsson, P.-G., Tropp, J.A.: Finding structure with randomness probabilistic algorithms for constructing approximate matrix decompositions. SIAM Rev. 53, 217–288 (2011)

    Article  MATH  MathSciNet  Google Scholar 

  4. Horn, R.A., Johnson, C.R.: Matrix analysis. Cambridge University Press, Cambridge (1985)

    Book  MATH  Google Scholar 

  5. Kent, J.T., Mardia, K.V., Bibby, J.M.: Multivariate analysis. Academic Press, New York (1976)

    Google Scholar 

  6. Martinsson, P.-G., Rokhlin, V., Tygert, M.: A randomized algorithm for the decomposition of matrices. Appl. Comput. Harmon. Anal. 30, 47–68 (2011)

    Article  MATH  MathSciNet  Google Scholar 

  7. Rokhlin, V., Szlam, A., Tygert, M.: A randomized algorithm for principal component analysis. SIAM J. Matrix Anal. Appl. 31, 1100–1124 (2009)

    Article  MathSciNet  Google Scholar 

  8. Rudelson, M., Vershynin, R.: Non-asymptotic theory of random matrices: extreme singular values. In: Proceedings of the International Congress of Mathematicians, pp. 1576–1602 (2010)

  9. Sarlós, T.: Proceedings of the 47th annual ieee symposium on foundations of computer science. In: Proceedings of the International Congress of Mathematicians, pp. 143–152 (2006)

  10. Silverstein, J.W.: The smallest eigenvalue of a large dimensional wishart matrix. Ann. Probab. 13, 1364–1368 (1985)

    Article  MATH  MathSciNet  Google Scholar 

  11. Woolfe, F., Liberty, E., Rokhlin, V., Tygert, M.: A fast randomized algorithm for the approximation of matrices. Appl. Comput. Harmon. Anal. 25, 335–366 (2008)

    Article  MATH  MathSciNet  Google Scholar 

Download references

Acknowledgments

E. C. is partially supported by NSF via grant CCF-0963835 and by a gift from the Broadcom Foundation. We thank Carlos Sing-Long for useful feedback about an earlier version of the manuscript. These results were presented in July 2013 at the European Meeting of Statisticians. We would like to thank the reviewers for useful comments.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Emmanuel Candès.

Appendix

Appendix

We use well-known bounds to control the expectation of the extremal singular values of a Gaussian matrix. These bounds are recalled in [8], though known earlier.

Lemma 4.1

If \(m>n\) and \(A\) is a \(m \times n\) matrix with i.i.d. \(\mathcal {N}(0,1)\) entries, then

$$\begin{aligned} \sqrt{m} - \sqrt{n} \le \mathbb {E}\sigma _{\min }(A)\le \mathbb {E}\sigma _{\max }(A) \le \sqrt{m} + \sqrt{n}. \end{aligned}$$

Next, we control the expectation of the norm of the pseudo-inverse \(A^\dagger \) of a Gaussian matrix \(A\).

Lemma 4.2

In the setup of Lemma 4.1, we have

$$\begin{aligned} \frac{1}{\sqrt{m-n}} \le \mathbb {E}\Vert A^\dagger \Vert \le e \frac{\sqrt{m}}{m-n}. \end{aligned}$$

Proof

The upper bound is the same as is used in [3] and follows from the work of [1]. For the lower bound, set \(B = (A^*A)^{-1}\) which has an inverse Wishart distribution, and observe that

$$\begin{aligned} \Vert A^\dagger \Vert ^2 = \Vert B\Vert \ge B_{11}, \end{aligned}$$

where \(B_{11}\) is the entry in the \((1,1)\) position. It is known that \(B_{11} \mathop {=}\limits ^{d} 1/Y\), where \(Y\) is distributed as a chi-square variable with \(d = m - n + 1\) degrees of freedom [5, Page 72]. Hence,

$$\begin{aligned} \mathbb {E}\Vert A^\dagger \Vert \ge \mathbb {E}\frac{1}{\sqrt{Y}} \ge \frac{1}{\sqrt{\mathbb {E}Y}} = \frac{1}{\sqrt{m-n+1}}. \end{aligned}$$

\(\square \)

The limit laws below are taken from [10] and [2].

Lemma 4.3

Let \(A_{m,n}\) be a sequence of \(m \times n\) matrix with i.i.d. \(\mathcal {N}(0,1)\) entries such that \(\lim _{n \rightarrow \infty } m/n = c \ge 1\). Then

$$\begin{aligned} \frac{1}{\sqrt{n}} \sigma _{\min }(A_{m,n})&\,{a.s. \over \rightarrow }\,\sqrt{c} - 1\\ \frac{1}{\sqrt{n}} \sigma _{\max }(A_{m,n})&\,{a.s. \over \rightarrow }\,\sqrt{c}+1. \end{aligned}$$

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Witten, R., Candès, E. Randomized Algorithms for Low-Rank Matrix Factorizations: Sharp Performance Bounds. Algorithmica 72, 264–281 (2015). https://doi.org/10.1007/s00453-014-9891-7

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00453-014-9891-7

Keywords

Navigation