Log in

Linear low-rank approximation and nonlinear dimensionality reduction

  • Published:
Science in China Series A: Mathematics Aims and scope Submit manuscript

Abstract

We present our recent work on both linear and nonlinear data reduction methods and algorithms: for the linear case we discuss results on structure analysis of SVD of columnpartitioned matrices and sparse low-rank approximation; for the nonlinear case we investigate methods for nonlinear dimensionality reduction and manifold learning. The problems we address have attracted great deal of interest in data mining and machine learning.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
EUR 32.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or Ebook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Bishop, M., Svensen, M., Williams, K. I., GTM: the generative topographie map**, Neural Computation, 1998, 10: 215–234.

    Article  Google Scholar 

  2. Preedman, D., Efficient simplicial reconstructions of manifolds from their samples, IEEE Transactions on Pattern Analysis and Machine Intelligence, 2002, 24: 1349–1357.

    Article  Google Scholar 

  3. Hinton, G., Roweis, S., Stochastic neighbor embedding, Neural Information Processing Systems, 2003, 15: 833–840.

    Google Scholar 

  4. Kohonen, T., Self-organizing Maps, 3rd ed., Berlin: Springer-Verlag, 2000.

    Google Scholar 

  5. Ramsay, J. Silverman, W., Applied Functional Data Analysis, Berlin: Springer-Verlag, 2002.

    MATH  Google Scholar 

  6. Roweis, S., Saul, L., Nonlinear dimensionality reduction by locally linear embedding, Science, 2000, 290: 2323–2326.

    Article  Google Scholar 

  7. Tenenbaum, J., De Silva, V., Langford, J., A global geometric framework for nonlinear dimension reduction, Science, 2000, 290: 2319–2323.

    Article  Google Scholar 

  8. Xu, G., Kailath, T., Fast subspace decompsotion, IEEE Transactions on Signal Processing, 1994, 42: 539–551.

    Article  Google Scholar 

  9. Xu, G., Zha, H., Golub, G. et al., Fast algorithms for updating signal subspaces, IEEE Transactions on Circuits and Systems, 1994, 41: 537–549.

    Article  MATH  Google Scholar 

  10. Zha, H., Marques, O., Simon, H., Large-scale SVD and subspace-based methods for information retrieval, Proceedings of Irregular ’98, Lecture Notes in Computer Science, 1998, 1457: 29–42.

    Article  MathSciNet  Google Scholar 

  11. Zhang, Z., Zha, H., Structure and perturbation analysis of truncated SVDs for column-partitioned matrices, SIAM Journal on Matrix Analysis and Applications, 2001, 22: 1245–1262.

    Article  MATH  MathSciNet  Google Scholar 

  12. Zhang, Z., Zha, H., Simon, H., Low-rank approximations with sparse factors I: basic algorithms and error analysis, SIAM Journal on Matrix Analysis and Applications, 2002, 23: 706–727.

    Article  MATH  MathSciNet  Google Scholar 

  13. Stewart, G. W., Four algorithms for the efficient computation of truncated pivoted QR approximation to a sparse matrix, Numerische Mathematik, 1999, 83: 313–323.

    Article  MATH  MathSciNet  Google Scholar 

  14. Golub, G., Van Loan, C., Matrix Computations, 3nd ed., Baltimore, Maryland: Johns Hopkins University Press, 1996.

    MATH  Google Scholar 

  15. Teh, Y., Roweis, S., Automatic alignment of hidden representations, Neural Information Processing Systems, 2002, 15: 841–848.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Zhenyue Zhang.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Zhang, Z., Zha, H. Linear low-rank approximation and nonlinear dimensionality reduction. Sci. China Ser. A-Math. 47, 908–920 (2004). https://doi.org/10.1360/04ys0016

Download citation

  • Received:

  • Revised:

  • Issue Date:

  • DOI: https://doi.org/10.1360/04ys0016

Keywords

Navigation