Clustered Nyström method for large scale manifold learning and dimension reduction

Kai Zhang, James T. Kwok

Research output: Contribution to journalArticlepeer-review

191 Scopus citations

Abstract

Kernel (or similarity) matrix plays a key role in many machine learning algorithms such as kernel methods, manifold learning, and dimension reduction. However, the cost of storing and manipulating the complete kernel matrix makes it infeasible for large problems. The Nystrm method is a popular sampling-based low-rank approximation scheme for reducing the computational burdens in handling large kernel matrices. In this paper, we analyze how the approximating quality of the Nystrm method depends on the choice of landmark points, and in particular the encoding powers of the landmark points in summarizing the data. Our (non-probabilistic) error analysis justifies a "clustered Nyström method" that uses the k-means clustering centers as landmark points. Our algorithm can be applied to scale up a wide variety of algorithms that depend on the eigenvalue decomposition of kernel matrix (or its variant), such as kernel principal component analysis, Laplacian eigenmap, spectral clustering, as well as those involving kernel matrix inverse such as least-squares support vector machine and Gaussian process regression. Extensive experiments demonstrate the competitive performance of our algorithm in both accuracy and efficiency.

Original languageEnglish
Article number5559473
Pages (from-to)1576-1587
Number of pages12
JournalIEEE Transactions on Neural Networks
Volume21
Issue number10
DOIs
StatePublished - Oct 2010
Externally publishedYes

Keywords

  • Dimension reduction
  • Nystrm method
  • eigenvalue decomposition
  • kernel matrix
  • low-rank approximation
  • manifold learning
  • sampling

Fingerprint

Dive into the research topics of 'Clustered Nyström method for large scale manifold learning and dimension reduction'. Together they form a unique fingerprint.

Cite this