A review of Nyström methods for large-scale machine learning

  • Shiliang Sun*
  • , Jing Zhao
  • , Jiang Zhu
  • *Corresponding author for this work

Research output: Contribution to journalReview articlepeer-review

60 Scopus citations

Abstract

Generating a low-rank matrix approximation is very important in large-scale machine learning applications. The standard Nyström method is one of the state-of-the-art techniques to generate such an approximation. It has got rapid developments since being applied to Gaussian process regression. Several enhanced Nyström methods such as ensemble Nyström, modified Nyström and SS-Nyström have been proposed. In addition, many sampling methods have been developed. In this paper, we review the Nyström methods for large-scale machine learning. First, we introduce various Nyström methods. Second, we review different sampling methods for the Nyström methods and summarize them from the perspectives of both theoretical analysis and practical performance. Then, we list several typical machine learning applications that utilize the Nyström methods. Finally, we make our conclusions after discussing some open machine learning problems related to Nyström methods.

Original languageEnglish
Pages (from-to)36-48
Number of pages13
JournalInformation Fusion
Volume26
DOIs
StatePublished - 16 May 2015

Keywords

  • Low-rank approximation
  • Machine learning
  • Nyström method
  • Sampling method

Fingerprint

Dive into the research topics of 'A review of Nyström methods for large-scale machine learning'. Together they form a unique fingerprint.

Cite this