跳到主要导航 跳到搜索 跳到主要内容

Scaling up graph-based semisupervised learning via prototype vector machines

  • Kai Zhang
  • , Liang Lan
  • , James T. Kwok
  • , Slobodan Vucetic
  • , Bahram Parvin

科研成果: 期刊稿件文章同行评审

摘要

When the amount of labeled data are limited, semisupervised learning can improve the learner's performance by also using the often easily available unlabeled data. In particular, a popular approach requires the learned function to be smooth on the underlying data manifold. By approximating this manifold as a weighted graph, such graph-based techniques can often achieve state-of-the-art performance. However, their high time and space complexities make them less attractive on large data sets. In this paper, we propose to scale up graph-based semisupervised learning using a set of sparse prototypes derived from the data. These prototypes serve as a small set of data representatives, which can be used to approximate the graph-based regularizer and to control model complexity. Consequently, both training and testing become much more efficient. Moreover, when the Gaussian kernel is used to define the graph affinity, a simple and principled method to select the prototypes can be obtained. Experiments on a number of real-world data sets demonstrate encouraging performance and scaling properties of the proposed approach. It also compares favorably with models learned via ℓ1-regularization at the same level of model sparsity. These results demonstrate the efficacy of the proposed approach in producing highly parsimonious and accurate models for semisupervised learning.

源语言英语
文章编号6803073
页(从-至)444-457
页数14
期刊IEEE Transactions on Neural Networks and Learning Systems
26
3
DOI
出版状态已出版 - 1 3月 2015
已对外发布

指纹

探究 'Scaling up graph-based semisupervised learning via prototype vector machines' 的科研主题。它们共同构成独一无二的指纹。

引用此