False positive rate control for positive unlabeled learning

  • Shuchen Kong
  • , Weiwei Shen
  • , Yingbin Zheng
  • , Ao Zhang
  • , Jian Pu*
  • , Jun Wang
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

8 Scopus citations

Abstract

Learning classifiers with false positive rate control have drawn intensive attention in applications over past years. While various supervised algorithms have been developed for obtaining low false positive rates, they commonly require the coexistence of both positive and negative samples in data. However, the scenario studied in positive unlabeled (PU) learning is more pervasive in practice. Namely, at inception, most of the data may not have known labels, and the data with known labels may only represent one type of samples. To tackle this challenge, in this paper we propose a new positive unlabeled learning classifier with false positive rate control. In particular, we first prove that in this context employing oft-adopted convex surrogate loss functions, such as the hinge loss function, begets a redundant penalty for false positive rates. Then, we present that the non-convex ramp loss surrogate function can overcome this barrier and show a concave-convex procedure can solve the associated non-convex optimization problem. Finally, we demonstrate the effectiveness of the proposed method through extensive experiments on multiple datasets.

Original languageEnglish
Pages (from-to)13-19
Number of pages7
JournalNeurocomputing
Volume367
DOIs
StatePublished - 20 Nov 2019

Keywords

  • Concave-convex procedure
  • False positive rate control
  • Positive unlabeled learning

Fingerprint

Dive into the research topics of 'False positive rate control for positive unlabeled learning'. Together they form a unique fingerprint.

Cite this