TY - JOUR
T1 - False positive rate control for positive unlabeled learning
AU - Kong, Shuchen
AU - Shen, Weiwei
AU - Zheng, Yingbin
AU - Zhang, Ao
AU - Pu, Jian
AU - Wang, Jun
N1 - Publisher Copyright:
© 2019 Elsevier B.V.
PY - 2019/11/20
Y1 - 2019/11/20
N2 - Learning classifiers with false positive rate control have drawn intensive attention in applications over past years. While various supervised algorithms have been developed for obtaining low false positive rates, they commonly require the coexistence of both positive and negative samples in data. However, the scenario studied in positive unlabeled (PU) learning is more pervasive in practice. Namely, at inception, most of the data may not have known labels, and the data with known labels may only represent one type of samples. To tackle this challenge, in this paper we propose a new positive unlabeled learning classifier with false positive rate control. In particular, we first prove that in this context employing oft-adopted convex surrogate loss functions, such as the hinge loss function, begets a redundant penalty for false positive rates. Then, we present that the non-convex ramp loss surrogate function can overcome this barrier and show a concave-convex procedure can solve the associated non-convex optimization problem. Finally, we demonstrate the effectiveness of the proposed method through extensive experiments on multiple datasets.
AB - Learning classifiers with false positive rate control have drawn intensive attention in applications over past years. While various supervised algorithms have been developed for obtaining low false positive rates, they commonly require the coexistence of both positive and negative samples in data. However, the scenario studied in positive unlabeled (PU) learning is more pervasive in practice. Namely, at inception, most of the data may not have known labels, and the data with known labels may only represent one type of samples. To tackle this challenge, in this paper we propose a new positive unlabeled learning classifier with false positive rate control. In particular, we first prove that in this context employing oft-adopted convex surrogate loss functions, such as the hinge loss function, begets a redundant penalty for false positive rates. Then, we present that the non-convex ramp loss surrogate function can overcome this barrier and show a concave-convex procedure can solve the associated non-convex optimization problem. Finally, we demonstrate the effectiveness of the proposed method through extensive experiments on multiple datasets.
KW - Concave-convex procedure
KW - False positive rate control
KW - Positive unlabeled learning
UR - https://www.scopus.com/pages/publications/85070515055
U2 - 10.1016/j.neucom.2019.08.001
DO - 10.1016/j.neucom.2019.08.001
M3 - 文章
AN - SCOPUS:85070515055
SN - 0925-2312
VL - 367
SP - 13
EP - 19
JO - Neurocomputing
JF - Neurocomputing
ER -