Variable selection of Kolmogorov-Smirnov maximization with a penalized surrogate loss

Xiefang Lin, Fang Fang*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

3 Scopus citations

Abstract

Kolmogorov-Smirnov (KS) statistic is quite popular in many areas as the major performance evaluation criterion for binary classification due to its explicit business intension. Fang and Chen (2019) proposed a novel DMKS method that directly maximizes the KS statistic and compares favorably with the popular existing methods. However, DMKS did not consider the critical problem of variable selection since the special form of KS brings great challenge to establish the DMKS estimator's asymptotic distribution which is most likely to be nonstandard. This intractable issue is handled by introducing a surrogate loss function which leads to a n-consistent estimator for the true parameter up to a multiplicative scalar. Then a nonconcave penalty function is combined to achieve the variable selection consistency and asymptotical normality with the oracle property. Results of empirical studies confirm the theoretical results and show advantages of the proposed SKS (Surrogated Kolmogorov-Smirnov) method compared to the original DMKS method without variable selection.

Original languageEnglish
Article number107944
JournalComputational Statistics and Data Analysis
Volume195
DOIs
StatePublished - Jul 2024

Keywords

  • Binary classification
  • Credit scoring
  • Nonconcave penalty
  • Oracle property
  • Variable selection consistency

Fingerprint

Dive into the research topics of 'Variable selection of Kolmogorov-Smirnov maximization with a penalized surrogate loss'. Together they form a unique fingerprint.

Cite this