Linear convergence of proximal gradient method for linear sparse SVM

  • Xiaoqi Jiao
  • , Heng Lian*
  • , Jiamin Liu
  • , Yingying Zhang
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Despite the hinge loss function being non-strongly-convex and non-strongly smooth, we establish the linear rate of convergence for sparse linear support vector machines (SVM) up to its statistical accuracy. The algorithm we use is the proximal gradient method for composite functions, applied to a sequence of regularization parameters to compute the approximate solution path on a grid. Unlike works on loss functions that are strongly convex and strongly smooth, here we do not have linear convergence to the exact solution, but we can demonstrate linear convergence to the population truth up to the statistical error (in particular, we simultaneously consider numerical convergence and statistical convergence). For any regularization parameter in the chosen decreasing sequence, we show that the estimator is in a small neighborhood of the exact solution after O(logs*) iterations, where s* is the sparsity of the true coefficient in the model, and a total number of O(logn) stages (i.e., using a sequence of regularization parameters of length O(logn)) are required to achieve the near-oracle statistical rate, with n the sample size.

Original languageEnglish
Article number108162
JournalNeural Networks
Volume194
DOIs
StatePublished - Feb 2026

Keywords

  • Geometric/Linear convergence
  • Proximal gradient descent
  • Sparsity
  • Support vector machines

Fingerprint

Dive into the research topics of 'Linear convergence of proximal gradient method for linear sparse SVM'. Together they form a unique fingerprint.

Cite this