The benefits of modeling slack variables in SVMs

Fengzhen Tang, Peter Tiňo, Pedro Antonio Gutiérrez, Huanhuan Chen

Research output: Contribution to journalArticlepeer-review

3 Citations (Scopus)

Abstract

In this letter, we explore the idea of modeling slack variables in support vector machine (SVM) approaches. The study is motivated by SVM+, which models the slacks through a smooth correcting function that is determined by additional (privileged) information about the training examples not available in the test phase. We take a closer look at the meaning and consequences of smooth modeling of slacks, as opposed to determining them in an unconstrained manner through the SVM optimization program. To better understand this difference we only allow the determination and modeling of slack values on the same information- that is, using the same training input in the original input space.We also explore whether it is possible to improve classification performance by combining (in a convex combination) the original SVM slacks with the modeled ones.We show experimentally that this approach not only leads to improved generalization performance but also yields more compact, lower-complexity models. Finally, we extend this idea to the context of ordinal regression, where a natural order among the classes exists. The experimental results confirm principal findings from the binary case.

Original languageEnglish
Pages (from-to)954-981
Number of pages28
JournalNeural Computation
Volume27
Issue number4
DOIs
Publication statusPublished - 19 Apr 2015

ASJC Scopus subject areas

  • Cognitive Neuroscience
  • Arts and Humanities (miscellaneous)

Fingerprint

Dive into the research topics of 'The benefits of modeling slack variables in SVMs'. Together they form a unique fingerprint.

Cite this