Indefinite proximity learning: A review

Research output: Contribution to journalArticle

Standard

Indefinite proximity learning : A review. / Schleif, Frank Michael; Tino, Peter.

In: Neural Computation, Vol. 27, No. 10, 21.10.2015, p. 2039-2096.

Research output: Contribution to journalArticle

Harvard

APA

Vancouver

Author

Bibtex

@article{5b3c8f082f0749bba5ec0ce99ca70160,
title = "Indefinite proximity learning: A review",
abstract = "Efficient learning of a data analysis task strongly depends on the data representation. Most methods rely on (symmetric) similarity or dissimilarity representations by means of metric inner products or distances, providing easy access to powerful mathematical formalisms like kernel or branch-and-bound approaches. Similarities and dissimilarities are, however, often naturally obtained by nonmetric proximity measures that cannot easily be handled by classical learning algorithms. Major efforts have been undertaken to provide approaches that can either directly be used for such data or to make standard methods available for these types of data. We provide a comprehensive survey for the field of learning with nonmetric proximities. First, we introduce the formalism used in nonmetric spaces and motivate specific treatments for nonmetric proximity data. Second, we provide a systematization of the various approaches. For each category of approaches, we provide a comparative discussion of the individual algorithms and address complexity issues and generalization properties. In a summarizing section, we provide a larger experimental study for the majority of the algorithms on standard data sets. We also address the problem of large-scale proximity learning, which is often overlooked in this context and of major importance to make the method relevant in practice. The algorithms we discuss are in general applicable for proximity-based clustering, one-class classification, classification, regression, and embedding approaches. In the experimental part, we focus on classification tasks.",
author = "Schleif, {Frank Michael} and Peter Tino",
year = "2015",
month = oct,
day = "21",
doi = "10.1162/NECO_a_00770",
language = "English",
volume = "27",
pages = "2039--2096",
journal = "Neural Computation",
issn = "0899-7667",
publisher = "Massachusetts Institute of Technology Press",
number = "10",

}

RIS

TY - JOUR

T1 - Indefinite proximity learning

T2 - A review

AU - Schleif, Frank Michael

AU - Tino, Peter

PY - 2015/10/21

Y1 - 2015/10/21

N2 - Efficient learning of a data analysis task strongly depends on the data representation. Most methods rely on (symmetric) similarity or dissimilarity representations by means of metric inner products or distances, providing easy access to powerful mathematical formalisms like kernel or branch-and-bound approaches. Similarities and dissimilarities are, however, often naturally obtained by nonmetric proximity measures that cannot easily be handled by classical learning algorithms. Major efforts have been undertaken to provide approaches that can either directly be used for such data or to make standard methods available for these types of data. We provide a comprehensive survey for the field of learning with nonmetric proximities. First, we introduce the formalism used in nonmetric spaces and motivate specific treatments for nonmetric proximity data. Second, we provide a systematization of the various approaches. For each category of approaches, we provide a comparative discussion of the individual algorithms and address complexity issues and generalization properties. In a summarizing section, we provide a larger experimental study for the majority of the algorithms on standard data sets. We also address the problem of large-scale proximity learning, which is often overlooked in this context and of major importance to make the method relevant in practice. The algorithms we discuss are in general applicable for proximity-based clustering, one-class classification, classification, regression, and embedding approaches. In the experimental part, we focus on classification tasks.

AB - Efficient learning of a data analysis task strongly depends on the data representation. Most methods rely on (symmetric) similarity or dissimilarity representations by means of metric inner products or distances, providing easy access to powerful mathematical formalisms like kernel or branch-and-bound approaches. Similarities and dissimilarities are, however, often naturally obtained by nonmetric proximity measures that cannot easily be handled by classical learning algorithms. Major efforts have been undertaken to provide approaches that can either directly be used for such data or to make standard methods available for these types of data. We provide a comprehensive survey for the field of learning with nonmetric proximities. First, we introduce the formalism used in nonmetric spaces and motivate specific treatments for nonmetric proximity data. Second, we provide a systematization of the various approaches. For each category of approaches, we provide a comparative discussion of the individual algorithms and address complexity issues and generalization properties. In a summarizing section, we provide a larger experimental study for the majority of the algorithms on standard data sets. We also address the problem of large-scale proximity learning, which is often overlooked in this context and of major importance to make the method relevant in practice. The algorithms we discuss are in general applicable for proximity-based clustering, one-class classification, classification, regression, and embedding approaches. In the experimental part, we focus on classification tasks.

UR - http://www.scopus.com/inward/record.url?scp=84941882719&partnerID=8YFLogxK

U2 - 10.1162/NECO_a_00770

DO - 10.1162/NECO_a_00770

M3 - Article

AN - SCOPUS:84941882719

VL - 27

SP - 2039

EP - 2096

JO - Neural Computation

JF - Neural Computation

SN - 0899-7667

IS - 10

ER -