Between Two Extremes: Examining Decompositions of the Ensemble Objective Function

G Brown, Jeremy Wyatt, Ping Sun

Research output: Contribution to journalArticlepeer-review

4 Citations (Scopus)


We study how the error of an ensemble regression estimator can be decomposed into two components: one accounting for the individual errors and the other accounting for the correlations within the ensemble. This is the well known Ambiguity decomposition; we show an alternative way to decompose the error, and show how both decompositions have been exploited in a learning scheme. Using a scaling parameter in the decomposition we can blend the gradient (and therefore the learning process) smoothly between two extremes, from concentrating on individual accuracies and ignoring diversity, up to a full non-linear optimization of all parameters, treating the ensemble as a single learning unit. We demonstrate how this also applies to ensembles using a soft combination of posterior probability estimates, so can be utilised for classifier ensembles.
Original languageEnglish
Pages (from-to)296-305
Number of pages10
JournalLecture Notes in Computer Science
Publication statusPublished - 1 Jan 2005
Event6th International Workshop on Multiple Classifier Systems -
Duration: 1 Jan 2005 → …


Dive into the research topics of 'Between Two Extremes: Examining Decompositions of the Ensemble Objective Function'. Together they form a unique fingerprint.

Cite this