Selective negative correlation learning approach to incremental learning

K Tang, M Lin, Leandro Minku, Xin Yao

Research output: Contribution to journalArticle

23 Citations (Scopus)

Abstract

Negative correlation learning (NCL) is a successful approach to constructing neural network ensembles. In batch learning mode, NCL outperforms many other ensemble learning approaches. Recently, NCL has also shown to be a potentially powerful approach to incremental learning, while the advantages of NCL have not yet been fully exploited. In this paper, we propose a selective NCL (SNCL) algorithm for incremental learning. Concretely, every time a new training data set is presented, the previously trained neural network ensemble is cloned. Then the cloned ensemble is trained on the new data set. After that, the new ensemble is combined with the previous ensemble and a selection process is applied to prune the whole ensemble to a fixed size. This paper is an extended version of our preliminary paper on SNCL. Compared to the previous work, this paper presents a deeper investigation into SNCL, considering different objective functions for the selection process and comparing SNCL to other NCL-based incremental learning algorithms on two more real world bioinformatics data sets. Experimental results demonstrate the advantage of SNCL Further, comparisons between SNCL and other existing incremental learning algorithms, such Learn + + and ARTMAP are also presented. (C) 2009 Elsevier B.V. All rights reserved.
Original languageEnglish
Pages (from-to)2796-2805
Number of pages10
JournalNeurocomputing
Volume72
Issue number13-15
DOIs
Publication statusPublished - 1 Aug 2009

Keywords

  • Selective ensemble
  • Negative correlation learning
  • Incremental learning
  • Neural network ensemble

Fingerprint

Dive into the research topics of 'Selective negative correlation learning approach to incremental learning'. Together they form a unique fingerprint.

Cite this