Diversity-based pool of models for dealing with recurring concepts

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Authors

Colleges, School and Institutes

External organisations

  • University of Leicester

Abstract

Several data stream applications involve recurring concepts, i.e., concept drifts that change the underlying distribution of the data to a distribution previously seen in the data stream. Examples include electricity price prediction and tweet topic classification. In such scenario, it is useful to maintain a pool of old models that could be recovered if their knowledge matches the recurring concept well. A few existing online learning approaches maintain such pools. However, there has been a little investigation on what is the best strategy to maintain an online learning pool with a limited size. We propose to make use of diversity to decide which models to keep in the pool once the pool reaches the maximum size. The motivation behind is that a diverse pool is more likely to maintain a set of representative models with considerably different concepts, helping to handle recurring concepts. We perform experiments to investigate if, when and why maintaining a diverse pool is helpful. The results show that the use of diversity to maintain pools can indeed be helpful to handle recurring concepts. However, the relationship between diversity and accuracy in the presence of concept drift is not straightforward. In particular, an initially good accuracy obtained when using diversity can lead to a stronger subsequent drop in accuracy than other strategies.

Details

Original languageEnglish
Title of host publication2018 International Joint Conference on Neural Networks (IJCNN)
Publication statusPublished - 15 Oct 2018
Event2018 International Joint Conference on Neural Networks (IJCNN) - Rio de Janeiro, Brazil
Duration: 8 Jul 201813 Jul 2018

Publication series

NameInternational Joint Conference on Neural Networks (IJCNN)
PublisherIEEE
Volume2018
ISSN (Print)2161-4393
ISSN (Electronic)2161-4407

Conference

Conference2018 International Joint Conference on Neural Networks (IJCNN)
Abbreviated titleIJCNN 2018
CountryBrazil
CityRio de Janeiro
Period8/07/1813/07/18

Keywords

  • Predictive models, Data models, Error analysis, Training, Prediction algorithms, Probability distribution, Streaming media