Corpus ID: 237532382

On-the-Fly Ensemble Pruning in Evolving Data Streams

@article{Elbasi2021OntheFlyEP,
  title={On-the-Fly Ensemble Pruning in Evolving Data Streams},
  author={Sanem Elbasi and Alican Buyukccakir and Hamed Bonab and Fazli Can},
  journal={ArXiv},
  year={2021},
  volume={abs/2109.07611}
}
Ensemble pruning is the process of selecting a subset of component classifiers from an ensemble which performs at least as well as the original ensemble while reducing storage and computational costs. Ensemble pruning in data streams is a largely unexplored area of research. It requires analysis of ensemble components as they are running on the stream, and differentiation of useful classifiers from redundant ones. We present CCRP, an on-the-fly ensemble pruning method for multi-class dataโ€ฆย Expand

Figures and Tables from this paper

References

SHOWING 1-10 OF 19 REFERENCES
GOOWE: Geometrically Optimum and Online-Weighted Ensemble Classifier for Evolving Data Streams
TLDR
This work proposes a novel data stream ensemble classifier, called Geometrically Optimum and Online-Weighted Ensemble (GOOWE), which assigns optimum weights to the component classifiers using a sliding window containing the most recent data instances, and presents a novel, dynamic, and online weighting approach. Expand
Ensemble learning for data stream analysis: A survey
TLDR
This paper surveys research on ensembles for data stream classification as well as regression tasks and discusses advanced learning concepts such as imbalanced data streams, novelty detection, active and semi-supervised learning, complex data representations and structured outputs. Expand
Mining concept-drifting data streams using ensemble classifiers
TLDR
This paper proposes a general framework for mining concept-drifting data streams using weighted ensemble classifiers, and shows that the proposed methods have substantial advantage over single-classifier approaches in prediction accuracy, and the ensemble framework is effective for a variety of classification models. Expand
Efficient data stream classification via probabilistic adaptive windows
TLDR
A probabilistic adaptive window (PAW) for data-stream learning is introduced, which improves this windowing technique with a mechanism to include older examples as well as the most recent ones, thus maintaining information on past concept drifts while being able to adapt quickly to new ones. Expand
Less Is More: A Comprehensive Framework for the Number of Components of Ensemble Classifiers
  • Hamed Bonab, F. Can
  • Medicine, Computer Science
  • IEEE Transactions on Neural Networks and Learning Systems
  • 2019
TLDR
A proposed geometric framework provides a theoretical explanation of diversity and its impact on the accuracy of predictions, and proves the existence of an ideal number of components, which is equal to the number of class labels. Expand
Reacting to Different Types of Concept Drift: The Accuracy Updated Ensemble Algorithm
TLDR
A new data stream classifier, called the Accuracy Updated Ensemble (AUE2), which aims at reacting equally well to different types of drift, and combines accuracy-based weighting mechanisms known from block-based ensembles with the incremental nature of Hoeffding Trees. Expand
FIRE-DES++: Enhanced Online Pruning of Base Classifiers for Dynamic Ensemble Selection
TLDR
The FIRE-DES++ is proposed, an enhanced FIRE- DES that removes noise and reduces the overlap of classes in the validation set; and defines the region of competence using an equal number of samples of each class, avoiding selecting a region of Competence with samples of a single class. Expand
Issues in evaluation of stream learning algorithms
TLDR
This paper proposes a general framework for assessing predictive stream learning algorithms, and defends the use of Predictive Sequential methods for error estimate - the prequential error. Expand
KNN Classifier with Self Adjusting Memory for Heterogeneous Concept Drift
TLDR
The Self Adjusting Memory (SAM) model for the k Nearest Neighbor (kNN) algorithm is proposed since kNN constitutes a proven classifier within the streaming setting and it can be easily applied in practice since an optimization of the meta parameters is not necessary. Expand
One-class classifier ensemble pruning and weighting with firefly algorithm
  • B. Krawczyk
  • Mathematics, Computer Science
  • Neurocomputing
  • 2015
TLDR
Experimental analysis proves the quality of the proposed method and its ability to outperform state-of-the-art algorithms for selecting one-class classifiers for the classification committees. Expand
...
1
2
...