Skip to main content
Log in

An incremental ensemble of classifiers

  • Published:
Artificial Intelligence Review Aims and scope Submit manuscript

Abstract

Along with the increase of data and information, incremental learning ability turns out to be more and more important for machine learning approaches. The online algorithms try not to remember irrelevant information instead of synthesizing all available information (as opposed to classic batch learning algorithms). Today, combining classifiers is proposed as a new road for the improvement of the classification accuracy. However, most ensemble algorithms operate in batch mode. For this reason, we propose an incremental ensemble that combines five classifiers that can operate incrementally: the Naive Bayes, the Averaged One-Dependence Estimators (AODE), the 3-Nearest Neighbors, the Non-Nested Generalised Exemplars (NNGE) and the Kstar algorithms using the voting methodology. We performed a large-scale comparison of the proposed ensemble with other state-of-the-art algorithms on several datasets and the proposed method produce better accuracy in most cases.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  • Auer P, Warmuth M (1998) Tracking the best disjunction. Mach Learn 32: 127–150, Kluwer Academic Publishers

    Google Scholar 

  • Chu F, Zaniolo C (2004) Fast and light boosting for adaptive mining of data streams. In: Advances in knowledge discovery and data mining, Lecture notes in computer science 3056. pp 282–292

  • Cleary JG, Trigg LE (1995) K*: an instance-based learner using an entropic distance measure. In: 12th international conference on machine learning. pp 108–114

  • Cohen W (1995) Fast effective rule induction. In: Proceedings of International Conference of ML-95. pp 115–123

  • Dietterich T. (2000) Ensemble Methods in Machine Learning, Lecture Notes in Computer Science, vol 1857, pp 1–15

  • Domingos P, Pazzani M (1997) On the optimality of the simple Bayesian classifier under zero-one loss. Mach Learn 29: 103–130

    Article  MATH  Google Scholar 

  • Fan W, Stolfo S, Zhang J (1999) The application of AdaBoost for distributed, scalable and on-line learning. In: Proceedings of the fifth ACM SIGKDD international conference on knowledge discovery and data mining. ACM Press, New York, pp 362–366

  • Fern A, Givan R (2000) Online ensemble learning: an empirical study. In: Proceedings of the seventeenth international conference on ML. Morgan Kaufmann, pp 279–286

  • Frank A, Asuncion A (2010) UCI Machine learning repository. [http://archive.ics.uci.edu/ml]. Irvine, CA: University of California, School of Information and Computer Science

  • Freund Y, Schapire R (1999) Large margin classification using the perceptron algorithm. Mach Learn 37: 277–296. Kluwer Academic Publishers

    Google Scholar 

  • Gangardiwala, A.; Polikar, R.; Dynamically weighted majority voting for incremental learning and comparison of three boosting based approaches, 2005 IEEE international joint conference on neural networks, IJCNN ’05, vol 2, pp. 1131–1136, 31 July–4 Aug 2005

  • Janssens D, Brijs T, Vanhoof K, Wets G (2006) Evaluating the performance of cost-based discretization versus entropy- and error-based discretization. Comput Oper Res 33(11): 3107–3123

    Article  MATH  Google Scholar 

  • Katagiri S, Abe S (2006) Incremental training of support vector machines using hyperspheres. Pattern Recognit Lett 27(13): 1495–1507

    Article  Google Scholar 

  • Kotsiantis S, Zaharakis I, intelas P (2006) Machine learning: a review of classification and combining techniques. Artificial Intell Rev 26(3): 159–190 (Springer)

    Article  Google Scholar 

  • Kuncheva LI (2004) Classifier ensembles for changing environments. In: Multiple classifier systems (MCS 2004), Lecture notes in computer science 3077. pp 1–15

  • Littlestone N, Warmuth M (1994) The weighted majority algorithm. Inf Comput 108: 212–261

    Article  MathSciNet  MATH  Google Scholar 

  • Menahem E, Rokach L, Elovici Y (2009) Troika—an improved stacking schema for classification tasks. Inf Sci. doi:10.1016/j.ins.2009.08.025

  • Oza NC, Russell S (2001) Online bagging and boosting. In: Richardson T, Jaakkola T (eds) Artificial intelligence and statistics. pp 105–112

  • Quinlan JR (1993) C4.5: programs for machine learning. Morgan Kaufmann, San Francisco

    Google Scholar 

  • Rokach L (2009) Taxonomy for characterizing ensemble methods in classification tasks: a review and annotated bibliography. Comput Statist Data Anal 53(12): 4046–4072

    Article  MathSciNet  MATH  Google Scholar 

  • Rokach L (2010) Ensemble-based classifiers. Artif Intell Rev 33(1–2): 1–39

    Article  Google Scholar 

  • Roy S (2002) Nearest neighbor with generalization. Christchurch, New Zealand

    Google Scholar 

  • Saad D (1998) Online learning in neural networks. Cambridge University Press, London

    Google Scholar 

  • Sahami M (1996) Learning limited dependence Bayesian classifiers. In: Proceedings of the 2nd international conference on knowledge discovery in databases. pp 335–338

  • Salzberg S (1997) On comparing classifiers: pitfalls to avoid and a recommended approach. Data Min Knowl Discov 1: 317–328

    Article  Google Scholar 

  • Ulaş A, Semerci M, Yıldız O T, Alpaydın E (2009) Incremental construction of classifier and discriminant ensembles. Inf Sci 179(9): 1298–1318

    Article  Google Scholar 

  • Utgoff P, Berkman N, Clouse J (1997) Decision tree induction based on efficient tree restructuring. Mach Learn 29: 5–44

    Article  MATH  Google Scholar 

  • Webb GI, Boughton JR, Wang Z (2005) Not so naive Bayes: aggregating one-dependence estimators. Mach Learn 58: 5–24

    Article  MATH  Google Scholar 

  • Widmer G, Kubat M (1996) Learning in the presence of concept drift and hidden contexts. Mach Learn 23: 69–101

    Google Scholar 

  • Witten I, Frank E (2005) Data mining: practical machine learning tools and techniques. 2. Morgan Kaufmann, San Francisco

    MATH  Google Scholar 

  • Wu Y, Ianakiev K, Govindaraju V (2002) Improved k-nearest neighbor classification. Pattern Recognit 35(10): 2311–2318

    Article  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to S. B. Kotsiantis.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Kotsiantis, S.B. An incremental ensemble of classifiers. Artif Intell Rev 36, 249–266 (2011). https://doi.org/10.1007/s10462-011-9211-4

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10462-011-9211-4

Keywords

Navigation