Skip to main content

A Comparison of Three Voting Methods for Bagging with the MLEM2 Algorithm

  • Conference paper
Intelligent Data Engineering and Automated Learning – IDEAL 2010 (IDEAL 2010)

Abstract

This paper presents results of experiments on some data sets using bagging on the MLEM2 rule induction algorithm. Three different methods of ensemble voting, based on support (a non-democratic voting in which ensembles vote with their strengths), strength only (an ensemble with the largest strength decides to which concept a case belongs) and democratic voting (each ensemble has at most one vote) were used. Our conclusions are that though in most cases democratic voting was the best, it is not significantly better than voting based on support. The strength voting was the worst voting method.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Breiman, L.: Bagging predictors. Machine Learning 24, 123–140 (1996)

    MATH  MathSciNet  Google Scholar 

  2. Freund, Y., Schapire, R.E.: Experiments with a new boosting algorithm. In: Proceedings of the 13th International Conference on Machine Learning, pp. 148–156 (1996)

    Google Scholar 

  3. Breiman, L.: Random forests. Machine Learning 45, 5–32 (2001)

    Article  MATH  Google Scholar 

  4. Bauer, E., Kohavi, R.: An empirical comparison of voting classification algorithms: bagging, boosting, and variants. Machine Learning 36, 105–139 (1999)

    Article  Google Scholar 

  5. Blaszczynski, J., Stefanowski, J., Zajac, M.: Ensembles of abstaining classifiers based on rule sets. In: Proceedings of the International Symposium on Foundations of Intelligent Systems, pp. 382–391 (2009)

    Google Scholar 

  6. Kuncheva, L.I.: Combining Pattern Classifiers. Methods and Algorithms. John Wiley & Sons, Hoboken (2004)

    Book  MATH  Google Scholar 

  7. Stefanowski, J.: The bagging and n 2-classifiers based on rules induced by MODLEM. In: Proceedings of the Fourth International Conference on Rough Sets and Current Trends in Computing, pp. 488–497 (2004)

    Google Scholar 

  8. Stefanowski, J.: On combined classifiers, rule induction and rough sets. Transactions on Rough Sets 6, 329–350 (2007)

    Article  Google Scholar 

  9. Zenko, B., Todorovski, L., Dzeroski, S.: On comparison of stacking with MDTs to bagging, boosting, and other stacking methods. In: Proceedings of the ECML/PKDD 01 Workshop on Integrating Aspects of Data Mining, Decision Support and Meta-Learning, pp. 163–175 (2001)

    Google Scholar 

  10. Wolpert, D.: Stacked generalization. Neural Networks 5, 241–260 (1992)

    Article  Google Scholar 

  11. Gama, J.: Combining classifiers by constructive induction. In: Proceedings of the 10th European Conference on Machine Learning, pp. 178–189 (1998)

    Google Scholar 

  12. Hall, L.O., Bowyer, K.W., Banfield, R.E., Bhadoria, D., Kegelmeyer, W.P., Eschrich, S.: Comparing pure parallel ensemble creation techniques against bagging. In: Proceedings of the IEEE International Conference on Data Mining, pp. 533–536 (2003)

    Google Scholar 

  13. Dietterich, T.G.: An experimental comparison of three methods for constructing ensembles of decision trees: bagging, boosting, and randomization. Machine Learning 40, 139–157 (2000)

    Article  Google Scholar 

  14. Pawlak, Z.: Rough Sets. Theoretical Aspects of Reasoning about Data. Kluwer Academic Publishers, Dordrecht (1991)

    MATH  Google Scholar 

  15. Chan, C.C., Grzymala-Busse, J.W.: On the attribute redundancy and the learning programs ID3, PRISM, and LEM2. Technical report, Department of Computer Science, University of Kansas (1991)

    Google Scholar 

  16. Grzymala-Busse, J.W.: A new version of the rule induction system LERS. Fundamenta Informaticae 31, 27–39 (1997)

    MATH  Google Scholar 

  17. Grzymala-Busse, J.W.: MLEM2: A new algorithm for rule induction from imperfect data. In: Proceedings of the 9th International Conference on Information Processing and Management of Uncertainty in Knowledge-Based Systems, pp. 243–250 (2002)

    Google Scholar 

  18. Holland, J.H., Holyoak, K.J., Nisbett, R.E.: Induction. Processes of Inference, Learning, and Discovery. MIT Press, Boston (1986)

    Google Scholar 

  19. Chmielewski, M.R., Grzymala-Busse, J.W.: Global discretization of continuous attributes as preprocessing for machine learning. International Journal of Approximate Reasoning 15(4), 319–331 (1996)

    Article  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2010 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Cohagan, C., Grzymala-Busse, J.W., Hippe, Z.S. (2010). A Comparison of Three Voting Methods for Bagging with the MLEM2 Algorithm. In: Fyfe, C., Tino, P., Charles, D., Garcia-Osorio, C., Yin, H. (eds) Intelligent Data Engineering and Automated Learning – IDEAL 2010. IDEAL 2010. Lecture Notes in Computer Science, vol 6283. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-15381-5_15

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-15381-5_15

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-15380-8

  • Online ISBN: 978-3-642-15381-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics