Skip to main content

Part of the book series: The Springer International Series in Engineering and Computer Science ((SECS,volume 454))

Abstract

With a unified model of feature selection, we are ready to discuss in detail different aspects of feature selection. The major aspects of feature selection are (1) search directions (feature subset generation), (2) search strategies, and (3) evaluation measures. The objective of this chapter is two-fold: (a) to study the various options for each aspect in a systematic and principled way and (b) to identify the essential and different characteristics of various feature selection systems.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 299.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 379.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 379.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  • Almuallim, H. and Dietterich, T. (1994). Learning boolean concepts in the presence of many irrelevant features. Artificial Intelligence, 69(1-2):279–305.

    Article  MathSciNet  MATH  Google Scholar 

  • Ben-Bassat, M. (1982). Pattern recognition and reduction of dimensionality. In Krishnaiah, P. R. and Kanal, L. N., editors, Handbook of statistics-II, pages 773–791. North Holland.

    Google Scholar 

  • Boddy, M. and Dean, T. (1994). Deliberation scheduling for problem solving in time-constrained environments. Artificial Intelligence, 67(2):245–285.

    Article  MATH  Google Scholar 

  • Catlett, J. (1991). On changing continuous attributes into ordered discrete attributes. In European Working Session on Learning.

    Google Scholar 

  • Cover, T. (1974). The best two independent measurements are not the two best. IEEE Trans. Systems, Man and Cybernictics, 4:116–117.

    MATH  Google Scholar 

  • Dash, M. and Liu, H. (1997). Feature selection methods for classifications. Intelligent Data Analysis: An International Journal, 1(3).

    Google Scholar 

  • Dietterich, T. (1997). Machine learning research: Four current directions. AI Magazine, pages 97–136.

    Google Scholar 

  • Domingos, P. (1997). Why does bagging work? a Bayesian account and its implications. In Proceedings of the Third International Conference on Knowledge Discovery and Data Mining, pages 155–158. AAAI Press.

    Google Scholar 

  • Goldberg, D. (1989). Genetic Algorithms in Search, Optimization and Machine Learning. Addison-Wesley Publishing Company, Inc.

    Google Scholar 

  • Jain, A. and Zongker, D. (1997). Feature selection: Evaluation, application, and small sample performance. IEEE Trans, on Pattern Analysis and Machine Intelligence, 19(2):153–158.

    Article  Google Scholar 

  • John, G., Kohavi, R., and Pfleger, K. (1994). Irrelevant feature and the subset selection problem. In Machine Learning: Proceedings of the Eleventh International Conference, pages 121–129. Morgan Kaufmann Publisher.

    Google Scholar 

  • Kerber, R. (1992). ChiMerge: Discretization of numeric attributes. In AAAI-92, Proceedings of the Ninth National Conference on Artificial Intelligence, pages 123–128. AAAI Press/The MIT Press.

    Google Scholar 

  • Kruse, R., Tondo, C., and Leung, B. (1997). Data structures & program design in C. International Edition.

    Google Scholar 

  • Liu, H. and Wen, W. (1993). Concept learning through feature selection. In Proceedings of the First Australian and New Zealand Conference on Intelligent Information Systems, pages 293–297.

    Google Scholar 

  • Mitch, T. (1997). Machine Learning. McGraw-Hill.

    Google Scholar 

  • Narendra, P. and Fukunaga, K. (1977). A branch and bound algorithm for feature subset selection. IEEE Trans, on Computer, C-26(9):917–922.

    Article  Google Scholar 

  • Press, W., Teukolsky, S., Vetterling, W., and Flannery, B. (1992). Numerical Recipes in C. Cambridge University Press, Cambridge.

    Google Scholar 

  • Quinlan, J. (1988). Decision trees and multi-values attributes. In J.E., H., Michie, D., and J., R., editors, Machine Intelligence, volume 11. Oxford University Press.

    Google Scholar 

  • Quinlan, J. (1993). C4.5: Programs for Machine Learning. Morgan Kaufmann.

    Google Scholar 

  • Setiono, R. and Liu, H. (1997). Neural network feature selectors. IEEE Trans. on Neural Networks, 8(3):654–662.

    Article  Google Scholar 

  • Siedlecki, W. and Sklansky, J. (1988). On automatic feature selection. International Journal of Pattern Recognition and Artificial Intelligence, 2:197–220.

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

Copyright information

© 1998 Springer Science+Business Media New York

About this chapter

Cite this chapter

Liu, H., Motoda, H. (1998). Feature Selection Aspects. In: Feature Selection for Knowledge Discovery and Data Mining. The Springer International Series in Engineering and Computer Science, vol 454. Springer, Boston, MA. https://doi.org/10.1007/978-1-4615-5689-3_3

Download citation

  • DOI: https://doi.org/10.1007/978-1-4615-5689-3_3

  • Publisher Name: Springer, Boston, MA

  • Print ISBN: 978-1-4613-7604-0

  • Online ISBN: 978-1-4615-5689-3

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics