Skip to main content

Oblique linear tree

  • Conference paper
  • First Online:
Advances in Intelligent Data Analysis Reasoning about Data (IDA 1997)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 1280))

Included in the following conference series:

Abstract

In this paper we present system Ltree for proposicional supervised learning. Ltree is able to define decision surfaces both orthogonal and oblique to the axes defined by the attributes of the input space. This is done combining a decision tree with a linear discriminant by means of constructive induction. At each decision node Ltree defines a new instance space by insertion of new attributes that are projections of the examples that fall at this node over the hyper-planes given by a linear discriminant function. This new instance space is propagated down through the tree. Tests based on those new attributes are oblique with respect to the original input space. Ltree is a probabilistic tree in the sense that it outputs a class probability distribution for each query example. The class probability distribution is computed at learning time, taking into account the different class distributions on the path from the root to the actual node. We have carried out experiments on sixteen benchmark datasets and compared our system with other well known decision tree systems (orthogonal and oblique) like C4.5, OC1 and LMDT. On these datasets we have observed that our system has advantages in what concerns accuracy and tree size at statistically significant confidence levels.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Breiman, L.: Bias, Variance and Arcing Classifiers, Technical Report 460, Statistics Department, University of California

    Google Scholar 

  2. Breiman, L., Friedman, J., Olshen, R., Stone, C.: Classification and Regression Trees, Wadsworth International Group, 1984.

    Google Scholar 

  3. Brodley, C., Utgoff, P.: Multivariate Decision Trees, in Machine Learning, 19, Kluwer Academic Press, 1995.

    Google Scholar 

  4. Buntime, W.: A theory of learning Classification rules, PhD thesis, University of Sydney, 1990

    Google Scholar 

  5. Dillon, W., Goldstein, M.: Multivariate analysis, Methods and Applications, John Willey & Sons, 1984

    Google Scholar 

  6. Esposito, F., Malerba, D., Semeraro, G.: Decision Tree Pruning as a Search is the State Space, in Machine Learning: ECML93, Ed. Pavel Brazdil, 1993

    Google Scholar 

  7. Henery, B.: FORTRAN programs for Discriminant Analysis, Internal report, Dep. Statistics and Modelling Science, University of Strathclyde, 1993

    Google Scholar 

  8. Jordan, M. Jacob, R.: Hierarchical mixtures of experts and the EM algorithm, Neural Computing, n.6, 1994

    Google Scholar 

  9. Kohavi, R., Wolpert, D.: Bias plus variance decomposition for zero-one loss functions, in Proceedings of 13 International Conference on Machine Learning — IML96, Ed. Lorenza Saitta, 1996

    Google Scholar 

  10. Loh W., Vanichsetakul N.: Tree-Structured Classification Via Generalized Discriminant Analysis, Journal of the American Statistical Association, 1988

    Google Scholar 

  11. Matheus,C., Rendell, L.: Constructive Induction on Decision Trees, in Proceedings of IJCAI 89, 1989

    Google Scholar 

  12. Michie, D., Spiegelhalter,J. Taylor,C.: Machine Learning, Neural and Statistical Classification, Ellis Horwood, 1994

    Google Scholar 

  13. Murthy, S., Kasif, S., Salzberg, S.: A system for Induction of Oblique Decision Trees, Journal of Artificial Intelligence Research, 1994

    Google Scholar 

  14. Press, W., Teukolsky, S., Vetterling, W. and Flannery, B.: Numerical Recipes in C: the art of scientific computing, 2 Ed. University of Cambridge, 1992

    Google Scholar 

  15. Quinlan R.: Learning with continuous classes, in Proceedings of AI92

    Google Scholar 

  16. Quinlan, R.: C4.5: Programs for Machine Learning, Morgan Kaufmann, 1993

    Google Scholar 

  17. Wolpert, D.: Stacked Generalisation, Neural Networks Vol.5, 1992

    Google Scholar 

  18. Yip, S., Webb, G.: Incorporating canonical discriminant attributes in classification learning, in Proceedings of the tenth Canadian Conference on Artificial Intelligence, Morgan Kaufmann, 1994

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Xiaohui Liu Paul Cohen Michael Berthold

Rights and permissions

Reprints and permissions

Copyright information

© 1997 Springer-Verlag

About this paper

Cite this paper

Gama, J. (1997). Oblique linear tree. In: Liu, X., Cohen, P., Berthold, M. (eds) Advances in Intelligent Data Analysis Reasoning about Data. IDA 1997. Lecture Notes in Computer Science, vol 1280. Springer, Berlin, Heidelberg. https://doi.org/10.1007/BFb0052840

Download citation

  • DOI: https://doi.org/10.1007/BFb0052840

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-63346-4

  • Online ISBN: 978-3-540-69520-2

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics