Skip to main content
Log in

Lightweight surrogate random forest support for model simplification and feature relevance

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

In this study, we propose a lightweight surrogate random forest (L-SRF) algorithm that can be interpreted through a new rule distillation method. The common surrogate models replace the existing heavy and deep but high-performance black box model using a teacher–student learning framework. However, the student model obtained in this way must maintain the performance of the teacher model, and thus the degree of model simplification and transparency is extremely limited. Therefore, to increase model transparency while maintaining the performance of the surrogate model, we propose two methods. First, we propose a cross-entropy Shapley value to evaluate the contribution of each rule in the student surrogate model. Second, a random mini-grouping method is devised to effectively distilless important rules while minimizing the overfitting problem caused by a model simplification. The proposed L-SRF based on a rule contribution has the advantage of improving the degree of simplification and transparency of the model by realizing the large distillation ratio against the initial SRF model. In addition, because the proposed L-SRF removes unnecessary rules, it is possible to minimize the loss of the importance and relevance of each feature. To demonstrate the superior performance of the proposed L-SRF method, several comparative experiments were conducted on various data sets. We proved experimentally that the proposed method achieves a more effective performance than black box AI models in terms of model transparency and memory requirement, as well as the interpretation of the feature relevance.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4

Similar content being viewed by others

Notes

  1. A shorter version of this paper was presented at the NeurIPS2020 Workshop.

References

  1. Adadi A, Berrada M (2018) Peeking inside the black-box: A survey on explainable artificial intelligence (XAI). IEEE Access 6:52138–52160

    Article  Google Scholar 

  2. Arrieta AB, et al. (2020) Explainable artificial intelligence (XAI): concepts, taxonomies, opportunities and challenges toward responsible AI. ELSEVIER Inf Fusion 58:82–115

    Article  Google Scholar 

  3. Tan S et al (2018) Distill-and-compare: auditing black-box models using transparent model distillation. In: 2018 AAAI/ACM conference on AI, ethics and society. pp 303–310

  4. Bastani O, Kim C, Bastani H. (2017) Interpretability via model extraction. arXiv:1706.09773

  5. Zagoruyko S, Komodakis N (2017) Paying more attention to attention: improving the performance of convolutional neural networks via attention transfer. In: ICLR, pp 1–11

  6. Xu K et al (2018) Interpreting deep classifier by visual distillation of dark knowledge. arXiv:1803.04042

  7. Kim S, Jeong M, Ko BC (2020) Interpretation and simplification of deep forest. TechRxiv, techrxiv. 11661246.v1

  8. Kim S, Boukouvala F (2020) Machine learning-based surrogate modeling for data-driven optimization: a comparison of subset selection for regression techniques. Springer Optim Lett 14:989–1010

    Article  MathSciNet  Google Scholar 

  9. Kim S, Jeong M, Ko BC (2020) Energy efficient pupil tracking based on rule distillation of cascade regression forest. MDPI Sensors 20:1–17

    Google Scholar 

  10. Kim S, Jeong M, Ko BC (2020) Is the surrogate model interpretable?. In: NeurIPS workshops. pp 1–5

  11. Kim SJ, Kwak SY, Ko BC (2019) Fast pedestrian detection in surveillance video based on soft target training of shallow random forest. IEEE ACCESS 7:12415–12426

    Article  Google Scholar 

  12. Breiman L (2001) Random forest. Springer Mach Learn 45:5–32

    Article  Google Scholar 

  13. Friedman J (2001) Greedy function approximation: a gradient boosting machine. Ann Stat 29:1189–1232

    Article  MathSciNet  Google Scholar 

  14. Chen T, Guestrin C (2016) Xgboost: A scalable tree boosting system. In: 22nd ACM SIGKDD International conference on knowledge discovery and data mining. pp 785–794

  15. Dorogush AV, Ershov V, Gulin A (2018) CatBoost: gradient boosting with categorical features support. arXiv:1810.11363

  16. Lundberg SM, et al. (2020) From local explanations to global understanding with explainable AI for trees. Nature Mach Intell 2:56–67

    Article  Google Scholar 

  17. Shapley LS (1953) A value for n-person games. In: Contributions to the theory of games, vol 2, pp 307–317

  18. Dua D, Graff C (2019) UCI Machine learning repository

  19. Olson RS, et al. (2017) PMLB: a large benchmark suite for machine learning evaluation and comparison. BioData mining 10:1–13

    Article  Google Scholar 

  20. Erickson N et al (2020) AutoGluon-tabular: robust and accurate automl for structured data. arXiv:2003.06505

  21. Geurts P, Ernst D, Wehenkel L (2006) Extremely randomized trees. Mach Learn 63:3–42

    Article  Google Scholar 

  22. Wilson DL (1972) Asymptotic properties of nearest neighbor rules using edited data. IEEE Trans Syst Man Cybern 3:408–421

    Article  MathSciNet  Google Scholar 

  23. Cortes C, Vapnik VN (1995) Support-vector networks. Mach Learn 20:273–297

    MATH  Google Scholar 

  24. Zhou ZH, Feng J (2017) Deep forest: towards an alternative to deep neural networks. arXiv:1702.08835

  25. Freund Y, Schapire R (1995) A decision-theoretic generalization of on-line learning and an application to boosting. J Comput Syst Sci 55:119–139

    Article  MathSciNet  Google Scholar 

  26. Ke G et al (2017) Lightgbm: A highly efficient gradient boosting decision tree. In: NeurIPS, pp 3146–3154

  27. Duan T et al (2020) Ngboost: Natural gradient boosting for probabilistic prediction. In: ICML, pp 2690–2700

  28. Kokel H et al (2020) A unified framework for knowledge intensive gradient boosting: leveraging human experts for noisy sparse domains. In: AAAI. pp 4460–4468

Download references

Acknowledgements

This research was supported by the Bisa Research Grant of Keimyung University in 2021.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Byoung Chul Ko.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Kim, S., Jeong, M. & Ko, B.C. Lightweight surrogate random forest support for model simplification and feature relevance. Appl Intell 52, 471–481 (2022). https://doi.org/10.1007/s10489-021-02451-x

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-021-02451-x

Keywords

Navigation