Abstract
Random forest is an ensemble supervised machine learning technique. The principle of ensemble suggests that to yield better accuracy, the base classifiers in the ensemble should be diverse and accurate. Random forest uses decision tree as base classifier. In this paper, we have done theoretical and empirical comparison of different split measures for induction of decision tree in Random forest and tested if there is any effect on the accuracy of Random forest.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Breiman, L.: Bagging predictors, Technical report No 421, September (1994).
Opitz, David: Maclin, richard: popular ensemble methods: an empirical study. J. Arti. Intel. 11, 169–198 (1999)
Brieman, Leo: Random forests. Machine Learning. 45, 5–32 (2001)
Sikonja, M.R.: Improving random forests. In: Boulicaut, J.F., et al. (eds): Machine Learning, ECML 2004 Proceedings, LNCS, vol. 3201, PP. 359–370, Springer, Berlin (2004).
Rokach, Lior: Maimon, oded: top-down induction of decision trees classifiers-a survey. IEEE trans. syst. man. cyber. part c: appli. rev. 35(4), 476–487 (2005).
Badulescu, L.A.: The choice of the best attribute selection measure in DecisionTree induction, Annals of University of Craiova, Math. Comp. Sci. Ser. Vol. 34 (1) (2007).
Mingers, J.: An empirical comparison of selection measures for decision tree induction. Mach. Learn. 3, 319–342 (1989)
Robnik-Sikonja, M., Kononenko, I.: Attribute dependencies, understandability, and split selection in tree based models, Machine Learning: Proceedings of the 6th International Conference (ICML), 344–353 (1999).
Brieman, Leo: Technical note-some properties of splitting criteria. Mach. Learn. 24, 41–47 (1996)
Han, J., Kamber, M.: Data mining: concepts and techniques, 2nd edn. Morgan Kaufmann Publisher, San Francisco (2006)
Buntine, Wray: Niblet, tim: a further comparison of splitting rules for decision tree induction. Mach. learn. 8, 75–85 (1992)
Kulkarni, V.Y., pradeep, K.S.: Random forest classifiers: a survey and future research directions. Int. J. Adv. Comput. ISSN 2051–0845. 36(1), 1144–1153 (2013).
Liu, W.Z., White, A.P.: The importance of attribute selection measures in decision tree induction. Mach. Learn. 15, 25–41 (1994)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer India
About this paper
Cite this paper
Kulkarni, V.Y., Petare, M., Sinha, P.K. (2014). Analyzing Random Forest Classifier with Different Split Measures. In: Babu, B., et al. Proceedings of the Second International Conference on Soft Computing for Problem Solving (SocProS 2012), December 28-30, 2012. Advances in Intelligent Systems and Computing, vol 236. Springer, New Delhi. https://doi.org/10.1007/978-81-322-1602-5_74
Download citation
DOI: https://doi.org/10.1007/978-81-322-1602-5_74
Published:
Publisher Name: Springer, New Delhi
Print ISBN: 978-81-322-1601-8
Online ISBN: 978-81-322-1602-5
eBook Packages: EngineeringEngineering (R0)