Abstract
This paper proposes a new method to qualify the result given by a decision tree when it is used as a decision aid system. When the data are numerical, we compute the distance of a case from the decision surface. This distance measures the sensitivity of the result to a change in the input data. With a different distance it is also possible to measure the sensitivity of the result to small changes in the tree. The distance from the decision surface can also be combined to the error rate in order to provide a context-dependent information to the end-user.
Chapter PDF
Similar content being viewed by others
References
Murthy, S.K.: Automatic Construction of Decision Trees from Data: A Multi- Disciplinary Survey. Data Mining and Knowledge Discovery 2(4), 345–389 (1998)
Nelson, L.M., Bloch, D.A., Longstreth, W.T., Shib, H.: Recursive Partitioning for the Identification of Disease Risk Subgroups: A Case-Control Study of Subarachnoid Hemorrhage. Journal of Clinical Epidemiology 51(3), 199–209 (1998)
West, D.: Neural network credit scoring models. Computers & Operations Research 27, 1131–1152 (2000)
Domingos, P.: MetaCost: A General Method for Making Classifiers Cost-Sensitive. In: Proc. of the Fifth Int. Conf. on K.D. and Data Mining, pp. 155–164 (1999)
Sigillito, V., Blake, C.L., Merz, C.J.: Pima Indian Diabetes. UCI Repository of machine learning databases. University of California, Irvine (1990)
Burges, C.: A tutorial on support vector machines for pattern recognition. Data Mining and Knowledge Discovery 2(2), 955–974 (1998)
Platt, J.: Probabilistic outputs for support vector machines. In: Smola, A.J., Bartlett, P., Schoelkopf, B., Schuurmans, D. (eds.) Advances in Large Margin Classifiers, pp. 61–74. MIT Press, Cambridge (2000)
Domingos, P.: A Unified Bias-Variance Decomposition and its Applications. In: Proc. of the 17th Int. Conf. on Machine Learning, pp. 231–238. Morgan Kaufmann, San Francisco (2000)
Kohavi, R.: A study of cross-validation and bootstrap for accuracy estimation and model selection. In: Proc. Int. Joint Conf. on Artificial Intelligence, pp. 1137–1143 (1995)
Provost, F., Fawcett, T.: Analysis and visualization of classifier performance: Comparison under imprecise class and cost distributions. In: Proc. Third Int. Conf. on Knowledge Discovery and Data Mining, pp. 43–48 (1997)
Jordan, M.I., Jacobs, R.A.: Hierarchical mixtures of experts and the EM algorithm. Neural Computation 6, 181–214 (1994)
Umano, M., Okomato, K., Hatono, I., Tamura, H., Kawachi, F., Umezu, S., Kinoshita, J.: Proc. of the 3rd IEEE Int. Conf. on Fuzzy Systems, pp. 2113–2118 (1994)
Quinlan, J.R.: C4.5: Programs for Machine Learning. Morgan Kaufmann, San Francisco (1993)
Breiman, L., Friedman, J.H., Olshen, R.A., Stone, C.J.: Classification and Regression Trees. Wadsworth, Belmont (1984)
Esposito, F., Malerba, D., Semeraro, G.: A comparative analysis of methods for pruning decision trees. IEEE Trans. on Pattern Analysis and Machine Intelligence 19(5), 476–491 (1997)
Cestnik, B.: Estimating probabilities: A crucial task in machine learning. In: Proceedings of the European Conference on Artificial Intelligence, pp. 147–149 (1990)
Domingos, P., Provost, F.: Well-trained PETs: Improving probability estimation trees. CDER Working Paper #00-04-IS Stern School of Business (2000)
Zadrozny, B., Elkan, C.: Obtaining calibrated probability estimates from decision trees and naive bayesian classifiers. In: Proc. of the 18th Int. Conf. on Machine Learning, pp. 609–616 (2001)
Kearns, M.J., Ron, D.: Algorithmic stability and sanity-check bounds for leaveone-out cross-validation. In: Proc. of the Tenth Conf. on Computational Learning Theory, pp. 152–162 (1997)
Bauschke, H., Borwein, J.M.: On projection algorithms for solving convex feasibility problems. SIAM Review 38(3), 367–426 (1996)
Bennett, K., Blue, J.: Optimal decision trees. Technical Report 214. R.P.I. Math. Science Dept., Troy, NY 12180 (1996)
Witten, I., Frank, E.: Data Mining: Practical Machine Learning Tools and Techniques with Java Implementation. Morgan Kaufmann, San Francisco (2000)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2004 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Alvarez, I. (2004). Sensitivity Analysis of the Result in Binary Decision Trees. In: Boulicaut, JF., Esposito, F., Giannotti, F., Pedreschi, D. (eds) Machine Learning: ECML 2004. ECML 2004. Lecture Notes in Computer Science(), vol 3201. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-30115-8_8
Download citation
DOI: https://doi.org/10.1007/978-3-540-30115-8_8
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-23105-9
Online ISBN: 978-3-540-30115-8
eBook Packages: Springer Book Archive