Abstract
An explanation capability is crucial in security-sensitive domains, such as medical applications. Although support vector machines (SVMs) have shown superior performance in a range of classification and regression tasks, SVMs, like artificial neural networks (ANNs), lack an explanatory capability. There is a significant literature on obtaining human-comprehensible rules from SVMs and ANNs in order to explain “how” a decision was made or “why” a certain result was achieved. This chapter proposes a novel approach to SVM classifiers. The experiments described in this chapter involve a first attempt to generate textual and visual explanations for classification results using multimedia content of various type: poems expressing positive or negative emotion, autism descriptions, and facial expressions, including those with medical relevance (facial palsy). Learned model parameters are analyzed to select important features, and filtering is applied to select feature subsets of explanatory value. The explanation components are used to generate textual summaries of classification results. We show that the explanations are consistent and that the accuracy of SVM models is bounded by the accuracy of explanation components. The results show that the generated explanations display a high level of fidelity and can generate textual summaries with an error rate of less than 35 %.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Used ConceptNet v2.1 from the Common Sense Computing Initiative at the MIT Media Lab (http://csc.media.mit.edu).
References
Afifi, N., Diederich, J., Shanableh, T.: Computational methods for the detection of facial palsy. J. Telemedicine Telecare 12(SUPPL. 3), 3–7 (2006)
Aliguliyev, R.M.: Clustering techniques and discrete particle swarm optimization algorithm for multi-document summarization. Comput. Intell. 26(4), 420–448 (2010)
Andrews, R., Diederich, J., Tickle, A.B.: Survey and critique of techniques for extracting rules from trained artificial neural networks. Knowl.-Based Syst. 8(6), 373–389 (1995)
Blum, A.L., Langley, P.: Selection of relevant features and examples in machine learning. Artif. Intell. 97(1–2), 245–271 (1997). doi:10.1016/s0004-3702(97)00063-5
Dorr, B.J., Monz, C., President, S., Schwartz, R., Zajic, D.: A methodology for extrinsic evaluation of text summarization: Does rouge correlate?. In: Proceedings of the ACL 2005. pp. 1–8
Cawsey, A.: Explanation and interaction: the computer generation of explanatory dialogues. MIT Press, USA (1993)
Cortes, C., Vladimir, V.: Support-vector networks. Mach. Learn. 20(3), 273–297 (1995)
Diederich, J.: Rule extraction from support vector machines: An introduction. Stud. Comput. Intell. 80 (2008)
DSM-IV: Diagnostic and statistical manual of mental disorders. American Psychiatric Association (1994)
Féraud, R., Clérot, F.: A methodology to explain neural network classification. Neural Networks 15(2), 237–246 (2002)
Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. J Mach Learn Res 3, 1157–1182 (2003)
ICD10: International Statistical Classification of Disease and Related Health. World Health Organization, Geneva (1992)
Joachims, T.: Making Large-Scale SVM Learning Practical. Advances in Kernel Methods—Support Vector Learning. MIT-Press, NY (1999)
Morik, K., Brockhausen, P., Joachims, T.: Combining statistical learning with a knowledge-based approach—A case study in intensive care monitoring. In: Proceedings of the 16th International Conference on Machine Learning, Morgan Kaufmann Publishers, USA (1999)
Lin, C.-Y.: Rouge: A package for automatic evaluation of summaries. In: Proceedings of ACL workshop on Text Summarization Branches Out (2004)
Liu, H., Push, S.: Conceptnet: A practical commonsense reasoning toolkit. BT Technol. J. 22, 211–226 (2004)
Menkovski, V., Christou, I.T., Efremidis, S.: Oblique decision trees using embedded support vector machines in classifier ensembles (2008)
Mitra, S., Hayashi, Y.: Neuro-fuzzy rule generation: survey in soft computing framework. IEEE Trans. Neural Networks 11(3), 748–768 (2000)
Thagard, P., Little, A.: Models of Scientific Explanation. Cambridge University Press, Cambridge (2007)
Quinlan, J.R.: Induction of decision trees. Mach. Learn. 1(1), 81–106 (1986)
Setiono, R., Liu, H.: A connectionist approach to generating oblique decision trees. IEEE Trans. Syst. Man Cybern. B Cybern. 29(3), 440–444 (1999)
Tickle, A.B., Andrews, R., Golea, M., Diederich, J.: The truth will come to light: directions and challenges in extracting the knowledge embedded within trained artificial neural networks. IEEE Trans. Neural Networks 9(6), 1057–1068 (1998)
Yan, J., Liu, N., Yan, S., Yang, Q., Fan, W., Wei, W., Chen, Z.: Trace-oriented feature analysis for large-scale text data dimension reduction. IEEE Trans. Knowl. Data Eng. 23(7), 1103–1117 (2011). doi:10.1109/tkde.2010.34
Yang, Y: A study on thresholding strategies for text categorization. In: Proceedings of the 24th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, New Orleans, LA, 9–13 September 2001, pp. 137–145 (2001)
Yu, L., Liu, H.: Efficient feature selection via analysis of relevance and redundancy. J Mach Learn Res 5, 1205–1224 (2004)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
Song, I., Diederich, J. (2014). Generating Explanations from Support Vector Machines for Psychological Classifications. In: Lech, M., Song, I., Yellowlees, P., Diederich, J. (eds) Mental Health Informatics. Studies in Computational Intelligence, vol 491. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-38550-6_7
Download citation
DOI: https://doi.org/10.1007/978-3-642-38550-6_7
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-38549-0
Online ISBN: 978-3-642-38550-6
eBook Packages: EngineeringEngineering (R0)