Abstract
As we mentioned in Chap. 16, variable selection is very important when dealing with bioinformatics, healthcare, and biomedical data, where we may have more features than observations. Variable selection, or feature selection, can help us focus only on the core important information contained in the observations, instead of every piece of information. Due to presence of intrinsic and extrinsic noise, the volume and complexity of big health data, and different methodological and technological challenges, this process of identifying the salient features may resemble finding a needle in a haystack. Here, we will illustrate alternative strategies for feature selection using filtering (e.g., correlation-based feature selection), wrapping (e.g., recursive feature elimination), and embedding (e.g., variable importance via random forest classification) techniques.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Guyon, E, Gunn, S, Nikravesh, M, Zadeh, LA (eds.) (2008) Feature Extraction: Foundations and Applications, Springer, ISBN 3540354883, 9783540354888
Liu, H and Motoda, H (eds.) (2007) Computational Methods of Feature Selection, Chapman & Hall/CRC, ISBN 1584888792, 9781584888796
Pacheco, ER (2015) Unsupervised Learning with R, Packt Publishing, ISBN 1785885812, 9781785885815
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2018 Ivo D. Dinov
About this chapter
Cite this chapter
Dinov, I.D. (2018). Variable/Feature Selection. In: Data Science and Predictive Analytics. Springer, Cham. https://doi.org/10.1007/978-3-319-72347-1_17
Download citation
DOI: https://doi.org/10.1007/978-3-319-72347-1_17
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-72346-4
Online ISBN: 978-3-319-72347-1
eBook Packages: Computer ScienceComputer Science (R0)