Abstract
This paper presents a wearable navigation assistive system for the blind and the visually impaired built with off-the-shelf technology. Microsoft Kinect’s on board depth sensor is used to extract Red, Green, Blue and Depth (RGB-D) data of the indoor environment. Speeded-Up Robust Features (SURF) and Bag-of-Visual-Words (BOVW) model is used to extract features and reduce generic indoor object detection into a machine learning problem. A Support Vector Machine classifier is used to classify scene objects and obstacles to issue critical real-time information to the user through an external aid (earphone) for safe navigation. We performed a user-study with blind-fold users to measure the efficiency of the overall framework.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Caltech101 dataset, http://www.vision.caltech.edu/Image_Datasets/Caltech101/
Kinect for Windows, http://www.microsoft.com/en-us/kinectforwindows/
Visually Impairment and Blindness. WHO Fact sheet–282 (2014), http://www.who.int/mediacentre/factsheets/fs282/en/
Bay, H., Ess, A., Tuytelaars, T., Van Gool, L.: Speeded-Up Robust Features (SURF). Computer Vision and Image Understanding (CVIU) 110, 346–359 (2008)
Bernabei, D., Ganovelli, F., Benedetto, M.D., Dellepiane, M., Scopigno, R.: A Low-Cost Time-Critical Obstacle Avoidance System for the Visually Impaired. In: International Conference on Indoor Positioning and Indoor Navigation, Portugal, pp. 21–23 (September 2011)
Brock, M.: Supporting Blind Navigation using Depth Sensing and Sonification. In: ACM Conference on Pervasive and Ubiquitous Computing - UbiComp 2013, pp. 255–258. ACM, New York (2013)
Burges, C.J.C.: A Tutorial on Support Vector Machines for Pattern Recognition. Data Mining and Knowledge Discovery 2(2), 121–167 (1998)
Csurka, G., Dance, C.R., Fan, L., Willamowski, J., Bray, C., Maupertuis, D.: Visual Categorization with Bags of Keypoints. In: Workshop on Statistical Leanring in Computer Vision, pp. 1–22 (2004)
Eichhorn, J., Chapelle, O.: Object categorization with SVM: Kernels for Local Features. Tech. Rep. 137, Max Planck Institute for Biological Cybernetics, Tubingen, Germany (2004)
Filipe, V., Fernandes, F., Fernandes, H., Sousa, A., Paredes, H., Barroso, J.A.: Blind Navigation Support System based on Microsoft Kinect. Procedia Computer Science 14, 94–101 (2012)
Greenwood, C., Nirjon, S., Stankovic, J., Yoon, H.J., Ra, H.-K., Son, S., Park, T.: KinSpace: Passive Obstacle Detection via Kinect. In: Krishnamachari, B., Murphy, A.L., Trigoni, N. (eds.) EWSN 2014. LNCS, vol. 8354, pp. 182–197. Springer, Heidelberg (2014)
Hicks, S.L., Wilson, I., Muhammed, L., Worsfold, J., Downes, S.M., Kennard, C.: A depth-based head-mounted visual display to aid navigation in partially sighted individuals. PloS One 8(7), e67695 (2013)
Khan, A., Moideen, F., Lopez, J., Khoo, W.L., Zhu, Z.: KinDectect: Kinect Detecting Objects. In: Miesenberger, K., Karshmer, A., Penaz, P., Zagler, W. (eds.) ICCHP 2012, Part II. LNCS, vol. 7383, pp. 588–595. Springer, Heidelberg (2012)
Mann, S., Huang, J., Janzen, R.: Blind Navigation with a Wearable Range Camera and Vibrotactile Helmet. In: Proceedings of 19th ACM International Conference on Multimedia, pp. 1325–1328. ACM, New York (2011)
Sales, D., Correa, D., Osório, F.S., Wolf, D.F.: 3D Vision-Based Autonomous Navigation System Using ANN and Kinect Sensor. In: Jayne, C., Yue, S., Iliadis, L. (eds.) EANN 2012. CCIS, vol. 311, pp. 305–314. Springer, Heidelberg (2012)
Wang, S., Pan, H., Zhang, C., Tian, Y.: RGB-D Image-Based Detection of Stairs, Pedestrian Crosswalks and Traffic Signs. Journal of Visual Communication and Image Representation 25(2), 263–272 (2014)
Yang, J., Hauptmann, A.G.: Evaluating Bag-of-Visual-Words Representations in Scene Classification. In: International Workshop on Multimedia Information Retrieval (MIR 2007), pp. 197–206. ACM, New York (2007)
Zöllner, M., Huber, S., Jetter, H.-C., Reiterer, H.: NAVI – A Proof-of-Concept of a Mobile Navigational Aid for Visually Impaired Based on the Microsoft Kinect. In: Campos, P., Graham, N., Jorge, J., Nunes, N., Palanque, P., Winckler, M. (eds.) INTERACT 2011, Part IV. LNCS, vol. 6949, pp. 584–587. Springer, Heidelberg (2011)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer International Publishing Switzerland
About this paper
Cite this paper
Bhowmick, A., Prakash, S., Bhagat, R., Prasad, V., Hazarika, S.M. (2014). IntelliNavi: Navigation for Blind Based on Kinect and Machine Learning. In: Murty, M.N., He, X., Chillarige, R.R., Weng, P. (eds) Multi-disciplinary Trends in Artificial Intelligence. MIWAI 2014. Lecture Notes in Computer Science(), vol 8875. Springer, Cham. https://doi.org/10.1007/978-3-319-13365-2_16
Download citation
DOI: https://doi.org/10.1007/978-3-319-13365-2_16
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-13364-5
Online ISBN: 978-3-319-13365-2
eBook Packages: Computer ScienceComputer Science (R0)