Abstract
This paper illustrates an alternative approach towards mission control of UAV beyond the usage of radio data link. It is based on bidirectional visual communication employing onboard computer vision to recognize the operator’s gestures and providing visible feedback to him. This work presents use cases and summarizes the achievements made so far at the Institute of Flight Systems with respect to functional concepts, system architectures, sensor technology, data processing and experimental validation.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Nesta: Flying High: shaping the future of drones in UK cities. Full report (2018). https://media.nesta.org.uk/documents/Flying-High-full-report-and-appendices.pdf. Accessed 5 Nov 2018
Monajjemi, V.M., et al.: HRI in the sky: creating and commanding teams of UAVs with a vision-mediated gestural interface. In: 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 03 November 2013, pp. 617–623. IEEE (2013)
Nagi, J., et al.: HRI in the sky: controlling UAVs using face poses and hand gestures. In: HRI, pp. 252–253 (2014)
Monajjemi, M., et al.: UAV, do you see me? Establishing mutual attention between an uninstrumented human and an outdoor UAV in flight. In: 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, pp. 3614–3620. IEEE (2015). https://doi.org/10.1109/iros.2015.7353882
Schelle, A., Stütz, P.: Modelling visual communication with UAS. In: Hodicky, J. (ed.) MESAS 2016. LNCS, vol. 9991, pp. 81–98. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-47605-6_7
Schelle, A., Stütz, P.: Visual communication with UAS: recognizing gestures from an airborne platform. In: Lackey, S., Chen, J. (eds.) VAMR 2017. LNCS, vol. 10280, pp. 173–184. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-57987-0_14
Schelle, A., Stütz, P.: Gestural transmission of tasking information to an airborne uav. In: Yamamoto, S., Mori, H. (eds.) HIMI 2018. LNCS, vol. 10904, pp. 318–335. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-92043-6_27
Cao, Z., Simon, T., Wei, S.-E., Sheikh, Y.: OpenPose: real-time multi-person keypoint detection library for body, face, and hands estimation. Carnegie Mellon University, Perceptual Computing Laboratory (2018). https://github.com/CMU-Perceptual-Computing-Lab/openpose
Kirk, D.: NVIDIA CUDA software and GPU parallel computing architecture. In: Proceedings of the 6th International Symposium on Memory Management, Montreal, Quebec, Canada, pp. 103–104. ACM, New York (2007). https://doi.org/10.1145/1296907.1296909
Yeo, H.-S., Lee, B.-G., Lim, H.: Hand tracking and gesture recognition system for human-computer interaction using low-cost hardware. Multimedia Tools Appl. (2015). https://doi.org/10.1007/s11042-013-1501-1
Schelle, A., Stütz, P.: Evaluierung eines symbolisch motivierten Gestenwortschatzes zur visuellen Kommandierung von unbemannten Flugsystemen (2019, manuscript submitted for publication)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Schelle, A., Stütz, P. (2019). Visual Communication with UAV: Use Cases and Achievements. In: Vento, M., et al. Computer Analysis of Images and Patterns. CAIP 2019. Communications in Computer and Information Science, vol 1089. Springer, Cham. https://doi.org/10.1007/978-3-030-29930-9_12
Download citation
DOI: https://doi.org/10.1007/978-3-030-29930-9_12
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-29929-3
Online ISBN: 978-3-030-29930-9
eBook Packages: Computer ScienceComputer Science (R0)