Abstract
In this paper we describe a close-to-sensor low latency visual processing system. We show that by adaptively sampling visual information, low level tracking can be achieved at high temporal frequencies with no increase in bandwidth and using very little memory. By having close-to-sensor processing, image regions can be captured and processed at millisecond sub-frame rates. If spatiotemporal regions have little useful information in them they can be discarded without further processing. Spatiotemporal regions that contain ‘interesting’ changes are further processed to determine what the interesting changes are. Close-to-sensor processing enables low latency programming of the image sensor such that interesting parts of a scene are sampled more often than less interesting parts. Using a small set of low level rules to define what is interesting, early visual processing proceeds autonomously. We demonstrate system performance with two applications. Firstly, to test the absolute performance of the system, we show low level visual tracking at millisecond rates and secondly a more general recursive Baysian tracker.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Rucci, M., Iovin, R., Poletti, M., Santini, F.: Miniature eye movements enhance fine spatial detail. Nature (2007)
Santini, F., Nambisan, R., Rucci, M.: Active 3d vision through gaze relocation in a humanoid robot. International Journal of Humanoid Robotics (2009)
Ko, H., Poletti, M., Rucci, M.: Microsaccades precisely relocate gaze in a high visual acuity task. Nature Neuroscience (2010)
Poletti, M., Listorti, C., Rucci, M.: Stability of the visual world during eye drift. Journal of Neuroscience (2010)
Poletti, M., Rucci, M.: Eye movements under various conditions of image fading. Journal of Vision (2010)
Graetzel, C.F., Fry, S.N., Nelson, B.J.: A 6000 hz computer vision system for real-time wing beat analysis of drosophila. In: International Conference on Biomedical Robotics and Biomechatronics, vol. 1 (2006)
Kirsch, G.: Interest point and local descriptor generation in silicon. In: International Solid-State Circuits Conference (2012)
Lichtsteiner, P., Posch, C., Delbruck, T.: An 128x128 120db 15us-latency temporal contrast vision sensor. IEEE Journal Solid State Circuits (2007)
Shraml, S., Belbachir, A.N.: A spatio-temporal clustering method using real-time motion analysis on event-based 3d vision. In: International Conference on Computer Vision and Pattern Recognition (2010)
Sonnleithner, D., Indiveri, G.: A neuromorphic saliency-map based active vision system. In: International Conference on Information Sciences and Systems (2011)
Dowski, E.: A new paradigm for future application-specific imaging systems. In: International Conference on Computational Photography (2011)
Reddy, D., Veeraraghavan, A., Chellapa, R.: P2c2: Programmable pixel compressive camera for high speed imaging. In: International Conference on Computer Vision and Pattern Recognition (2011)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Gibson, D., Muller, H., Campbell, N., Bull, D. (2013). Adaptive Sampling for Low Latency Vision Processing. In: Park, JI., Kim, J. (eds) Computer Vision - ACCV 2012 Workshops. ACCV 2012. Lecture Notes in Computer Science, vol 7728. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-37410-4_17
Download citation
DOI: https://doi.org/10.1007/978-3-642-37410-4_17
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-37409-8
Online ISBN: 978-3-642-37410-4
eBook Packages: Computer ScienceComputer Science (R0)