Trajectory reconstruction by means of an event-camera-based visual odometry method and machine learned features

Trajectory reconstruction by means of an event-camera-based visual odometry method and machine learned features

S. Chiodini, G. Trevisanuto, C. Bettanini, G. Colombatti, M. Pertile

download PDF

Abstract. This paper presents a machine learned feature detector targeted to event-camera based visual odometry methods for unmanned aerial vehicles trajectory reconstruction. The proposed method uses machine-learned features to enhance the accuracy of the trajectory reconstruction. Traditional visual odometry methods suffer from poor performance in low light conditions and high-speed motion. The event-camera-based approach overcomes these limitations by detecting and processing only the changes in the visual scene. The machine-learned features are crafted to capture the unique characteristics of the event-camera data, enhancing the accuracy of the trajectory reconstruction. The inference pipeline is composed of a module repeated twice in sequence, formed by a Squeeze-and-Excite block and a ConvLSTM block with residual connection; it is followed by a final convolutional layer that provides the trajectories of the corners as a sequence of heatmaps. In the experimental part, a sequence of images was collected using an event-camera in outdoor environments for training and test.

Keywords
Visual Odometry, Computer Vision, Machine Learning

Published online 11/1/2023, 4 pages
Copyright © 2023 by the author(s)
Published under license by Materials Research Forum LLC., Millersville PA, USA

Citation: S. Chiodini, G. Trevisanuto, C. Bettanini, G. Colombatti, M. Pertile, Trajectory reconstruction by means of an event-camera-based visual odometry method and machine learned features, Materials Research Proceedings, Vol. 37, pp 705-708, 2023

DOI: https://doi.org/10.21741/9781644902813-150

The article was published as article 150 of the book Aeronautics and Astronautics

Content from this work may be used under the terms of the Creative Commons Attribution 3.0 license. Any further distribution of this work must maintain attribution to the author(s) and the title of the work, journal citation and DOI.

References
[1] Lichtsteiner, P., Posch, C., & Delbruck, T. (2008). A 128$\times $128 120 dB 15$\mu $ s latency asynchronous temporal contrast vision sensor. IEEE journal of solid-state circuits, 43(2), 566-576. https://doi.org/10.1109/JSSC.2007.914337
[2] Mueggler, E., Rebecq, H., Gallego, G., Delbruck, T., & Scaramuzza, D. (2017). The event-camera dataset and simulator: Event-based data for pose estimation, visual odometry, and SLAM. The International Journal of Robotics Research, 36(2), 142-149. https://doi.org/10.1177/0278364917691115
[3] Sikorski, O., Izzo, D., & Meoni, G. (2021). Event-based spacecraft landing using time-to-contact. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 1941-1950). https://doi.org/10.1109/CVPRW53098.2021.00222
[4] Vasco, V., Glover, A., Bartolozzi, C.: Fast Event-Based Harris Corner Detection Exploiting the Advantages of Event-Driven Cameras. In: IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). pp. 4144–4149 (2016). https://doi.org/10.1109/IROS.2016.7759610
[5] Chiberre, P., Perot, E., Sironi, A., & Lepetit, V. (2022). Long-Lived Accurate Keypoint in Event Streams. arXiv preprint arXiv:2209.10385.