Multi Camera Deep Tracking and Mapping

Visual Odometry methods based on classical 3D geometry have been around for years, using either indirect feature matching or direct visual error minimization. Lately,learning-based methods that combine both matching and geometry estimation in a single network have achieved impressive results. One such method is DeepTAM. Further, it has been shown that classical methods benefit from the extended field of view provided by using multiple cameras. However, these setups have been ignored by current learning-based methods.

In this work, we extend the existing DeepTAM pipeline to leverage a multi-camera setup with known geometry. We demonstrate the generalizability of DeepTAM to other monocular setups and highlight the scenarios in which it performs poorly. We show the efficacy of our proposed multi-camera VO pipeline to receive better pose estimates using experiments based on simulation.

Contributors: Mayank Mittal, Rohit Suri, Fadhil Ginting, Parker Ewen

https://github.com/surirohit/multi-camera-deeptam

Avatar
M. Fadhil Ginting
PhD Student in Aeronautics and Astronautics

PhD Student in Aero Astro at Stanford University.

Related