ISPRS Ann. Photogramm. Remote Sens. Spatial Inf. Sci., IV-2/W3, 67-74, 2017
https://doi.org/10.5194/isprs-annals-IV-2-W3-67-2017
© Author(s) 2017. This work is distributed under
the Creative Commons Attribution 4.0 License.
 
18 Aug 2017
END-TO-END DEPTH FROM MOTION WITH STABILIZED MONOCULAR VIDEOS
C. Pinard1,2, L. Chevalley1, A. Manzanera2, and D. Filliat2 1Parrot, Paris, France
2U2IS, ENSTA ParisTech, Universite Paris-Saclay, Palaiseau, France
Keywords: Dataset, Navigation, Monocular, Depth from Motion, End-to-end, Deep Learning Abstract. We propose a depth map inference system from monocular videos based on a novel dataset for navigation that mimics aerial footage from gimbal stabilized monocular camera in rigid scenes. Unlike most navigation datasets, the lack of rotation implies an easier structure from motion problem which can be leveraged for different kinds of tasks such as depth inference and obstacle avoidance. We also propose an architecture for end-to-end depth inference with a fully convolutional network. Results show that although tied to camera inner parameters, the problem is locally solvable and leads to good quality depth prediction.
Conference paper (PDF, 1633 KB)


Citation: Pinard, C., Chevalley, L., Manzanera, A., and Filliat, D.: END-TO-END DEPTH FROM MOTION WITH STABILIZED MONOCULAR VIDEOS, ISPRS Ann. Photogramm. Remote Sens. Spatial Inf. Sci., IV-2/W3, 67-74, https://doi.org/10.5194/isprs-annals-IV-2-W3-67-2017, 2017.

BibTeX EndNote Reference Manager XML