Multi range Real-time depth inference from a monocular stabilized footage using a Fully Convolutional Neural Network

Abstract : We propose a neural network architecture for depth map inference from monocular stabilized videos with application to UAV videos in rigid scenes. Training is based on a novel synthetic dataset for navigation that mimics aerial footage from gimbal stabilized monocular camera in rigid scenes. Based on this network, we propose a multi-range architecture for unconstrained UAV flight, leveraging flight data from sensors to make accurate depth maps for uncluttered outdoor environment. We try our algorithm on both synthetic scenes and real UAV flight data. Quantitative results are given for synthetic scenes with a slightly noisy orientation, and show that our multi-range architecture improves depth inference. Along with this article is a video that present our results more thoroughly.
Complete list of metadatas

Cited literature [23 references]  Display  Hide  Download

https://hal.archives-ouvertes.fr/hal-01587658
Contributor : Clément Pinard <>
Submitted on : Thursday, September 14, 2017 - 3:15:34 PM
Last modification on : Wednesday, July 3, 2019 - 10:48:05 AM

File

Article ECMR.pdf
Files produced by the author(s)

Identifiers

  • HAL Id : hal-01587658, version 1

Citation

Clément Pinard, Laure Chevalley, Antoine Manzanera, David Filliat. Multi range Real-time depth inference from a monocular stabilized footage using a Fully Convolutional Neural Network. European Conference on Mobile Robotics, ENSTA ParisTech, Sep 2017, Paris, France. ⟨hal-01587658⟩

Share

Metrics

Record views

429

Files downloads

598