27 April 2022 MLFFNet: multilevel feature fusion network for monocular depth estimation from aerial images
Huihui Xu, Fei Li, Zhiquan Feng
Author Affiliations +
Abstract

Convincing depth estimation from monocular aerial images is a fundamental task in environment perception. We propose an efficient multilevel feature fusion network (MLFFNet) for estimating depths from aerial images. Specifically, the proposed MLFFNet consists of a multi-level attention pyramid (MAP) module and an adaptive feature fusion (AFF) module. The MAP module extracts the low-level and high-level useful information from the perspective of nonlocal spatial attention and channel attention, while the AFF module adaptively integrates the extracted multilevel features to enhance the estimated effect. Moreover, since images taken by drones have a large depth range, we designed a loss function suitable for aerial images. The evaluation experiments are performed on the MidAir dataset. Experimental results denote that our MLFFNet outperforms other depth estimation methods in predicting the depths. We also test several images from real-life scenarios, and our method can obtain reasonable outputs.

© 2022 Society of Photo-Optical Instrumentation Engineers (SPIE) 1931-3195/2022/$28.00 © 2022 SPIE
Huihui Xu, Fei Li, and Zhiquan Feng "MLFFNet: multilevel feature fusion network for monocular depth estimation from aerial images," Journal of Applied Remote Sensing 16(2), 026506 (27 April 2022). https://doi.org/10.1117/1.JRS.16.026506
Received: 2 December 2021; Accepted: 31 March 2022; Published: 27 April 2022
Lens.org Logo
CITATIONS
Cited by 3 scholarly publications.
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Image fusion

Lithium

3D modeling

RGB color model

Unmanned aerial vehicles

Image resolution

Cameras

Back to Top