Boosting Monocular Depth Estimation Models to High-Resolution via Content-Adaptive Multi-Resolution Merging
Abstract
Neural networks have shown great abilities in estimating depth from a single image. However, the inferred depth maps are well below one-megapixel resolution and often lack fine-grained details, which limits their practicality. Our method builds on our analysis on how the input resolution and the scene structure affects depth estimation performance. We demonstrate that there is a trade-off between a consistent scene structure and the high-frequency details, and merge low- and high-resolution estimations to take advantage of this duality using a simple depth merging network. We present a double estimation method that improves the whole-image depth estimation and a patch selection method that adds local details to the final result. We demonstrate that by merging estimations at different resolutions with changing context, we can generate multi-megapixel depth maps with a high level of detail using a pre-trained model.
Implementation
Boosting Monocular Depth GitHub Repository
Boost Your Own Depth GitHub Repository
Video
Examples
Paper
Poster
Media Coverage
SFU Announcement - How SFU Researchers are Teaching AI to See Depth in Photographs and Paintings
PetaPixel - Scientists Teach AI Cameras to See Depth in Photos Better
Nature - Best Science Images of the Month - August 2021
ugocapeto3d YouTube Channel - I have tested Boosting Monocular Depth Estimation
BibTeX
author={S. Mahdi H. Miangoleh and Sebastian Dille and Long Mai and Sylvain Paris and Ya\u{g}{\i}z Aksoy},
title={Boosting Monocular Depth Estimation Models to High-Resolution via Content-Adaptive Multi-Resolution Merging},
journal={Proc. CVPR},
year={2021},
}
Related Publications