-
[pdf]
[supp]
[arXiv]
[bibtex]@InProceedings{Ganj_2025_WACV, author = {Ganj, Ashkan and Su, Hang and Guo, Tian}, title = {HybridDepth: Robust Metric Depth Fusion by Leveraging Depth from Focus and Single-Image Priors}, booktitle = {Proceedings of the Winter Conference on Applications of Computer Vision (WACV)}, month = {February}, year = {2025}, pages = {973-982} }
HybridDepth: Robust Metric Depth Fusion by Leveraging Depth from Focus and Single-Image Priors
Abstract
We propose HYBRIDDEPTH a robust depth estimation pipeline that addresses key challenges in depth estimation including scale ambiguity hardware heterogeneity and generalizability. HYBRIDDEPTH leverages focal stack data conveniently accessible in common mobile devices to produce accurate metric depth maps. By incorporating depth priors afforded by recent advances in single-image depth estimation our model achieves a higher level of structural detail compared to existing methods. We test our pipeline as an end-to-end system with a newly developed mobile client to capture focal stacks which are then sent to a GPU-powered server for depth estimation. Comprehensive quantitative and qualitative analyses demonstrate that HYBRIDDEPTH outperforms state-of-the-art (SOTA) models on common datasets such as DDFF12 and NYU Depth V2. HYBRIDDEPTH also shows strong zero-shot generalization. When trained on NYU Depth V2 HYBRIDDEPTH surpasses SOTA models in zero-shot performance on ARKitScenes and delivers more structurally accurate depth maps on Mobile Depth. The code is available at https://github.com/cake-lab/HybridDepth/.
Related Material