MSI-NeRF: Linking Omni-Depth with View Synthesis through Multi-Sphere Image Aided Generalizable Neural Radiance Field

Dongyu Yan, Guanyu Huang, Fengyu Quan, Haoyao Chen; Proceedings of the Winter Conference on Applications of Computer Vision (WACV), 2025, pp. 2517-2526

Abstract


Panoramic observation using fisheye cameras is significant in virtual reality (VR) and robot perception. However panoramic images synthesized by traditional methods lack depth information and can only provide three degrees-of-freedom (3DoF) rotation rendering in VR applications. To fully preserve and exploit the parallax information within the original fisheye cameras we introduce MSI-NeRF which combines deep learning omnidirectional depth estimation and novel view synthesis. We construct a multi-sphere image as a cost volume through feature extraction and warping of the input images. We further build an implicit radiance field using spatial points and interpolated 3D feature vectors as input which can simultaneously realize omnidirectional depth estimation and 6DoF view synthesis. Leveraging the knowledge from depth estimation task our method can learn scene appearance by source view supervision only. It does not require novel target views and can be trained conveniently on existing panorama depth estimation datasets. Our network has the generalization ability to reconstruct unknown scenes efficiently using only four images. Experimental results show that our method outperforms existing methods in both depth estimation and novel view synthesis tasks.

Related Material


[pdf] [supp]
[bibtex]
@InProceedings{Yan_2025_WACV, author = {Yan, Dongyu and Huang, Guanyu and Quan, Fengyu and Chen, Haoyao}, title = {MSI-NeRF: Linking Omni-Depth with View Synthesis through Multi-Sphere Image Aided Generalizable Neural Radiance Field}, booktitle = {Proceedings of the Winter Conference on Applications of Computer Vision (WACV)}, month = {February}, year = {2025}, pages = {2517-2526} }