E2NeRF: Event Enhanced Neural Radiance Fields from Blurry Images

Yunshan Qi, Lin Zhu, Yu Zhang, Jia Li; Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2023, pp. 13254-13264

Abstract


Neural Radiance Fields (NeRF) achieves impressive ren-dering performance by learning volumetric 3D representation from several images of different views. However, it is difficult to reconstruct a sharp NeRF from blurry input as often occurred in the wild. To solve this problem, we propose a novel Event-Enhanced NeRF (E2NeRF) by utilizing the combination data of a bio-inspired event camera and a standard RGB camera. To effectively introduce event stream into the learning process of neural volumetric representation, we propose a blur rendering loss and an event rendering loss, which guide the network via modelling real blur process and event generation process, respectively. Moreover, a camera pose estimation framework for real-world data is built with the guidance of event stream to generalize the method to practical applications. In contrast to previous image-based or event-based NeRF, our framework effectively utilizes the internal relationship between events and images. As a result, E2NeRF not only achieves image deblurring but also achieves high-quality novel view image generation. Extensive experiments on both synthetic data and real-world data demonstrate that E2NeRF can effectively learn a sharp NeRF from blurry images, especially in complex and low-light scenes. Our code and datasets are publicly available at https://github.com/iCVTEAM/E2NeRF.

Related Material


[pdf] [supp]
[bibtex]
@InProceedings{Qi_2023_ICCV, author = {Qi, Yunshan and Zhu, Lin and Zhang, Yu and Li, Jia}, title = {E2NeRF: Event Enhanced Neural Radiance Fields from Blurry Images}, booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)}, month = {October}, year = {2023}, pages = {13254-13264} }