NPC: Neural Point Characters from Video

Shih-Yang Su, Timur Bagautdinov, Helge Rhodin; Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2023, pp. 14795-14805

Abstract


High-fidelity human 3D models can now be learned directly from videos, typically by combining a template-based surface model with neural representations. However, obtaining a template surface requires expensive multi-view capture systems, laser scans, or strictly controlled conditions. Previous methods avoid using a template but rely on a costly or ill-posed mapping from observation to canonical space. We propose a hybrid point-based representation for animatable humans that does not require an explicit surface model, while being generalizable to novel poses. For a given video, our method automatically produces an explicit set of 3D points representing approximate canonical geometry, and learns an articulated deformation model that produces pose-dependent point transformations. The points serve both as a scaffold for high-frequency neural features and an anchor for efficiently mapping between observation and canonical space. We demonstrate on established benchmarks that our representation overcomes limitations of prior work operating in either canonical or in observation space. Moreover, our automatic point extraction approach enables learning models of human and animal characters alike, matching the performance of the methods using rigged surface templates despite being more general. Project website: https: //lemonatsu.github.io/npc/.

Related Material


[pdf] [supp] [arXiv]
[bibtex]
@InProceedings{Su_2023_ICCV, author = {Su, Shih-Yang and Bagautdinov, Timur and Rhodin, Helge}, title = {NPC: Neural Point Characters from Video}, booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)}, month = {October}, year = {2023}, pages = {14795-14805} }