-
[pdf]
[supp]
[bibtex]@InProceedings{Tian_2023_CVPR, author = {Tian, Stephen and Cai, Yancheng and Yu, Hong-Xing and Zakharov, Sergey and Liu, Katherine and Gaidon, Adrien and Li, Yunzhu and Wu, Jiajun}, title = {Multi-Object Manipulation via Object-Centric Neural Scattering Functions}, booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)}, month = {June}, year = {2023}, pages = {9021-9031} }
Multi-Object Manipulation via Object-Centric Neural Scattering Functions
Abstract
Learned visual dynamics models have proven effective for robotic manipulation tasks. Yet, it remains unclear how best to represent scenes involving multi-object interactions. Current methods decompose a scene into discrete objects, yet they struggle with precise modeling and manipulation amid challenging lighting conditions since they only encode appearance tied with specific illuminations. In this work, we propose using object-centric neural scattering functions (OSFs) as object representations in a model-predictive control framework. OSFs model per-object light transport, enabling compositional scene re-rendering under object rearrangement and varying lighting conditions. By combining this approach with inverse parameter estimation and graph-based neural dynamics models, we demonstrate improved model-predictive control performance and generalization in compositional multi-object environments, even in previously unseen scenarios and harsh lighting conditions.
Related Material