ReinDiffuse: Crafting Physically Plausible Motions with Reinforced Diffusion Model

Gaoge Han, Mingjiang Liang, Jinglei Tang, Yongkang Cheng, Wei Liu, Shaoli Huang; Proceedings of the Winter Conference on Applications of Computer Vision (WACV), 2025, pp. 2218-2227

Abstract


Generating human motion from textual descriptions is a challenging task. Existing methods either struggle with physical credibility or are limited by the complexities of physics simulations. In this paper we present ReinDiffuse that combines reinforcement learning with motion diffusion model to generate physically credible human motions that align with textual descriptions. Our method adapts Motion Diffusion Model to output a parameterized distribution of actions making them compatible with reinforcement learning paradigms. We employ reinforcement learning with the objective of maximizing physically plausible rewards to optimize motion generation for physical fidelity. Our approach outperforms existing state-of-the-art models on two major datasets HumanML3D and KIT-ML achieving significant improvements in physical plausibility and motion quality. Project: https://reindiffuse.github.io/

Related Material


[pdf] [supp] [arXiv]
[bibtex]
@InProceedings{Han_2025_WACV, author = {Han, Gaoge and Liang, Mingjiang and Tang, Jinglei and Cheng, Yongkang and Liu, Wei and Huang, Shaoli}, title = {ReinDiffuse: Crafting Physically Plausible Motions with Reinforced Diffusion Model}, booktitle = {Proceedings of the Winter Conference on Applications of Computer Vision (WACV)}, month = {February}, year = {2025}, pages = {2218-2227} }