EgoCast: Forecasting Egocentric Human Pose in the Wild

Maria Escobar, Juanita Puentes, Cristhian Forigua, Jordi Pont-Tuset, Kevis-Kokitsi Maninis, Pablo Arbelaez; Proceedings of the Winter Conference on Applications of Computer Vision (WACV), 2025, pp. 5831-5841

Abstract


Accurately estimating and forecasting human body pose is important for enhancing the user's sense of immersion in Augmented Reality. Addressing this need our paper introduces EgoCast a bimodal method for 3D human pose forecasting using egocentric videos and proprioceptive data. We study the task of human pose forecasting in a realistic setting extending the boundaries of temporal forecasting in dynamic scenes and building on the current framework for current pose estimation in the wild. We introduce a current-frame estimation module that generates pseudo-groundtruth poses for inference eliminating the need for past groundtruth poses typically required by current methods during forecasting. Our experimental results on the recent Ego-Exo4D and Aria Digital Twin datasets validate EgoCast for real-life motion estimation. On the Ego-Exo4D Body Pose 2024 Challenge our method significantly outperforms the state-of-the-art approaches laying the groundwork for future research in human pose estimation and forecasting in unscripted activities with egocentric inputs.

Related Material


[pdf] [supp] [arXiv]
[bibtex]
@InProceedings{Escobar_2025_WACV, author = {Escobar, Maria and Puentes, Juanita and Forigua, Cristhian and Pont-Tuset, Jordi and Maninis, Kevis-Kokitsi and Arbelaez, Pablo}, title = {EgoCast: Forecasting Egocentric Human Pose in the Wild}, booktitle = {Proceedings of the Winter Conference on Applications of Computer Vision (WACV)}, month = {February}, year = {2025}, pages = {5831-5841} }