Learning Robot Activities From First-Person Human Videos Using Convolutional Future Regression

Jangwon Lee, Michael S. Ryoo; Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2017, pp. 1-2

Abstract


We design a new approach that allows robot learning of new activities from unlabeled human example videos. Given videos of humans executing an activity from their own viewpoint (i.e., first-person videos), our objective is to make the robot learn the temporal structure of the activity as its future regression network, and learn to transfer such model for its own motor execution. We present a new fully convolutional neural network architecture to regress the intermediate scene representation corresponding to the future frame, thereby enabling explicit forecasting of future hand locations given the current frame.

Related Material


[pdf] [arXiv]
[bibtex]
@InProceedings{Lee_2017_CVPR_Workshops,
author = {Lee, Jangwon and Ryoo, Michael S.},
title = {Learning Robot Activities From First-Person Human Videos Using Convolutional Future Regression},
booktitle = {Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops},
month = {July},
year = {2017}
}