D3D: Distilled 3D Networks for Video Action Recognition

Jonathan Stroud, David Ross, Chen Sun, Jia Deng, Rahul Sukthankar; The IEEE Winter Conference on Applications of Computer Vision (WACV), 2020, pp. 625-634

Abstract


State-of-the-art methods for action recognition commonly use two networks: the spatial stream, which takes RGB frames as input, and the temporal stream, which takes optical flow as input. In recent work, both streams are 3D Convolutional Neural Networks, which extract features using spatiotemporal filters. These filters can respond to motion, and therefore should allow the network to learn motion representations, removing the need for optical flow. However, we still see significant benefits in performance by feeding optical flow into the temporal stream, indicating that the spatial stream is "missing" some of the signal that the temporal stream captures. In this work, we first investigate whether motion representations are indeed missing in the spatial stream, and show that there is significant room for improvement. Second, we demonstrate that these motion representations can be improved using distillation, that is, by tuning the spatial stream to mimic the temporal stream, effectively combining both models into a single stream. Finally, we show that our Distilled 3D Network (D3D) achieves performance on par with the two-stream approach, with no need to compute optical flow during inference.

Related Material


[pdf]
[bibtex]
@InProceedings{Stroud_2020_WACV,
author = {Stroud, Jonathan and Ross, David and Sun, Chen and Deng, Jia and Sukthankar, Rahul},
title = {D3D: Distilled 3D Networks for Video Action Recognition},
booktitle = {The IEEE Winter Conference on Applications of Computer Vision (WACV)},
month = {March},
year = {2020}
}