AWSD: Adaptive Weighted Spatiotemporal Distillation for Video Representation

Mohammad Tavakolian, Hamed R. Tavakoli, Abdenour Hadid; The IEEE International Conference on Computer Vision (ICCV), 2019, pp. 8020-8029

Abstract


We propose an Adaptive Weighted Spatiotemporal Distillation (AWSD) technique for video representation by encoding the appearance and dynamics of the videos into a single RGB image map. This is obtained by adaptively dividing the videos into small segments and comparing two consecutive segments. This allows using pre-trained models on still images for video classification while successfully capturing the spatiotemporal variations in the videos. The adaptive segment selection enables effective encoding of the essential discriminative information of untrimmed videos. Based on Gaussian Scale Mixture, we compute the weights by extracting the mutual information between two consecutive segments. Unlike pooling-based methods, our AWSD gives more importance to the frames that characterize actions or events thanks to its adaptive segment length selection. We conducted extensive experimental analysis to evaluate the effectiveness of our proposed method and compared our results against those of recent state-of-the-art methods on four benchmark datatsets, including UCF101, HMDB51, ActivityNet v1.3, and Maryland. The obtained results on these benchmark datatsets showed that our method significantly outperforms earlier works and sets the new state-of-the-art performance in video classification. Code is available at the project webpage: https://mohammadt68.github.io/AWSD/

Related Material


[pdf]
[bibtex]
@InProceedings{Tavakolian_2019_ICCV,
author = {Tavakolian, Mohammad and Tavakoli, Hamed R. and Hadid, Abdenour},
title = {AWSD: Adaptive Weighted Spatiotemporal Distillation for Video Representation},
booktitle = {The IEEE International Conference on Computer Vision (ICCV)},
month = {October},
year = {2019}
}