LAMV: Learning to Align and Match Videos With Kernelized Temporal Layers

Lorenzo Baraldi, Matthijs Douze, Rita Cucchiara, Hervé Jégou; The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018, pp. 7804-7813

Abstract


This paper considers a learnable approach for comparing and aligning videos. Our architecture builds upon and revisits temporal match kernels within neural networks: we propose a new temporal layer that finds temporal alignments by maximizing the scores between two sequences of vectors, according to a time-sensitive similarity metric parametrized in the Fourier domain. We learn this layer with a temporal proposal strategy, in which we minimize a triplet loss that takes into account both the localization accuracy and the recognition rate. We evaluate our approach on video alignment, copy detection and event retrieval. Our approach outperforms the state on the art on temporal video alignment and video copy detection datasets in comparable setups. It also attains the best reported results for particular event search, while precisely aligning videos.

Related Material


[pdf]
[bibtex]
@InProceedings{Baraldi_2018_CVPR,
author = {Baraldi, Lorenzo and Douze, Matthijs and Cucchiara, Rita and Jégou, Hervé},
title = {LAMV: Learning to Align and Match Videos With Kernelized Temporal Layers},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2018}
}