A Simple and Efficient Pipeline To Build an End-to-End Spatial-Temporal Action Detector

Lin Sui, Chen-Lin Zhang, Lixin Gu, Feng Han; Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), 2023, pp. 5999-6008

Abstract


Spatial-temporal action detection is a vital part of video understanding. Current spatial-temporal action detection methods mostly use an object detector to obtain person candidates and classify these person candidates into different action categories. So-called two-stage methods are heavy and hard to apply in real-world applications. Some existing methods build one-stage pipelines, But a large performance drop exists with the vanilla one-stage pipeline and extra classification modules are needed to achieve comparable performance. In this paper, we explore a simple and effective pipeline to build a strong one-stage spatial-temporal action detector. The pipeline is composed by two parts: one is a simple end-to-end spatial-temporal action detector. The proposed end-to-end detector has minor architecture changes to current proposal-based detectors and does not add extra action classification modules. The other part is a novel labeling strategy to utilize unlabeled frames in sparse annotated data. We named our model as SE-STAD. The proposed SE-STAD achieves around 2% mAP boost and around 80% FLOPs reduction. Our code will be released at https://github.com/4paradigm-CV/SE-STAD.

Related Material


[pdf] [supp] [arXiv]
[bibtex]
@InProceedings{Sui_2023_WACV, author = {Sui, Lin and Zhang, Chen-Lin and Gu, Lixin and Han, Feng}, title = {A Simple and Efficient Pipeline To Build an End-to-End Spatial-Temporal Action Detector}, booktitle = {Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)}, month = {January}, year = {2023}, pages = {5999-6008} }