Tracking without Label: Unsupervised Multiple Object Tracking via Contrastive Similarity Learning

Sha Meng, Dian Shao, Jiacheng Guo, Shan Gao; Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2023, pp. 16264-16273

Abstract


Unsupervised learning is a challenging task due to the lack of labels. Multiple Object Tracking (MOT), which inevitably suffers from mutual object interference, occlusion, etc., is even more difficult without label supervision. In this paper, we explore the latent consistency of sample features across video frames and propose an Unsupervised Contrastive Similarity Learning method, named UCSL, including three contrast modules: self-contrast, cross-contrast, and ambiguity contrast. Specifically, i) self-contrast uses intra-frame direct and inter-frame indirect contrast to obtain discriminative representations by maximizing self-similarity. ii) Cross-contrast aligns cross- and continuous-frame matching results, mitigating the persistent negative effect caused by object occlusion. And iii) ambiguity contrast matches ambiguous objects with each other to further increase the certainty of subsequent object association through an implicit manner. On existing benchmarks, our method outperforms the existing unsupervised methods using only limited help from ReID head, and even provides higher accuracy than lots of fully supervised methods.

Related Material


[pdf] [supp] [arXiv]
[bibtex]
@InProceedings{Meng_2023_ICCV, author = {Meng, Sha and Shao, Dian and Guo, Jiacheng and Gao, Shan}, title = {Tracking without Label: Unsupervised Multiple Object Tracking via Contrastive Similarity Learning}, booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)}, month = {October}, year = {2023}, pages = {16264-16273} }