Beyond Local Search: Tracking Objects Everywhere With Instance-Specific Proposals

Gao Zhu, Fatih Porikli, Hongdong Li; The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 943-951

Abstract


Most tracking-by-detection methods employ a local search window around the predicted object location in the current frame assuming the previous location is accurate, the trajectory is smooth, and the computational capacity permits a search radius that can accommodate the maximum speed yet small enough to reduce mismatches. These, however, may not be valid always, in particular for fast and irregularly moving objects. Here, we present an object tracker that is not limited to a local search window and has ability to probe efficiently the entire frame. Our method generates a small number of "high-quality" proposals by a novel instance-specific objectness measure and evaluates them against the object model that can be adopted from an existing tracking-by-detection approach as a core tracker. During the tracking process, we update the object model concentrating on hard false-positives supplied by the proposals, which help suppressing distractors caused by difficult background clutters, and learn how to re-rank proposals according to the object model. Since we reduce significantly the number of hypotheses the core tracker evaluates, we can use richer object descriptors and stronger detector. Our method outperforms most recent state-of-the-art trackers on popular tracking benchmarks, and provides improved robustness for fast moving objects as well as for ultra low-frame-rate videos.

Related Material


[pdf]
[bibtex]
@InProceedings{Zhu_2016_CVPR,
author = {Zhu, Gao and Porikli, Fatih and Li, Hongdong},
title = {Beyond Local Search: Tracking Objects Everywhere With Instance-Specific Proposals},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2016}
}