Learning Target Candidate Association To Keep Track of What Not To Track


METADATA ONLY
Loading...

Date

2021

Publication Type

Conference Paper

ETH Bibliography

yes

Citations

Altmetric
METADATA ONLY

Data

Rights / License

Abstract

The presence of objects that are confusingly similar to the tracked target, poses a fundamental challenge in appearance-based visual tracking. Such distractor objects are easily misclassified as the target itself, leading to eventual tracking failure. While most methods strive to suppress distractors through more powerful appearance models, we take an alternative approach.We propose to keep track of distractor objects in order to continue tracking the target. To this end, we introduce a learned association network, allowing us to propagate the identities of all target candidates from frame-to-frame. To tackle the problem of lacking ground-truth correspondences between distractor objects in visual tracking, we propose a training strategy that combines partial annotations with self-supervision. We conduct comprehensive experimental validation and analysis of our approach on several challenging datasets. Our tracker sets a new state-of-the-art on six benchmarks, achieving an AUC score of 67.1% on LaSOT [21] and a +5.8% absolute gain on the OxUvA long-term dataset [41]. The code and trained models are available at https://github.com/visionml/pytracking

Publication status

published

Editor

Book title

2021 IEEE/CVF International Conference on Computer Vision (ICCV)

Journal / series

Volume

Pages / Article No.

13444 - 13454

Publisher

IEEE

Event

18th International Conference on Computer Vision (ICCV 2021)

Edition / version

Methods

Software

Geographic location

Date collected

Date created

Subject

Motion and tracking

Organisational unit

03514 - Van Gool, Luc (emeritus) / Van Gool, Luc (emeritus) check_circle

Notes

Funding

Related publications and datasets