In this paper, we propose to learn temporally invariant features from a large number of image sequences to represent objects for visual tracking. These features are trained on a convolutional neural network with temporal invariance constraints and robust to diverse motion transformations. We employ linear correlation filters to encode the appearance templates of targets and perform the tracking task by searching for the maximum responses at each frame. The learned filters are updated online and adapt to significant appearance changes during tracking. Extensive experimental results on challenging sequences show that the proposed algorithm performs favorably against state-of-the-art methods in terms of efficiency, accuracy, and robustness.
|Title of host publication||2015 IEEE International Conference on Image Processing, ICIP 2015 - Proceedings|
|Publisher||IEEE Computer Society|
|Number of pages||5|
|Publication status||Published - 2015 Dec 9|
|Event||IEEE International Conference on Image Processing, ICIP 2015 - Quebec City, Canada|
Duration: 2015 Sep 27 → 2015 Sep 30
|Name||Proceedings - International Conference on Image Processing, ICIP|
|Other||IEEE International Conference on Image Processing, ICIP 2015|
|Period||15/9/27 → 15/9/30|
Bibliographical notePublisher Copyright:
© 2015 IEEE.
All Science Journal Classification (ASJC) codes
- Computer Vision and Pattern Recognition
- Signal Processing