Back to Search Start Over

Hedging Deep Features for Visual Tracking.

Authors :
Qi, Yuankai
Zhang, Shengping
Qin, Lei
Huang, Qingming
Yao, Hongxun
Lim, Jongwoo
Yang, Ming-Hsuan
Source :
IEEE Transactions on Pattern Analysis & Machine Intelligence; May2019, Vol. 41 Issue 5, p1116-1130, 15p
Publication Year :
2019

Abstract

Convolutional Neural Networks (CNNs) have been applied to visual tracking with demonstrated success in recent years. Most CNN-based trackers utilize hierarchical features extracted from a certain layer to represent the target. However, features from a certain layer are not always effective for distinguishing the target object from the backgrounds especially in the presence of complicated interfering factors (e.g., heavy occlusion, background clutter, illumination variation, and shape deformation). In this work, we propose a CNN-based tracking algorithm which hedges deep features from different CNN layers to better distinguish target objects and background clutters. Correlation filters are applied to feature maps of each CNN layer to construct a weak tracker, and all weak trackers are hedged into a strong one. For robust visual tracking, we propose a hedge method to adaptively determine weights of weak classifiers by considering both the difference between the historical as well as instantaneous performance, and the difference among all weak trackers over time. In addition, we design a Siamese network to define the loss of each weak tracker for the proposed hedge method. Extensive experiments on large benchmark datasets demonstrate the effectiveness of the proposed algorithm against the state-of-the-art tracking methods. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
01628828
Volume :
41
Issue :
5
Database :
Complementary Index
Journal :
IEEE Transactions on Pattern Analysis & Machine Intelligence
Publication Type :
Academic Journal
Accession number :
135773549
Full Text :
https://doi.org/10.1109/TPAMI.2018.2828817