中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
Do not Lose the Details: Reinforced Representation Learning for High Performance Visual Tracking

文献类型:会议论文

作者Qiang Wang1,3; Mengdan Zhang3; Junliang Xing3; Jin Gao3; Weiming Hu3; Steve Maybank2; Hu, Weiming; Wang, Qiang; Wang, Qiang; Gao, Jin
出版日期2018-07
会议日期2018-7
会议地点Stockholm, Sweden
英文摘要

This work presents a novel end-to-end trainable CNN model for high performance visual object tracking. It learns both low-level fine-grained representations and a high-level semantic embedding space in a mutual reinforced way, and a multi-task learning strategy is proposed to perform the correlation analysis on representations from both levels. In particular, a fully convolutional encoderdecoder network is designed to reconstruct the original visual features from the semantic projections to preserve all the geometric information. Moreover, the correlation filter layer working on the finegrained representations leverages a global context constraint for accurate object appearance modeling. The correlation filter in this layer is updated online efficiently without network fine-tuning. Therefore, the proposed tracker benefits from two complementary effects: the adaptability of the fine-grained correlation analysis and the generalization capability of the semantic embedding. Extensive experimental evaluations on four popular benchmarks demonstrate its state-of-the-art performance.

源URL[http://ir.ia.ac.cn/handle/173211/39071]  
专题自动化研究所_模式识别国家重点实验室_视频内容安全团队
中国科学院自动化研究所
作者单位1.University of Chinese Academy of Sciences
2.Department of Computer Science and Information Systems, Birkbeck College, University of London
3.National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy of Sciences
推荐引用方式
GB/T 7714
Qiang Wang,Mengdan Zhang,Junliang Xing,et al. Do not Lose the Details: Reinforced Representation Learning for High Performance Visual Tracking[C]. 见:. Stockholm, Sweden. 2018-7.

入库方式: OAI收割

来源:自动化研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。