中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
High-Performance Discriminative Tracking with Transformers

文献类型:会议论文

作者Bin, Yu; Ming, Tang; Linyu, Zheng; Guibo, Zhu; Jinqiao, Wang; Hao, Feng; Xuetao, Feng; Hanqing, Lu
出版日期2021-10-11
会议日期2021-10-11--2021-10-17
会议地点online
DOI10.1109/ICCV48922.2021.00971
英文摘要

End-to-end discriminative trackers improve the state of the art significantly, yet the improvement in robustness and efficiency is restricted by the conventional discriminative model, i.e., least-squares based regression. In this paper, we present DTT, a novel single-object discriminative tracker, based on an encoder-decoder Transformer architecture. By self- and encoder-decoder attention mechanisms, our approach is able to exploit the rich scene information in an
end-to-end manner, effectively removing the need for hand-designed discriminative models. In online tracking, given a new test frame, dense prediction is performed at all spatial positions. Not only location, but also bounding box of the target object is obtained in a robust fashion, streamlining
the discriminative tracking pipeline. DTT is conceptually simple and easy to implement. It yields state-of-the-art performance on four popular benchmarks including GOT-10k, LaSOT, NfS, and TrackingNet while running at over 50 FPS, confirming its effectiveness and efficiency. We hope DTT may provide a new perspective for single-object visual tracking.

会议录出版者IEEE
语种英语
源URL[http://ir.ia.ac.cn/handle/173211/48791]  
专题自动化研究所_模式识别国家重点实验室_图像与视频分析团队
作者单位School of Artificial Intelligence, UCAS, China
推荐引用方式
GB/T 7714
Bin, Yu,Ming, Tang,Linyu, Zheng,et al. High-Performance Discriminative Tracking with Transformers[C]. 见:. online. 2021-10-11--2021-10-17.

入库方式: OAI收割

来源:自动化研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。