中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
Joint Anchor-Feature Refinement for Real-Time Accurate Object Detection in Images and Videos

文献类型:期刊论文

作者Xingyu Chen3,4; Junzhi Yu2,3; Shihan Kong3,4; Zhengxing Wu3,4; Li Wen1; Chen, Xingyu; Yu, Junzhi; Wu, Zhengxing; Kong, Shihan
刊名IEEE Transactions on Circuits and Systems for Video Technology
出版日期2020-03
卷号期号:页码:
关键词Object detection Neural networks Computer vision Deep learning
英文摘要

Object detection has been vigorously investigated for years but fast accurate detection for real-world scenes remains a very challenging problem. Overcoming drawbacks of single-stage detectors, we take aim at precisely detecting objects for static and temporal scenes in real time. Firstly, as a dual refinement mechanism, a novel anchor-offset detection is designed, which includes an anchor refinement, a feature location refinement, and a deformable detection head. This new detection mode is able to simultaneously perform two-step regression and capture accurate
object features. Based on the anchor-offset detection, a dual refinement network (DRNet) is developed for high-performance static detection, where a multi-deformable head is further designed to leverage contextual information for describing objects. As for temporal detection in videos, temporal refinement networks (TRNet) and temporal dual refinement networks (TDRNet) are developed by propagating the refinement information across time. We also propose a soft refinement strategy to temporally match object motion with the previous refinement. Our proposed methods are evaluated on PASCAL VOC, COCO, and ImageNet
VID datasets. Extensive comparisons on static and temporal detection verify the superiority of DRNet, TRNet, and TDRNet. Consequently, our developed approaches run in a fairly fast speed, and in the meantime achieve a significantly enhanced detection accuracy, i.e., 84.4% mAP on VOC 2007, 83.6% mAP on VOC 2012, 69.4% mAP on VID 2017, and 42.4% AP on COCO. Ultimately, producing encouraging results, our methods
are applied to online underwater object detection and grasping with an autonomous system. Codes are publicly available at https://github.com/SeanChenxy/TDRN.

源URL[http://ir.ia.ac.cn/handle/173211/39064]  
专题自动化研究所_复杂系统管理与控制国家重点实验室_先进机器人控制团队
通讯作者Junzhi Yu; Yu, Junzhi
作者单位1.Beihang University
2.Peking University
3.Institute of Automation, Chinese Academy of Science
4.University of Chinese Academy of Sciences
推荐引用方式
GB/T 7714
Xingyu Chen,Junzhi Yu,Shihan Kong,et al. Joint Anchor-Feature Refinement for Real-Time Accurate Object Detection in Images and Videos[J]. IEEE Transactions on Circuits and Systems for Video Technology,2020,无(无):无.
APA Xingyu Chen.,Junzhi Yu.,Shihan Kong.,Zhengxing Wu.,Li Wen.,...&Kong, Shihan.(2020).Joint Anchor-Feature Refinement for Real-Time Accurate Object Detection in Images and Videos.IEEE Transactions on Circuits and Systems for Video Technology,无(无),无.
MLA Xingyu Chen,et al."Joint Anchor-Feature Refinement for Real-Time Accurate Object Detection in Images and Videos".IEEE Transactions on Circuits and Systems for Video Technology 无.无(2020):无.

入库方式: OAI收割

来源:自动化研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。