中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
Directional Deep Embedding and Appearance Learning for Fast Video Object Segmentation

文献类型:期刊论文

作者Yin, Yingjie1,2,3; Xu, De1,2; Wang, Xingang1,2; Zhang, Lei3
刊名IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS
出版日期2021-02-12
页码11
关键词Feature extraction Kernel Object segmentation Faces Probabilistic logic Learning systems Image segmentation Deep appearance learning directional deep embedding learning directional statistics-based learning video object segmentation (VOS)
ISSN号2162-237X
DOI10.1109/TNNLS.2021.3054769
通讯作者Yin, Yingjie(haidaying.jie@163.com)
英文摘要Most recent semisupervised video object segmentation (VOS) methods rely on fine-tuning deep convolutional neural networks online using the given mask of the first frame or predicted masks of subsequent frames. However, the online fine-tuning process is usually time-consuming, limiting the practical use of such methods. We propose a directional deep embedding and appearance learning (DDEAL) method, which is free of the online fine-tuning process, for fast VOS. First, a global directional matching module (GDMM), which can be efficiently implemented by parallel convolutional operations, is proposed to learn a semantic pixel-wise embedding as an internal guidance. Second, an effective directional appearance model-based statistics is proposed to represent the target and background on a spherical embedding space for VOS. Equipped with the GDMM and the directional appearance model learning module, DDEAL learns static cues from the labeled first frame and dynamically updates cues of the subsequent frames for object segmentation. Our method exhibits the state-of-the-art VOS performance without using online fine-tuning. Specifically, it achieves a J & F mean score of 74.8% on DAVIS 2017 data set and an overall score G of 71.3% on the large-scale YouTube-VOS data set, while retaining a speed of 25 fps with a single NVIDIA TITAN Xp GPU. Furthermore, our faster version runs 31 fps with only a little accuracy loss.
资助项目National Natural Science Foundation of China[61703398] ; Science and Technology Program of Beijing Municipal Science and Technology Commission[Z191100008019004] ; Hong Kong Research Grants Council (RGC) General Research Fund (GRF)[PolyU 152135/16E] ; Hong Kong Scholars Program[XJ2017031]
WOS研究方向Computer Science ; Engineering
语种英语
WOS记录号WOS:000733511700001
出版者IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
资助机构National Natural Science Foundation of China ; Science and Technology Program of Beijing Municipal Science and Technology Commission ; Hong Kong Research Grants Council (RGC) General Research Fund (GRF) ; Hong Kong Scholars Program
源URL[http://ir.ia.ac.cn/handle/173211/47109]  
专题精密感知与控制研究中心_精密感知与控制
通讯作者Yin, Yingjie
作者单位1.Chinese Acad Sci, Res Ctr Precis Sensing & Control, Inst Automat, Beijing 100190, Peoples R China
2.Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing 100049, Peoples R China
3.Hong Kong Polytech Univ, Dept Comp, Hong Kong, Peoples R China
推荐引用方式
GB/T 7714
Yin, Yingjie,Xu, De,Wang, Xingang,et al. Directional Deep Embedding and Appearance Learning for Fast Video Object Segmentation[J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS,2021:11.
APA Yin, Yingjie,Xu, De,Wang, Xingang,&Zhang, Lei.(2021).Directional Deep Embedding and Appearance Learning for Fast Video Object Segmentation.IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS,11.
MLA Yin, Yingjie,et al."Directional Deep Embedding and Appearance Learning for Fast Video Object Segmentation".IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS (2021):11.

入库方式: OAI收割

来源:自动化研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。