中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
Find Who to Look at: Turning From Action to Saliency

文献类型:期刊论文

作者Xu, Mai2; Liu, Yufan1,2; Hu, Roland3; He, Feng2
刊名IEEE TRANSACTIONS ON IMAGE PROCESSING
出版日期2018-09-01
卷号27期号:9页码:4529-4544
关键词Video analysis saliency prediction face
ISSN号1057-7149
DOI10.1109/TIP.2018.2837106
通讯作者He, Feng(robinleo@buaa.edu.cn)
英文摘要The past decade has witnessed the use of high-level features in saliency prediction for both videos and images. Unfortunately, the existing saliency prediction methods only handle high-level static features, such as face. In fact, high-level dynamic features (also called actions), such as speaking or head turning, are also extremely attractive to visual attention in videos. Thus, in this paper, we propose a data-driven method for learning to predict the saliency of multiple-face videos, by leveraging both static and dynamic features at high-level. Specifically, we introduce an eye-tracking database, collecting the fixations of 39 subjects viewing 65 multiple-face videos. Through analysis on our database, we find a set of high-level features that cause a face to receive extensive visual attention. These high-level features include the static features of face size, center-bias and head pose, as well as the dynamic features of speaking and head turning. Then, we present the techniques for extracting these high-level features. Afterwards, a novel model, namely multiple hidden Markov model (M-HMM), is developed in our method to enable the transition of saliency among faces. In our M-HMM, the saliency transition takes into account both the state of saliency at previous frames and the observed high-level features at the current frame. The experimental results show that the proposed method is superior to other state-of-the-art methods in predicting visual attention on multiple-face videos. Finally, we shed light on a promising implementation of our saliency prediction method in locating the region-of-interest, for video conference compression with high efficiency video coding.
WOS关键词VIDEO CODING HEVC ; VISUAL-ATTENTION ; SPATIOTEMPORAL SALIENCY ; MODEL ; FACE ; EFFICIENCY ; IMAGE ; SCENE ; GAZE ; COMPRESSION
资助项目Natural Key R&D Program of China[2017YFB1002400] ; NSFC projects[61573037] ; Fok Ying-Tong Education Foundation[151061] ; Zhejiang Public Welfare Research Program[2016C31062] ; Natural Science Foundation of Zhejiang Province[LY16F010004]
WOS研究方向Computer Science ; Engineering
语种英语
WOS记录号WOS:000435518500008
出版者IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
资助机构Natural Key R&D Program of China ; NSFC projects ; Fok Ying-Tong Education Foundation ; Zhejiang Public Welfare Research Program ; Natural Science Foundation of Zhejiang Province
源URL[http://ir.ia.ac.cn/handle/173211/27995]  
专题自动化研究所_模式识别国家重点实验室_视频内容安全团队
通讯作者He, Feng
作者单位1.Chinese Acad Sci, Inst Automat, Beijing 100190, Peoples R China
2.Beihang Univ, Sch Elect & Informat Engn, Beijing 100191, Peoples R China
3.Zhejiang Univ, Coll Informat Sci & Elect Engn, Hangzhou 310027, Zhejiang, Peoples R China
推荐引用方式
GB/T 7714
Xu, Mai,Liu, Yufan,Hu, Roland,et al. Find Who to Look at: Turning From Action to Saliency[J]. IEEE TRANSACTIONS ON IMAGE PROCESSING,2018,27(9):4529-4544.
APA Xu, Mai,Liu, Yufan,Hu, Roland,&He, Feng.(2018).Find Who to Look at: Turning From Action to Saliency.IEEE TRANSACTIONS ON IMAGE PROCESSING,27(9),4529-4544.
MLA Xu, Mai,et al."Find Who to Look at: Turning From Action to Saliency".IEEE TRANSACTIONS ON IMAGE PROCESSING 27.9(2018):4529-4544.

入库方式: OAI收割

来源:自动化研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。