中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
Boosted Transformer for Image Captioning

文献类型:期刊论文

作者Li, Jiangyun1,2; Yao, Peng1,2,4; Guo, Longteng3; Zhang, Weicun1,2
刊名APPLIED SCIENCES-BASEL
出版日期2019-08-01
卷号9期号:16页码:15
关键词image captioning self-attention deep learning transformer
DOI10.3390/app9163260
通讯作者Zhang, Weicun(weicunzhang@ustb.edu.cn)
英文摘要Image captioning attempts to generate a description given an image, usually taking Convolutional Neural Network as the encoder to extract the visual features and a sequence model, among which the self-attention mechanism has achieved advanced progress recently, as the decoder to generate descriptions. However, this predominant encoder-decoder architecture has some problems to be solved. On the encoder side, without the semantic concepts, the extracted visual features do not make full use of the image information. On the decoder side, the sequence self-attention only relies on word representations, lacking the guidance of visual information and easily influenced by the language prior. In this paper, we propose a novel boosted transformer model with two attention modules for the above-mentioned problems, i.e., Concept-Guided Attention (CGA) and Vision-Guided Attention (VGA). Our model utilizes CGA in the encoder, to obtain the boosted visual features by integrating the instance-level concepts into the visual features. In the decoder, we stack VGA, which uses the visual information as a bridge to model internal relationships among the sequences and can be an auxiliary module of sequence self-attention. Quantitative and qualitative results on the Microsoft COCO dataset demonstrate the better performance of our model than the state-of-the-art approaches.
资助项目National Nature Science Foundation of China[61671054] ; Beijing Natural Science Foundation[4182038]
WOS研究方向Chemistry ; Materials Science ; Physics
语种英语
WOS记录号WOS:000484444100054
出版者MDPI
资助机构National Nature Science Foundation of China ; Beijing Natural Science Foundation
源URL[http://ir.ia.ac.cn/handle/173211/27241]  
专题自动化研究所_模式识别国家重点实验室_图像与视频分析团队
通讯作者Zhang, Weicun
作者单位1.Univ Sci & Technol Beijing, Sch Automat & Elect Engn, Beijing 100083, Peoples R China
2.Minist Educ, Key Lab Knowledge Automat Ind Proc, Beijing 100083, Peoples R China
3.Chinese Acad Sci, Natl Lab Pattern Recognit, Inst Automat, Beijing 100190, Peoples R China
4.Univ Sci & Technol Beijing, Beijing 100083, Peoples R China
推荐引用方式
GB/T 7714
Li, Jiangyun,Yao, Peng,Guo, Longteng,et al. Boosted Transformer for Image Captioning[J]. APPLIED SCIENCES-BASEL,2019,9(16):15.
APA Li, Jiangyun,Yao, Peng,Guo, Longteng,&Zhang, Weicun.(2019).Boosted Transformer for Image Captioning.APPLIED SCIENCES-BASEL,9(16),15.
MLA Li, Jiangyun,et al."Boosted Transformer for Image Captioning".APPLIED SCIENCES-BASEL 9.16(2019):15.

入库方式: OAI收割

来源:自动化研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。