|
作者 | Huang, Yan ; Wang, Liang
|
出版日期 | 2019-11
|
会议日期 | 2019.10.26-2019.11.2
|
会议地点 | Seoul
|
关键词 | Image And Sentence Matching
|
卷号 | 0
|
期号 | 0
|
DOI | 0
|
页码 | 0
|
英文摘要 | Image and sentence matching has drawn much attention
recently, but due to the lack of sufficient pairwise data for
training, most previous methods still cannot well associate
those challenging pairs of images and sentences containing
rarely appeared regions and words, i.e., few-shot content.
In this work, we study this challenging scenario as few-shot
image and sentence matching, and accordingly propose an
Aligned Cross-Modal Memory (ACMM) model to memorize
the rarely appeared content. Given a pair of image and sentence,
the model first includes an aligned memory controller
network to produce two sets of semantically-comparable interface
vectors through cross-modal alignment. Then the
interface vectors are used by modality-specific read and update
operations to alternatively interact with shared memory
items. The memory items persistently memorize crossmodal
shared semantic representations, which can be addressed
out to better enhance the representation of few-shot
content. We apply the proposed model to both conventional
and few-shot image and sentence matching tasks, and
demonstrate its effectiveness by achieving the state-of-theart
performance on two benchmark datasets. |
源文献作者 | Kyoung Mu Lee
|
会议录出版者 | IEEE
|
会议录出版地 | USA
|
URL标识 | 查看原文
|
源URL | [http://ir.ia.ac.cn/handle/173211/25797]  |
专题 | 自动化研究所_智能感知与计算研究中心
|
作者单位 | 中科院自动化所
|
推荐引用方式 GB/T 7714 |
Huang, Yan,Wang, Liang. ACMM: Aligned Cross-Modal Memory for Few-Shot Image and Sentence Matching[C]. 见:. Seoul. 2019.10.26-2019.11.2.
|