中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
X-GACMN: An X-Shaped Generative Adversarial Cross-Modal Network with Hypersphere Embedding

文献类型:会议论文

作者Weikuo Guo1; Jian Liang2,3; Xiangwei Kong1; Lingxiao Song2; Ran He2,3
出版日期2018
会议日期2018
会议地点Australia
英文摘要

How to bridge heterogeneous gap between different modalities is one of the main challenges in cross-modal retrieval task. Most existing methods try to tackle this problem by projecting data from different modalities into a common space. In this paper, we introduce a novel X-Shaped Generative Adversarial Cross-Modal Network (X-GACMN) to learn a better common space between different modalities. Specifically, the proposed architecture combines the process of synthetic data generation and distribution adapting into a unified framework to make sure the heterogeneous modality distributions similar to each other in the learned common subspace. To promote the discriminative ability, a new loss function that combines intra-modality angular softmax loss and cross-modality pair-wise consistent loss is further imposed on the common space, hence the learned features can well preserve both intermodality structure and intra-modality structure on a hypersphere manifold. Extensive experiments on three benchmark datasets show the effectiveness of the proposed approach.

源URL[http://ir.ia.ac.cn/handle/173211/23807]  
专题自动化研究所_智能感知与计算研究中心
通讯作者Xiangwei Kong
作者单位1.Dalian University of Technology
2.University of Chinese Academy of Science(UCAS)
3.CRIPAC and NLPR, CASIA
推荐引用方式
GB/T 7714
Weikuo Guo,Jian Liang,Xiangwei Kong,et al. X-GACMN: An X-Shaped Generative Adversarial Cross-Modal Network with Hypersphere Embedding[C]. 见:. Australia. 2018.

入库方式: OAI收割

来源:自动化研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。