X-GACMN: An X-Shaped Generative Adversarial Cross-Modal Network with Hypersphere Embedding
文献类型:会议论文
作者 | Weikuo Guo1; Jian Liang2,3![]() ![]() ![]() |
出版日期 | 2018 |
会议日期 | 2018 |
会议地点 | Australia |
英文摘要 | How to bridge heterogeneous gap between different modalities is one of the main challenges in cross-modal retrieval task. Most existing methods try to tackle this problem by projecting data from different modalities into a common space. In this paper, we introduce a novel X-Shaped Generative Adversarial Cross-Modal Network (X-GACMN) to learn a better common space between different modalities. Specifically, the proposed architecture combines the process of synthetic data generation and distribution adapting into a unified framework to make sure the heterogeneous modality distributions similar to each other in the learned common subspace. To promote the discriminative ability, a new loss function that combines intra-modality angular softmax loss and cross-modality pair-wise consistent loss is further imposed on the common space, hence the learned features can well preserve both intermodality structure and intra-modality structure on a hypersphere manifold. Extensive experiments on three benchmark datasets show the effectiveness of the proposed approach. |
源URL | [http://ir.ia.ac.cn/handle/173211/23807] ![]() |
专题 | 自动化研究所_智能感知与计算研究中心 |
通讯作者 | Xiangwei Kong |
作者单位 | 1.Dalian University of Technology 2.University of Chinese Academy of Science(UCAS) 3.CRIPAC and NLPR, CASIA |
推荐引用方式 GB/T 7714 | Weikuo Guo,Jian Liang,Xiangwei Kong,et al. X-GACMN: An X-Shaped Generative Adversarial Cross-Modal Network with Hypersphere Embedding[C]. 见:. Australia. 2018. |
入库方式: OAI收割
来源:自动化研究所
其他版本
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。