Partial Visual-Tactile Fused Learning for Robotic Object Recognition
文献类型:期刊论文
作者 | Zhang T(张涛)1,2,3![]() ![]() ![]() ![]() |
刊名 | IEEE Transactions on Systems, Man, and Cybernetics: Systems
![]() |
出版日期 | 2021 |
页码 | 1-13 |
关键词 | Intelligent robots partial multiview learning (PMVL) visual-tactile fused sensing (VTFS) |
ISSN号 | 2168-2216 |
产权排序 | 1 |
英文摘要 | Currently, visual-tactile fusion learning for robotic object recognition has achieved appealing performance, due to the fact that visual and tactile data can offer complementary information. However: 1) the distinct gap between vision and touch makes it difficult to fully explore the complementary information, which would further lead to performance degradation and 2) most of the existing visual-tactile fused learning methods assume that visual and tactile data are complete, which is often difficult to be satisfied in many real-world applications. In this article, we propose a partial visual-tactile fused (PVTF) framework for robotic object recognition to address these challenges. Specifically, we first employ two modality-specific (MS) encoders to encode partial visual-tactile data into two incomplete subspaces (i.e., visual subspace and tactile subspace). Then, a modality gap mitigated (MGM) network is adopted to discover modality-invariant high-level label information, which is utilized to generate gap loss and further help updating the MS encoders for relatively consistent visual and tactile subspaces generation. In this way, the huge gap between vision and touch is mitigated, which would further contribute to mine the complementary visual-tactile information. Finally, to achieve data completeness and complementary visual-tactile information exploration simultaneously, a cycle subspace leaning technique is proposed to project the incomplete subspaces into a complete subspace by fully exploiting all the obtainable samples, where complete latent representations with maximum complementary information can be learned. A lot of comparative experiments conducted on three visual-tactile datasets validate the advantage of the proposed PVTF framework, by comparing with state-of-the-art baselines. |
资助项目 | National Key Research and Development Program of China[2019YFB1310300] ; National Nature Science Foundation of China[61821005] ; Liaoning Revitalization Talents Program[XLYC1807053] |
WOS研究方向 | Automation & Control Systems ; Computer Science |
语种 | 英语 |
WOS记录号 | WOS:000732098300001 |
资助机构 | National Key Research and Development Program of China under Grant 2019YFB1310300 ; National Nature Science Foundation of China under Grant 61821005 ; Liaoning Revitalization Talents Program under Grant XLYC1807053 |
源URL | [http://ir.sia.cn/handle/173321/29385] ![]() |
专题 | 沈阳自动化研究所_机器人学研究室 |
通讯作者 | Cong Y(丛杨) |
作者单位 | 1.Institute for Robotics and Intelligent Manufacturing, Chinese Academy of Sciences, Shenyang 110016, China 2.University of Chinese Academy of Sciences, Beijing 100049, China 3.State Key Laboratory of Robotics, Shenyang Institute of Automation, Shenyang 110016, China |
推荐引用方式 GB/T 7714 | Zhang T,Cong Y,Dong JH,et al. Partial Visual-Tactile Fused Learning for Robotic Object Recognition[J]. IEEE Transactions on Systems, Man, and Cybernetics: Systems,2021:1-13. |
APA | Zhang T,Cong Y,Dong JH,&Hou DD.(2021).Partial Visual-Tactile Fused Learning for Robotic Object Recognition.IEEE Transactions on Systems, Man, and Cybernetics: Systems,1-13. |
MLA | Zhang T,et al."Partial Visual-Tactile Fused Learning for Robotic Object Recognition".IEEE Transactions on Systems, Man, and Cybernetics: Systems (2021):1-13. |
入库方式: OAI收割
来源:沈阳自动化研究所
其他版本
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。