Visual-Tactile Fused Graph Learning for Object Clustering
文献类型:期刊论文
作者 | Zhang T(张涛)1,2,3![]() ![]() ![]() ![]() |
刊名 | IEEE Transactions on Cybernetics
![]() |
出版日期 | 2021 |
页码 | 1-15 |
关键词 | Clustering graph learning unsupervised learning visual-tactile fused sensing |
ISSN号 | 2168-2267 |
产权排序 | 1 |
英文摘要 | Object clustering has received considerable research attention most recently. However, 1) most existing object clustering methods utilize visual information while ignoring important tactile modality, which would inevitably lead to model performance degradation and 2) simply concatenating visual and tactile information via multiview clustering method can make complementary information to not be fully explored, since there are many differences between vision and touch. To address these issues, we put forward a graph-based visual-tactile fused object clustering framework with two modules: 1) a modality-specific representation learning module MR and 2) a unified affinity graph learning module MU. Specifically, MR focuses on learning modality-specific representations for visual-tactile data, where deep non-negative matrix factorization (NMF) is adopted to extract the hidden information behind each modality. Meanwhile, we employ an autoencoder-like structure to enhance the robustness of the learned representations, and two graphs to improve its compactness. Furthermore, MU highlights how to mitigate the differences between vision and touch, and further maximize the mutual information, which adopts a minimizing disagreement scheme to guide the modality-specific representations toward a unified affinity graph. To achieve ideal clustering performance, a Laplacian rank constraint is imposed to regularize the learned graph with ideal connected components, where noises that caused wrong connections are removed and clustering labels can be obtained directly. Finally, we propose an efficient alternating iterative minimization updating strategy, followed by a theoretical proof to prove framework convergence. Comprehensive experiments on five public datasets demonstrate the superiority of the proposed framework. |
WOS关键词 | FUSION ; RECOGNITION |
资助项目 | Major Project of the New Generation of Artificial Intelligence[2018AAA0102905] ; National Natural Science Foundation of China[61821005] ; National Natural Science Foundation of China[62003336] ; Liaoning Revitalization Talents Program[XLYC1807053] ; Nature Foundation of Liaoning Province of China[2020KF-11-01] |
WOS研究方向 | Automation & Control Systems ; Computer Science |
语种 | 英语 |
WOS记录号 | WOS:000732294000001 |
资助机构 | Major Project of the New Generation of Artificial Intelligence under Grant 2018AAA0102905 ; National Natural Science Foundation of China under Grant 61821005 and Grant 62003336 ; Liaoning Revitalization Talents Program under Grant XLYC1807053 ; Nature Foundation of Liaoning Province of China under Grant 2020-KF-11-01 |
源URL | [http://ir.sia.cn/handle/173321/29420] ![]() |
专题 | 沈阳自动化研究所_机器人学研究室 |
通讯作者 | Cong Y(丛杨) |
作者单位 | 1.Institutes for Robotics and Intelligent Manufacturing, Chinese Academy of Sciences, Shenyang 110016, China 2.University of Chinese Academy of Sciences, Beijing 100049, China 3.State Key Laboratory of Robotics, Shenyang Institute of Automation, Shenyang 110016, China |
推荐引用方式 GB/T 7714 | Zhang T,Cong Y,Sun G,et al. Visual-Tactile Fused Graph Learning for Object Clustering[J]. IEEE Transactions on Cybernetics,2021:1-15. |
APA | Zhang T,Cong Y,Sun G,&Dong JH.(2021).Visual-Tactile Fused Graph Learning for Object Clustering.IEEE Transactions on Cybernetics,1-15. |
MLA | Zhang T,et al."Visual-Tactile Fused Graph Learning for Object Clustering".IEEE Transactions on Cybernetics (2021):1-15. |
入库方式: OAI收割
来源:沈阳自动化研究所
其他版本
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。