中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
Grasp State Assessment of Deformable Objects Using Visual-Tactile Fusion Perception

文献类型:会议论文

作者Cui SW(崔少伟)1,2; Wang R(王睿)2; Wei JH(魏俊杭)1,2; Li FR(李繁荣)1,2; Wang S(王硕)1,2
出版日期2020-05
会议日期2022-5-31
会议地点Paris, France
DOI10.1109/ICRA40945.2020.9196787
英文摘要

Humans can quickly determine the force required to grasp a deformable object to prevent its sliding or excessive deformation through vision and touch, which is still a challenging task for robots. To address this issue, we propose a novel 3D convolution-based visual-tactile fusion deep neural network (C3D-VTFN) to evaluate the grasp state of various deformable objects in this paper. Specifically, we divide the grasp states of deformable objects into three categories of sliding, appropriate and excessive. Also, a dataset for training and testing the proposed network is built by extensive grasping and lifting experiments with different widths and forces on 16 various deformable objects with a robotic arm equipped with a wrist camera and a tactile sensor. As a result, a classification accuracy as high as 99.97% is achieved. Furthermore, some delicate grasp experiments based on the proposed network are implemented in this paper. The experimental results demonstrate that the C3D-VTFN is accurate and efficient enough for grasp state assessment, which can be widely applied to automatic force control, adaptive grasping, and other visual-tactile spatiotemporal sequence learning problems.

语种英语
URL标识查看原文
源URL[http://ir.ia.ac.cn/handle/173211/48970]  
专题智能机器人系统研究
通讯作者Wang S(王硕)
作者单位1.中国科学院大学
2.中国科学院自动化研究所
推荐引用方式
GB/T 7714
Cui SW,Wang R,Wei JH,et al. Grasp State Assessment of Deformable Objects Using Visual-Tactile Fusion Perception[C]. 见:. Paris, France. 2022-5-31.

入库方式: OAI收割

来源:自动化研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。