Grasp State Assessment of Deformable Objects Using Visual-Tactile Fusion Perception
文献类型:会议论文
| 作者 | Cui SW(崔少伟)1,2 ; Wang R(王睿)2 ; Wei JH(魏俊杭)1,2 ; Li FR(李繁荣)1,2 ; Wang S(王硕)1,2
|
| 出版日期 | 2020-05 |
| 会议日期 | 2022-5-31 |
| 会议地点 | Paris, France |
| DOI | 10.1109/ICRA40945.2020.9196787 |
| 英文摘要 | Humans can quickly determine the force required to grasp a deformable object to prevent its sliding or excessive deformation through vision and touch, which is still a challenging task for robots. To address this issue, we propose a novel 3D convolution-based visual-tactile fusion deep neural network (C3D-VTFN) to evaluate the grasp state of various deformable objects in this paper. Specifically, we divide the grasp states of deformable objects into three categories of sliding, appropriate and excessive. Also, a dataset for training and testing the proposed network is built by extensive grasping and lifting experiments with different widths and forces on 16 various deformable objects with a robotic arm equipped with a wrist camera and a tactile sensor. As a result, a classification accuracy as high as 99.97% is achieved. Furthermore, some delicate grasp experiments based on the proposed network are implemented in this paper. The experimental results demonstrate that the C3D-VTFN is accurate and efficient enough for grasp state assessment, which can be widely applied to automatic force control, adaptive grasping, and other visual-tactile spatiotemporal sequence learning problems. |
| 语种 | 英语 |
| URL标识 | 查看原文 |
| 源URL | [http://ir.ia.ac.cn/handle/173211/48970] ![]() |
| 专题 | 智能机器人系统研究 |
| 通讯作者 | Wang S(王硕) |
| 作者单位 | 1.中国科学院大学 2.中国科学院自动化研究所 |
| 推荐引用方式 GB/T 7714 | Cui SW,Wang R,Wei JH,et al. Grasp State Assessment of Deformable Objects Using Visual-Tactile Fusion Perception[C]. 见:. Paris, France. 2022-5-31. |
入库方式: OAI收割
来源:自动化研究所
其他版本
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。

