Grasp State Assessment of Deformable Objects Using Visual-Tactile Fusion Perception
文献类型:会议论文
作者 | Cui SW(崔少伟)1,2![]() ![]() ![]() ![]() ![]() |
出版日期 | 2020-05 |
会议日期 | 2022-5-31 |
会议地点 | Paris, France |
DOI | 10.1109/ICRA40945.2020.9196787 |
英文摘要 | Humans can quickly determine the force required to grasp a deformable object to prevent its sliding or excessive deformation through vision and touch, which is still a challenging task for robots. To address this issue, we propose a novel 3D convolution-based visual-tactile fusion deep neural network (C3D-VTFN) to evaluate the grasp state of various deformable objects in this paper. Specifically, we divide the grasp states of deformable objects into three categories of sliding, appropriate and excessive. Also, a dataset for training and testing the proposed network is built by extensive grasping and lifting experiments with different widths and forces on 16 various deformable objects with a robotic arm equipped with a wrist camera and a tactile sensor. As a result, a classification accuracy as high as 99.97% is achieved. Furthermore, some delicate grasp experiments based on the proposed network are implemented in this paper. The experimental results demonstrate that the C3D-VTFN is accurate and efficient enough for grasp state assessment, which can be widely applied to automatic force control, adaptive grasping, and other visual-tactile spatiotemporal sequence learning problems. |
语种 | 英语 |
URL标识 | 查看原文 |
源URL | [http://ir.ia.ac.cn/handle/173211/48970] ![]() |
专题 | 智能机器人系统研究 |
通讯作者 | Wang S(王硕) |
作者单位 | 1.中国科学院大学 2.中国科学院自动化研究所 |
推荐引用方式 GB/T 7714 | Cui SW,Wang R,Wei JH,et al. Grasp State Assessment of Deformable Objects Using Visual-Tactile Fusion Perception[C]. 见:. Paris, France. 2022-5-31. |
入库方式: OAI收割
来源:自动化研究所
其他版本
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。