6D Object Pose Estimation using Few-Shot Instance Segmentation and 3D Matching
文献类型:会议论文
作者 | Wanyi Li; Jia Sun; Yongkang Luo; Peng Wang |
出版日期 | 2019-09-25 |
会议日期 | 2019.12.6-9 |
会议地点 | Xiamen, China |
英文摘要 | 6D object pose estimation is an important but difficult computer vision task. It has many applications such as robotic manipulation and augmented reality. Although a large number of 6D object pose estimation methods have been developed, there are still many challenges, for example, background clutter, foreground occlusion, and lack of annotated training samples. To deal with these difficulties, a compact and effective algorithm for 6D pose estimation using RGB-D data under few-shot condition is presented in this paper. The proposed algorithm consists of two stages. The first stage is few shot instance segmentation, which segments known objects from RGB image. The second stage is 3D matching, which recovers the poses of objects from cropped point clouds. Proposed segmentation method can achieve satisfactory performance using only few labeled samples. Comparison experiments on two challenging datasets are carried out, and the results demonstrate that the proposed method outperforms the state-of-the-art greatly. Recall scores obtained by the proposed method are 74.91% and 55.44%, while of the state-of-the-art are 61.87% and 44.92%, obtaining 13.04% and 10.52% improvement respectively. |
语种 | 英语 |
资助项目 | National Natural Science Foundation of China[91748131] ; National Natural Science Foundation of China[U1613213] ; National Natural Science Foundation of China[61771471] ; Youth Innovation Promotion Association of CAS[2015112] |
源URL | [http://ir.ia.ac.cn/handle/173211/25850] |
专题 | 智能机器人系统研究 |
通讯作者 | Yongkang Luo |
作者单位 | Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China |
推荐引用方式 GB/T 7714 | Wanyi Li,Jia Sun,Yongkang Luo,et al. 6D Object Pose Estimation using Few-Shot Instance Segmentation and 3D Matching[C]. 见:. Xiamen, China. 2019.12.6-9. |
入库方式: OAI收割
来源:自动化研究所
其他版本
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。