中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
Learning Explicit Contact for Implicit Reconstruction of Hand-Held Objects from Monocular Images

文献类型:会议论文

作者Junxing Hu2,3; Hongwen Zhang4; Zerui Chen1; Mengcheng Li5; Yunlong Wang2; Yebin Liu5; Zhenan Sun2
出版日期2024-03-24
会议日期February 22 – February 25, 2024
会议地点Vancouver, Canada
卷号38
期号3
DOIhttps://doi.org/10.1609/aaai.v38i3.27995
页码2220-2228
英文摘要

Reconstructing hand-held objects from monocular RGB images is an appealing yet challenging task. In this task, contacts between hands and objects provide important cues for recovering the 3D geometry of the hand-held objects. Though recent works have employed implicit functions to achieve impressive progress, they ignore formulating contacts in their frameworks, which results in producing less realistic object meshes. In this work, we explore how to model contacts in an explicit way to benefit the implicit reconstruction of hand-held objects. Our method consists of two components: explicit contact prediction and implicit shape reconstruction. In the first part, we propose a new subtask of directly estimating 3D hand-object contacts from a single image. The part-level and vertex-level graph-based transformers are cascaded and jointly learned in a coarse-to-fine manner for more accurate contact probabilities. In the second part, we introduce a novel method to diffuse estimated contact states from the hand mesh surface to nearby 3D space and leverage diffused contact probabilities to construct the implicit neural representation for the manipulated object. Benefiting from estimating the interaction patterns between the hand and the object, our method can reconstruct more realistic object meshes, especially for object parts that are in contact with hands. Extensive experiments on challenging benchmarks show that the proposed method outperforms the current state of the arts by a great margin. Our code is publicly available at https://junxinghu.github.io/projects/hoi.html.

会议录AAAI Technical Track on Computer Vision II
语种英语
源URL[http://ir.ia.ac.cn/handle/173211/56691]  
专题自动化研究所_智能感知与计算研究中心
通讯作者Zhenan Sun
作者单位1.Inria, DI ENS, CNRS, PSL Research University
2.State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences
3.School of Artificial Intelligence, University of Chinese Academy of Sciences
4.School of Artificial Intelligence, Beijing Normal University
5.Tsinghua University
推荐引用方式
GB/T 7714
Junxing Hu,Hongwen Zhang,Zerui Chen,et al. Learning Explicit Contact for Implicit Reconstruction of Hand-Held Objects from Monocular Images[C]. 见:. Vancouver, Canada. February 22 – February 25, 2024.

入库方式: OAI收割

来源:自动化研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。