2021 International Conference on 3D Vision (3DV) 2021
DOI: 10.1109/3dv53792.2021.00012
|View full text |Cite
|
Sign up to set email alerts
|

A Skeleton-Driven Neural Occupancy Representation for Articulated Hands

Abstract: adrian.spurr,zicong.fan,otmar.hilliges,siyu.tang}@inf.ethz.ch https://korrawe.github.io/halo Figure 1. We introduce a novel neural implicit surface representation of human hands (HALO) that is fully driven by keypoint-based skeleton articulation. Taking 3D keypoints as input, a fully differentiable implicit occupancy representation produces high-fidelity reconstruction of the hand surface (top row). We show that HALO facilitates the conditional generation of articulated hands that grasp 3D objects in a real… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
11
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1

Relationship

2
6

Authors

Journals

citations
Cited by 35 publications
(12 citation statements)
references
References 62 publications
0
11
0
Order By: Relevance
“…A key problem in computer vision is to understand how humans interact with their surroundings. Because hands are our primary means of manipulation with the physical world, there has been an intense interest in hand-object pose estimation [5, 13-15, 19, 38, 39] and the synthesis of static grasps for a given object [19,21,24,38]. However, human grasping is not limited to a single time instance, but involves a continuous interaction with objects in order to move them.…”
Section: Introductionmentioning
confidence: 99%
“…A key problem in computer vision is to understand how humans interact with their surroundings. Because hands are our primary means of manipulation with the physical world, there has been an intense interest in hand-object pose estimation [5, 13-15, 19, 38, 39] and the synthesis of static grasps for a given object [19,21,24,38]. However, human grasping is not limited to a single time instance, but involves a continuous interaction with objects in order to move them.…”
Section: Introductionmentioning
confidence: 99%
“…We use the trimesh library to detect whether there exists a collision between the hand mesh and the object mesh and compute the max penetration depth between two meshes. We follow the same process as [25,26] to compute I v .…”
Section: Discussionmentioning
confidence: 99%
“…For more accurate reconstruction of hands and manipulated objects, we attempt to combine the advantages of the parametric models and SDFs. Along this direction, previous works [10,14,25,56] attempt to leverage parametric models to learn SDFs from 3D poses or raw scans. In our work, we address a different and more challenging setup of reconstructing hands and objects from monocular RGB images.…”
Section: Introductionmentioning
confidence: 99%
“…More directly related are methods that attempt to generate static grasps given an object and sometimes also information about the hand [1, 2, 20-22, 39, 46]. Generally, these approaches either predict a contact map on the object [1,2,20] or synthesize the joint-angle configuration of the grasping hand [21,22,39,46]. [20] propose a hybrid method, where predicted contact maps on objects are used to refine an initial grasp prediction.…”
Section: Related Workmentioning
confidence: 99%
“…A key problem in computer vision is to understand how humans interact with their surroundings. Because hands are our primary means of manipulation with the physical world, there has been an intense interest in hand-object pose estimation [5, 14-16, 19, 39, 40] and the synthesis of static grasps for a given object [19,21,25,39]. However, human grasping is not limited to a single time instance, but involves a continuous interaction with objects in order to move them.…”
Section: Introductionmentioning
confidence: 99%