🤲 Im2Hands

Learning Attentive Implicit Representation of Interacting Two-Hand Shapes

CVPR 2023

KAIST
KAIST
KAIST
KAIST, Imperial College London

Paper

Code

Presentation

Abstract

We present Implicit Two Hands (Im2Hands), the first neural implicit representation of two interacting hands. Unlike existing methods on two-hand reconstruction that rely on a parametric hand model and/or low-resolution meshes, Im2Hands can produce fine-grained geometry of two hands with high hand-to-hand and hand-to-image coherency. To handle the shape complexity and interaction context between two hands, Im2Hands models the occupancy volume of two hands - conditioned on an RGB image and coarse 3D keypoints - by two novel attention-based modules responsible for initial occupancy estimation and context-aware occupancy refinement, respectively. Im2Hands first learns per-hand neural articulated occupancy in the canonical space designed for each hand using query-image attention. It then refines the initial two-hand occupancy in the posed space to enhance the coherency between the two hand shapes using query-anchor attention. In addition, we introduce an optional keypoint refinement module to enable robust two-hand shape estimation from predicted hand keypoints in a single-image reconstruction scenario. We experimentally demonstrate the effectiveness of Im2Hands on two-hand reconstruction in comparison to related methods, where ours achieves state-of-the-art results.

Video Results (InterHand2.6M)

Image Results (InterHand2.6M)

Green boxes show penetrations, brown boxes show non-smooth shapes, and purple boxes show shapes with bad image alignment. Our method produces two-hand shapes with better hand-to-image and hand-to-hand coherency, less penetrations, and a higher resolution.

Generalizability Test (RGB2Hands and EgoHands)

Citation

@inproceedings{lee2023im2hands,
    title={Im2Hands: Learning Attentive Implicit Representation of Interacting Two-Hand Shapes},
    author={Lee, Jihyun and Sung, Minhyuk and Choi, Honggyu and Kim, Tae-Kyun},
    booktitle={CVPR},
    year={2023}
  }
      

Acknowledgement

This work is in part sponsored by NST grant (CRC 21011, MSIT) and KOCCA grant (R2022020028, MCST). Minhyuk Sung acknowledges the support of the NRF grant (RS-2023-00209723) and IITP grant (2022-0-00594) funded by the Korean government (MSIT), and grants from Adobe, ETRI, KT, and Samsung Electronics.


Inspired by HumanNeRF