Hands23







Towards A Richer 2D Understanding of Hands at Scale


Tianyi Cheng1, *
Dandan Shan1, *
Ayda Hassen2
Richard Higgins1
David Fouhey3

University of Michigan1, Addis Ababa University2, New York University3
NeurIPS 2023


Paper
Github code
Code
Dataset
Data
Github code
Data Repo


Abstract

As humans, we learn a lot about how to interact with the world by observing others interacting with their hands. To help AI systems obtain a better understanding of hand interactions, we introduce a new model that produces a rich understanding of hand interaction. Our system produces a richer output than past systems at a larger scale. Our outputs include boxes and segments for hands, in-contact objects, and second objects touched by tools as well as contact and grasp type. Supporting this method are annotations of 257K images, 401K hands, 288K objects, and 19K second objects spanning four datasets. We show that our method provides rich information and performs and generalizes well.


Method



Our approach predicts: (1) rich information per-region of interest (ROI) that is done with MLPs on top of standard instance segmentation machinery; and (2) an interaction between pairs of ROIs (here, a hand plus a held object). The final results merge these outputs with a simple algorithm.


Results



Citation



Acknowledgement

This material is based upon work supported by the National Science Foundation under Grant No. 2006619. We thank the University of Michigan DCO for their tireless continued support.