Leveraging Photometric Consistency over Time for Sparsely Supervised Hand-Object Reconstruction
- Yana Hasson ,
- Bugra Tekin ,
- Federica Bogo ,
- Ivan Laptev ,
- Marc Pollefeys ,
- Cordelia Schmid
2020 Conference on Computer Vision and Pattern Recognition (CVPR) |
Organized by IEEE
Modeling hand-object manipulations is essential for understanding how humans interact with their environment. While of practical importance, estimating the pose of hands and objects during interactions is challenging due to the large mutual occlusions that occur during manipulation.Recent efforts have been directed towards fully-supervised methods that require large amounts of labeled training samples. Collecting 3D ground-truth data for hand-object interactions, however, is costly, tedious, and error-prone. To overcome this challenge we present a method to leverage photometric consistency across time when annotations are only available for a sparse subset of frames in a video. Our model is trained end-to-end on color images to jointly re-construct hands and objects in 3D by inferring their poses.Given our estimated reconstructions, we differentiably render the optical flow between pairs of adjacent images and use it within the network to warp one frame to another. We then apply a self-supervised photometric loss that relies on the visual consistency between nearby images. We achieve state-of-the-art results on 3D hand-object reconstruction benchmarks and demonstrate that our approach allows us to improve the pose estimation accuracy by leveraging in-formation from neighboring frames in low-data regimes.