File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

백승렬

Baek, Seungryul
UNIST VISION AND LEARNING LAB.
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.conferencePlace US -
dc.citation.conferencePlace Long Beach -
dc.citation.title IEEE Conference on Computer Vision and Pattern Recognition -
dc.contributor.author Baek, Seungryul -
dc.contributor.author Kim, Kwang In -
dc.contributor.author Kim, Tae-Kyun -
dc.date.accessioned 2024-02-01T00:08:32Z -
dc.date.available 2024-02-01T00:08:32Z -
dc.date.created 2019-11-30 -
dc.date.issued 2019-06-18 -
dc.description.abstract Estimating 3D hand meshes from single RGB images is challenging, due to intrinsic 2D-3D mapping ambiguities and limited training data. We adopt a compact parametric 3D hand model that represents deformable and articulated hand meshes. To achieve the model fitting to RGB images, we investigate and contribute in three ways: 1) Neural rendering: inspired by recent work on human body, our hand mesh estimator (HME) is implemented by a neural network and a differentiable renderer, supervised by 2D segmentation masks and 3D skeletons. HME demonstrates good performance for estimating diverse hand shapes and improves pose estimation accuracies. 2) Iterative testing refinement: Our fitting function is differentiable. We iteratively refine the initial estimate using the gradients, in the spirit of iterative model fitting methods like ICP. The idea is supported by the latest research on human body. 3) Self-data augmentation: collecting sized RGB-mesh (or segmentation mask)-skeleton triplets for training is a big hurdle. Once the model is successfully fitted to input RGB images, its meshes i.e. shapes and articulations, are realistic, and we augment view-points on top of estimated dense hand poses. Experiments using three RGB-based benchmarks show that our framework offers beyond state-of-the-art accuracy in 3D pose estimation, as well as recovers dense 3D hand shapes. Each technical component above meaningfully improves the accuracy in the ablation study. -
dc.identifier.bibliographicCitation IEEE Conference on Computer Vision and Pattern Recognition -
dc.identifier.doi 10.1109/CVPR.2019.00116 -
dc.identifier.scopusid 2-s2.0-85071928369 -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/79654 -
dc.publisher Institute of Electrical and Electronics Engineers Inc. -
dc.title Pushing the envelope for RGB-based dense 3D hand pose estimation via neural rendering -
dc.type Conference Paper -
dc.date.conferenceDate 2019-06-16 -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.