File Download

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

백승렬

Baek, Seungryul
UNIST VISION AND LEARNING LAB.
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.number 20 -
dc.citation.startPage 9724 -
dc.citation.title APPLIED SCIENCES-BASEL -
dc.citation.volume 11 -
dc.contributor.author Cha, Junuk -
dc.contributor.author Saqlain, Muhammad -
dc.contributor.author Lee, Changhwa -
dc.contributor.author Lee, Seongyeong -
dc.contributor.author Lee, Seungeun -
dc.contributor.author Kim, Donguk -
dc.contributor.author Park, Won-Hee -
dc.contributor.author Baek, Seungryul -
dc.date.accessioned 2023-12-21T15:09:52Z -
dc.date.available 2023-12-21T15:09:52Z -
dc.date.created 2021-12-09 -
dc.date.issued 2021-10 -
dc.description.abstract Three-dimensional human pose and shape estimation is an important problem in the computer vision community, with numerous applications such as augmented reality, virtual reality, human computer interaction, and so on. However, training accurate 3D human pose and shape estimators based on deep learning approaches requires a large number of images and corresponding 3D ground-truth pose pairs, which are costly to collect. To relieve this constraint, various types of weakly or self-supervised pose estimation approaches have been proposed. Nevertheless, these methods still involve supervision signals, which require effort to collect, such as unpaired large-scale 3D ground truth data, a small subset of 3D labeled data, video priors, and so on. Often, they require installing equipment such as a calibrated multi-camera system to acquire strong multi-view priors. In this paper, we propose a self-supervised learning framework for 3D human pose and shape estimation that does not require other forms of supervision signals while using only single 2D images. Our framework inputs single 2D images, estimates human 3D meshes in the intermediate layers, and is trained to solve four types of self-supervision tasks (i.e., three image manipulation tasks and one neural rendering task) whose ground-truths are all based on the single 2D images themselves. Through experiments, we demonstrate the effectiveness of our approach on 3D human pose benchmark datasets (i.e., Human3.6M, 3DPW, and LSP), where we present the new state-of-the-art among weakly/self-supervised methods. -
dc.identifier.bibliographicCitation APPLIED SCIENCES-BASEL, v.11, no.20, pp.9724 -
dc.identifier.doi 10.3390/app11209724 -
dc.identifier.issn 2076-3417 -
dc.identifier.scopusid 2-s2.0-85117613139 -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/55176 -
dc.identifier.url https://www.mdpi.com/2076-3417/11/20/9724 -
dc.identifier.wosid 000716409000001 -
dc.language 영어 -
dc.publisher MDPI -
dc.title Towards Single 2D Image-Level Self-Supervision for 3D Human Pose and Shape Estimation -
dc.type Article -
dc.description.isOpenAccess TRUE -
dc.relation.journalWebOfScienceCategory Chemistry, Multidisciplinary; Engineering, Multidisciplinary; Materials Science, Multidisciplinary; Physics, Applied -
dc.relation.journalResearchArea Chemistry; Engineering; Materials Science; Physics -
dc.type.docType Article -
dc.description.journalRegisteredClass scie -
dc.description.journalRegisteredClass scopus -
dc.subject.keywordAuthor deep learning -
dc.subject.keywordAuthor human body pose estimation -
dc.subject.keywordAuthor human body mesh estimation -
dc.subject.keywordAuthor neural rendering -
dc.subject.keywordAuthor self-supervised learning -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.