File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.endPage 1044 -
dc.citation.number 9 -
dc.citation.startPage 1027 -
dc.citation.title INTERNATIONAL JOURNAL OF COMPUTER VISION -
dc.citation.volume 126 -
dc.contributor.author Hattori, Hironori -
dc.contributor.author Lee, Namhoon -
dc.contributor.author Boddeti, Vishnu Naresh -
dc.contributor.author Beainy, Fares -
dc.contributor.author Kitani, Kris M. -
dc.contributor.author Kanade, Takeo -
dc.date.accessioned 2023-12-21T20:11:43Z -
dc.date.available 2023-12-21T20:11:43Z -
dc.date.created 2020-12-02 -
dc.date.issued 2018-09 -
dc.description.abstract We consider scenarios where we have zero instances of real pedestrian data (e.g., a newly installed surveillance system in a novel location in which no labeled real data or unsupervised real data exists yet) and a pedestrian detector must be developed prior to any observations of pedestrians. Given a single image and auxiliary scene information in the form of camera parameters and geometric layout of the scene, our approach infers and generates a large variety of geometrically and photometrically accurate potential images of synthetic pedestrians along with purely accurate ground-truth labels through the use of computer graphics rendering engine. We first present an efficient discriminative learning method that takes these synthetic renders and generates a unique spatially-varying and geometry-preserving pedestrian appearance classifier customized for every possible location in the scene. In order to extend our approach to multi-task learning for further analysis (i.e., estimating pose and segmentation of pedestrians besides detection), we build a more generalized model employing a fully convolutional neural network architecture for multi-task learning leveraging the "free" ground-truth annotations that can be obtained from our pedestrian synthesizer. We demonstrate that when real human annotated data is scarce or non-existent, our data generation strategy can provide an excellent solution for an array of tasks for human activity analysis including detection, pose estimation and segmentation. Experimental results show that our approach (1) outperforms classical models and hybrid synthetic-real models, (2) outperforms various combinations of off-the-shelf state-of-the-art pedestrian detectors and pose estimators that are trained on real data, and (3) surprisingly, our method using purely synthetic data is able to outperform models trained on real scene-specific data when data is limited. -
dc.identifier.bibliographicCitation INTERNATIONAL JOURNAL OF COMPUTER VISION, v.126, no.9, pp.1027 - 1044 -
dc.identifier.doi 10.1007/s11263-018-1077-3 -
dc.identifier.issn 0920-5691 -
dc.identifier.scopusid 2-s2.0-85044024177 -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/48885 -
dc.identifier.url https://link.springer.com/article/10.1007%2Fs11263-018-1077-3 -
dc.identifier.wosid 000441553300009 -
dc.language 영어 -
dc.publisher SPRINGER -
dc.title Synthesizing a Scene-Specific Pedestrian Detector and Pose Estimator for Static Video Surveillance -
dc.type Article -
dc.description.isOpenAccess FALSE -
dc.relation.journalWebOfScienceCategory Computer Science, Artificial Intelligence -
dc.relation.journalResearchArea Computer Science -
dc.type.docType Article -
dc.description.journalRegisteredClass scie -
dc.description.journalRegisteredClass scopus -
dc.subject.keywordAuthor Training with synthetic data -
dc.subject.keywordAuthor Pedestrian detection -
dc.subject.keywordAuthor Pose estimation -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.