File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

유현우

Yu, Hyeonwoo
Lab. of AI and Robotics
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.endPage 2169 -
dc.citation.number 2 -
dc.citation.startPage 2162 -
dc.citation.title IEEE ROBOTICS AND AUTOMATION LETTERS -
dc.citation.volume 7 -
dc.contributor.author Yu, Hyeonwoo -
dc.contributor.author Oh, Jean -
dc.date.accessioned 2023-12-21T14:19:05Z -
dc.date.available 2023-12-21T14:19:05Z -
dc.date.created 2022-02-16 -
dc.date.issued 2022-04 -
dc.description.abstract For effective human-robot teaming, it is important for the robots to be able to share their visual perception with the human operators. In a harsh remote collaboration setting, data compression techniques such as autoencoder can be utilized to obtain and transmit the data in terms of latent variables in a compact form. In addition, to ensure real-time runtime performance even under unstable environments, an anytime estimation approach is desired that can reconstruct the full contents from incomplete information. In this context, we propose a method for imputation of latent variables whose elements are partially lost. To achieve the anytime property with only a few dimensions of variables, exploiting prior information of the category-level is essential. A prior distribution used in variational autoencoders is simply assumed to be isotropic Gaussian regardless of the labels of each training datapoint. This type of flattened prior makes it difficult to perform imputation from the category-level distributions. We overcome this limitation by exploiting a category-specific multi-modal prior distribution in the latent space. The missing elements of the partially transferred data can be sampled, by finding a specific modal according to the remaining elements. Since the method is designed to use partial elements for anytime estimation, it can also be applied for data over-compression. Based on the experiments on the ModelNet and Pascal3D datasets, the proposed approach shows consistently superior performance over autoencoder and variational autoencoder up to 70% data loss. The software is open source and is available from our repository -
dc.identifier.bibliographicCitation IEEE ROBOTICS AND AUTOMATION LETTERS, v.7, no.2, pp.2162 - 2169 -
dc.identifier.doi 10.1109/lra.2022.3142439 -
dc.identifier.issn 2377-3766 -
dc.identifier.scopusid 2-s2.0-85123322313 -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/57246 -
dc.identifier.url https://ieeexplore.ieee.org/document/9681277 -
dc.identifier.wosid 000748560800031 -
dc.language 영어 -
dc.publisher IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC -
dc.title Anytime 3D Object Reconstruction Using Multi-Modal Variational Autoencoder -
dc.type Article -
dc.description.isOpenAccess FALSE -
dc.relation.journalWebOfScienceCategory Robotics -
dc.relation.journalResearchArea Robotics -
dc.type.docType Article -
dc.description.journalRegisteredClass scie -
dc.description.journalRegisteredClass scopus -
dc.subject.keywordAuthor 3D object reconstruction -
dc.subject.keywordAuthor multi-modal variational autoencoder -
dc.subject.keywordAuthor anytime algorithm -
dc.subject.keywordAuthor data imputation -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.