File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

윤성환

Yoon, Sung Whan
Machine Intelligence and Information Learning Lab.
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.conferencePlace US -
dc.citation.conferencePlace Online (due to COVID-19 outbreak) -
dc.citation.title International Conference on Machine Learning -
dc.contributor.author Yoon, Sung Whan -
dc.contributor.author Kim, Do-Yeon -
dc.contributor.author Seo, Jun -
dc.contributor.author Moon, Jaekyun -
dc.date.accessioned 2024-01-31T23:06:05Z -
dc.date.available 2024-01-31T23:06:05Z -
dc.date.created 2020-12-01 -
dc.date.issued 2020-07-15 -
dc.description.abstract Learning novel concepts while preserving prior knowledge is a long-standing challenge in machine learning. The challenge gets greater when a novel task is given with only a few labeled examples, a problem known as incremental few-shot learning. We propose XtarNet, which learns to extract task-adaptive representation (TAR) for facilitating incremental few-shot learning. The method utilizes a backbone network pretrained on a set of base categories while also employing additional modules that are meta-trained across episodes. Given a new task, the novel feature extracted from the meta-trained modules is mixed with the base feature obtained from the pretrained model. The process of combining two different features provides TAR and is also controlled by meta-trained modules. The TAR contains effective information for classifying both novel and base categories. The base and novel classifiers quickly adapt to a given task by utilizing the TAR. Experiments on standard image datasets indicate that XtarNet achieves state-of-the-art incremental few-shot learning performance. The concept of TAR can also be used in conjunction with existing incremental few-shot learning methods; extensive simulation results in fact show that applying TAR enhances the known methods significantly. -
dc.identifier.bibliographicCitation International Conference on Machine Learning -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/78402 -
dc.language 영어 -
dc.publisher International Machine Learning Society (IMLS) -
dc.title XtarNet: Learning to Extract Task-Adaptive Representation for Incremental Few-Shot Learning -
dc.type Conference Paper -
dc.date.conferenceDate 2020-07-12 -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.