Text2Action: Generative Adversarial Synthesis from Language to Action
|dc.identifier.citation||IEEE International Conference on Robotics and Automation, pp.5915 - 5920||ko|
|dc.description.abstract||In this paper, we propose a generative model which learns the relationship between language and human action in order to generate a human action sequence given a sentence describing human behavior. The proposed generative model is a generative adversarial network (GAN), which is based on the sequence to sequence (SEQ2SEQ) model. Using the proposed generative network, we can synthesize various actions for a robot or a virtual agent using a text encoder recurrent neural network (RNN) and an action decoder RNN. The proposed generative network is trained from 29,770 pairs of actions and sentence annotations extracted from MSR-Video-to-Text (MSR-VTT), a large-scale video dataset. We demonstrate that the network can generate human-like actions which can be transferred to a Baxter robot, such that the robot performs an action based on a provided sentence. Results show that the proposed generative network correctly models the relationship between language and action and can generate a diverse set of actions from the same sentence.||ko|
|dc.publisher||IEEE COMPUTER SOC||ko|
|dc.title||Text2Action: Generative Adversarial Synthesis from Language to Action||ko|
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.