File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

나형호

Na, Hyungho
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.conferencePlace SI -
dc.citation.title International Conference on Learning Representations -
dc.contributor.author Na, Hyungho -
dc.contributor.author Lee, Kwanghyeon -
dc.contributor.author Lee, Sumin -
dc.contributor.author Moon, Il-Chul -
dc.date.accessioned 2026-04-09T15:00:05Z -
dc.date.available 2026-04-09T15:00:05Z -
dc.date.created 2026-04-09 -
dc.date.issued 2025-04-25 -
dc.description.abstract In the context of multi-agent reinforcement learning, generalization is a challenge to solve various tasks that may require different joint policies or coordination without relying on policies specialized for each task. We refer to this type of problem as a multi-task, and we train agents to be versatile in this multi-task setting through a single training process. To address this challenge, we introduce TRajectoryclass-Aware Multi-Agent reinforcement learning (TRAMA). In TRAMA, agents recognize a task type by identifying the class of trajectories they are experiencing through partial observations, and the agents use this trajectory awareness or prediction as additional information for action policy. To this end, we introduce three primary objectives in TRAMA: (a) constructing a quantized latent space to generate trajectory embeddings that reflect key similarities among them; (b) conducting trajectory clustering using these trajectory embeddings; and (c) building a trajectory-class-aware policy. Specifically for (c), we introduce a trajectory-class predictor that performs agent-wise predictions on the trajectory class; and we design a trajectory-class representation model for each trajectory class. Each agent takes actions based on this trajectory-class representation along with its partial observation for task-aware execution. The proposed method is evaluated on various tasks, including multi-task problems built upon StarCraft II. Empirical results show further performance improvements over state-of-the-art baselines. -
dc.identifier.bibliographicCitation International Conference on Learning Representations -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/91318 -
dc.language 영어 -
dc.publisher International Conference on Learning Representations -
dc.title Trajectory-Class-Aware Multi-agent Reinforcement Learning -
dc.type Conference Paper -
dc.date.conferenceDate 2025-04-24 -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.