| dc.citation.conferencePlace |
CN |
- |
| dc.citation.conferencePlace |
Vancouver, Canada |
- |
| dc.citation.title |
International Conference on Machine Learning |
- |
| dc.contributor.author |
Kim, Jeongmo |
- |
| dc.contributor.author |
Park, Yisak |
- |
| dc.contributor.author |
Kim, Minung |
- |
| dc.contributor.author |
Han, Seungyul |
- |
| dc.date.accessioned |
2025-07-21T09:30:05Z |
- |
| dc.date.available |
2025-07-21T09:30:05Z |
- |
| dc.date.created |
2025-07-19 |
- |
| dc.date.issued |
2025-07-17 |
- |
| dc.description.abstract |
Meta reinforcement learning aims to develop policies that generalize to unseen tasks sampled from a task distribution. While context-based meta-RL methods improve task representation using task latents, they often struggle with out-of-distribution (OOD) tasks. To address this, we propose Task- Aware Virtual Training (TAVT), a novel algorithm that accurately captures task characteristics for both training and OOD scenarios using metric-based representation learning. Our method successfully preserves task characteristics in virtual tasks and employs a state regularization technique to mitigate overestimation errors in state-varying environments. Numerical results demonstrate that TAVT significantly enhances generalization to OOD tasks across various MuJoCo and Meta-World environments. Our code is available at https://github.com/JM-Kim-94/tavt.git. |
- |
| dc.identifier.bibliographicCitation |
International Conference on Machine Learning |
- |
| dc.identifier.uri |
https://scholarworks.unist.ac.kr/handle/201301/87472 |
- |
| dc.identifier.url |
https://icml.cc/virtual/2025/poster/44998 |
- |
| dc.language |
영어 |
- |
| dc.publisher |
International Conference on Machine Learning |
- |
| dc.title |
Task-Aware Virtual Training: Enhancing Generalization in Meta-Reinforcement Learning for Out-of-Distribution Tasks |
- |
| dc.type |
Conference Paper |
- |
| dc.date.conferenceDate |
2025-07-13 |
- |