There are no files associated with this item.
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.citation.startPage | 108004 | - |
dc.citation.title | COMPUTERS & CHEMICAL ENGINEERING | - |
dc.citation.volume | 167 | - |
dc.contributor.author | Kim, Jong Woo | - |
dc.contributor.author | Oh, Tae Hoon | - |
dc.contributor.author | Son, Sang Hwan | - |
dc.contributor.author | Lee, Jong Min | - |
dc.date.accessioned | 2024-03-13T10:05:11Z | - |
dc.date.available | 2024-03-13T10:05:11Z | - |
dc.date.created | 2024-03-13 | - |
dc.date.issued | 2022-11 | - |
dc.description.abstract | The main objective of this study is to develop primal-dual differential dynamic programming (DDP), a model -based reinforcement learning (RL) framework that can handle constrained dynamic optimization problems. DDP has advantages of being able to provide a closed-loop policy and having computational complexity that grows linearly with respect to the time horizon. To take advantage, the DDP should consider optimality and feasibility for the disturbed state during closed-loop operations. Previous DDPs consider the feasibility only for the nominal state condition and can handle limited types of constraints. In this paper, we propose a primal- dual DDP incorporating modified augmented Lagrangian that can handle general nonlinear constraints. We pay special attention to obtain the feasible policy when active set changes due to the state perturbations, using path-following predictor-corrector approach. The developed framework method was applied to van der Pol oscillator and batch crystallization process, thereby validating the key aspects of this study. | - |
dc.identifier.bibliographicCitation | COMPUTERS & CHEMICAL ENGINEERING, v.167, pp.108004 | - |
dc.identifier.doi | 10.1016/j.compchemeng.2022.108004 | - |
dc.identifier.issn | 0098-1354 | - |
dc.identifier.scopusid | 2-s2.0-85139734485 | - |
dc.identifier.uri | https://scholarworks.unist.ac.kr/handle/201301/81573 | - |
dc.identifier.wosid | 000872394000001 | - |
dc.language | 영어 | - |
dc.publisher | PERGAMON-ELSEVIER SCIENCE LTD | - |
dc.title | Primal-dual differential dynamic programming: A model-based reinforcement learning for constrained dynamic optimization | - |
dc.type | Article | - |
dc.description.isOpenAccess | FALSE | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Interdisciplinary Applications; Engineering, Chemical | - |
dc.relation.journalResearchArea | Computer Science; Engineering | - |
dc.type.docType | Article; Early Access | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.subject.keywordAuthor | Reinforcement learning | - |
dc.subject.keywordAuthor | Differential dynamic programming | - |
dc.subject.keywordAuthor | Constrained optimization | - |
dc.subject.keywordAuthor | Augmented Lagrangian | - |
dc.subject.keywordAuthor | Path-following | - |
dc.subject.keywordPlus | ALGORITHM | - |
Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Tel : 052-217-1404 / Email : scholarworks@unist.ac.kr
Copyright (c) 2023 by UNIST LIBRARY. All rights reserved.
ScholarWorks@UNIST was established as an OAK Project for the National Library of Korea.