File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

오태훈

Oh, Tae Hoon
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Automatic control of simulated moving bed process with deep Q-network

Author(s)
Oh, Tae HoonKim, Jong WooSon, Sang HwanKim, HosooLee, KyungmooLee, Jong Min
Issued Date
2021-06
DOI
10.1016/j.chroma.2021.462073
URI
https://scholarworks.unist.ac.kr/handle/201301/81578
Citation
JOURNAL OF CHROMATOGRAPHY A, v.1647, pp.462073
Abstract
Optimal control of a simulated moving bed (SMB) process is challenging because the system dynamics is represented as nonlinear partial differential-algebraic equations combined with discrete events. In addition, product purity constraints are active at the optimal operating condition, which implies that these constraints can be easily violated by disturbance. Recently, artificial intelligence techniques have received significant attention for their ability to address complex problems, involving a large number of state variables. In this study, a data-based deep Q-network, which is a model-free reinforcement learning method, is applied to the SMB process to train a near-optimal control policy. Using a deep Q-network, the control policy of a complex dynamic system can be trained off-line as long as a sufficient number of data is provided. These data can be efficiently generated by performing numerical simulations in parallel on multiple machines. The on-line computation of the control input using a trained Q-network is fast enough to satisfy the computational time limit for the SMB process. However, because the Q-network does not predict the future state, it is not possible to explicitly impose state constraints. Instead, the state constraints are indirectly imposed by providing a relatively large penalty (negative reward) when the constraints are violate. Furthermore, logic-based switching control is utilized to limit the ranges of the extract and raf-finate purities, which helps to satisfy the state constraints and reduce the regions in the state space for reinforcement learning to explore. The simulation results demonstrate the advantages of applying deep reinforcement learning to control the SMB process. (C) 2021 Elsevier B.V. All rights reserved.
Publisher
ELSEVIER
ISSN
0021-9673
Keyword (Author)
Simulated moving bedReinforcement learningDeep neural networkOptimal control
Keyword
PREDICTIVE CONTROLCHROMATOGRAPHIC-SEPARATIONNEURAL-NETWORKSMODELSTRATEGIESDESIGNCOUNTERCURRENTOPERATIONSYSTEMS

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.