DC Field | Value | Language |
---|---|---|
dc.contributor.author | Yoo, Haeun | ko |
dc.contributor.author | Kim, Boeun | ko |
dc.contributor.author | Kim, Jong Woo | ko |
dc.contributor.author | Lee, Jay Hyung | ko |
dc.date.accessioned | 2021-01-28T05:52:31Z | - |
dc.date.available | 2021-01-28T05:52:31Z | - |
dc.date.created | 2020-11-04 | - |
dc.date.issued | 2021-01 | - |
dc.identifier.citation | COMPUTERS & CHEMICAL ENGINEERING, v.144, pp.107133 | - |
dc.identifier.issn | 0098-1354 | - |
dc.identifier.uri | http://hdl.handle.net/10203/280027 | - |
dc.description.abstract | Batch process control represents a challenge given its dynamic operation over a large operating envelope. Nonlinear model predictive control (NMPC) is the current standard for optimal control of batch processes. The performance of conventional NMPC can be unsatisfactory in the presence of uncertainties. Reinforcement learning (RL) which can utilize simulation or real operation data is a viable alternative for such problems. To apply RL to batch process control effectively, however, choices such as the reward function design and value update method must be made carefully. This study proposes a phase segmentation approach for the reward function design and value/policy function representation. In addition, the deep deterministic policy gradient algorithm (DDPG) is modified with Monte-Carlo learning to ensure more stable and efficient learning behavior. A case study of a batch polymerization process producing polyols is used to demonstrate the improvement brought by the proposed approach and to highlight further issues. | - |
dc.language | English | - |
dc.publisher | PERGAMON-ELSEVIER SCIENCE LTD | - |
dc.title | Reinforcement Learning Based Optimal Control of Batch Processes Using Monte-Carlo Deep Deterministic Policy Gradient with Phase Segmentation | - |
dc.type | Article | - |
dc.identifier.wosid | 000598170500004 | - |
dc.identifier.scopusid | 2-s2.0-85096193107 | - |
dc.type.rims | ART | - |
dc.citation.volume | 144 | - |
dc.citation.beginningpage | 107133 | - |
dc.citation.publicationname | COMPUTERS & CHEMICAL ENGINEERING | - |
dc.identifier.doi | 10.1016/j.compchemeng.2020.107133 | - |
dc.contributor.localauthor | Lee, Jay Hyung | - |
dc.contributor.nonIdAuthor | Kim, Boeun | - |
dc.contributor.nonIdAuthor | Kim, Jong Woo | - |
dc.description.isOpenAccess | Y | - |
dc.type.journalArticle | Article | - |
dc.subject.keywordAuthor | Batch process | - |
dc.subject.keywordAuthor | Reinforcement learning | - |
dc.subject.keywordAuthor | Optimal control | - |
dc.subject.keywordAuthor | Actor-Critic | - |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.