Deep reinforcement learning in an ultrafiltration system: Optimizing operating pressure and chemical cleaning conditions

Enhancing engineering efficiency and reducing operating costs are permanent subjects that face all engineers over the world. To effectively improve the performance of filtration systems, it is necessary to determine an optimal operating condition beyond conventional methods of periodic and empirical...

Full description

Saved in:
Bibliographic Details
Published inChemosphere (Oxford) Vol. 308; p. 136364
Main Authors Park, Sanghun, Shim, Jaegyu, Yoon, Nakyung, Lee, Sungman, Kwak, Donggeun, Lee, Seungyong, Kim, Young Mo, Son, Moon, Cho, Kyung Hwa
Format Journal Article
LanguageEnglish
Published Elsevier Ltd 01.12.2022
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Enhancing engineering efficiency and reducing operating costs are permanent subjects that face all engineers over the world. To effectively improve the performance of filtration systems, it is necessary to determine an optimal operating condition beyond conventional methods of periodic and empirical operation. Herein, this paper proposes an effective approach to finding an optimal operating strategy using deep reinforcement learning (DRL), particularly for an ultrafiltration (UF) system. Deep learning was developed to represent the UF system utilizing a long-short term memory and provided an environment for DRL. DRL was designed to control three actions; operating pressure, cleaning time, and cleaning concentration. Ultimately, DRL proposed the UF system to actively change the operating pressure and cleaning conditions over time toward better water productivity and operating efficiency. DRL denoted ∼20.9% of specific energy consumption can be reduced by increasing average water flux (39.5–43.7 L m−2 h−1) and reducing operating pressure (0.617–0.540 bar). Moreover, the optimal action of DRL was reasonable to achieve better performance beyond the conventional operation. Crucially, this study demonstrated that due to the nature of DRL, the approach is tractable for engineering systems that have structurally complex relationships among operating conditions and resultants. [Display omitted] •LSTM model represented the performance of the UF with a high accuracy.•DRL used the LSTM model as the environment of reinforcement learning.•DRL controlled operating pressure, chemical cleaning time, and concentration.•DRL proposed the optimal operational strategy for the UF reducing 20.9% of SEC.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 23
ISSN:0045-6535
1879-1298
DOI:10.1016/j.chemosphere.2022.136364