A Deep Reinforcement Learning Approach to Improve the Learning Performance in Process Control

Advanced model-based control methods have been widely used in industrial process control, but excellent performance requires regular maintenance of its model. Reinforcement learning can online update its policy through the observed data by interacting with the environment. Since a fast and stable le...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Industrial & engineering chemistry research 2021-04, Vol.60 (15), p.5504-5515
Hauptverfasser: Bao, Yaoyao, Zhu, Yuanming, Qian, Feng
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Advanced model-based control methods have been widely used in industrial process control, but excellent performance requires regular maintenance of its model. Reinforcement learning can online update its policy through the observed data by interacting with the environment. Since a fast and stable learning process is required to improve the adaptability of the controller, we propose an improved deep deterministic actor critic predictor in this paper, where the immediate reward is separated from the action-value function to provide the actor with reliable gradient information at early stages. Then, an expectation form of policy gradient is developed based on the assumption that the state obeys the normal distribution. Simulation results show that the proposed algorithm achieves a more stable and faster learning procedure than those state-of-art deep reinforcement learning (DRL) algorithms. Meanwhile, the obtained policy achieves a more advantageous performance than the fine-tuned proportional integral and derivative (PID) and linear model predictive controllers, especially for those processes with nonlinearity. These indicate that the improved DRL controller has the potential to become an important tool in practical applications.
ISSN:0888-5885
1520-5045
DOI:10.1021/acs.iecr.0c05678