framework demonstrates an improvement in gas
holdup, liquid velocity profiles, and energy
consumption, which allows for a scalable approach to
optimizing industrial processes. Experiment results
showed that SAC was the best algorithm with the
optimal helping accuracy and the fastest convergence.
Also, while other trained models rely exclusively on
experience replay, in this high-dimensional,
continuous-action environment, we found that
entropic regularization, which allows balancing
exploration and exploitation, allowed the trained
model to achieve better performance in the reactor.
PPO also did well, but took longer to converge.
Although, DQN works fast on discrete environments
and does not work able to work with continuous
action spaces, which in turn makes it less suitable for
the dynamic needs of the reactor. These findings
highlight the versatility and feasibility of DRL agents
in industrial applications, thus enabling rapid and
efficient reactors management. In future efforts, we
will look forward to verifying the proposed method in
experimental setups, and merging DRL with hybrid
CFD models. This work lays the groundwork for
future improvements in smart process optimization
and offers value to sustainable and efficient industrial
operations
REFERENCES
Degaleesan, S., Dudukovic, M. P., & Pan, Y. (2001).
"Experimental study of gas-liquid flow in bubble
columns." Chemical Engineering Science, 56(21),
5681-5692.
Gao, F., Liu, W., & Sun, Z. (2021). "Real-time optimization
of bioreactors using proximal policy optimization."
Biochemical Engineering Journal, 173, 108085.
Guo, J., Huang, S., & Chen, F. (2019). "Deep Q-learning
for gas flow rate control in multiphase reactors."
Computers & Chemical Engineering, 126, 220-231.
Haarnoja, T., et al. (2018). "Soft actor-critic algorithms and
applications." arXiv preprint arXiv:1812.05905.
Haarnoja, T., et al. (2018). "Soft actor-critic algorithms and
applications." arXiv preprint arXiv:1812.05905.
Kober, J., Bagnell, J. A., & Peters, J. (2013).
"Reinforcement learning in robotics: A survey." The
International Journal of Robotics Research, 32(11),
1238-1274.
Krishna, R., & Van Baten, J. M. (2001). "Scaling up bubble
column reactors with process intensification."
Chemical Engineering Science, 56(21), 6237-6245.
Krishna, R., & Van Baten, J. M. (2001). "Scaling up bubble
column reactors with process intensification."
Chemical Engineering Science, 56(21), 6237-6245.
Li, X., Zhang, Q., & Chen, G. (2020). "Optimization of
chemical reaction conditions using deep reinforcement
learning." AIChE Journal, 66(1), e16885.
Li, Y., Xu, D., & Zhang, P. (2022). "Application of soft
actor-critic in hydrodynamic optimization of
multiphase reactors." Chemical Engineering Journal,
430, 132876.
Mnih, V., et al. (2015). "Human-level control through deep
reinforcement learning." Nature, 518(7540), 529-533.
Ranade, V. V., & Chaudhari, R. V. (2014). Multiphase
Reactors: Theory, Design, and Scaleup. Academic
Press.
Santosh. W. and Sathe, V., Study on the Gas Holdup of
Triangular Pitch and Square Pitch Sparger Geometry in
Bubble Columns, International Journal of Fluid
Mechanics Research, Vol.39, No.1, 2012, pp. 85-97.
Schulman, J., et al. (2017). "Proximal policy optimization
algorithms." arXiv preprint arXiv:1707.06347.
Schulman, J., et al. (2017). "Proximal policy optimization
algorithms." arXiv preprint arXiv:1707.06347.
Sutton, R. S., & Barto, A. G. (2018). Reinforcement
Learning: An Introduction. MIT Press.
Van Hasselt, H., Guez, A., & Silver, D. (2016). "Deep
reinforcement learning with double Q-learning."
Proceedings of the AAAI Conference on Artificial
Intelligence, 30(1).
Versteeg, H. K., & Malalasekera, W. (2007). An
Introduction to Computational Fluid Dynamics: The
Finite Volume Method. Pearson Education.
Wang, Y., Sun, Y., & Zhang, H. (2018). "Application of
reinforcement learning in optimizing stirred tank
reactors." Journal of Process Control, 67, 16-23.
Zhou, R., Liu, J., & Wang, Y. (2016). "Prediction of
hydrodynamic characteristics in bubble columns using
artificial neural networks." Industrial & Engineering
Chemistry Research, 55(9), 2510-2520.