论文标题
使用深入增强学习训练的人工神经网络对雷诺数字进行强大的主动流量控制
Robust active flow control over a range of Reynolds numbers using an artificial neural network trained through deep reinforcement learning
论文作者
论文摘要
本文着重于使用深钢筋学习(DRL)在一系列雷诺数上进行计算流体动力学模拟的主动流控制。更准确地说,近端策略优化(PPO)方法用于控制四个合成喷气机的质量流量,该质量对称地位于嵌入二维流动域的圆柱体的上和下侧。学习环境分别支持雷诺数字100、200、300和400的四种流程配置。提出了一种新的平滑插值功能,以帮助PPO算法学习设定连续的动作,这对于有效抑制有问题的提升跃升非常重要,并为训练过程提供更好的收敛。结果表明,DRL控制器能够显着减少升力和阻力波动,并在$ $ = 100、200、200、300和400上分别将阻力降低约5.7%,21.6%,32.7%和38.7%。更重要的是,它还可以有效地减少雷诺数的任何以前看不见的价值在60到400之间的阻力。这突出了深神经网络的概括能力,并且是主动流控制的重要里程碑。
This paper focuses on the active flow control of a computational fluid dynamics simulation over a range of Reynolds numbers using deep reinforcement learning (DRL). More precisely, the proximal policy optimization (PPO) method is used to control the mass flow rate of four synthetic jets symmetrically located on the upper and lower sides of a cylinder immersed in a two-dimensional flow domain. The learning environment supports four flow configurations with Reynolds numbers 100, 200, 300 and 400, respectively. A new smoothing interpolation function is proposed to help the PPO algorithm to learn to set continuous actions, which is of great importance to effectively suppress problematic jumps in lift and allow a better convergence for the training process. It is shown that the DRL controller is able to significantly reduce the lift and drag fluctuations and to actively reduce the drag by approximately 5.7%, 21.6%, 32.7%, and 38.7%, at $Re$=100, 200, 300, and 400 respectively. More importantly, it can also effectively reduce drag for any previously unseen value of the Reynolds number between 60 and 400. This highlights the generalization ability of deep neural networks and is an important milestone to active flow control.