This paper presents a novel framework for the active control of transonic airfoil flutter using synthetic jets through deep reinforcement learning (DRL). The research, conducted in a wide range of Mach numbers and flutter velocities, involves an elastically mounted airfoil with two degrees of freedom of pitching and plunging oscillations, subjected to transonic flow conditions at varying Mach numbers. Synthetic jets with zero-mass flux are strategically placed on the airfoil's upper and lower surfaces. This fluid–structure interaction (FSI) problem is treated as the learning environment and is addressed by using the arbitrary Lagrangian–Eulerian lattice Boltzmann flux solver (ALE-LBFS) coupled with a structural solver on dynamic meshes. DRL strategies with proximal policy optimization agents are introduced and trained, based on the velocities probed around the airfoil and the dynamic responses of the structure. The results demonstrate that the pitching and plunging motions of the airfoil in the limited cycle oscillation (LCO) can be effectively alleviated across an extended range of Mach numbers and critical flutter velocities beyond the initial training conditions for control onset. Furthermore, the aerodynamic performance of the airfoil is also enhanced, with an increase in lift coefficient and a reduction in drag coefficient. Even in previously unseen environments with higher flutter velocities, the present strategy is achievable satisfactory control results, including an extended flutter boundary and a reduction in the transonic dip phenomenon. This work underscores the potential of DRL in addressing complex flow control challenges and highlights its potential to expedite the application of DRL in transonic flutter control for aeronautical applications.
Read full abstract