Reinforcement learning (RL) has emerged as a promising approach to automating decision processes. This paper explores the application of RL techniques to optimise the polynomial order in the computational mesh when using high-order solvers. Mesh adaptation plays a crucial role in improving the efficiency of numerical simulations by increasing accuracy while reducing the cost. Here, actor-critic RL models based on Proximal Policy Optimization offer an approach for agents to learn optimal mesh modifications based on evolving conditions.The paper provides a strategy for p-adaptation in high-order solvers and includes insights into the main aspects of RL-based mesh adaptation, including the formulation of appropriate reward structures and the interaction between the RL agent and the simulation environment. The proposed strategy does not require a high-fidelity solution during the training process and the formulation is general for any computational mesh and partial differential equation (PDE), solved in a discontinuous Galerkin solver. We discuss the impact of RL-based mesh p-adaptation on computational efficiency and accuracy. We apply the RL p-adaptation strategy to a one-dimensional inviscid Burgers' equation, focusing our analysis on smooth solutions of the equation to showcase its effectiveness. The RL strategy reduces the computational cost and improves accuracy over uniform adaptation, while minimising human intervention.
Read full abstract