Abstract
In this work, we address a new feedforward control scheme for the normalized beta (β N) in tokamak plasmas, using the deep reinforcement learning (RL) technique. The deep RL algorithm optimizes an artificial decision-making agent that adjusts the discharge scenario to obtain a given target β N from the state–action–reward sets explored by its own trial and error in a virtual tokamak environment. The virtual environment for the RL training is constructed using a long short-term memory (LSTM) network that imitates the plasma responses to external actuator controls, which is trained using five years’ worth of KSTAR experimental data. The RL agent then experiences numerous discharges with different actuator controls in the LSTM simulator, and its internal parameters are optimized in the direction of maximizing the reward. We analyze a series of KSTAR experiments conducted with the RL-determined scenarios to validate the feasibility of the beta control scheme in a real device. We discuss the successes and limitations of feedforward beta control by RL, and suggest a future research path for this area of study.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.