Abstract
This study introduces an innovative approach that harnesses the Stackelberg differential game framework in combination with advantage actor-critic (A2C) reinforcement learning (RL) to attain adaptive and dynamic control, in the presence of complex faults within spacecraft systems. By reconfiguring the problem of fault tolerance into a strategic interaction involving two pivotal players, specifically, the leader (fault-tolerant control, FTC) and the follower (fault detection observer, FDO), our approach facilitates the dynamic adaptation of their control strategies in response to the system's state, effectively providing a solution for addressing faults within spacecraft systems. This approach overcomes the challenges of cross-optimal indexes commonly found in traditional FTC, resulting in improved efficiency and stability of RL. It harbors substantial potential for applications requiring robust and adaptive FTC strategies. Simulation validation, both the leader and follower can successfully control the spacecraft's velocity before and after a fault occurs, ultimately achieving optimal fault estimation (FE) and FTC objectives.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.