Abstract

To realize rapid, safe and optimal avoidance for spacecraft in complex threat situations and incomplete-information conditions, and considering the problems in existing hierarchical decision-making modes, a learning-based spacecraft reactive anti-hostile-rendezvous (AHR) maneuver control framework is proposed in this paper. First, a fundamental AHR action planner based on the interfered fluid dynamical system (IFDS) is proposed, in which the avoidance strategies for hostile and non-adversarial threats are appropriately integrated. An avoidance strategy matched with the IFDS is further designed by specifically considering high-intensity confrontation situations where a threat maintains persistent pursuit. Then, the “state-action” solution methods based on deep reinforcement learning (DRL) and multi-agent-DRL (MADRL) methods for orbital avoidance maneuvers are proposed to control the avoidance time and directions by optimizing the IFDS parameters. In the offline training stage, the DRL method will first simulate the interactive processes between the spacecraft and non-adversarial threats so that the spacecraft can gradually adapt to complex space environments and generate the relative primary avoidance maneuver strategies. On this basis, for hostile threats, the MADRL-based solution method of orbit game strategies is further established, in which the spacecraft and hostile threats are constructed as two types of opposite agents to simulate the game processes so that the strategy levels can be bilaterally evolved and the AHR success rates can be improved. After training, the corresponding neural networks are extracted from the agents in the DRL-based and MADRL-based methods for reactive AHR maneuver control. Moreover, a progressive agent training mechanism matched with the above DRL-based and MADRL-based solution methods is proposed, which can reduce the invalid interactions in initial training episodes, improve the training efficiency, and enhance the generalization ability of orbital game strategies. Finally, the effectiveness of the proposed framework is demonstrated by simulations.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call