Unmanned aerial vehicle (UAV) swarm confrontation jamming offers a cost-effective and long-range countermeasure against hostile swarms. Intelligent decision-making is a key factor in ensuring its effectiveness. In response to the low-timeliness problem caused by linear programming in current algorithms, this paper proposes an intelligent decision-making algorithm for UAV swarm confrontation jamming based on the multi-agent actor–critic (M2AC) model. First, based on Markov games, an intelligent mathematical decision-making model is constructed to transform the confrontation jamming scenario into a symbolized mathematical problem. Second, the indicator function under this learning paradigm is designed by combining the actor–critic algorithm with Markov games. Finally, by employing a reinforcement learning algorithm with multithreaded parallel training–contrastive execution for solving the model, a Markov perfect equilibrium solution is obtained. The experimental results indicate that the algorithm based on M2AC can achieve faster training and decision-making speeds, while effectively obtaining a Markov perfect equilibrium solution. The training time is reduced to less than 50% compared to the baseline algorithm, with decision times maintained below 0.05 s across all simulation conditions. This helps alleviate the low-timeliness problem of UAV swarm confrontation jamming intelligent decision-making algorithms under highly dynamic real-time conditions, leading to more effective and efficient UAV swarm operations in various jamming and electronic warfare scenarios.