Abstract

This study presents a novel multi-agent reinforcement learning (MARL) framework for optimizing high-frequency trading strategies. The proposed approach leverages the StarCraft Multi-Agent Challenge (SMAC) environment, adapted for financial markets, to simulate complex trading scenarios. We implement a Value Decomposition Network (VDN) architecture combined with the Multi-Agent Proximal Policy Optimization (MAPPO) algorithm to coordinate multiple trading agents. The framework is evaluated using high-frequency limit order book data from the FI-2010 dataset, augmented with derived features to capture market microstructure dynamics. Experimental results demonstrate that our MARL-based strategy significantly outperforms traditional algorithmic trading approaches and single-agent reinforcement learning models. The strategy achieves a Sharpe ratio of 2.87 and a maximum drawdown of 12.3%, showcasing superior risk-adjusted returns and robust risk management. Comparative analysis reveals a 9.8% improvement in annualized returns over a single-agent Deep Q-Network approach. Furthermore, the implementation of our strategy shows a positive impact on market quality metrics, including a 2.3% decrease in effective spread and a 15% reduction in price impact. These findings suggest that the proposed MARL framework not only enhances trading performance but also contributes to market stability and efficiency in high-frequency trading environments.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.