Abstract

Traffic forecasting is essential in the development of intelligent transportation systems, as it enables the formulation of effective traffic dispatching strategies and contributes to the reduction of traffic congestion. The abundance of research focused on modeling complex spatiotemporal correlations for accurate traffic prediction, however many of these prior works perform feature extraction based solely on prior graph structures, thereby overlooking the latent graph connectivity inherent in the data and degrading a decline in prediction accuracy. In this study, we present a novel Attention-based Multiple Graph Convolutional Recurrent Network (AMGCRN) to capture dynamic and latent spatiotemporal correlations in traffic data. The proposed model comprises two spatial feature extraction modules. Firstly, a dot product attention mechanism is utilized to construct an adaptive graph to extract the similarity of road structure. Secondly, the graph attention mechanism is leveraged to enhance the extraction of local traffic flow features. The outputs of these two spatial feature extraction modules are integrated through a gating mechanism and fed into a Gated Recurrent Unit (GRU) to make spatiotemporal interaction predictions. Experimental results on two real-world traffic datasets demonstrate the superiority of the proposed AMGCRN over state-of-the-art baselines. The results suggest that the proposed model is effective in capturing complex spatiotemporal correlations and achieving about 1% improvements in traffic forecasting.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.