The escalating usage of automated guided vehicles (AGVs) on a grand scale accentuates the scalability limitations inherent in traditional, centrally-controlled path planning algorithms. Addressing this challenge, we introduce a pioneering decentralized path planning model, DPPDRL, built upon the foundations of deep reinforcement learning. Drawing inspiration from human path planning and dynamic coupling techniques, DPPDRL imbues AGVs with the autonomy to independently dictate and dynamically adapt their paths in real time. Notably, DPPDRL champions the adoption of diverse policies, thereby facilitating efficient collision avoidance and exploration of potential routes. Simulation-based evidence substantiates critical findings, including remarkable improvements in training speed and scalability, heightened adaptability to environmental variations and agent number changes, and applicability across any number of AGVs, provided a reasonable task distribution. However, DPPDRL's performance becomes limited when tasks are densely concentrated. Addressing this issue through task distribution and hierarchical reinforcement learning is a promising direction for future research. These findings underscore the transformative potential of DPPDRL in managing extensive, dynamic AGV operations.
Read full abstract