Abstract

Autonomous path planning of automated guided vehicles (AGVs) is an important part of the intelligent logistics systems (ILS), which can greatly improve the abilities of intelligence and automation. Traditional AGVs navigation, electromagnetic navigation, and tape navigation, only navigate on metal belts or magnetic tapes, and the freedom of navigation is greatly reduced. This paper models the AGVs path planning problem as a reinforcement learning model to improve the capabilities of autonomous path planning and freedom of navigation. The path planning method based on the Dueling Double Deep Q Network with Prioritized Experience Reply (Dueling DDQN-PER) is used to learn the AGVs control in a simulation environment of an ILS. Using multi-modal sensory environment information, the AGV's ability to approach the target location and avoid obstacles has been significantly improved. Experimental results show that the AGVs can autonomously plan the path in the environment with obstacles, and has a high success rate and obstacle avoidance ability.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.