Abstract

This article addresses the robot pathfinding problem with environmental disturbances, where a solution to this problem must consider potential risks inherent in an uncertain and stochastic environment. For example, the movements of an underwater robot can be seriously disturbed by ocean currents, and thus any applied control actions to the robot cannot exactly lead to the desired locations. Reinforcement learning is a formal methodology that has been extensively studied in many sequential decision-making domains with uncertainty, but most reinforcement learning algorithms consider only a single objective encoded by a scalar reward. However, the robot pathfinding problem with environmental disturbances naturally promotes multiple conflicting objectives. Specifically, in this work, the robot has to minimise its moving distance so as to save energy, and, moreover, it has to keep away from unsafe regions as far as possible. To this end, we first propose a multiobjective model-free learning framework, and then proceed to investigate an appropriate action selection strategy by improving a baseline with respect to two dimensions. To demonstrate the effectiveness of the proposed learning framework and evaluate the performance of three action selection strategies, we also carry out an empirical study in a simulated environment.

Highlights

  • The pathfinding problem has been extensively studied in the robot domain, where the robot has to generate a collision-free path in a given or an unknown environment.[1,2] In this work, we seek to address the robot pathfinding problem with environmental disturbances

  • D2: the learning agent should pay more attention to potentially promising actions so that the learning process can quickly coverage to the optimal policy); and an empirical study to demonstrate the effectiveness of the proposed model-free learning framework, as well as to evaluate the performance of three action selection strategies in a simulated environment

  • We focus on the robot pathfinding problem in an initially unknown and stochastic environment with conflicting objectives

Read more

Summary

Introduction

A multiobjective model-free learning framework that can handle multiple conflicting objectives; an action selection strategy for the pathfinding problem with respect to two dimensions D2: the learning agent should pay more attention to potentially promising actions so that the learning process can quickly coverage to the optimal policy); and an empirical study to demonstrate the effectiveness of the proposed model-free learning framework, as well as to evaluate the performance of three action selection strategies in a simulated environment.

Related work
Ãðs0ÞÃ ð3Þ
6: Update
16: Update
Conclusion

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.