Abstract

AbstractThis paper presents a Reinforcement Learning application using a recursive least squares (RLS) with an exponential forgetting (EF) factor to solve the Discrete Linear Quadratic Regulator problem. Temporal Difference learning based RLS algorithm is implemented to find a kernel matrix of the action value function (or Q-function) approximated by neural network. Based on the EF RLS, a New Exponential Forgetting (New EF) factor algorithm is developed by adding a covariance term to the forgetting factor to prevent the estimator windup problem. Numerical simulations on a fixed-wing aircraft are performed to show the effectiveness of the new EF RLS.KeywordsReinforcement learningQ-LearningOptimal controlAdaptive controlRecursive least squaresForgetting factor

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call