Abstract

The paper deals with a novel design of an approximate active fault detector for discrete-time stochastic linear Markovian switching systems on the infinite-time horizon. The problem is formulated as an optimization problem with the aim to minimize a general discounted detection cost criterion. The proposed solution is inspired by approximate dynamic programming and reinforcement learning. The active fault detector is trained by a temporal-difference Q-learning algorithm with a linear parametric Q-function approximation adjusted to fit the true Q-function. The main advantage is that this approach is computationally less expensive than a temporal-difference learning with a value function.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call