Abstract
In this letter, we study the control of probabilistic Boolean control networks (PBCNs) by leveraging a model-free reinforcement learning (RL) technique. In particular, we propose a Q-learning (QL) based approach to address the feedback stabilization problem of PBCNs, and we design optimal state feedback controllers such that the PBCN is stabilized at a given equilibrium point. The optimal controllers are designed for both finite-time stability and asymptotic stability of PBCNs. In order to verify the convergence of the proposed QL algorithm, the obtained optimal policy is compared with the optimal solutions of model-based techniques, namely value iteration (VI) and semi-tensor product (STP) methods. Finally, some PBCN models of gene regulatory networks (GRNs) are considered to verify the obtained results.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.