Abstract

The paper addresses the issue of extended dissipative learning for a class of delayed recurrent neural networks. Both time-varying delay and time-invariant delay are taken into account. By choosing appropriate Lyapunov–Krasovkii functionals and utilizing some inequalities, several weight learning rules are developed for ensuring the network to be asymptotically stable and extended dissipative. The existence conditions for these learning strategies consist of a few linear matrix inequalities, which are able to be verified readily by Matlab software. Two numerical examples are employed to show the effectiveness and low conservatism of the proposed learning rules.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call