Please use this identifier to cite or link to this item:
Title: Exponential stability of delayed recurrent neural networks with Markovian jumping parameters
Authors: Wang, Z
Liu, Y
Yu, L
Iu, X
Keywords: Recurrent neural networks;Markovian jumping parameters;Time delays;Stochastic systems;Global exponential stability in the mean square;Linear matrix inequality
Issue Date: 2006
Publisher: Elsevier
Citation: Physics Letters A, 356(4-5): 346-352, Aug 2006
Abstract: In this Letter, the global exponential stability analysis problem is considered for a class of recurrent neural networks (RNNs) with time delays and Markovian jumping parameters. The jumping parameters considered here are generated from a continuous-time discrete-state homogeneous Markov process, which are governed by a Markov process with discrete and finite state space. The purpose of the problem addressed is to derive some easy-to-test conditions such that the dynamics of the neural network is stochastically exponentially stable in the mean square, independent of the time delay. By employing a new Lyapunov–Krasovskii functional, a linear matrix inequality (LMI) approach is developed to establish the desired sufficient conditions, and therefore the global exponential stability in the mean square for the delayed RNNs can be easily checked by utilizing the numerically efficient Matlab LMI toolbox, and no tuning of parameters is required. A numerical example is exploited to show the usefulness of the derived LMI-based stability conditions.
Description: This is the post print version of the article. The official published version can be obtained from the link below - Copyright 2006 Elsevier Ltd.
ISSN: 0375-9601
Appears in Collections:Computer Science
Dept of Computer Science Research Papers

Files in This Item:
File Description SizeFormat 
Fulltext.pdf138.08 kBAdobe PDFView/Open

Items in BURA are protected by copyright, with all rights reserved, unless otherwise indicated.