+44 (0)1509 263171
Please use this identifier to cite or link to this item:
|Title: ||On global asymptotic stability of fully connected recurrent neural networks|
|Authors: ||Mandic, Danilo P.|
Bozic, Milorad M.
|Issue Date: ||2000|
|Publisher: ||© IEEE|
|Citation: ||MANDIC, D.P., CHAMBERS, J.A. and BOZIC, M.M., 2000. On global asymptotic stability of fully connected recurrent neural networks. IN: IEEE International Conference on Acoustics, Speech, and Signal Processing, (ICASSP '00), Istanbul, 5-9 June, Vol 6, pp. 3406 - 3409|
|Abstract: ||Conditions for global asymptotic stability (GAS) of a nonlinear relaxation process realized by a recurrent neural network (RNN) are provided. Existence, convergence, and robustness of such a process are analyzed. This is undertaken based upon the contraction mapping theorem (CMT) and the corresponding fixed point iteration (FPI). Upper bounds for such a process are shown to be the conditions of convergence for a commonly analyzed RNN with a linear state dependence|
|Description: ||This is a conference paper [© IEEE]. It is also available at: http://ieeexplore.ieee.org/ Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.|
|Appears in Collections:||Conference Papers and Presentations (Mechanical, Electrical and Manufacturing Engineering)|
Files associated with this item:
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.