Abstract
Dynamic neural networks (DNNs), which are also known as recurrent neural networks, are often used for nonlinear system identification. The main contribution of this letter is the introduction of an efficient parameterization of a class of DNNs. Having to adjust less parameters simplifies the training problem and leads to more parsimonious models. The parameterization is based on approximation theory dealing with the ability of a class of DNNs to approximate finite trajectories of nonautonomous systems. The use of the proposed parameterization is illustrated through a numerical example, using data from a nonlinear model of a magnetic levitation system.
Original language | English |
---|---|
Pages (from-to) | 983 - 988 |
Journal | IEEE Transactions on Neural Networks |
Volume | 16 |
Issue number | 4 |
DOIs | |
Publication status | Published - 2005 |