A training algorithm and stability analysis for recurrent neural networks
Training of recurrent neural networks (RNNs) introduces considerable computational complexities due to the need for gradient evaluations. How to get fast convergence speed and low computational complexity remains a challenging and open topic. Besides, the transient response of learning process of RN...
Saved in:
Main Authors: | , , , |
---|---|
Other Authors: | |
Format: | Conference or Workshop Item |
Language: | English |
Published: |
2014
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/101742 http://hdl.handle.net/10220/19738 http://ieeexplore.ieee.org/xpl/login.jsp?tp=&arnumber=6290583&url=http%3A%2F%2Fieeexplore.ieee.org%2Fiel5%2F6269381%2F6289713%2F06290583.pdf%3Farnumber%3D6290583 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-101742 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1017422019-12-06T20:43:45Z A training algorithm and stability analysis for recurrent neural networks Xu, Zhao Song, Qing Wang, Danwei Fan, Haijin School of Electrical and Electronic Engineering International Conference on Information Fusion (FUSION) (15th : 2012) DRNTU::Engineering::Electrical and electronic engineering Training of recurrent neural networks (RNNs) introduces considerable computational complexities due to the need for gradient evaluations. How to get fast convergence speed and low computational complexity remains a challenging and open topic. Besides, the transient response of learning process of RNNs is a critical issue, especially for on-line applications. Conventional RNNs training algorithms such as the backpropagation through time (BPTT) and real-time recurrent learning (RTRL) have not adequately satisfied these requirements because they often suffer from slow convergence speed. If a large learning rate is chosen to improve performance, the training process may become unstable in terms of weight divergence. In this paper, a novel training algorithm of RNN, named robust recurrent simultaneous perturbation stochastic approximation (RRSPSA), is developed with a specially designed recurrent hybrid adaptive parameter and adaptive learning rates. RRSPSA is a powerful novel twin-engine simultaneous perturbation stochastic approximation (SPSA) type of RNN training algorithm. It utilizes specific designed three adaptive parameters to maximize training speed for recurrent training signal while exhibiting certain weight convergence properties with only two objective function measurements as the original SPSA algorithm. The RRSPSA is proved with guaranteed weight convergence and system stability in the sense of Lyapunov function. Computer simulations were carried out to demonstrate applicability of the theoretical results. Published version 2014-06-13T03:15:19Z 2019-12-06T20:43:45Z 2014-06-13T03:15:19Z 2019-12-06T20:43:45Z 2012 2012 Conference Paper Xu, Z., Song, Q., Wang, D., & Fan, H. (2012). A training algorithm and stability analysis for recurrent neural networks. 2012 15th International Conference on Information Fusion (FUSION), 2285-2292. https://hdl.handle.net/10356/101742 http://hdl.handle.net/10220/19738 http://ieeexplore.ieee.org/xpl/login.jsp?tp=&arnumber=6290583&url=http%3A%2F%2Fieeexplore.ieee.org%2Fiel5%2F6269381%2F6289713%2F06290583.pdf%3Farnumber%3D6290583 en © 2012 International Society of Information Fusion. This paper was published in 2012 15th International Conference on Information Fusion (FUSION) and is made available as an electronic reprint (preprint) with permission of International Society of Information Fusion. The paper can be found at the following official URL: http://ieeexplore.ieee.org/xpl/login.jsp?tp=&arnumber=6290583&url=http%3A%2F%2Fieeexplore.ieee.org%2Fiel5%2F6269381%2F6289713%2F06290583.pdf%3Farnumber%3D6290583 . One print or electronic copy may be made for personal use only. Systematic or multiple reproduction, distribution to multiple locations via electronic or other means, duplication of any material in this paper for a fee or for commercial purposes, or modification of the content of the paper is prohibited and is subject to penalties under law. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
country |
Singapore |
collection |
DR-NTU |
language |
English |
topic |
DRNTU::Engineering::Electrical and electronic engineering |
spellingShingle |
DRNTU::Engineering::Electrical and electronic engineering Xu, Zhao Song, Qing Wang, Danwei Fan, Haijin A training algorithm and stability analysis for recurrent neural networks |
description |
Training of recurrent neural networks (RNNs) introduces considerable computational complexities due to the need for gradient evaluations. How to get fast convergence speed and low computational complexity remains a challenging and open topic. Besides, the transient response of learning process of RNNs is a critical issue, especially for on-line applications. Conventional RNNs training algorithms such as the backpropagation through time (BPTT) and real-time recurrent learning (RTRL) have not adequately satisfied these requirements because they often suffer from slow convergence speed. If a large learning rate is chosen to improve performance, the training process may become unstable in terms of weight divergence. In this paper, a novel training algorithm of RNN, named robust recurrent simultaneous perturbation stochastic approximation (RRSPSA), is developed with a specially designed recurrent hybrid adaptive parameter and adaptive learning rates. RRSPSA is a powerful novel twin-engine simultaneous perturbation stochastic approximation (SPSA) type of RNN training algorithm. It utilizes specific designed three adaptive parameters to maximize training speed for recurrent training signal while exhibiting certain weight convergence properties with only two objective function measurements as the original SPSA algorithm. The RRSPSA is proved with guaranteed weight convergence and system stability in the sense of Lyapunov function. Computer simulations were carried out to demonstrate applicability of the theoretical results. |
author2 |
School of Electrical and Electronic Engineering |
author_facet |
School of Electrical and Electronic Engineering Xu, Zhao Song, Qing Wang, Danwei Fan, Haijin |
format |
Conference or Workshop Item |
author |
Xu, Zhao Song, Qing Wang, Danwei Fan, Haijin |
author_sort |
Xu, Zhao |
title |
A training algorithm and stability analysis for recurrent neural networks |
title_short |
A training algorithm and stability analysis for recurrent neural networks |
title_full |
A training algorithm and stability analysis for recurrent neural networks |
title_fullStr |
A training algorithm and stability analysis for recurrent neural networks |
title_full_unstemmed |
A training algorithm and stability analysis for recurrent neural networks |
title_sort |
training algorithm and stability analysis for recurrent neural networks |
publishDate |
2014 |
url |
https://hdl.handle.net/10356/101742 http://hdl.handle.net/10220/19738 http://ieeexplore.ieee.org/xpl/login.jsp?tp=&arnumber=6290583&url=http%3A%2F%2Fieeexplore.ieee.org%2Fiel5%2F6269381%2F6289713%2F06290583.pdf%3Farnumber%3D6290583 |
_version_ |
1681047861298462720 |