A robust recurrent simultaneous perturbation stochastic approximation training algorithm for recurrent neural networks

Training of recurrent neural networks (RNNs) introduces considerable computational complexities due to the need for gradient evaluations. How to get fast convergence speed and low computational complexity remains a challenging and open topic. Besides, the transient response of learning process of RN...

Full description

Saved in:
Bibliographic Details
Main Authors: Xu, Zhao, Song, Qing, Wang, Danwei
Other Authors: School of Electrical and Electronic Engineering
Format: Article
Language:English
Published: 2013
Subjects:
Online Access:https://hdl.handle.net/10356/106964
http://hdl.handle.net/10220/17513
http://dx.doi.org/10.1007/s00521-013-1436-5
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-106964
record_format dspace
spelling sg-ntu-dr.10356-1069642019-12-06T22:22:02Z A robust recurrent simultaneous perturbation stochastic approximation training algorithm for recurrent neural networks Xu, Zhao Song, Qing Wang, Danwei School of Electrical and Electronic Engineering DRNTU::Engineering::Electrical and electronic engineering Training of recurrent neural networks (RNNs) introduces considerable computational complexities due to the need for gradient evaluations. How to get fast convergence speed and low computational complexity remains a challenging and open topic. Besides, the transient response of learning process of RNNs is a critical issue, especially for online applications. Conventional RNN training algorithms such as the backpropagation through time and real-time recurrent learning have not adequately satisfied these requirements because they often suffer from slow convergence speed. If a large learning rate is chosen to improve performance, the training process may become unstable in terms of weight divergence. In this paper, a novel training algorithm of RNN, named robust recurrent simultaneous perturbation stochastic approximation (RRSPSA), is developed with a specially designed recurrent hybrid adaptive parameter and adaptive learning rates. RRSPSA is a powerful novel twin-engine simultaneous perturbation stochastic approximation (SPSA) type of RNN training algorithm. It utilizes three specially designed adaptive parameters to maximize training speed for a recurrent training signal while exhibiting certain weight convergence properties with only two objective function measurements as the original SPSA algorithm. The RRSPSA is proved with guaranteed weight convergence and system stability in the sense of Lyapunov function. Computer simulations were carried out to demonstrate applicability of the theoretical results. 2013-11-08T07:15:14Z 2019-12-06T22:22:02Z 2013-11-08T07:15:14Z 2019-12-06T22:22:02Z 2013 2013 Journal Article Xu, Z., Song, Q., & Wang, D. A robust recurrent simultaneous perturbation stochastic approximation training algorithm for recurrent neural networks. Neural Computing and Application,in press. 0941-0643 https://hdl.handle.net/10356/106964 http://hdl.handle.net/10220/17513 http://dx.doi.org/10.1007/s00521-013-1436-5 en Neural computing and applications
institution Nanyang Technological University
building NTU Library
country Singapore
collection DR-NTU
language English
topic DRNTU::Engineering::Electrical and electronic engineering
spellingShingle DRNTU::Engineering::Electrical and electronic engineering
Xu, Zhao
Song, Qing
Wang, Danwei
A robust recurrent simultaneous perturbation stochastic approximation training algorithm for recurrent neural networks
description Training of recurrent neural networks (RNNs) introduces considerable computational complexities due to the need for gradient evaluations. How to get fast convergence speed and low computational complexity remains a challenging and open topic. Besides, the transient response of learning process of RNNs is a critical issue, especially for online applications. Conventional RNN training algorithms such as the backpropagation through time and real-time recurrent learning have not adequately satisfied these requirements because they often suffer from slow convergence speed. If a large learning rate is chosen to improve performance, the training process may become unstable in terms of weight divergence. In this paper, a novel training algorithm of RNN, named robust recurrent simultaneous perturbation stochastic approximation (RRSPSA), is developed with a specially designed recurrent hybrid adaptive parameter and adaptive learning rates. RRSPSA is a powerful novel twin-engine simultaneous perturbation stochastic approximation (SPSA) type of RNN training algorithm. It utilizes three specially designed adaptive parameters to maximize training speed for a recurrent training signal while exhibiting certain weight convergence properties with only two objective function measurements as the original SPSA algorithm. The RRSPSA is proved with guaranteed weight convergence and system stability in the sense of Lyapunov function. Computer simulations were carried out to demonstrate applicability of the theoretical results.
author2 School of Electrical and Electronic Engineering
author_facet School of Electrical and Electronic Engineering
Xu, Zhao
Song, Qing
Wang, Danwei
format Article
author Xu, Zhao
Song, Qing
Wang, Danwei
author_sort Xu, Zhao
title A robust recurrent simultaneous perturbation stochastic approximation training algorithm for recurrent neural networks
title_short A robust recurrent simultaneous perturbation stochastic approximation training algorithm for recurrent neural networks
title_full A robust recurrent simultaneous perturbation stochastic approximation training algorithm for recurrent neural networks
title_fullStr A robust recurrent simultaneous perturbation stochastic approximation training algorithm for recurrent neural networks
title_full_unstemmed A robust recurrent simultaneous perturbation stochastic approximation training algorithm for recurrent neural networks
title_sort robust recurrent simultaneous perturbation stochastic approximation training algorithm for recurrent neural networks
publishDate 2013
url https://hdl.handle.net/10356/106964
http://hdl.handle.net/10220/17513
http://dx.doi.org/10.1007/s00521-013-1436-5
_version_ 1681043222856466432