Abstract
This paper proposes a new hybrid approach for recurrent neural networks (RNN). The basic idea of this approach is to train an input layer by unsupervised learning and an output layer by supervised learning. In this method, the Kohonen algorithm is used for unsupervised learning, and dynamic gradient descent method is used for supervised learning. The performances of the proposed algorithm are compared with backpropagation through time (BPTT) on three benchmark problems. Simulation results show that the performances of the new proposed algorithm exceed the standard backpropagation through time in the reduction of the total number of iterations and in the learning time required in the training process.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Nerrand O, Roussel-Ragot P, Urbani D, Personnaz L, Dreyfus G (1994) Training recurrent neural networks: why and How? An illustration in dynamical process modeling. IEEE Trans Neural Netw 5(2):178–184
Narendra KS, Parthasarathy K (1990) Identification and control of dynamical systems using neural networks. IEEE Trans Neural Netw 1(1):4–27
Li X, Yu W (2002) Dynamic system identification via recurrent multilayer perception. Inf Sci 147:45–63
Guler I, Ubeyli ED (2006) A recurrent neural networks classifier for doppler ultrasound blood flow signal. Pattern Recogn Lett 27:1560–1571
Puskorius GV, Feldkamp LA (1998) A signal processing framework based on dynamic neural networks with application to problems in adaptation, filtering, and classification. Proc IEEE 86:2259–2277
Chtourou S, Chtourou M, Hammami O (2008) A hybrid approach for training recurrent neural networks: application to multi-Step-ahead prediction of noisy and large data sets. Neural Comput Appl 17:245–254
Connor T, Douglas R (1994) Recurrent neural networks and robust time series prediction. IEEE Trans Neural Netw 5(2):240–254
Graves D, Pedrycz W (2009) Fuzzy prediction architecture using recurrent neural networks. Neurocomputing 72:1668–1678
Funahashi KI, Nakamura Y (1993) Approximation of dynamical systems by continuous networks. Neural Netw 6:801–806
Kosmatopoulos EA, Polycarpou MM, Christodoulou MA, Oannou PI (1995) High-order neural network structures for identification of dynamical systems. IEEE Trans Neural Netw 6:422–431
Werbos PJ (1990) Backpropagation through time: What it does and how to do it. IEEE Proc 78(10):1550–1560
Rumelhart DE, Hinton GE, Williams RJ (1986) Learning internal representations by error propagation. Parallel distributed processing: explorations in the microstructures of cognition, vol 1. MIT Press, Cambridge, pp 318–362
Williams RJ, Zipser D (1989) A learning algorithm for continually running fully recurrent neural networks. Neural Comput 1:270–280
Toomarian N, Barben J (1991) Adjoint-function and temporal learning algorithms in neural networks. Adv Neural Inf Process Syst 3:113–120
Toomarian N, Barben J (1992) Learning a trajectory using adjoint functions and teacher forcing. Neural Netw 5(3):473–484
Sim GZ, Chen HH, Lee YC (1992) Green’s function method for fast on-line learning algorithm of recurrent neural networks. Adv Neural Inf Process Syst 4:333–340
Schmidhuber J (1992) A fixed storage O(N3) time complexity learning algorithm for fully recurrent continually running networks. Neural Comput 4(2):243–248
Williams R (1989) Complexity of exact gradient computation algorithms for recurrent neural networks. Northeasterm University, College Comp. Sci., Boston, MA, Tech. Rep. NU-CCS-89-27
Feldkamp LA, Prokhorov DV (1998) Phased backpropagation: a hybrid of BPTT and temporal BP. Proc IEEE Int Joint Conf Neural Netw 3:2262–2267
Wan EA (1990) Temporal backpropagation for FIR neural networks. Proc IEEE Int Joint Conf Neural Netw 1:575–580
Hochneiter S, Schmidhuber J (1997) Long short term memory. Neural Comput 9(8):1735–1780
Atiya AF, Parlos AG (2000) New results on recurrent network training: Unifying the algorithms and accelerating convergence. IEEE Trans Neural Netw 11(3):697–709
De Jesus O, Hagan MT (2007) Backpropagation algorithms for a broad class of dynamics networks. IEEE Trans Neural Netw 18(1):14–27
Song Q, Wu Y, Soh YC (2008) Robust adaptive gradient descent training algorithm for recurrent neural networks in discrete time domain. IEEE Trans Neural Netw 19(11):1841–1853
Mandic DP, Chambers JA (2001) Recurrent neural networks for prediction: learning algorithm, architecture and stability. Wiley, Chichester, UK
Williams R (1992) Training recurrent networks using the extended kalman filter, international joint conference on neural networks, Baltimore, ΙV, pp 241–246
Puskorius GV, Feldkamp LA (1992) Recurrent networks training with decoupled extended Kalman Filter algorithm. In: Proceedings of the 1992 SPIE conference on the science of artificial neural networks, Orlando 1710, pp 461–473
Vartak AA, Georgiopoulos M, Anagnostopoulos GC (2005) On line gauss- Newton—based learning for fully recurrent neural networks. Nonlinear Anal 63:867–876
Ben Nasr M, Chtourou M (2006) A hybrid training algorithm for feedforward neural networks. Neural Process Lett 24(2):107–117
Ben Nasr M, Chtourou M (2009) A fuzzy neighborhood-based training algorithm for feedforward neural networks. Neural Comput Appl 18(2):127–133
Kohonen T (1990) The self- organizing map. Proc IEEE 78(9):1464–1480
Box GE, Jenkins GM (1970) Time series analysis, forecasting and control, San Francisco, Holden Day
Li M, Mechrotra K, Mohan C, Ranka S (1990) Sunspot numbers forecasting using neural network. In: Proc. IEEE Int. Symp. Intell. Contr., pp 524–528
Kim J (1999) Adaptive neuro-fuzzy inference system and their application to non linear dynamical system. Neural Netw 12:1301–1319
Acknowledgments
The authors would like to thank all the anonymous reviewers for their useful suggestions, which improved the quality of this paper.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Nasr, M.B., Chtourou, M. Training recurrent neural networks using a hybrid algorithm. Neural Comput & Applic 21, 489–496 (2012). https://doi.org/10.1007/s00521-010-0506-1
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-010-0506-1