Abstract
In this chapter, we present three different recurrent neural network architectures that we employ for the prediction of real-valued time series. All the models reviewed in this chapter can be trained through the previously discussed backpropagation through time procedure. First, we present the most basic version of recurrent neural networks, called Elman recurrent neural network. Then, we introduce two popular gated architectures, which are long short-term memory and the gated recurrent units. We discuss the main advantages of these more sophisticated architectures, especially regarding their capability to process much longer dependencies in time by maintaining an internal memory for longer periods. For each one of the reviewed network, we provide the details and we show the equations for updating the internal state and computing the output at each time step. Then, for each recurrent neural network we also provide a quick overview of its main applications in previous works in the context of real-valued time series forecasting.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
The logistic sigmoid is defined as \(\sigma (x) = \frac{1}{1+e^{-x}}\).
References
Bianchi FM, Kampffmeyer M, Maiorino E, Jenssen R (2017) Temporal overdrive recurrent neural network. arXiv:170105159
Cai X, Zhang N, Venayagamoorthy GK, Wunsch DC (2007) Time series prediction with recurrent neural networks trained by a hybrid PSO-EA algorithm. Neurocomputing 70(13–15):2342–2353. https://doi.org/10.1016/j.neucom.2005.12.138
Chitsaz H, Shaker H, Zareipour H, Wood D, Amjady N (2015) Short-term electricity load forecasting of buildings in microgrids. Energy Build 99:50–60. https://doi.org/10.1016/j.enbuild.2015.04.011
Cho V (2003) A comparison of three different approaches to tourist arrival forecasting. Tour Manag 24(3):323–330. https://doi.org/10.1016/S0261-5177(02)00068-7
Cho K, Van Merriënboer B, Gulcehre C, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014) Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv:14061078
Chung J, Gülçehre Ç, Cho K, Bengio Y (2014) Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv:1412.3555
Eck D, Schmidhuber J (2002) Finding temporal structure in music: Blues improvisation with LSTM recurrent networks. In: Proceedings of the 2002 12th IEEE workshop on neural networks for signal processing, 2002. IEEE, pp 747–756
Elman JL (1995) Language as a dynamical system. In: Mind as motion: explorations in the dynamics of cognition, pp 195–223
Felder M, Kaifel A, Graves A (2010) Wind power prediction using mixture density recurrent neural networks. Poster P0, vol 153, pp 1–7
Gers FA, Schmidhuber J (2001) Lstm recurrent networks learn simple context-free and context-sensitive languages. IEEE Trans Neural Netw 12(6):1333–1340
Graves A, Mohamed AR, Hinton G (2013) Speech recognition with deep recurrent neural networks. In: 2013 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, pp 6645–6649
Graves A, Schmidhuber J (2005) Framewise phoneme classification with bidirectional LSTM and other neural network architectures. iJCNN 2005, Neural Netw 18(56):602–610. https://doi.org/10.1016/j.neunet.2005.06.042
Graves A, Schmidhuber J (2009) Offline handwriting recognition with multidimensional recurrent neural networks. In: Advances in neural information processing systems, pp 545–552
Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780
Mandal P, Senjyu T, Urasaki N, Funabashi T (2006) A neural network based several-hour-ahead electric load forecasting using similar days approach. International J Electr Power Energy Syst 28(6):367–373. https://doi.org/10.1016/j.ijepes.2005.12.007
Ma X, Tao Z, Wang Y, Yu H, Wang Y (2015) Long short-term memory neural network for traffic speed prediction using remote microwave sensor data. Transp Res Part C Emerg Technol 54:187–197. https://doi.org/10.1016/j.trc.2015.03.014
Mori HMH, Ogasawara TOT (1993) A recurrent neural network for short-term load forecasting. In: 1993 proceedings of the second international forum on applications of neural networks to power systems, vol 31, pp 276–281. https://doi.org/10.1109/ANN.1993.264315
Ogata T, Murase M, Tani J, Komatani K, Okuno HG (2007) Two-way translation of compound sentences and arm motions by recurrent neural networks. In: IROS 2007, IEEE/RSJ International conference on intelligent robots and systems, 2007. IEEE, pp 1858–1863
Pawlowski K, Kurach K (2015) Detecting methane outbreaks from time series data with deep neural networks. In: Proceedings rough sets, fuzzy sets, data mining, and granular computing—15th international conference, RSFDGrC 2015, Tianjin, China, November 20–23, 2015, vol 9437, pp 475–484. https://doi.org/10.1007/978-3-319-25783-9_42
Vinyals O, Toshev A, Bengio S, Erhan D (2017) Show and tell: lessons learned from the 2015 MSCOCO image captioning challenge. IEEE Trans Pattern Anal Mach Intell 39(4):652–663. https://doi.org/10.1109/TPAMI.2016.2587640
Zaremba W (2015) An empirical exploration of recurrent network architectures. In: Proceedings of the 32nd international conference on machine learning, Lille, France
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Copyright information
© 2017 The Author(s)
About this chapter
Cite this chapter
Bianchi, F.M., Maiorino, E., Kampffmeyer, M.C., Rizzi, A., Jenssen, R. (2017). Recurrent Neural Network Architectures. In: Recurrent Neural Networks for Short-Term Load Forecasting. SpringerBriefs in Computer Science. Springer, Cham. https://doi.org/10.1007/978-3-319-70338-1_3
Download citation
DOI: https://doi.org/10.1007/978-3-319-70338-1_3
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-70337-4
Online ISBN: 978-3-319-70338-1
eBook Packages: Computer ScienceComputer Science (R0)