Broad echo state network for multivariate time series prediction

https://doi.org/10.1016/j.jfranklin.2019.01.027Get rights and content

Abstract

In this paper, a broad echo state network with multiple reservoirs in parallel configuration (Broad-ESN) is proposed for a class of multivariate time series prediction. Firstly, through the unsupervised learning algorithm of restricted Boltzmann machine (RBM), the number of reservoirs of Broad-ESN can be determined, such that the dynamic characteristics of a class of multivariate time series can be fully reflected. Secondly, a parameter optimization method based on Davidon–Fletcher–Powell (DFP) quasi-Newton algorithm is proposed to optimize the reservoir parameters of Broad-ESN. Meanwhile, an output weights learning method based on output error is given to train the output weights of Broad-ESN. Thirdly, a sufficient condition for the echo state property of Broad-ESN is given. Finally, four examples are given to verify the effectiveness of Broad-ESN.

Introduction

Echo state network (ESN) [1], [2] is a class of recurrent neural network (RNN) [3], [4], whose hidden layer is replaced by a dynamical reservoir. Compared with RNN, the advantage of ESN is that only the output weights need to be trained, while the reservoir internal weights and input weights are given randomly. Nowadays, ESN has been paid much attention in many input-driven applications, for example, time-series prediction and classification [5], [6], [7], [8], [9], [10], [11], [12], dynamic pattern recognition [13], [14], system modeling or identification [15], [16], [17], [18], filtering or control [19], [20], [21], big data applications [22], [23], etc.

For the multivariate time series [24], [25], [26], due to the increasing of feature information, the traditional ESN can not meet the requirements of prediction performance. Therefore, many improved ESNs have been presented [5], [6], [12], [16], [17], [27], for example, in [12], a new model called adaptive elastic ESN (AEESN) is proposed to overcome the collinearity problem and a sparse solution for multivariate time series is obtained. In [6], a fast subspace decomposition echo state network (FSDESN) is given to solve the ill-posed problem of multivariate time series prediction. In [27], a multi-reservoir Echo State Network based on Sparse Bayesian method (MrBESN) is proposed to solve the problem of selecting reservoir weights. In order to obtain the higher prediction accuracy of multivariate time series, the reservoir sizes of these improved ESNs are usually larger. However, how to use a smaller size reservoir to achieve the higher prediction accuracy is an interesting work.

In addition, for the reservoir parameter optimization problem, the batch gradient descent (BGD) algorithm [5] is usually used to optimize the reservoir parameters of the existing ESNs. When optimization problem is a multi-peak optimization problem, the BGD algorithm may be trapped into the local minimum. For the output weights training problem, the matrix pseudo-inverse method [5], [6], [7] and gradient-based learning method [28] are usually used to train the output weights of the existing ESNs. However, the matrix pseudo-inverse method could cause a significant delay, such that a larger calculation error of computing prediction accuracy could be introduced. In [29], the authors claimed that the trained output weights by the gradient-based learning method could suffer from the danger of trapping into the local minima.

In order to overcome the aforementioned problems, considering the decomposition mechanism of the traditional reservoir, a novel echo state network with multiple reservoirs in parallel configuration, called broad echo state network (Broad-ESN), is proposed for a class of multivariate time series prediction. Compared with the existing ESNs, the dynamic feature of multivariate time series can be adequately reflected by virtue of the multi-reservoir of Broad-ESN. However, for the Broad-ESN, how to determine the number of reservoirs is a difficult problem. The existing literature does not give the corresponding selection rule. In order to determine the number of reservoirs, a restricted Boltzmann machine (RBM) [30], [31] is introduced into the Broad-ESN. Through the unsupervised learning algorithm of RBM, the feature of input information can be extracted, and thus the number of reservoirs will be determined. Because the reservoir parameters of Broad-ESN are significantly increased, the reservoir optimization problem will be transformed into multi-peak optimization problem. Thus, a new optimization method based on the Davidon–Fletcher–Powell (DFP) quasi-Newton algorithm [32], [33] is given to optimize the reservoir parameters of Broad-ESN.

The main contributions of this paper are as follows:

  • 1.

    A Broad-ESN is proposed for a class of multivariate time series prediction. Through the unsupervised learning algorithm of RBM, the number of reservoirs of Broad-ESN can be determined, such that the dynamic characteristics of a class of multivariate time series can be fully reflected.

  • 2.

    A parameter optimization method based on DFP quasi-Newton algorithm is proposed to optimize the reservoir parameters of Broad-ESN.

  • 3.

    An output weights learning method based on output error is given to train the output weights of Broad-ESN. In addition, a sufficient condition for the echo state property of Broad-ESN is given.

The remaining part of this paper is organized as follows. In Section 2, a Broad-ESN with multi-reservoir in parallel configuration is introduced. In Section 3, a parameter optimization method based on DFP quasi-Newton algorithm is given, and an output weights training method based on output error is given. Simulation examples are performed in Section 4. Finally, the conclusion and future work are given in Section 5.

Section snippets

Broad echo state network

In this section, we give an introduction to the multivariate time series prediction method based on Broad-ESN. In Section 2.1, the basic description of ESN is introduced. In Section 2.2, the basic concept of RBM is introduced. In Section 2.3, the proposed prediction model based on Broad-ESN is introduced. The echo state property of Broad-ESN is proved in Section 2.4.

Learning algorithm of reservoir parameters and output weights

In this section, some learning algorithms are given to train the Broad-ESN. In Section 3.1, a parameter optimization method based on DFP quasi-Newton algorithm is given. An output weights learning method based on output error is given in Section 3.2.

Simulation examples

In this section, the Broad-ESN with DFP quasi-Newton algorithm (27)–(32) and output weight learning method (33) is used for multivariate time series prediction. Two benchmark datasets and two real-world datasets are selected to verify the prediction performance of the Broad-ESN. The performance metric is the root mean squared errors (RMSE) of one-step ahead prediction. The definition of RMSE is given as follows:RMSE=n=1T(y(n)d(n))2/(T1)where T is the number of data points in the sample

Conclusion

In this paper, a Broad-ESN is proposed for a class of multivariate time series prediction. According to the unsupervised learning algorithm of RBM, the number of reservoirs of Broad-ESN can be determined. A parameter optimization method based on DFP quasi-Newton algorithm is given to optimize the reservoir parameters of Broad-ESN. Meanwhile, an output weights learning method is given to train the output weights of Broad-ESN. Compared with ESN, Leaky-ESN, AEESN, RVESN and MrBESN, the simulation

Acknowledgment

This work was supported in part by the National Natural Science Foundation of China under Grants 61473070, 61433004, 61627809, and 61773074, and in part by the Fundamental Research Funds for the State Key Laboratory of Synthetical Automation for Process Industries (SAPI) under Grant 2018ZCX22, and in part by the Fundamental Research Funds for the Central Universities under Grant N160406002.

References (37)

  • L. Bozhkov et al.

    Learning to decode human emotions with echo state networks

    Neural Netw.

    (2016)
  • S. Scardapane et al.

    A decentralized training algorithm for echo state networks in distributed big data applications

    Neural Netw.

    (2016)
  • R. Fletcher et al.

    A rapidly convergent descent method for minimization

    Comput. J.

    (1963)
  • P. Rebentrost et al.

    Quantum gradient descent and Newton’s method for constrained polynomial optimization

    (2016)
  • H. Jaeger et al.

    Harnessing nonlinearity: predicting chaotic systems and saving energy in wireless telecommunication

    Science

    (2004)
  • H. Jaeger

    The ‘echo state’ approach to analysing and training recurrent neural networks–with an erratum note

    GMD Report 148

    (2010)
  • H. Jaeger

    A tutorial on training recurrent neural networks, covering BPTT, RURL, EKF and the ‘Echo State Network’ Approach

    Technical Report GMD Report 159

    (2002)
  • H.G. Zhang et al.

    A comprehensive review of stability analysis of continuous-time recurrent neural networks

    IEEE Trans. Neural Netw. Learn. Syst.

    (2014)
  • Cited by (38)

    • Error-output recurrent multi-layer Kernel Reservoir Network for electricity load time series forecasting

      2023, Engineering Applications of Artificial Intelligence
      Citation Excerpt :

      At the same time, three types of classic randomization-based learning algorithms were employed to tackle the time series prediction problem and power systems. For example, ESN achieved good performance in time series prediction (Yao and Wang, 2019; Zheng et al., 2020). ELM played a vital role in short-term wind speed prediction (Wang et al., 2018; Mi et al., 2017).

    • Echo state network with multiple delayed outputs for multiple delayed time series prediction

      2022, Journal of the Franklin Institute
      Citation Excerpt :

      Case 2): For the NARMA-5 Compared with other methods (ESN [1], Leaky-ESN [9], RVESN [32], MAR-DESN [34] and Broad-ESN [33]), the prediction accuracy of these methods is shown in Table 6. The output curves of the MDO-ESN and ESN are shown in Fig. 5, and the error curves of the MDO-ESN and ESN are shown in Fig. 6.

    • Online robust echo state broad learning system

      2021, Neurocomputing
      Citation Excerpt :

      The weights used by their state matrix are randomly generated, which makes them have better generalization ability while improving the learning speed. Several ESN-based methods have been proposed for multivariate time series classification through ingenious constructions of reservoirs[14,15]. In recent years, broad learning system (BLS) is proposed as an emerging and efficient randomized neural network, which is established on the flat network architecture.

    View all citing articles on Scopus
    View full text