Next Article in Journal
Information Geometry Theoretic Measures for Characterizing Neural Information Processing from Simulated EEG Signals
Previous Article in Journal
(Non)Resonance Bonds in Molecular Dynamics Simulations: A Case Study concerning C60 Fullerenes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ensemble and Pre-Training Approach for Echo State Network and Extreme Learning Machine Models

1
School of Science, Civil Aviation Flight University of China, Guanghan 618307, China
2
Business School, Sichuan Normal University, Chengdu 610101, China
*
Author to whom correspondence should be addressed.
Entropy 2024, 26(3), 215; https://doi.org/10.3390/e26030215
Submission received: 25 January 2024 / Revised: 17 February 2024 / Accepted: 27 February 2024 / Published: 28 February 2024
(This article belongs to the Section Multidisciplinary Applications)

Abstract

:
The echo state network (ESN) is a recurrent neural network that has yielded state-of-the-art results in many areas owing to its rapid learning ability and the fact that the weights of input neurons and hidden neurons are fixed throughout the learning process. However, the setting procedure for initializing the ESN’s recurrent structure may lead to difficulties in designing a sound reservoir that matches a specific task. This paper proposes an improved pre-training method to adjust the model’s parameters and topology to obtain an adaptive reservoir for a given application. Two strategies, namely global random selection and ensemble training, are introduced to pre-train the randomly initialized ESN model. Specifically, particle swarm optimization is applied to optimize chosen fixed and global weight values within the network, and the reliability and stability of the pre-trained model are enhanced by employing the ensemble training strategy. In addition, we test the feasibility of the model for time series prediction on six benchmarks and two real-life datasets. The experimental results show a clear enhancement in the ESN learning results. Furthermore, the proposed global random selection and ensemble training strategies are also applied to pre-train the extreme learning machine (ELM), which has a similar training process to the ESN model. Numerical experiments are subsequently carried out on the above-mentioned eight datasets. The experimental findings consistently show that the performance of the proposed pre-trained ELM model is also improved significantly. The suggested two strategies can thus enhance the ESN and ELM models’ prediction accuracy and adaptability.

1. Introduction

Artificial intelligence-based prediction techniques, as a core class of forecasting methods, have been harnessed for tasks such as energy consumption estimation [1,2,3], electric load forecasting [4,5,6], wind speed forecasting [7,8,9], air pollution [10,11,12], traffic flow forecasting [13,14,15], COVID-19 forecasting [16,17,18], stock price prediction [19,20,21], natural language processing [22,23,24], and hydrological forecasting [25,26,27]. The feedforward neural network (FFNN) is a typical artificial intelligence-based prediction model showing research promise in multiple fields. A large body of empirical evidence suggests that these models perform well even when the quality of time series data is questionable. The extreme learning machine (ELM) is an emerging learning technique proposed for generalized single-hidden-layer FFNNs [28]. Different from the learning theory of conventional FFNNs, the hidden layer of a generalized single-layer feedforward network does not need to be tuned in the ELM. The method’s advantages (e.g., simplicity, interpretability, fast learning, and potential for high-quality forecasting results) have garnered close attention in the domains of medical diagnosis [29,30], signal analysis [31,32], tourism prediction [33,34], and image recognition [35,36]. However, due to a lack of recurrent or cyclic connections between neurons in the hidden layer or the output layer, an FFNN only provides a static mapping between input and output layers as information is spread throughout the network in a feedforward manner only. This constraint leads to a restricted memory and limited ability to store information. An FFNN is therefore inadequate for certain sequence tasks.
Recurrent neural networks (RNNs) are neural networks that can capture all information stored in sequence in the previous element. RNNs have feedback connections between their layers, in contrast to the FFNN model. These recurrent connections create rich dynamics and permit data flow across different layers. An RNN can accordingly make use of the information in a relatively long sequence. Specifically, RNNs perform the same tasks for every element in the sequence, with output dependent on all previous computations. These networks can hence reveal nonlinear system behavior. Because RNNs can effectively capture contextual data from a sequence, such networks have drawn wide attention and are being increasingly applied in several prediction contexts.
For example, the long short-term memory (LSTM) neural network is a popular variation of RNN-based representation learning and has been extensively adopted. Bi et al. [37] proposed LSTM networks including multivariate time series data to forecast daily visitation at tourist attractions. Qing et al. [38] devised a solar prediction scheme for hour-ahead solar irradiance prediction via an LSTM network, taking weather forecasting data into account. Abbasimehr et al. [39] constructed a demand forecasting method based on multi-layer LSTM networks. Bappy et al. [40] leveraged resampling features, LSTM cells, and an encoder–decoder network to detect image forgery. Ullah et al. [41] created a motion recognition method using a convolutional neural network and a deep bidirectional LSTM network to process video data.
Gated recurrent units (GRUs) are a common type of gated RNN. Fanta et al. [42] deemed a novel version of a GRU, called a single-tunneled GRU for abnormality detection. Ravanelli et al. [43] revised GRUs and crafted a simplified architecture for automatic speech recognition. Zhao et al. [44] proposed local feature-based GRU networks for machine health monitoring. The attention model is an appealing prospect in the field of deep learning and has been used to process sequence-related data; this model simulates the attention mechanism of the human brain. Zheng et al. [45] created a model, correlating time-series-oriented LSTM with an attention mechanism, to study multiple tourist attractions on an hourly scale. Mou et al. [46] presented a framework for driver stress detection through multimodal fusion using attention-based deep learning techniques. Relatedly, Y. Li et al. [47] put forth an evolutionary attention-based LSTM training method with a competitive random search. By transferring shared parameters, an evolutionary attention learning approach was introduced to LSTM. Song et al. [48] suggested an attention-based LSTM network for facilitating action analysis based on skeleton data.
Although RNNs are adept at dealing with temporal information and are particularly capable of approximating arbitrary nonlinear dynamic systems with precision in theory, practitioners have always struggled to train RNNs using gradient-based approaches due to high computational costs, slow convergence, and vanishing gradients. Reservoir computing is grounded in the idea of using a large randomly and sparsely connected recurrent layer called a reservoir. This method presents an efficient alternative to gradient-based learning algorithms for designing and training RNNs in most cases [49]. The echo state network (ESN) and its variants have stimulated researchers’ interest in recent decades owing to simple model training. Like the ELM model, only the weights feeding into output neurons need to be updated in ESNs. The weights can be calculated through a simple linear regression problem, while all other weights are generated randomly and remain unchanged. ESNs have been used in various tasks. For instance, Lv et al. [50] proposed a hybrid model integrating a stacked autoencoder with echo state regression to forecast tourist flows based on search query data. Hu et al. [51] constructed a hybrid model named VMD-DE-ESN by incorporating variational mode decomposition (VMD) and differential evolution (DE) into an ESN for wind speed projections. Ma et al. [52] put forward a novel ESN approach, functional ESN, for time series classification.
It should be noticed that, while most of the literature has revealed the merits of ESNs, the efforts are still unable to clearly recognize the adaptable implications for the models. Only training hidden-output layer weights can adversely affect the training results and compromise ESNs’ performance in many use cases [53]. The network’s parameters and topology therefore are required to be adjusted to create a sound reservoir for a given application [54]. Basterrech et al. [55] have suggested the use of particle swarm optimization (PSO) to identify initial hidden–hidden weights in the ESN model. The authors modified a subset of reservoir weights and kept the rest of the weights fixed during training. Naima et al. [53] also proposed a pre-trained ESN model. They posited that, for a given ESN structure, some of the fixed non-zero weights from input-to-hidden, hidden-to-hidden, and output-to-hidden connections could be randomly selected and optimized. Then, the obtained weights would be reinjected into the network and proceed to conduct normal learning once the optimization process was complete.
The study [53] draws our attention to two key and pertinent questions. First, is it possible to randomly select some weights for pre-training while maintaining performance consistency? In such a case, different subsets of weights for pre-training might result in distinct model performance. It is therefore important to determine ways to obtain reliable, stable performance for a given task based on a pre-trained ESN. Second, would the ESN model’s performance be improved by only taking non-zero elements as pre-training candidates? A neural network’s structure is primarily determined by the connections among network nodes, and the strength of these connections is reflected in the weight between two related nodes. Thus, updating the non-zero weights merely modifies the connection strength instead of changing the actual neural network. Network performance would be severely hampered if an initialized ESN with an unreasonable structure were generated for a specific training task. Although pre-optimization weighting could improve deficient network performance to some extent, the network structure’s innate inappropriateness would persist, as the structure of the network would not demonstrably change.
According to the questions, this paper introduces two strategies, namely ensemble training and global random selection, to improve model performance and address the above lines of inquiry. (1) The ensemble training is expected to increase performance robustness. As discussed, the fixed weights to be optimized will be randomly selected (i.e., it is unknown which weights are more important, and there is no guarantee that predominant weights will be chosen). The ensemble operates well in most situations, especially when small changes in the network structure might generate large changes in forecasting performance. (2) The global random selection strategy, which expands the selection range of pre-training weights from the non-zero range to all randomly initialized fixed weights (including zero candidates), might improve the adaptability of the network architecture. Introducing both strategies is hoped to result in a reasonable network structure with optimized weights for a specific task.
In accordance with the two introduced strategies and the basic model proposed in [53], a new model named EN-PSO-ESN (ensemble pre-trained ESN model by PSO) is proposed. The contributions of this paper are as follows. First, the proposed model adopts a pre-training strategy and improves the network’s adaptability when facing different tasks with unique data characteristics. That is, the network has suitable presuppositions when training specific data rather than treating all circumstances generically. Second, partially fixed weight optimization reduces calculation costs while keeping the ESN’s basic properties. In addition, considering comparability, we apply the PSO algorithm for pre-training that has appeared in [53]. Third, introducing an ensemble strategy could enhance the model’s generalizability. Fourth, we expand the selection range of pre-training weights from the non-zero range to all randomly initialized weights. Some zero weights may change to non-zero weights during pre-training, while some non-zero weights might be updated. The neural network structure will be modified as a result. We can therefore transform a network with an unreasonable structure into one with a reasonable structure. In short, all randomly initialized weights that serve as pre-training candidates will enrich the basic model’s diversity for the ensemble.
Moreover, we present a parallel study on the ELM model for a similar training process with the ESN model. The highlight of the ELM model is that input weights and biases are randomly generated, and only hidden-layer parameters should be tuned. Therefore, borrowing the premise from [53], we build a PSO-ELM model and introduce the above-discussed strategies to establish the EN-PSO-ELM model. A comparative study between PSO-ELM and EN-PSO-ELM and between PSO-ESN and EN-PSO-ESN is subsequently performed.
The remainder of this paper is organized as follows: Section 2 provides an overview of ESN, ELM, and PSO. Section 3 details the proposed EN-PSO-ESN model with ensemble training and global random selection. In Section 4, our approach is applied to eight datasets. Several simulation results confirm the efficiency of this method. Conclusions and directions for future work are outlined in Section 5.

2. Methodology

2.1. Basic Theory of Echo State Network

An ESN is a promising type of RNN that can be used to model and predict the temporal behavior of nonlinear dynamic systems [56]. The basic ESN includes an input layer, reservoir layer, and output layer. Assume that the ESN has N reservoir units, K input units, and H neurons in the output layer. At time step t , the input state matrix is given as u t = ( u 1 t , u 2 t , , u K t ) T R K , the reservoir state matrix is x t = ( x t , x 2 t , , x N t ) T R N , and the output state matrix is y t = ( y 1 t , y 2 t , , y H t ) T R H , where t = 1,2 , , T and t represents the total training time step.
The reservoir states are updated as
x t + 1 = f ( W i n u t + 1 + W x t + W f b · y ( t ) )
where f ( · ) is the internal activation function of the reservoir; the Sigmoid function or tanh function is usually selected, which makes the ESN have good nonlinear characteristics. W i n is the N × K input weight matrix; W is the N × N reservoir weight matrix, which is a randomly initialized sparse matrix; W f b is the N × H output feedback matrix. For tasks where no output feedback is required, W f b is nulled. Matrices W i n , W , and W f b are randomly generated obeying a uniform distribution and remain unchanged during the training process.
The output layer is defined as
y t + 1 = g ( W o u t ·   u t + 1 ;   x t + 1 )
where g ( · ) is the activation fucntion of output units; [ ; ] stands for a vertical vector (or matrix) concatenation; W o u t is an H × ( K + N ) matrix of output weights that need to be trained in ESN learning.
If the activation function g ( · ) is an identity function, the outputs from an ESN are typically linear and feedforward, Equation (2) can be written in a matrix form:
Y = W o u t · X
where Y R H × T and X R ( K + N ) × T is the matrix form of u t ;   x t . Then, W o u t can be represented as follows:
W o u t = Y X
where X is the pseudo-inverse of X .

2.2. Basic Theory of Extreme Learning Machine

The ELM, proposed in [57], is a single-hidden-layer FFNN that has been broadly applied given its swift learning speed and generalization ability. A highlight of the ELM model is that input weights and biases are randomly generated, and only hidden-layer parameters must be tuned (Figure 1b). ELM learning theory maintains that these random hidden neurons can often be generated independently of the training data and application environment. The ELM therefore offers many advantages such as rapid learning and minimal manual intervention.
For T arbitrary samples u t , y t , t = 1,2 , , T , u ( t ) = ( u 1 t , u 2 t , , u K t ) T R K , y ( t ) = ( y 1 t , y 2 t , , y H t ) T R H . A standard single-hidden-layer FFNN with L hidden nodes and activation function f ( · ) is modeled as
H t = f W i n · u t + β , t = 1,2 , , T
where H t is the output of the hidden layer with L × 1 dimensions corresponding to the input of the t-th sample; W i n is a L × K input weight matrix; β is a L × 1 bias vector; W i n · u ( t ) denotes the inner product of W i n and u ( t ) .
For the total T samples, Equation (5) can be written in a matrix form:
H = f W i n · u + β
where H is the output of the hidden layer with L × T dimensions; u is the input matrix of all samples. Then, the output of the ELM network Y can be expressed as
Y = W o u t · H
where Y is an H × T matrix and W o u t is an H × L output weight matrix that connects the hidden layer and the output layer; W o u t can be represented as
W o u t = Y H
where H is the pseudo-inverse of H .
Through the above description, a brief comparison can be made between ESNs and ELMs. ESNs are based on reservoir computing, which involves a randomly connected network with fixed weights that captures the temporal dynamics of the input data. Conversely, ELMs belong to feedforward neural networks with random hidden layer weights and a linear output layer. Despite differences in network structure, both ESN and ELM fix other parameters and only train the weights of the output layer during the training process. This makes both ESN and ELM networks capable of handling large-scale datasets and well suited for real-time applications with high efficiency. Therefore, in our papers, we have conducted corresponding research on both networks.

2.3. Basic Particle Swarm Optimization

Kennedy and Eberhart proposed PSO in 1995; it is an exemplar of swarm intelligence technology, inspired by birds’ foraging behavior [57]. The seminal PSO paper described a population-based optimization algorithm for searching for an optimum in a search space: a group of particles is randomly initiated in hyperspace, and then each particle moves in the hyperspace at a dynamically adjusted velocity that is determined by its own best experience as well as the entire population’s best experience. Mathematically, the PSO algorithm can be briefly described as follows:
To solve an optimization problem in a d-dimensional space, N particles are considered as initial solutions. The location of the ith particle in a d-dimensional space can be represented as X i = x i 1 , x i 1 , , x i d , and V i = ( v i 1 , v i 1 , , v i d ) represents the displacement of each iteration. Particles update themselves using two extremes: one is the optimal solution P i = p i 1 , p i d , , p i d   , calculated by itself, and the other is the optimal solution P g , calculated by the global group. The iteration formula is as follows:
v i d = ω × v i d + c 1 r 1 × p i d x i d + c 2 r 2 ( p g x i d )
x i d = x i d + v i d
where ω denotes the inertia weight, c 1 and c 2 denote acceleration factors, r 1 and r 2 are random numbers evenly distributed in the range [ 0 ,   1 ] , ω × v i d represents the effect of the previous speed on the current speed, c 1 r 1 × p i d x i d allows particles to adjust according to their optimal position, and c 2 r 2 ( p g x i d ) allows particles to adjust through the optimum of the population. The flowchart of the basic PSO algorithm is depicted in Figure 2.
We chose PSO for selected weight pre-training considering two considerations. First, PSO is an important means of addressing optimization; the algorithm has been applied in a range of fields since its conception, and a growing number of practical problems have been solved using PSO or its variants. The algorithm therefore possesses a strong ability to obtain the global optimum. Second, the model proposed in this study was inspired by [53], which adopted the PSO approach for pre-training. To make a fair comparison with the model put forth in [53] and to verify these models’ performance, using the same optimization technique is imperative. In addition, the initial values of the parameters c 1 ,   c 2 , and ω are also initialized as 1.193, 1.193, and 0.721, respectively, in this study, and then at each iteration, c 1 and c 2 take their values from a uniform distribution between 0 and their initial values as well.

3. Proposed EN-PSO-ESN Model

3.1. Motivation

As mentioned, [53] proposed a pre-training strategy to optimize some non-zero weights randomly collected in matrices W i n , W , and W f b before ESN learning to provide weights suited to the target task. Two concerns apply in this case: (1) how to maintain the robustness of the model’s performance and (2) which selection strategy can most effectively boost the model’s performance—only choosing non-zero weights as candidates, or considering both non-zero and zero weights?
Regarding the first issue, selecting some weights randomly for pre-training is a matter of trial and error to enhance network performance. The model may exhibit satisfactory performance with a randomly chosen set of weights; however, there is no guarantee that the model will suit the target application in another random selection. In essence, for any given task, it would be impossible to forecast which connection weights will significantly affect network performance. Similarly, it is not possible to project which weights have more potential for optimization. Ensuring the pre-trained ESN’s reliable and stable performance by randomly choosing fixed weights is correspondingly important.
For the second question, the connections among network nodes determine the neural network’s structure. The strength of connections is reflected by the weight between two related nodes. Thus, simply updating non-zero weights does not alter the network structure and only affects connection strength. Choosing non-zero elements for pre-training means that the network’s optimization is based on a fixed and pre-determined structure. The initialization creates unsuitable input-to-hidden and hidden-to-hidden links, and the basic network contains innate structural deficiencies. Weight pre-optimization can improve the network’s performance; however, the inherent inappropriateness of its structure persists.
Based on [53] and the above two issues, this paper introduces two strategies (i.e., ensemble training and global random selection) to the PSO-ESN model. A new model, EN-PSO-ESN, is described in this section. An ensemble training strategy is applied to increase the robustness of network performance. The global random selection strategy expands the range of possible pre-training weights from the non-zero range to all randomly initialized fixed weights. Figure 3 illustrates a flowchart of the overall process for the proposed model; details are provided in Section 3.2.

3.2. Main Steps of EN-PSO-ESN Model

As Figure 3 indicates, the proposed model comprises three main stages: (1) model training and selection (blue block); (2) ensemble prediction (green block); and (3) performance evaluation (orange block). The first two stages are the core tenets of the proposed approach. The full process can be divided into several steps.
Step 1. Construct n independent basic ESN models. n independent basic ESN models are initialized. Each model has the same amount of input, reservoir units, and output as well as the same connection rate for reservoir units. Then, the input-to-hidden and hidden-to-hidden link weights are randomly initialized separately.
Step 2. Select random weights. For each of the n independent initialized ESN models, the weights needing to be pre-optimized are randomly chosen. These weights include some from the input-to-hidden, hidden-to-hidden, and output-to-hidden connections and cover all elements belonging to the three weight matrices W i n , W , and W f b , including zero and non-zero elements. Suppose parameters α ,   β , and γ are the chosen rates for the three weight matrices. Then, the total number of pre-training weights can be represented as α ϕ ( W i n ) + β ϕ ( W ) + γ ϕ ( W f b ) , where ϕ ( ) denotes the number of elements in a matrix, and the parameters α ,   β , and γ belong to the interval 0,1 . It is advisable to assign these parameters small values to choose an array of weights to be pre-trained; just a few selected weights can preserve the model’s specificity while controlling the computational cost for weight optimization.
Step 3. Optimize selected weights. Because of the advantages of PSO and to offer a fair comparison with [53], PSO is chosen as the optimization technique. The training set is divided into two parts, namely Set 1 and Set 2. Set 1 is obtained by sampling the original training set: for the given sampling ratio s % , s % training samples are sampled to establish Set 1. Set 2 contains the remaining data from Set 1. If the size of the training set is insufficient, then sampling can be performed with the replacement method to ensure that Set 1 includes an adequate number of samples.
Consider a basic ESN model as an example of optimizing randomly chosen weights. Once the weights that need to be pre-optimized have been randomly selected, the total number of pre-training weights and their locations in the corresponding matrix are determined. PSO can then be used to optimize the selected weights; the optimization diagram is shown in Figure 4. Here, each particle represents a candidate solution vector (as pictured in Figure 5b), and the particle length represents the number of weights to be optimized. For a given initialized particle, the corresponding solution vector is injected into the three weight matrices W i n , W , and W f b (in Figure 5c). The basic ESN model is next trained based on Set 1 (Figure 5d). After training the basic ESN model, the training error (i.e., the root mean square error (RMSE); see Equation (11)) of the model is set as the fitness value to measure the current particle’s performance. If the fitness value does not meet the stop criterion or if the iteration epoch has been reached, then the particle is updated based on Equations (9) and (10). The updated particle is next reinjected into its corresponding position in the three weight matrices. Then, based on Set 1, the basic ESN model is re-trained, and the new fitness value is obtained. These steps are repeated until the target or maximum iteration is reached. The global best solution can be realized based on the best solution vector of each particle. Namely, the weight matrices are learned, and the basic ESN model is pre-trained. The n trained ESN model can be derived via the parallel optimization process in Figure 4.
The training framework for the EN-PSO-ELM model is the same as for the EN-PSO-ESN model. The canonical ELM model is a single-layer FFNN; as such, only two weight matrices ( W i n and W o u t ) are involved, where W i n is the input-to-hidden connection and W o u t is the hidden-to-output connection. Some weights in the sparse matrix W i n will be chosen for pre-training.
Step 4. Evaluate and select a model. For the pre-trained n ESN models, Set 2 is taken as validation data and fed into these models to evaluate their performance (i.e., the RMSE). The n models are then sorted by RMSE in ascending order. The top t models are finally chosen in line with practical application problems and used to constitute an ensemble model.
Step 5. Re-train the top t models. To ensure that the pre-trained models can accurately capture data patterns, all training datasets, including Sets 1 and 2, are applied to re-train the selected models independently.
Step 6. Conduct independent and integrated prediction. Next, corresponding prediction results are respectively derived from t trained models. Different integration strategies are used to obtain the final integrated prediction results, such as the optimistic choice for the largest prediction result, the negative choice for the smallest result, and the neutral choice for a weighted average result.
Step 7. Assess performance. To evaluate the prediction accuracy of the proposed approach, different performance evaluation criteria (e.g., root mean square error (RMSE), relative mean absolute error (RMAE), mean absolute percentage error (MAPE), root mean square percentage error (RMSPE), and directional accuracy (DA); see Equations (11)–(15)) can be used to measure prediction errors in the test data.

4. Experiments and Results

In this section, the forecasting ability of the proposed model is tested against that of several benchmark models. The data description, performance evaluation criteria, and benchmarks are presented in Section 4.1. Experiments for six synthetic benchmark time series predictions are summarized in Section 4.2, with experiments for the two real benchmark time series predictions appearing in Section 4.3. Experimental results are analyzed respectively. An additional experiment performed to further illustrate the proposed model’s effectiveness is presented in Section 4.4.

4.1. Datasets, Performance Evaluation Criteria, and Benchmarks

Six artificial datasets and two real-life datasets were chosen in this study to demonstrate improvements attributable to the proposed model. The six artificial datasets include two Mackey and Glass time series with different parameters, two nonlinear auto-regressive moving averages with different parameters, and the Henon attractor and Lorenz attractor data series (see Equations (16)–(19)). These datasets have been drawn from [53] to facilitate a fair comparison and a reliable demonstration (i.e., that our proposed model leads to improvements). Furthermore, two real-life datasets—the air quality (AQ) dataset and airfoil self-noise (ASN) dataset—are applied; both were acquired from the Machine Learning Repository at the University of California, Irvine (UCI) [58].
To evaluate models’ forecasting performance in terms of directional prediction and level prediction, five indicators (i.e., RMSE, RMAE, MAPE, RMSPE, and DA) that have been frequently used in recent years were selected. y ( t ) and y ^ ( t ) stand for the actual value and predicted value, respectively; a t = 1 , if ( y t + 1 y ( t ) ) y ^ t + 1 y ^ ( t ) 0 , or a t = 0 otherwise. N is the size of predictions. For the first four indicators, smaller index values represent higher forecasting accuracy; larger index values indicate higher forecasting accuracy for DA.
R M S E = 1 N t = 1 N y t y ^ ( t ) 2
R M A E = t = 1 N y t y ^ ( t ) t = 1 N y ( t )
M A P E = 1 N t = 1 N y t y ^ ( t ) y ( t )
R M S P E = 1 N t = 1 N y t y ^ t y t 2
D A = 1 N t = 1 N a ( t ) × 100 %
Two groups of models (Table 1) were compared to test the effectiveness of the proposed strategies.
Regarding the first group—ESN-based models—the canonical ESN model serves as the foundation for the PSO-ESN model proposed in [53] and the model proposed in this study. The basic ESN model was thus taken as a benchmark. Additionally, because the proposed model was derived from the PSO-ESN model, the PSO-ESN model should be selected.
For the second group, as noted, the ELM model shares similarities in model construction and training with the canonical ESN. Applying the two proposed strategies to the ELM-based model can further verify their effectiveness. The ELM-PSO model, in line with [53], was constructed accordingly. The corresponding EN-PSO-ELM model was established by applying ensemble training and global random selection to the PSO-ELM model.

4.2. Experiment for Synthetic Benchmark Time Series Prediction

(1) 
Mackey and Glass dataset
Mackey and Glass (MG) time series possess chaotic nonlinear behavior, rendering it difficult to identify underlying patterns from the training samples. The dynamic predictive system and corresponding parameters are described in Equation (16):
x ( t ) ˙ = a x ( t τ ) 1 + x c ( t ) b x ( t )
where a = 0.2 ,   b = 0.1 ,   c = 10 . According to the dynamic equation and the different values of τ   ( τ = 17   a n d   τ = 30 ), two series, MG(1) and MG(2), respectively, were generated. Each of the series contains 1000 samples, with 800 samples used as a training dataset and 200 samples used for testing.
The above-mentioned six models were applied to MG(1) and MG(2). Some hyperparameters should be defined before model training (e.g., hidden neuron number, selection rate of the weight matrix, ensemble number). Detailed parameter information is listed in Table 2. The running times, set to 50, indicate that every model is repeated 50 times to derive its average performance. The hidden neuron number indicates the number of neurons in the hidden layer of the ELM model and the reservoir size for the ESN model. The parameters α and β denote the selection ratios of the weight elements for pre-optimization in the weight matrices W i n and W ; the feedback weight is not considered in this study. For ELM and ELM-based models, only some of the weight elements in the weight matrix W i n are pre-trained. The parameter β is therefore not involved. The ensemble number is only for the ensemble model. The leak rate reflects the memory capacity of ESN-based models, and the collection rate of the reservoir indicates the sparsity of the ESN model.
The prediction results for the six models are displayed in Table 3. In general, ensemble training and global random selection substantially influenced both the ELM model and the ESN model. Taking the ELM-based model as an example, the proposed EN-PSO-ELM model was superior to the PSO-ELM model (see Figure 6a). The canonical ELM model showed a strong prediction effect for the two datasets but performed worse than the two improved models. The same conclusion applied to the ESN-based models (see Figure 6b).
For the MG(2) dataset, the EN-PSO-ELM model’s performance did not surpass all ELM-based models; it was slightly inferior to the PSO-ELM model. The accuracy of the directional prediction was identical to that of the best model, reaching 100%. Although the other four measurement indices slightly exceeded the best ones, all four indices maintained the same prediction accuracy level as the best-performing model (see Figure 7a,b).
(2) 
Nonlinear auto-regressive moving average dataset
The nonlinear auto-regressive moving average (NARMA) is a popular and hard predictive time series characterized by highly chaotic behavior. The dynamics of this dataset are summarized in Equation (17):
y t + 1 = c 1 y t + c 2 y t + i = 1 k y ( t i ) + c 3 x t k 1 x t + c 4
For the given parameters c 1 = 0.3 , c 2 = 0.05 , c 3 = 1.5 , c 4 = 0.1 , k = 6, and k = 10, two series, NARMA(1) and NARMA(2), were generated. Each of the data series contained 1000 samples: 800 for training and 200 for testing.
Using the same hyperparameters as in Table 2, we derived the prediction results for NARMA(1) and NARMA(2) with different lag orders. The performance evaluations are displayed in Table 4.
Table 4 reveals a similar conclusion to that revealed in Table 3; that is, the proposed EN-PSO model was superior to the PSO-based model and the canonical model, including the ELM-based and ESN-based models. Yet the PSO-ELM model’s performance also warrants a note: this model was inferior to the EN-PSO-ELM model and performed worse than the canonical ELM model for both the NARMA(1) and NARMA(2) datasets (see Figure 8 and Figure 9).
(3) 
The Henon attractor and Lorenz attractor data series
The Henon attractor and Lorenz attractor data series are dynamic systems frequently mentioned in relevant literature. The Henon map is a discrete-time dynamic attractor; Lorenz is a set of differential equations describing a fluid motion between a hot and a cool surface. These datasets can be generated by the corresponding Equations (18) and (19). As before, each series contained 1000 samples (800 for training and 200 for testing).
x t + 1 = y t a x 2 t + 1 , y t + 1 = b x t , a = 1.4 ,     b = 0.3 .
x ˙ = a ( y x ) y ˙ = y x z + r x z ˙ = x y b z a = 10 ,   b = 28 , c = 8 3
Experiments were conducted on both series. The associated performance measurement indices for the six models are listed in Table 5, mirroring the findings of prior experiments: the proposed EN-PSO-based model consistently performed best, followed by the PSO-based model (see Figure 10 and Figure 11).

4.3. Experiment for Real Benchmark Time Series Prediction

To further demonstrate the proposed model’s forecasting performance, two real-life datasets (an AQ dataset and an ASN dataset) were obtained from the UCI Machine Learning Repository. Several measurement indices, including RMSE, RMAE, MAPE, RMSPE, and DA, were used to compare the models’ forecasting accuracy. To eliminate fluctuations in forecasting performance, every model was run 50 times for both data series; the mean values of these 50 indices were calculated. Table 6 and Table 7 indicate the performance of different benchmarks.

4.3.1. Air Quality (AQ) Dataset

The AQ dataset contained 9358 instances of hourly averaged responses from an array of five metal oxide chemical sensors in the field in a significantly polluted area of an Italian city, collected from March 2004 to February 2005. The dataset included 13 attributes (e.g., temperature, true hourly averaged NO2 concentration in microg/m3) in addition to date and time. The true hourly averaged CO concentration in mg/m3 was taken as the target for prediction in this work. After a preprocessing procedure was performed to exclude missing values, 90% of the data were used for training with 10% reserved for testing. The results are listed in Table 6 and are illustrated in Figure 12.

4.3.2. Airfoil Self-Noise (ASN) Dataset

The ASN dataset included five features: (a) frequency in hertz; (b) angle of attack in degrees; (c) chord length in meters; (d) free-stream velocity in meters per second; and (e) suction side displacement thickness in meters. Scaled sound pressure level was the sole output variable. The testing results are listed in Table 7 and are visually depicted in Figure 13.
Based on the eight experiments presented above, the following conclusions can be drawn: First, like findings for the six synthetic benchmark time series, the prediction results clearly showed that the proposed EN-PSO-based models’ accuracy surpassed that of POS-based and canonical models with smaller standard deviations. Taking the AQ dataset for instance, the ensemble training and global random selection strategies effectively improved forecasting accuracy whether using the ESN-based or ELM-based model. More precisely, for the EN-PSO-ELM model, the first four measurement indices respectively declined by 15.3464%, 15.2731%, 6.8253%, and 7.2051%; the accuracy of direction prediction increased by 2.4060% compared with the canonical ELM model. The first four prediction error indices decreased by 1.0844%, 4.7800%, 2.7451%, and 3.9515% compared with the PSO-ELM model and rose by 3.6654% for direction prediction, respectively. The conclusions remained consistent for the EN-PSO-ESN model, such that this model’s prediction accuracy also rose considerably and maintained a minimum standard deviation. In detail, compared with the canonical ESN model, the error indices (i.e., RMSE, RMAE, MAPE, and RMSPE) declined by 13.0744%, 18.2907%, 8.7771%, and 31.5504%, respectively. The EN-PSO-ESN model also performed better than the PSO-ESN model. Furthermore, overall, both the PSO-ELM model and the PSO-ESN model appeared superior to the canonical ELM and ESN models.
Second, although the PSO-ELM model and PSO-ESN model each appeared superior to the canonical ELM and ESN models in most situations, the PSO-ELM model did not always outperform canonical models. For example, in terms of the ASN dataset, the PSO-ELM model performed worse than the canonical ELM model. The first four measurement indices exceeded those of the canonical ELM model by 1.9824%, 1.5625%, 1.5773%, and 2.1951%, respectively; the accuracy of direction prediction was 1.8012% lower than that for the ELM model (see Table 7). Similar findings emerged for prior experiments: as shown in Table 4, the PSO-ELM model’s performance was inferior to that of the canonical ELM for the NARMA(1) and NARMA(2) datasets. We carried out another experiment on the ASN dataset (see the next subsection) to explore why this phenomenon might have occurred.

4.4. Additional Experiment

This section presents an additional experiment on the ASN dataset, focusing on a comparison between the canonical ELM model and the PSO-ELM model. This experiment was intended to uncover (a) why the PSO-ELM model did not necessarily maintain superiority over the canonical ELM model and (b) why the EN-PSO-ELM model maintained sound performance with the above eight datasets. Four models (see Table 8) were considered.
To evaluate the average performance of the above four models, we first generated 100 basic canonical ELM models for training. Put simply, according to the datasets’ input and output, 100 weight matrices W i n were randomly generated. One hundred prediction results and corresponding measurement indices were obtained for the ASN dataset based on the basic ELM models.
Next, the pre-training strategy was applied to 100 initialized ELM models, and 100 corresponding PSO-ELM(I), PSO-ELM(II), and EN-PSO-ELM models were derived. For the PSO-ELM(I) model, some of the non-zero weights (equal to 10% of all weight elements) were randomly selected for pre-training. The same proportion was randomly selected for the PSO-ELM(II) and EN-PSO-ELM models but covered both non-zero and zero elements. The pre-optimization and network training were thus identical to the previous experiments.
To discern the performance of the compared models, DA is taken as a sample metric here to evaluate the prediction results. All models were derived from the 100 initialized basic model as indicated; doing so enabled us to observe the DA distribution and identify performance discrepancies. Figure 14 displays the DA distribution of the canonical ELM models, PSO-based ELM models (I), PSO-based ELM models (II), and ensemble PSO-based ELM models. Figure 14a, showing the DA distribution of the initialized 100 canonical ELM models, reveals that DA was distributed over a large interval (from 55% to 69%). Figure 14b,c contain the 100 DA distributions of the PSO-ELM(I) models and PSO-ELM(II) models, respectively. Both figures are similar in that the distribution is more concentrated than that in the canonical ELM models. This phenomenon suggests that the pre-trained algorithm can improve the stability of prediction performance to a certain extent. DA was also concentrated within a relatively small range in both instances—from 54% to 65% and from 56% to 64%, respectively. However, the largest DA of the two models was less than 65%. On the contrary, the largest DA of the canonical ELM models reached 69%. In sum, for the ASN dataset, the pre-training strategy seemed useful for boosting the model’s stability but had a less pronounced impact on prediction accuracy. This relatively lower improvement may explain why the PSO-ELM model performed slightly worse than the canonical ELM models. Meanwhile, for the ASN dataset, although we expanded the pre-training candidates from non-zero elements to all elements, prediction accuracy demonstrated no significant increase. Lastly, compared with Figure 14b,c, the DA distribution in Figure 14d clearly shifted to the right. The prediction accuracy therefore exhibited a significant improvement. The AD distribution in Figure 14d also maintained a relative concentration; that is, because the EN-PSO-ELM model applied the pre-training and ensemble strategy, both the prediction accuracy and robustness were elevated.

5. Conclusions

In this work, we proposed two strategies—global random selection and ensemble training—for ESN and ELM models in order to pre-train the initialized weight parameters and the network topology to improve the models’ adaptability. Global random selection expands the range of chosen pre-training weights from non-zero weights to all randomly initialized weights. This characteristic enriches the basic models’ diversity. We also adopted ensemble training to enhance the pre-trained models’ reliability and stability. To ensure a fair comparison between the proposed model and the one from [53], PSO was applied as an optimization technique to realize weight pre-training. The proposed approach was then tested for time series prediction based on six benchmarks and two real-life datasets. The experimental results highlighted notable enhancements in ESN and ELM learning results.
The experiments further show that, although the PSO-ELM model and the PSO-ESN model seemed superior to the canonical ELM and ESN models in most cases, the PSO-ELM model is not consistently superior to the canonical models. For example, the PSO-ELM model performs worse than the canonical ELM model on the ASN dataset and NARMA dataset. To investigate this outcome in greater depth, we performed a subsequent experiment with the ELM model using the ASN dataset. The four models’ DA distributions (i.e., ELM, PSO-ELM[I], PSO-ELM[II], and EN-PSO-ELM) reveal several interesting patterns. For example, with the ASN dataset, the pre-trained algorithm that only includes non-zero elements for pre-training may lead to more stable prediction performance but not better prediction accuracy. Global selection could further enhance the model’s robustness but had little effect on prediction accuracy. Yet introducing ensemble training significantly increases prediction accuracy, underscoring that these two strategies can jointly promote models’ prediction accuracy and robustness.
In conclusion, it is essential to acknowledge the limitations of this study, specifically the absence of inter-model comparisons. Our study originally intended to conduct two types of experiments: intra-model comparisons (ELM vs. EN-PSO-ELM and ESN vs. EN-PSO-ESN) and inter-model comparisons (EN-PSO-ELM vs. EN-PSO-ESN). Given that the ESN is a recurrent neural network and the ELM is a feedforward neural network, we hypothesized that EN-PSO-ESN would likely outperform EN-PSO-ELM. However, the experimental results based on the applied datasets did not support this assumption. Instead, EN-PSO-ELM performed better in some tasks, while EN-PSO-ESN excelled in others. We speculate that this outcome may be linked to two factors: (1) the initialization process of the ESN, where hyperparameters like the leaking rate, input scaling, spectral radius, and sparse degree could impact performance, and (2) the intrinsic characteristics of the dataset itself. Future work will focus on determining under which circumstances each model excels. Despite these limitations, the two proposed strategies in this study—global random selection and ensemble training—offer valuable insights into model pre-training, enhancing models’ adaptability across diverse task situations.

Author Contributions

Conceptualization, J.W.; Methodology, L.T.; Software, J.W. and M.W.; Validation, M.W. and C.Z.; Formal analysis, C.Z.; Writing—original draft, L.T. and J.W.; Writing—review & editing, J.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China (No. 71501138, 71971151).

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Panapakidis, I.P.; Dagoumas, A.S. Day-ahead natural gas demand forecasting based on the combination of wavelet transform and ANFIS/genetic algorithm/neural network model. Energy 2017, 118, 231–245. [Google Scholar] [CrossRef]
  2. Hu, H.; Wang, L.; Peng, L.; Zeng, Y.R. Effective energy consumption forecasting using enhanced bagged echo state network. Energy 2020, 193, 116778. [Google Scholar] [CrossRef]
  3. Tran, D.H.; Luong, D.L.; Chou, J.S. Nature-inspired metaheuristic ensemble model for forecasting energy consumption in residential buildings. Energy 2020, 191, 116552. [Google Scholar] [CrossRef]
  4. Sujjaviriyasup, T. A new class of MODWT-SVM-DE hybrid model emphasizing on simplification structure in data pre-processing: A case study of annual electricity consumptions. Appl. Soft Comput. 2017, 54, 150–163. [Google Scholar] [CrossRef]
  5. Chen, Y.; Yang, Y.; Liu, C.; Li, C.; Li, L. A hybrid application algorithm based on the support vector machine and artificial intelligence: An example of electric load forecasting. Appl. Math. Model. 2015, 39, 2617–2632. [Google Scholar] [CrossRef]
  6. Fan, C.; Ding, C.; Zheng, J.; Xiao, L.; Ai, Z. Empirical Mode Decomposition based Multi-objective Deep Belief Network for short-term power load forecasting. Neurocomputing 2020, 388, 110–123. [Google Scholar] [CrossRef]
  7. Wang, J.; Yang, Z. Ultra-short-term wind speed forecasting using an optimized artificial intelligence algorithm. Renew. Energy 2021, 171, 1418–1435. [Google Scholar] [CrossRef]
  8. Peng, T.; Zhang, C.; Zhou, J.; Nazir, M.S. Negative correlation learning-based RELM ensemble model integrated with OVMD for multi-step ahead wind speed forecasting. Renew. Energy 2020, 156, 804–819. [Google Scholar] [CrossRef]
  9. Zhang, L.; Wang, J.; Niu, X.; Liu, Z. Ensemble wind speed forecasting with multi-objective Archimedes optimization algorithm and sub-model selection. Appl. Energy. 2021, 301, 117449. [Google Scholar] [CrossRef]
  10. Mishra, D.; Goyal, P.; Upadhyay, A. Artificial intelligence based approach to forecast PM 2.5 during haze episodes: A case study of Delhi, India. Atmos. Environ. 2015, 102, 239–248. [Google Scholar] [CrossRef]
  11. Li, V.O.K.; Lam, J.C.K.; Han, Y.; Chow, K. A Big Data and Artificial Intelligence Framework for Smart and Personalized Air Pollution Monitoring and Health Management in Hong Kong. Environ. Sci. Policy 2021, 124, 441–450. [Google Scholar] [CrossRef]
  12. Shams, S.R.; Jahani, A.; Kalantary, S.; Moeinaddini, M.; Khorasani, N. Artificial intelligence accuracy assessment in NO2 concentration forecasting of metropolises air. Sci. Rep. 2021, 11, 1805. [Google Scholar] [CrossRef]
  13. Guo, S.; Lin, Y.; Feng, N.; Song, C.; Wan, H. Attention Based Spatial-Temporal Graph Convolutional Networks for Traffic Flow Forecasting. Proc. AAAI Conf. Artif. Intell. 2019, 33, 922–929. [Google Scholar] [CrossRef]
  14. Cai, W.; Yang, J.; Yu, Y.; Song, Y.; Zhou, T.; Qin, J. PSO-ELM: A Hybrid Learning Model for Short-Term Traffic Flow Forecasting. IEEE Access 2020, 8, 6505–6514. [Google Scholar] [CrossRef]
  15. Du, S.; Li, T.; Gong, X.; Horng, S.-J. A Hybrid Method for Traffic Flow Forecasting Using Multimodal Deep Learning. Int. J. Comput. Intell. Syst. 2020, 13, 85. [Google Scholar] [CrossRef]
  16. Lalmuanawma, S.; Hussain, J.; Chhakchhuak, L. Applications of machine learning and artificial intelligence for COVID-19 (SARS-CoV-2) pandemic: A review. Chaos Solitons Fractals 2020, 139, 110059. [Google Scholar] [CrossRef]
  17. da Silva, R.G.; Ribeiro, M.H.D.M.; Mariani, V.C.; Coelho, L.D.S. Forecasting Brazilian and American COVID-19 cases based on artificial intelligence coupled with climatic exogenous variables. Chaos Solitons Fractals 2020, 139, 110027. [Google Scholar] [CrossRef]
  18. Vaishya, R.; Javaid, M.; Khan, I.H.; Haleem, A. Artificial Intelligence (AI) applications for COVID-19 pandemic. Diabetes Metab. Syndr. Clin. Res. Rev. 2020, 14, 337–339. [Google Scholar] [CrossRef]
  19. Yu, P.; Yan, X. Stock price prediction based on deep neural networks. Neural Comput. Appl. 2020, 32, 1609–1628. [Google Scholar] [CrossRef]
  20. Ding, G.; Qin, L. Study on the prediction of stock price based on the associated network model of LSTM. Int. J. Mach. Learn. Cybern. 2020, 11, 1307–1317. [Google Scholar] [CrossRef]
  21. Yun, K.K.; Yoon, S.W.; Won, D. Prediction of stock price direction using a hybrid GA-XGBoost algorithm with a three-stage feature engineering process. Expert Syst. Appl. 2021, 186, 115716. [Google Scholar] [CrossRef]
  22. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM. 2017, 60, 84–90. [Google Scholar] [CrossRef]
  23. Juhn, Y.; Liu, H. Artificial intelligence approaches using natural language processing to advance EHR-based clinical research. J. Allergy Clin. Immunol. 2020, 145, 463–469. [Google Scholar] [CrossRef]
  24. Trappey, A.J.C.; Trappey, C.V.; Wu, J.L.; Wang, J.W.C. Intelligent compilation of patent summaries using machine learning and natural language processing techniques. Adv. Eng. Inform. 2020, 43, 101027. [Google Scholar] [CrossRef]
  25. Feng, Z.; Niu, W.; Tang, Z.; Xu, Y.; Zhang, H. Evolutionary artificial intelligence model via cooperation search algorithm and extreme learning machine for multiple scales nonstationary hydrological time series prediction. J. Hydrol. 2021, 595, 126062. [Google Scholar] [CrossRef]
  26. Hussain, D.; Hussain, T.; Khan, A.A.; Naqvi, S.A.A.; Jamil, A. A deep learning approach for hydrological time-series prediction: A case study of Gilgit river basin. Earth Sci. Inform. 2020, 13, 915–927. [Google Scholar] [CrossRef]
  27. Sahoo, B.B.; Jha, R.; Singh, A.; Kumar, D. Long short-term memory (LSTM) recurrent neural network for low-flow hydrological time series forecasting. Acta Geophys. 2019, 67, 1471–1481. [Google Scholar] [CrossRef]
  28. Huang, G.B.; Zhu, Q.-Y.; Siew, C.-K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  29. Zhao, X.; Zhang, X.; Cai, Z.; Tian, X.; Wang, X.; Huang, Y.; Chen, H.; Hu, L. Chaos enhanced grey wolf optimization wrapped ELM for diagnosis of paraquat-poisoned patients. Comput. Biol. Chem. 2019, 78, 481–490. [Google Scholar] [CrossRef] [PubMed]
  30. Muduli, D.; Dash, R.; Majhi, B. Automated breast cancer detection in digital mammograms: A moth flame optimization based ELM approach. Biomed. Signal Process. Control 2020, 59, 101912. [Google Scholar] [CrossRef]
  31. Diker, A.; Sönmez, Y.; Özyurt, F.; Avcı, E.; Avcı, D. Examination of the ECG signal classification technique DEA-ELM using deep convolutional neural network features. Multimed. Tools Appl. 2021, 80, 24777–24800. [Google Scholar] [CrossRef]
  32. Virgeniya, S.C.; Ramaraj, E. A Novel Deep Learning based Gated Recurrent Unit with Extreme Learning Machine for Electrocardiogram (ECG) Signal Recognition. Biomed. Signal Process. Control 2021, 68, 102779. [Google Scholar] [CrossRef]
  33. Lingyu, T.; Jun, W.; Chunyu, Z. Mode decomposition method integrating mode reconstruction, feature extraction, and ELM for tourist arrival forecasting. Chaos Solitons Fractals 2021, 143, 110423. [Google Scholar] [CrossRef]
  34. Sun, S.; Wei, Y.; Tsui, K.L.; Wang, S. Forecasting tourist arrivals with machine learning and internet search index. Tour. Manag. 2019, 70, 1–10. [Google Scholar] [CrossRef]
  35. Maliha, A.; Yusof, R.; Shapiai, M.I. Extreme learning machine for structured output spaces. Neural Comput. Appl. 2018, 30, 1251–1264. [Google Scholar] [CrossRef]
  36. Patil, A.; Shen, S.; Yao, E.; Basu, A. Hardware architecture for large parallel array of Random Feature Extractors applied to image recognition. Neurocomputing 2017, 261, 193–203. [Google Scholar] [CrossRef]
  37. Bi, J.W.; Liu, Y.; Li, H. Daily tourism volume forecasting for tourist attractions. Ann. Tour. Res. 2020, 83, 102923. [Google Scholar] [CrossRef]
  38. Qing, X.; Niu, Y. Hourly day-ahead solar irradiance prediction using weather forecasts by LSTM. Energy 2018, 148, 461–468. [Google Scholar] [CrossRef]
  39. Abbasimehr, H.; Shabani, M.; Yousefi, M. An optimized model using LSTM network for demand forecasting. Comput. Ind. Eng. 2020, 143, 106435. [Google Scholar] [CrossRef]
  40. Bappy, J.H.; Simons, C.; Nataraj, L.; Manjunath, B.S.; Roy-Chowdhury, A.K. Hybrid LSTM and Encoder–Decoder Architecture for Detection of Image Forgeries. IEEE Trans. Image Process. 2019, 28, 3286–3300. [Google Scholar] [CrossRef]
  41. Ullah, A.; Ahmad, J.; Muhammad, K.; Sajjad, M.; Baik, S.W. Action Recognition in Video Sequences using Deep Bi-Directional LSTM With CNN Features. IEEE Access 2018, 6, 1155–1166. [Google Scholar] [CrossRef]
  42. Fanta, H.; Shao, Z.; Ma, L. SiTGRU: Single-Tunnelled Gated Recurrent Unit for Abnormality Detection. Inf. Sci. 2020, 524, 15–32. [Google Scholar] [CrossRef]
  43. Ravanelli, M.; Brakel, P.; Omologo, M.; Bengio, Y. Light Gated Recurrent Units for Speech Recognition. IEEE Trans. Emerg. Top. Comput. Intell. 2018, 2, 92–102. [Google Scholar] [CrossRef]
  44. Zhao, R.; Wang, D.; Yan, R.; Mao, K.; Shen, F.; Wang, J. Machine Health Monitoring Using Local Feature-Based Gated Recurrent Unit Networks. IEEE Trans. Ind. Electron. 2018, 65, 1539–1548. [Google Scholar] [CrossRef]
  45. Zheng, W.; Huang, L.; Lin, Z. Multi-attraction, hourly tourism demand forecasting. Ann. Tour. Res. 2021, 90, 103271. [Google Scholar] [CrossRef]
  46. Mou, L.; Zhou, C.; Zhao, P.; Nakisa, B.; Rastgoo, M.N.; Jain, R.; Gao, W. Driver stress detection via multimodal fusion using attention-based CNN-LSTM. Expert Syst. Appl. 2021, 173, 114693. [Google Scholar] [CrossRef]
  47. Li, Y.; Zhu, Z.; Kong, D.; Han, H.; Zhao, Y. EA-LSTM: Evolutionary attention-based LSTM for time series prediction. Knowl. Based Syst. 2019, 181, 104785. [Google Scholar] [CrossRef]
  48. Song, S.; Lan, C.; Xing, J.; Zeng, W.; Liu, J. Spatio-Temporal Attention-Based LSTM Networks for 3D Action Recognition and Detection. IEEE Trans. Image Process. 2018, 27, 3459–3471. [Google Scholar] [CrossRef] [PubMed]
  49. Li, F.; Wang, X.; Li, Y. Effects of singular value spectrum on the performance of echo state network. Neurocomputing 2019, 358, 414–423. [Google Scholar] [CrossRef]
  50. Lv, S.-X.; Peng, L.; Wang, L. Stacked autoencoder with echo-state regression for tourism demand forecasting using search query data. Appl. Soft Comput. 2018, 73, 119–133. [Google Scholar] [CrossRef]
  51. Hu, H.; Wang, L.; Tao, R. Wind speed forecasting based on variational mode decomposition and improved echo state network. Renew. Energy 2021, 164, 729–751. [Google Scholar] [CrossRef]
  52. Ma, Q.; Shen, L.; Chen, W.; Wang, J.; Wei, J.; Yu, Z. Functional echo state network for time series classification. Inf. Sci. 2016, 373, 1–20. [Google Scholar] [CrossRef]
  53. Chouikhi, N.; Ammar, B.; Rokbani, N.; Alimi, A.M. PSO-based analysis of Echo State Network parameters for time series forecasting. Appl. Soft Comput. 2017, 55, 211–225. [Google Scholar] [CrossRef]
  54. Ferreira, A.A.; Ludermir, T.B. Comparing evolutionary methods for reservoir computing pre-training. In Proceedings of the 2011 International Joint Conference on Neural Networks, San Jose, CA, USA, 31 July–5 August 2011; pp. 283–290. [Google Scholar] [CrossRef]
  55. Basterrech, S.; Alba, E.; Snášel, V. An Experimental Analysis of the Echo State Network Initialization Using the Particle Swarm Optimization. arXiv 2015, arXiv:150100436. [Google Scholar]
  56. Jaeger, H.; Haas, H. Harnessing Nonlinearity: Predicting Chaotic Systems and Saving Energy in Wireless Communication. Science 2004, 304, 78–80. [Google Scholar] [CrossRef]
  57. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95—International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; Volume 4, pp. 1942–1948. [Google Scholar] [CrossRef]
  58. Dua, D.; Graff, C. UCI Machine Learning Repository; University of California, School of Information and Computer Science: Irvine, CA, USA, 2019; Available online: http://archive.ics.uci.edu/ml (accessed on 14 April 2022).
Figure 1. (a) Basic structure of an ESN model and (b) basic structure of an ELM model. Solid bold arrows represent fixed connections, and dashed arrows represent connections to be trained.
Figure 1. (a) Basic structure of an ESN model and (b) basic structure of an ELM model. Solid bold arrows represent fixed connections, and dashed arrows represent connections to be trained.
Entropy 26 00215 g001
Figure 2. Flowchart of basic PSO algorithm.
Figure 2. Flowchart of basic PSO algorithm.
Entropy 26 00215 g002
Figure 3. Flowchart of EN-PSO-ESN model.
Figure 3. Flowchart of EN-PSO-ESN model.
Entropy 26 00215 g003
Figure 4. Pre-training framework of N basic ESNs based on PSO algorithm.
Figure 4. Pre-training framework of N basic ESNs based on PSO algorithm.
Entropy 26 00215 g004
Figure 5. Iteration procedure of a given particle.
Figure 5. Iteration procedure of a given particle.
Entropy 26 00215 g005
Figure 6. Visual representation of prediction performance for six models (MG(1) dataset). (a) The comparison of the average measure indices of the ELM-based models. (b) The comparison of the average measure indices of the ESN-based models.
Figure 6. Visual representation of prediction performance for six models (MG(1) dataset). (a) The comparison of the average measure indices of the ELM-based models. (b) The comparison of the average measure indices of the ESN-based models.
Entropy 26 00215 g006
Figure 7. Visual representation of prediction performance for six models (MG(2) dataset). (a) The comparison of the average measure indices of the ELM-based models. (b) The comparison of the average measure indices of the ESN-based models.
Figure 7. Visual representation of prediction performance for six models (MG(2) dataset). (a) The comparison of the average measure indices of the ELM-based models. (b) The comparison of the average measure indices of the ESN-based models.
Entropy 26 00215 g007
Figure 8. Visual representation of prediction performance for six models (NARMA(1) dataset). (a) The comparison of the average measure indices of the ELM-based models. (b) The comparison of the average measure indices of the ESN-based models.
Figure 8. Visual representation of prediction performance for six models (NARMA(1) dataset). (a) The comparison of the average measure indices of the ELM-based models. (b) The comparison of the average measure indices of the ESN-based models.
Entropy 26 00215 g008
Figure 9. Visual representation of prediction performance for six models (NARMA(2) dataset). (a) The comparison of the average measure indices of the ELM-based models. (b) The comparison of the average measure indices of the ESN-based models.
Figure 9. Visual representation of prediction performance for six models (NARMA(2) dataset). (a) The comparison of the average measure indices of the ELM-based models. (b) The comparison of the average measure indices of the ESN-based models.
Entropy 26 00215 g009
Figure 10. Visual representation of prediction performance for six models (Henon attractor dataset). (a) The comparison of the average measure indices of the ELM-based models. (b) The comparison of the average measure indices of the ESN-based models.
Figure 10. Visual representation of prediction performance for six models (Henon attractor dataset). (a) The comparison of the average measure indices of the ELM-based models. (b) The comparison of the average measure indices of the ESN-based models.
Entropy 26 00215 g010
Figure 11. Visual representation of prediction performance for six models (Lorenz attractor dataset). (a) The comparison of the average measure indices of the ELM-based models. (b) The comparison of the average measure indices of the ESN-based models.
Figure 11. Visual representation of prediction performance for six models (Lorenz attractor dataset). (a) The comparison of the average measure indices of the ELM-based models. (b) The comparison of the average measure indices of the ESN-based models.
Entropy 26 00215 g011
Figure 12. Visual representation of prediction results for six models (AQ dataset). (a) The comparison of the average measure indices of the ELM-based models. (b) The comparison of the average measure indices of the ESN-based models. (c) The fitting performances on the test data. (d) The fitting results on the test data.
Figure 12. Visual representation of prediction results for six models (AQ dataset). (a) The comparison of the average measure indices of the ELM-based models. (b) The comparison of the average measure indices of the ESN-based models. (c) The fitting performances on the test data. (d) The fitting results on the test data.
Entropy 26 00215 g012
Figure 13. Visual representation of prediction results for six models (ASN dataset). (a) The comparison of the average measure indices of the ELM-based models. (b) The comparison of the average measure indices of the ESN-based models. (c) The fitting performances on the test data. (d) The fitting results on the test data.
Figure 13. Visual representation of prediction results for six models (ASN dataset). (a) The comparison of the average measure indices of the ELM-based models. (b) The comparison of the average measure indices of the ESN-based models. (c) The fitting performances on the test data. (d) The fitting results on the test data.
Entropy 26 00215 g013
Figure 14. The DA distribution for four models. (a) The distribution of DA values for 100 initialized canonical ELM models. (b) The DA distribution for 100 PSO-ELM(I) models. (c) The DA distribution for 100 PSO-ELM(II) models. (d) The DA distribution for ensemble PSO-based ELM models.
Figure 14. The DA distribution for four models. (a) The distribution of DA values for 100 initialized canonical ELM models. (b) The DA distribution for 100 PSO-ELM(I) models. (c) The DA distribution for 100 PSO-ELM(II) models. (d) The DA distribution for ensemble PSO-based ELM models.
Entropy 26 00215 g014
Table 1. Models compared in this study.
Table 1. Models compared in this study.
ESN-Based ModelsELM-Based Models
Compared Models
(1)
Canonical ESN
(1)
Canonical ELM
(2)
PSO-ESN
(2)
PSO-ELM
(3)
EN-PSO-ESN
(3)
EN-PSO-ELM
Table 2. Parameter settings for model comparison.
Table 2. Parameter settings for model comparison.
Running TimesHidden Neuron NumberαβEnsemble NumberLeak RateThe Collection Rate of the Reservoir
ELM50100\\\\\
PSO-ELM501000.1\\\\
EN-PSO-ELM501000.1\50\\
Canonical ESN501000.10.01\0.30.01
PSO-ESN501000.10.01\0.30.01
EN-PSO-ESN501000.10.01500.30.01
Table 3. Prediction performance of six models for datasets MG(1) and MG(2); number of running times is 50.
Table 3. Prediction performance of six models for datasets MG(1) and MG(2); number of running times is 50.
ModelsRMSERMAEMAPERMSPEDAPerformance
Order
Mean ValueStandard DeviationMean ValueStandard DeviationMean ValueStandard DeviationMean ValueStandard DeviationMean ValueStandard Deviation
MG(1)ELM4.2333 × 10−43.6354 × 10−52.7920 × 10−41.6822 × 10−53.5139 × 10−42.7389 × 10−56.6799 × 10−49.5655 × 10−5100.00%03
PSO-ELM4.0988 × 10−51.3745 × 10−52.1236 × 10−55.6891 × 10−62.5438 × 10−58.2240 × 10−65.8877 × 10−53.1339 × 10−5100.00%02
EN-PSO-ELM3.3164 × 10−51.3182 × 10−51.5342 × 10−55.4203 × 10−62.1275 × 10−58.1346 × 10−66.4802 × 10−53.1238 × 10−5100.00%01
ESN2.0595 × 10−37.0927 × 10−41.3627 × 10−35.0073 × 10−41.6014 × 10−35.7275 × 10−42.9686 × 10−31.0756 × 10−399.0201%5.1431 × 10−33
PSO-ESN2.0336 × 10−34.7435 × 10−41.2950 × 10−33.6885 × 10−41.5015 × 10−34.4408 × 10−42.7100 × 10−38.3517 × 10−498.9950%7.2647 × 10−32
EN-PSO-ESN1.0973 × 10−34.6523 × 10−49.7777 × 10−43.5712 × 10−41.0609 × 10−34.2366 × 10−41.3206 × 10−37.8256 × 10−498.9950%5.3627 × 10−31
MG(2)ELM3.7327 × 10−41.8502 × 10−53.0227 × 10−41.3882 × 10−54.0961 × 10−42.1957 × 10−56.6574 × 10−45.1463 × 10−5100.00%03
PSO-ELM2.9236 × 10−56.7239 × 10−62.3141 × 10−55.2118 × 10−62.8240 × 10−57.4557 × 10−64.3585 × 10−52.2779 × 10−5100.00%01
EN-PSO-ELM3.4774 × 10−56.6342 × 10−62.7965 × 10−55.0656 × 10−63.7718 × 10−57.3756 × 10−67.3590 × 10−52.0232 × 10−5100.00%02
ESN3.7092 × 10−22.6262 × 10−22.9828 × 10−21.9397 × 10−23.1434 × 10−21.9213 × 10−24.3163 × 10−22.8004 × 10−276.8342%7.7177 × 10−23
PSO-ESN1.7372 × 10−27.2545 × 10−31.5625 × 10−26.6169 × 10−31.5896 × 10−28.3831 × 10−31.9330 × 10−21.1820 × 10−379.8995%1.3707 × 10−22
EN-PSO-ESN2.6844 × 10−37.2331 × 10−32.2819 × 10−36.4356 × 10−32.7199 × 10−38.2356 × 10−34.0754 × 10−31.0298 × 10−497.9899%1.1635 × 10−21
Table 4. Prediction performance of six models for NARMA time series with k = 6 and k = 10; number of running times is 50.
Table 4. Prediction performance of six models for NARMA time series with k = 6 and k = 10; number of running times is 50.
ModelsRMSERMAEMAPERMSPEDAPerformance Order
Mean ValueStandard DeviationMean ValueStandard DeviationMean ValueStandard DeviationMean ValueStandard DeviationMean ValueStandard Deviation
NARMA(1)ELM8.4527 × 10−21.5725 × 10−32.0259 × 10−14.2477 × 10−31.9763 × 10−14.2896 × 10−32.4655 × 10−15.0867 × 10−368.1658%1.7849 × 10−22
PSO-ELM8.7526 × 10−23.0443 × 10−32.1204 × 10−17.0243 × 10−32.1022 × 10−16.5851 × 10−32.6206 × 10−18.3883 × 10−366.3317%1.5415 × 10−23
EN-PSO-ELM8.3184 × 10−22.6577 × 10−32.0105 × 10−16.2344 × 10−31.9733 × 10−15.8975 × 10−32.4445 × 10−17.5654 × 10−371.3568%1.1103 × 10−21
ESN1.39263.6549 × 10−13.34877.6720 × 10−13.66719.0628 × 10−15.03131.476551.2563%2.3409 × 10−23
PSO-ESN2.5375 × 10−11.6762 × 10−26.5980 × 10−15.2713 × 10−27.1371 × 10−17.0387 × 10−29.0056 × 10−17.8975 × 10−257.2864%1.3531 × 10−22
EN-PSO-ESN8.8658 × 10−21.4525 × 10−22.2177 × 10−14.7286 × 10−22.2927 × 10−16.6895 × 10−22.8936 × 10−16.5657 × 10−268.8442%1.0126 × 10−21
NARMA(2)ELM7.5817 × 10−21.8216 × 10−31.6056 × 10−12.9249 × 10−31.6436 × 10−12.6978 × 10−32.0151 × 10−14.9365 × 10−365.5276%1.9358 × 10−22
PSO-ELM7.9777 × 10−21.7392 × 10−31.7079 × 10−13.6645 × 10−31.7622 × 10−14.4762 × 10−32.1918 × 10−16.4204 × 10−365.3266%2.0750 × 10−23
EN-PSO-ELM7.6952 × 10−21.2366 × 10−31.6418 × 10−13.0012 × 10−31.6772 × 10−13.8745 × 10−32.0773 × 10−14.8673 × 10−367.8392%1.5266 × 10−21
ESN9.1040 × 10−12.2495 × 10−11.89003.9288 × 10−12.13314.7139 × 10−12.94877.9041 × 10−151.9598%2.5594 × 10−23
PSO-ESN2.1788 × 10−12.5335 × 10−34.4598 × 10−16.2503 × 10−34.9090 × 10−17.9616 × 10−36.6947 × 10−19.9815 × 10−356.7839%2.0675 × 10−22
EN-PSO-ESN7.5906 × 10−22.4659 × 10−31.5905 × 10−16.0325 × 10−31.6497 × 10−17.2377 × 10−32.1019 × 10−19.3257 × 10−367.3367%1.6885 × 10−21
Table 5. Prediction performance of six models for Henon attractor and Lorenz attractor; number of running times is 50.
Table 5. Prediction performance of six models for Henon attractor and Lorenz attractor; number of running times is 50.
ModelsRMSERMAEMAPERMSPEDAPerformance
Order
Mean ValueStandard
Deviation
Mean ValueStandard
Deviation
Mean ValueStandard DeviationMean ValueStandard
Deviation
Mean ValueStandard Deviation
Henon attractorELM3.3204 × 10−61.9751 × 10−64.4166 × 10−52.6936 × 10−53.7655 × 10−52.3881 × 10−51.5742 × 10−41.0435 × 10−4100.00%03
PSO-ELM1.4408 × 10−61.7335 × 10−61.8179 × 10−52.4365 × 10−51.7493 × 10−51.4665 × 10−58.0322 × 10−53.2457 × 10−5100.00%02
EN-PSO-ELM6.8053 × 10−71.6568 × 10−68.0781 × 10−62.4415 × 10−68.3876 × 10−61.3622 × 10−54.3527 × 10−53.1757 × 10−5100.00%01
ESN1.2784 × 10−55.6772 × 10−61.2730 × 10−45.3143 × 10−51.3102 × 10−45.8792 × 10−58.0622 × 10−45.8563 × 10−4100.00%03
PSO-ESN4.0724 × 10−61.9122 × 10−74.4212 × 10−51.9947 × 10−66.4647 × 10−52.1198 × 10−65.6427 × 10−41.3948 × 10−5100.00%02
EN-PSO-ESN9.4210 × 10−81.8652 × 10−79.2122 × 10−71.8257 × 10−61.1861 × 10−61.9121 × 10−69.5494 × 10−61.2398 × 10−5100.00%01
Lorenz attractorELM8.9132 × 10−56.7269 × 10−53.9429 × 10−62.9819 × 10−64.6187 × 10−63.4981 × 10−65.0824 × 10−63.8467 × 10−6100.00%03
PSO-ELM4.4364 × 10−53.0233 × 10−51.1664 × 10−61.0605 × 10−61.3840 × 10−61.2885 × 10−62.4607 × 10−62.0413 × 10−6100.00%02
EN-PSO-ELM1.5507 × 10−52.9626 × 10−54.9162 × 10−79.2377 × 10−74.9223 × 10−71.0132 × 10−66.4194 × 10−71.8652 × 10−6100.00%01
ESN9.0528 × 10−37.2441 × 10−32.5788 × 10−42.1277 × 10−43.9498 × 10−43.5720 × 10−47.2311 × 10−46.5796 × 10−499.6315%4.2861 × 10−33
PSO-ESN8.0337 × 10−32.3815 × 10−39.6566 × 10−56.0522 × 10−52.0888 × 10−46.1014 × 10−47.5513 × 10−41.0421 × 10−498.4925%1.1511 × 10−32
EN-PSO-ESN6.4638 × 10−42.2356 × 10−32.4464 × 10−55.8677 × 10−52.6527 × 10−55.8206 × 10−43.1241 × 10−58.9885 × 10−5100.00%01
Table 6. Prediction performance of six models (AQ dataset); number of running times is 50.
Table 6. Prediction performance of six models (AQ dataset); number of running times is 50.
ModelsRMSERMAEMAPERMSPEDAPerformance Order
Mean ValueStandard
Deviation
Mean ValueStandard
Deviation
Mean ValueStandard
Deviation
Mean ValueStandard DeviationMean ValueStandard
Deviation
AQELM6.0336 × 10−13.1730 × 10−22.0693 × 10−11.0420 × 10−23.1946 × 10−11.5002 × 10−29.2986 × 10−17.0943 × 10−282.3036%1.3441 × 10−23
PSO-ELM5.1643 × 10−12.5506 × 10−21.8407 × 10−11.0414 × 10−23.0602 × 10−12.3974 × 10−28.9843 × 10−19.2220 × 10−281.3022%1.7108 × 10−22
EN-PSO-ELM5.1077 × 10−12.2681 × 10−21.7526 × 10−11.0287 × 10−22.9757 × 10−12.0834 × 10−28.6287 × 10−18.9669 × 10−284.2763%1.4262 × 10−21
ESN5.2005 × 10−16.1308 × 10−11.8365 × 10−18.0975 × 10−24.0563 × 10−11.1006 × 10−11.88084.5650 × 10−183.6772%2.0616 × 10−23
PSO-ESN4.7754 × 10−12.0790 × 10−21.5994 × 10−18.6829 × 10−33.9858 × 10−12.5555 × 10−21.38181.0542 × 10−186.2597%1.1972 × 10−22
EN-PSO-ESN4.5206 × 10−11.8512 × 10−21.5012 × 10−17.8701 × 10−33.7002 × 10−12.3122 × 10−21.28748.2653 × 10−288.0127%8.2701 × 10−31
Table 7. Prediction performance of six models (ASN dataset); number of running times is 50.
Table 7. Prediction performance of six models (ASN dataset); number of running times is 50.
ModelsRMSERMAEMAPERMSPEDAPerformance Order
Mean ValueStandard
Deviation
Mean ValueStandard
Deviation
Mean ValueStandard
Deviation
Mean ValueStandard
Deviation
Mean ValueStandard Deviation
ASNELM5.27134.7943 × 10−13.2036 × 10−22.4332 × 10−33.1748 × 10−22.4495 × 10−34.1038 × 10−23.7342 × 10−361.0726%3.3335 × 10−22
PSO-ELM5.37581.09273.2501 × 10−26.0878 × 10−33.2192 × 10−26.3756 × 10−34. 1941 × 10−29.3557 × 10−359.9668%3.4171 × 10−23
EN-PSO-ELM5.27489.0837 × 10−13.0147 × 10−25.7950 × 10−33.0231 × 10−26.2611 × 10−34.1590 × 10−28.9755 × 10−369.0036%3.2435 × 10−21
ESN8.65834.3130 × 10−14.8673 × 10−29.7585 × 10−24.8916 × 10−23.8668 × 10−26.9217 × 10−23.5095 × 10−261.8763%3.4994 × 10−23
PSO-ESN5.64475.5497 × 10−13.3215 × 10−23.3487 × 10−33.2834 × 10−23.4597 × 10−34.4033 × 10−24.5414 × 10−364.1582%4.2528 × 10−22
EN-PSO-ESN5.07175.3557 × 10−12.9766 × 10−23.1642 × 10−32.9418 × 10−23.2883 × 10−33.9145 × 10−24.3776 × 10−365.0027%3.0217 × 10−21
Table 8. Models compared in additional experiment.
Table 8. Models compared in additional experiment.
Description
ModelsCanonical ELM modelThe traditional ELM model without pre-training
PSO-ELM(I) modelPart of the non-zero elements in the Win will be selected as a candidate for pre-training by PSO
PSO-ELM(II) modelPart of the element in the Win, including non-zero and zero elements, will be selected as candidates for pre-training by PSO
EN-PSO-ELM modelEnsemble strategy combined with PSO-ELM(II) model
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tang, L.; Wang, J.; Wang, M.; Zhao, C. Ensemble and Pre-Training Approach for Echo State Network and Extreme Learning Machine Models. Entropy 2024, 26, 215. https://doi.org/10.3390/e26030215

AMA Style

Tang L, Wang J, Wang M, Zhao C. Ensemble and Pre-Training Approach for Echo State Network and Extreme Learning Machine Models. Entropy. 2024; 26(3):215. https://doi.org/10.3390/e26030215

Chicago/Turabian Style

Tang, Lingyu, Jun Wang, Mengyao Wang, and Chunyu Zhao. 2024. "Ensemble and Pre-Training Approach for Echo State Network and Extreme Learning Machine Models" Entropy 26, no. 3: 215. https://doi.org/10.3390/e26030215

APA Style

Tang, L., Wang, J., Wang, M., & Zhao, C. (2024). Ensemble and Pre-Training Approach for Echo State Network and Extreme Learning Machine Models. Entropy, 26(3), 215. https://doi.org/10.3390/e26030215

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop