Next Article in Journal
Investigation and Validation of Numerical Models for Composite Wind Turbine Blades
Next Article in Special Issue
Verification and Validation of a Methodology to Numerically Generate Waves Using Transient Discrete Data as Prescribed Velocity Boundary Condition
Previous Article in Journal
Static Characteristics and Leakage Rates of Smooth Annular Seals Based on a New Solution Method for Gas-Liquid Two-Phase Conditions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhanced Weight-Optimized Recurrent Neural Networks Based on Sine Cosine Algorithm for Wave Height Prediction

by
Alawi Alqushaibi
1,2,
Said Jadid Abdulkadir
1,2,*,
Helmi Md Rais
1,
Qasem Al-Tashi
1,2,3,
Mohammed G. Ragab
1,2 and
Hitham Alhussian
1,2
1
Department of Computer and Information Sciences, Universiti Teknologi PETRONAS, Seri Iskandar 32610, Perak, Malaysia
2
Centre for Research in Data Science (CERDAS), Universiti Teknologi PETRONAS, Seri Iskandar 32610, Perak, Malaysia
3
Faculty of Administrative and Computer Sciences, University of Albaydha, Rada’a CV46+6X, Yemen
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2021, 9(5), 524; https://doi.org/10.3390/jmse9050524
Submission received: 30 March 2021 / Revised: 13 April 2021 / Accepted: 14 April 2021 / Published: 12 May 2021
(This article belongs to the Special Issue Wind and Wave Climate)

Abstract

:
Constructing offshore and coastal structures with the highest level of stability and lowest cost, as well as the prevention of faulty risk, is the desired plan that stakeholders seek to obtain. The successful construction plans of such projects mostly rely on well-analyzed and modeled metocean data that yield high prediction accuracy for the ocean environmental conditions including waves and wind. Over the past decades, planning and designing coastal projects have been accomplished by traditional static analytic, which requires tremendous efforts and high-cost resources to validate the data and determine the transformation of metocean data conditions. Therefore, the wind plays an essential role in the oceanic atmosphere and contributes to the formation of waves. This paper proposes an enhanced weight-optimized neural network based on Sine Cosine Algorithm (SCA) to accurately predict the wave height. Three neural network models named: Long Short-Term Memory (LSTM), Vanilla Recurrent Neural Network (VRNN), and Gated Recurrent Network (GRU) are enhanced, instead of random weight initialization, SCA generates weight values that are adaptable to the nature of the data and model structure. Besides, a Grid Search (GS) is utilized to automatically find the best models’ configurations. To validate the performance of the proposed models, metocean datasets have been used. The original LSTM, VRNN, and GRU are implemented and used as benchmarking models. The results show that the optimized models outperform the original three benchmarking models in terms of mean squared error (MSE), root mean square error (RMSE), and mean absolute error (MAE).

1. Introduction

In coastal and marine structural engineering, wave height is the main factor that needs to be considered. The wave conditions affect several marine activities and have significant consequences for marine industries. Nevertheless, it is challenging to accurately predict such tasks because the ocean waves are stochastic by nature [1,2]. Thus, estimating the wave heights and their trends was, and still is, a big challenge that has to be tackled [3,4].
At early stages, semi-analytic models, such as the Pierson–Neumann–James and Sverdroup–Munk–Bretscheider models, were used to predict the height of the waves; still, such models are inefficient at clearly describing in detail the wave conditions of the sea surface [5,6]. After that, numerical models have been widely used for wave height prediction; nevertheless, such models require a high computational resource when dealing with large amount of data [7,8].
Metocean data are the combination of meteorology and oceanography parameters that need to be studied in order to predict the oceanic environmental change [9]. The meteorology parameters are wind speed, direction, humidity, and air temperature. The oceanography parameters are wave height, period, current, and tides [10,11]. The prediction of oceanic environmental change is essential when planning coastal and offshore constructions, maintenance, or operation. A high-accuracy prediction of these parameters in the offshore environment reduces cost, time consumed, and fault risk. It also helps in determining the required weather window to achieve the projects in time.
Deep learning models were recently applied to atmospheric observation [12,13]. First, the height of the wave has been predicted based on Artificial Neural Network (ANN) models. For instance, a feedforward network model was utilized to forecast the real-time wave height [14]. Similarly, Mandal et al. [15] proposed a Recurrent Neural Network (RNN) for forecast wave height and showed better correlation coefficient compared to feedforward network. Another study conducted by Mahjoobi et al. [16] predicted wave height based on regressive support vector machine (SVM). Their outcomes demonstrated that the proposed model outperforms ANN model in terms of accuracy and computational time. Likewise, ANNs backpropagation and feedforward were compared against model trees to indicate the superiority of predicating wind speed as well as wave height and found that model trees is superior [17].
The ultimate objective of sea wave prediction modeling is to find precise short- or long-term forecasts of the studied variables at a certain time and location [18,19]. The literature of oceanic modeling characteristic is categorized into physical-based and model-based methods [20,21]. These methods have been utilized to forecast ocean waves. According to the work in [3], which was based on the work in [21], these approaches are further graded based on their attempts to specifically parameterize ocean wave interactions. Physical-based methods mimic sea waves by finding the appropriate equation solutions and have been proven to be beneficial over longer time periods for forecasting, whereas model-based methods can be considered as time-series or statistical methods and work well for short-term prediction [22,23,24,25,26]. Moreover, machine learning (or statistical) models can be applied to postprocess physics-based models [27,28,29,30,31].
The main contribution of this study is to predict the wave height with high accuracy and less computational cost. We precisely investigated the following objectives:
  • To propose an enhanced weight-optimized RNN based on SCA optimization to process time-series data with high accuracy.
  • Three variants of RNNs are enhanced based on SCA: VRNN-SCA, LSTM-SCA, and GRU-SCA.
  • To update and tune the learning rate based on grid search mechanism.
  • To compare the proposed models against three well-regarded prediction models: LSTM, GRU, and VRNN, then investigate whether the proposed models outperforms in terms of mean squared error (MSE), root mean square error (RMSE), and mean absolute error (MAE).
The rest of this paper is categorized as follows. Section 2 reviews the recent related work on recurrent neural network models for predictions. Section 3 describes metocean data properties and four stations information in different locations. Section 4 explains the methodology, and Section 5 illustrates the experimental setup. Results and discussion are explained in Section 7. Last, this paper is concluded in Section 8.

2. Related Work

Artificial neural network models are labeled to fill within the model-based category, and several neural networks have been utilized to forecast and reconstruct wave characteristics. Artificial Intelligence (AI) models have become an alternative to numerical models in recent years. These models are relatively simple to assemble and have outperformed computational and statistical models for a site-specific wave parameter [32]. The metocean data can be modeled and featured by applying the feature selection algorithms [20,33,34,35,36,37]. Fully connected neural networks in [38] were used to forecast wave heights. RNNs were used in [32] to predict the waves and showed a better correlation Coefficient forecasting. A fuzzy logic modeling machine was also used in [32] to predict the ocean wave energy. An extreme learning machine was applied in [39].
Several studies have been conducted using meta-heuristic algorithms to improve neural network prediction. For instance, Zhang et al. [40] used a fruit fly optimization algorithm and kernel extreme learning machine for bankruptcy prediction. Zhao et al. [41] improved ant colony optimization by utilizing a chaotic intensification strategy and random spare strategy for multi-threshold image segmentation. Tu et al. [42] proposed an improved whale optimization algorithm to overcome the local optimum stagnation problem as well as slow convergence speed. Similarly, the whale optimizer was improved using chaotic multi-swarm to boost support vector machine for medical diagnosis [43]. Shan et al. [44] proposed an improved moth flame optimizer based on a mechanism of adaptive weight. Besides, an enhanced moth flame optimizer-based Gaussian mutation, Levy mutation, as well as Cauchy mutation was proposed for global optimization [45]. Similarly, chaotic moth flame optimizer was utilized in [46] to boost kernel extreme learning machine for medical diagnoses. Chen et al. [47] proposed a new variant of Harris hawks optimizer by integrating several strategies such as topological multi-population, chaos, as well as differential evolution.
Comparative studies have been carried out in [48], taking into account geographical differences and using different neutral networks. The findings show that networks with less allocation of resources have a better structure and adaptability to different situations and conditions [1]. Wang et al. [7] used an evolutionary hybrid algorithm to simulate the ocean waves. This method demonstrated better results than other neural networks in which fewer weather data were available. A symbiotic organisms hunt in [49] was proposed to forecast ocean wave heights in two time zones based on a large number of accurate weather data including wave heights measured by buoys. The proposed model performed better than other state-of-the-art models. In [50], the efficiency of optimization algorithms in resolving real-world complex problems such as the wave height problem is discussed by proposing hybrid approach based on accent-based multi-objective particle swarm optimization algorithm, whereas in [20], a sequence-to-sequence neural network, feature selection, and bayesian hyperparameter are the techniques that were used and applied to forecast the height of the wave and reconstruct a prediction for neighboring buoys stations.
For time series forecasting, researchers apply cross-validation as a technique to distribute the data as in [51,52]. However, cross-validation approaches can be applied to stationary synthetic time series. Besides, using cross-validation depends mainly on the type of the data. Therefore, in our study we used the Out-of-sample (OOS) method, which is traditionally used to estimate predictive performance in time-dependent data.
Despite the good performance of such models, they still face some limitations. For example, feedforward models do not perform well with big time-series data. The RNNs face the phenomenal issue of vanishing gradient. Recently, Rashid et al. [53] attempted to use a set of meta-heuristics algorithms, such as Harmony search, GWO, as well as SCA, in order to overcome the vanishing gradient in LSTM. Similarly, Somu et al. [54] proposed an improved SCA using a Haar wavelet-based mutation operator to optimize the hyperparameters of LSTM for energy consumption prediction [55]. In this paper, the main contribution is to utilize the SCA to enhance the weight of three different types of RNN: LSTM, VRNN, and GRU to forecast wave heights with high accuracy, so that instead of random weight initialization, SCA generates weight values that are adaptable to the nature of the data and model structure. Furthermore, GS is employed to automatically find the best models’ configurations. Besides, according to the No free Lunch theorem [56], there is no specific model to solve all forecasting problems. As such, improvements can be made to the existing models to enhance the performance of such models [57].

3. Study Area

In order to measure the efficiency of the proposed models, four stations were selected from different locations, having several climate conditions as well as variants depths of water. The hourly data that have been used in this study were recorded by floating buoys located in the North Atlantic Ocean. Table 1 contains the details of the four stations, and Figure 1 depicts their distribution.
The selected data obtained from the NOAA (https://www.ndbc.noaa.gov/, accessed on 24 December 2020). Four stations are used in this study: Station (41010) (https://www.ndbc.noaa.gov/station_page.php?station=41010, accessed on 24 December 2020); Station (41040) (https://www.ndbc.noaa.gov/station_page.php?station=41040, accessed on 11 January 2021); and Stations (41043) (https://www.ndbc.noaa.gov/station_page.php?station=41043, accessed on 5 January 2021). The details of these stations are described in Table 1. The stations are owned and maintained by National Data Buoy Center National Data Buoy Center, the world’s largest environmental data buoys network [58].
The fourth station (41060) (https://www.ndbc.noaa.gov/station_page.php?station=41060, accessed on 17 Janury 2021) is owned and maintained by Woods Hole Northwest Tropical Atlantic Wave Station (http://uop.whoi.edu/currentprojects/currentprojects.html, accessed on 17 January 2021), which mainly focuses on studying the physical process in ocean surface using moored surface buoys that have metrological and oceanographic sensors.
The wave height is influenced by several factors such as direction and speed of wind, past wave height, and temperature of ocean surface [59,60]. The dataset of each station contains the following features: wind speed, significant wave height, wind direction, gust speed, dominant wave period, the direction from which the waves at the dominant period are coming, average wave period, air temperature, dewpoint temperature, sea surface temperature, sea level pressure, the water level in feet above or below mean lower low water, station visibility, and pressure tendency.

4. Methods

In this section, the main methods used in this paper are explained. First, the original VRNN is outlined, followed by LSTM then GRU. After that, the SCA algorithm is mathematically presented and finally the GS mechanism is elaborated.

4.1. Vanilla Recurrent Neural Network (VRNN)

The simplest structure of VRNN is initialized with one single input layer, one single hidden layer, and one single output layer [61]. Simple architecture is constructed of three layers that can process the sequence of T inputs through time t, which is the vector of , x t 1 , x t , x t + 1 , . The x t = x 1 , x 2 , x T , where T is the different length series of the x t inputs.
All three layers are hierarchically joined from input layer to the hidden layers and from the hidden layer to the output layer. The link between the layers of the network is called a weight matrix. W W defines the weight connection between the hidden layer and the input layer at each time-step. Equation (1) computes the hidden layer recursively to measure the current state of the network. W V is the weight matrix reference that links the hidden layer units to each other. The number of hidden units H is h t = h 1 , h 2 , h H .
s t = W W x t + W V h t 1 + b h
The W W matrix is multiplied by the x t inputs and summed up with the product of W V and the previous state h t 1 . Then, this result is added to bias b h of the hidden layer. Equation (1) defines this process. Equation (2) define the networks’ current state.
h t = f H ( s t )
f H ( . ) represents the nonlinear activation function that converts the result of s t to values that depend on the selected activation function. Meanwhile, several activation functions on the state-of-the-art such as sigmoid, tanh, relu, leaky relu, and many more [62,63,64].
W U is the weight matrix that connects hidden and output layers. The output numbers is N units and can be represented as y t = y 1 , y 2 , , y N , while y ^ t is the network prediction that can be obtained by Equation (3).
y ^ t = f s ( W U h t + b o )
The result of the output layer is the sum of product of weights W U and current hidden state h t form Equation (2), added to bias of output layer b o , then it is transformed by activation function f s ( . ) . The output layer predicts the results at each time step based on the results of the hidden layers calculations and input values. Y ^ t defines the total prediction length parameter. The y ^ t = y ^ t 1 , y ^ t 2 , , y ^ Y ^ .

4.2. Long Short-Term Memory (LSTM)

LSTM is a new improved variant of RNN that has been proposed to solve the classical version VRNN in terms of gradient problem [65]. Nevertheless, because of different memory cells, LSTM is more expensive in terms of computational resources and requires extra memory compared to RNN [65]. Typical LSTMs have almost four times more parameters than simple VRNN; thus, they suffer from high complexity in hidden layers. The main goal of designing LSTM is to solve the difficulties of learning long-term dependencies, regardless of the uncertainty of their costs [61,66].
The LSTM cell is constructed of several main gates, which are the input gate, forget gate, and the output gate. Irreverent information that has less importance on the prediction is dropped out by the forget gate. By this mechanism, LSTM determines which new data are going to be processed in the cell state [61]. The cell state is altered by the forget gate positioned below the cell state and by the input gate. The previous cell state forgets and adds new information through the output of input gates. The forget gate decides which data should be dropped. It forgets the irrelevant details coming from the previous state with the following calculation in Equation (8) [61,67]. The input gates decide which cell state or long-term memory information can enter. There are two sections to this layer: One is the sigmoid function, while the other is the function of tanh ( ) . Typically, the target vector is called the “output gate”. The following equations explain the parameter notations in LSTM cell.
i t = σ ( W x i x t + W h i h t 1 + W c i c t 1 + b i )
f t = σ ( W x f x t + W h f h t 1 + W c f c t 1 + b f )
g t = tanh ( W x g x t + W h g h t 1 + b g )
o t = σ ( W x o x t + W h o h t 1 + W c o c t + b o )
c t = f t c t 1 + i t g t
h t = o t tanh ( c t )
y ^ t = h t
The f ( t ) is the forget gate, i ( t ) is the input gate, and o ( t ) is the output gate. g ( t ) is the method that checks the current input of x ( t ) . The previous short-term memory is h ( t 1 ) . C ( t 1 ) is the previous long-term memory state. The logistic sigmoid function is σ and the tanh ( ) is the tanh function. y ^ t is the expected prediction output of the LSTM cell. The c ( t ) is the long-term state for the next cell, and the h ( t ) is the short-term state for the next cell.
W x i , W x f , W x g , and W x o are the connecting weight matrices between the input x t and the four layers. On the other hand, the connections between the previous state h ( t 1 ) and the four layers are W h i , W h f , W h g , and W h o . Every layer has its own bias as b i , b f , b g , and b o , respectively.

4.3. Gated Recurrent Units (GRU)

Cho et al. [68,69] introduced GRU to solve the problem of decay of information that occurs in the traditional recurrent neural network and minimize the computational complexity in LSTM. Two controlling gates are used by GRU, which are the updating and resetting gates that track the data forwarded to the output gate. The update gates decide the past and present data that must be moved to the new state, while the reset gate determines the previous data that must be dropped at every time step. The parameters of the GRU cell are defined by the following equations.
Z t = σ ( W x z x t + W h z h t 1 + b z )
R t = σ ( W x r x t + W h r h t 1 + b r )
g t = tanh ( W x g x t + W h g ( R t h t 1 ) + b g )
h t = Z t h t 1 + ( 1 Z t g t )
where Z t is the update gate and W x z is the weight matrix between input layer and the update gate, whereas W h z is the weights connecting the update gate with the hidden state h t 1 . b z is the bias of the updating gate.
R t representing the reset gate which is connected to the input layer by W x r , while W h r is the weight matrix that connects it to the hidden state h t 1 . The bias that belongs to this gate is b z .

4.4. Sine Cosine Algorithm (SCA)

SCA is a population-based optimization technique that randomly generates multiple possible solutions for optimization problems. Using the mathematical Sine-Cosine equations to oscillate towards or outwards in order to find the optimal solutions. It uses random variables to adaptively ensure this technique emphasizes on the exploitation and exploration to finding the possible global optima on the search space [70,71]. SCA has been found more efficient than other population-based algorithms in achieving an optimal global solution [54]. Different millstones on the search area are investigated when the sine and cosine functions return values greater than one or less than one. Equations (15) and (16) demonstrate the formation of sine and cosine functions.
X i t + 1 = X i t + r 1 * sin ( r 1 ) * | r 3 P i t X i t |
X i t + 1 = X i t + r 1 * cos ( r 1 ) * | r 3 P i t X i t |
where X i t is the position of the current candidate at the tth iteration in the ith dimension AND P i is the position of the best candidate at the tth iteration in the ith dimension. The random agents are r 1 , r 2 , r 3 , and r 4 . The ( * ) is the multiplication sign. Equations (15) and (16) are combined in Equation (17).
X i t + 1 = X i t + r 1 * sin ( r 1 ) * | r 3 P i t X i t | , r 4 < 0.5 X i t + r 1 * cos ( r 1 ) * | r 3 P i t X i t | , r 4 0.5
The first agent ( r 1 ) is responsible for defining the afterward search space, located between the solution region or outside it. The second operator ( r 2 ) defines the amount of distance in the search space that should be in or out of the destination
r i = a t a T
where T indicates the maximum iterations number and t is the currently running iteration. The a is a constant variable. Algorithm 1 shows the main pseudocode of SCA algorithm.

4.5. Grid Search (GS)

In order to produce good accurate results, deep learning models require many parameters that need to be predefined before the training takes place [72,73]. These hyperparameters have to be suitable for the network structure and the nature of the dataset. Setting these hyperparameters can be accomplished manually by trial and error until the best results are achieved; however, this is an inefficient method as it consumes time and may not work as expected [68,74]. Therefore, there are many hyperparameter tuning techniques that have automatically configured models to make them suitable to the network’s structure [75,76].
Grid Search is a hyperparameter tuning technique that can be applied to find the best model configuration [77]; it produces more accurate results [75,78]. Therefore, in this paper, GS has been utilized to obtain the best learning rate values.
Algorithm 1: Algorithm SCA
Input: 
      • Set the lower bound and upper bound of X solutions
      • Set the population size
      • Initialize the agents of the search space randomly
      • Specify the maximum number of iterations
Output: 
       • The best-selected solution ( X * )
LOOP Process
1:
while ( t Max number of iterations T ) do
2:
   Calculate every single solutions candidate
3:
   Define the best-selected solution ( X * )
4:
   Update r 1 , r 2 , r 3 , a n d r 4
5:
   Update agents’ locations in the search space with (18)
6:
end while
Return (X*) 

5. The Proposed Enhanced Weight Optimized Recurrent Neural Networks

The models developed in this paper have been achieved by integrating the sine cosine algorithm [70] to optimize and update the model weights of simple recurrent neural networks aiming to overcome the problem of vanishing gradient in prediction wave heights. Besides, we integrated an effective grid search mechanism to find the optimal configuration of models’ structures.
The main contribution is to generate weight values that are adaptable to the nature of dataset and models’ structure. To achieve that, instead of initializing the weights of recurrent neural networks randomly, the weight initialization is adapted using SCA algorithm. As explained in the literature, the traditional training of RNNs is usually based on back propagation so that parameters, such as weights and leaning rate, would be updated either by increasing their values or decreasing it until finding the optimal values, resulting in minimizing the error value. The random initialization of the weights was developed to overcome the drawbacks of the back propagation and to enhance the convergence speed with less time. However, the randomness of weight initialization might not be adaptive with every data type and size [79]. Therefore, we proposed an enhanced weight optimized model to effectively overcome the limitations of the state-of-the-art methods, as well as to be more adaptive to any type of dataset.
The general structure of the proposed model is illustrated in Figure 2. The following subsections demonstrate the phases of the proposed model to analyze the metocean data and predict the wave heights.

5.1. Data Preprocessing

As explained in Section 3, the four datasets that have been used in this paper were selected from different locations based on Table 1. The datasets have been preprocessed as follows:
  • The linear interpolation is used to handle the missing values.
  • The time series dataset has been partitioned into 70% for training data and 30% for testing data; 25% of the training portion was assigned for validation.
  • After splitting the dataset, the input features and target labels are identified.
  • Finally, the MinMaxscaler function from a Python library named Scikit-Learn has been utilized as a normalization technique to scale the data into a suitable form within the range (0, 1).

5.2. Grid Search Mechanism

The optimal hyperparameters are essential in any neural network to obtain sufficient performance. Traditionally, trial and error was used to find the best hyperparameters; however, this technique is not effective as it takes a long time and, in most cases, the optimal values are not guaranteed. It needs to try to set the parameters, train the model, validate, test, and compute the error. This cycling process needs to be repeated many times until it begins getting better results.
Therefore, we integrate an effective technique known as grid search to efficiently find the best hyperparameters for the proposed models. Table 2 shows the parameter settings of the grid search.
Table 2 shows the values of the search space that allows the GS to find and selects the optimal values that are suitable to dataset size and model structure. To avoid the vanishing and exploding gradient problem, the optimal learning rate should not be too small nor too big. Therefore, we set the search space to be between 0.001 and 0.3.

5.3. Weight Optimization Process Using SCA

SCA is a recent effective optimization algorithm capable of exposing an efficient performance and proved to be more effective than several optimization algorithms for attaining optimum or near optimum solution; therefore, it has been selected in this study due to its special characteristics, which are summarized as follows.
  • The potential to escape the local optima, besides the high exploration inheritance of sine and cosine functions.
  • The basic sine cosine functions enable this algorithm to adaptively shift from exploration [<1, >1] to exploitation [ 1 , 1 ] .
  • The inclination towards the finest region of the search space as the solution modifies its location around the finest solutions gained so far.
The three basic recurrent neural network models (VRNN, LSTM, and GRU) are enhanced based on sine cosine algorithm, and new models named VRNN-SCA, LSTM-SCA, and GRU-SCA were developed. A clear description of these models is explained as follows.

5.3.1. Optimizing VRNN-SCA

The VRNN-SCA model’s inputs are multiplied with weights that are generated by the SCA algorithm. The results of the multiplication are then fed as input vector to the hidden state and multiplied with the hidden state weight matrix generated by SCA as well. All products’ results are summed up with additional bias values that came from the SCA. The weights of the basic VRNN presented in Equation (1) are updated by integrating the SCA as shown in Equation (19).
s t = W scaW x t + W scaV h t 1 + b scah
where W scaW indicates the generated weight based SCA in the input layer, whereas x t represents the inputs. W scaV indicates the generated weight based SCA in the hidden layer, whereas, h t 1 is the hidden state. Besides, b scah represents the generated bias based on SCA.

5.3.2. Optimizing LSTM-SCA

The cell state c ( t ) and the three gates of LSTM—input gate i ( t ) , forget gate f ( t ) , and output gate o ( t ) —have their own weights, as explained in Section 4.2. Instead of generating these weights randomly, they have been generated based on SCA to be more adaptive to the input dataset. Equations (4), (5) and (7) have been updated as
i t = σ ( W scaXi x t + W scaHi h t 1 + W scaCi c t 1 + b scai )
where W scaXi represents the generated weights (based on SCA) that connect the input layer to input gate i ( t ) , while hidden state h t 1 and input gate are connected by W scaHi , which represents the updated weight matrix-based SCA. The cell state c t 1 is connected by W scaCi generated SCA weights to input gate and b scai is the input gate bias updated by SCA.
f t = σ ( W scaXf x t + W scaHf h t 1 + W scaCf c t 1 + b scaf )
The forget gate f t is connected to the input layer by generated SCA weights W scaXf and connected to the hidden state h t 1 by W scaHf weights which generated by SCA as well. It is also connected to the cell state c t 1 by SCA generated weights W scaCf . The forget gate’s bias, which is generated by SCA, is b scaf .
o t = σ ( W scaXo x t + W scaHo h t 1 + W scaCo c t + b scao )
where W scaXo represents the generated weights based on SCA that connect the output gate o t to the input layer. The generated SCA weight W scaHo is between the output gate and hidden state h t 1 , while the cell state is connected with the output gate by W scaCo , which indicates SCA generated weights. The generated SCA bias of the output gate is b scao .

5.3.3. Optimizing GRU-SCA

GRU only has two gates: the update gate Z t and reset gate R t . These gates have the weights matrix as described in Section 4.3. These weights have been updated based on SCA. Equations (11) and (12) have been updated as follows:
Z t = σ ( W scaXz x t + W scaHz h t 1 + b scaz )
W scaXz is the weight matrix between the input layer and the update gate Z t , whereas W scaHz is generated SCA weights that connecting the update gate with the hidden state h t 1 . b scaz is the updating gate bias generated based on SCA.
R t = σ ( W scaXr x t + W scaHr h t 1 + b scar )
Similarly, the reset gate R t is connected to the input layer by SCA generated weights ( W scaXr ), while W scaHr is the weight matrix obtained by SCA, and it connects the reset gate to the hidden state h t 1 . The bias that belongs to this gate is b scaz , and it is generated based on SCA as well.

6. Experimental Setup

This section explains the experimental settings used in this paper; it starts with the evaluation measures used to validate the proposed models then the parameter setting. The implementation is performed on Intel(R) Core(TM) i7-9700 CPU @ 3.00 GHz 3.00 GHz, 0 16.0 GB RAM. The experiments were implemented using the Python 3 programming language and the libraries below:
  • Environment setup:
    • Python version (Python 3.7.9).
    • Virtual environment from Anaconda.
    • TensorFlow (2.3.0) as backend.
  • Library setup:
    • scikit-learn (0.23.2).
    • scipy (1.4.1).
    • pandas (1.1.3).
    • numpy (1.18.5).
    • matplotlib (3.3.2).
    • Keras (2.4.3)

6.1. Evaluation Measures

In order to comprehensively evaluate the effectiveness and prediction of the proposed models, three common metrics are used in this paper: Mean Square Error (MSE), Root Mean Square Error (RMSE), and Mean Absolute Error (MAE). All of these error evaluation indices have been extensively applied in the forecasting model estimation. These three metrics are defined as:
M S E = 1 N i = 1 N y i y ^ i 2
R M S E = i = 1 N y i y ^ i 2 N
M A E = 1 N i = 1 N y i y ^ i 2

6.2. Parameter Settings

Setting up the experimental environments has been done by first defining the parameters of the SCA as illustrated in Table 3; the parameters are based on the default setting of the algorithm itself. Then, we defined the parameters of the models that are explained in Table 4. Note that the three recurrent neural network models (VRNN, LSTM, and GRU) are implemented as benchmarking models for comparison purposes.
For all the proposed models, the model structure consists of an input layer which includes 12 input features, a hidden layer consisting of 32 hidden units, and an output layer. The dataset has been partitioned into 70% for training data and 30% for testing data; 25% of the training portion was assigned for validation.

7. Results and Discussion

This section demonstrates and discusses the results of the enhanced models. To ensure the stability of the proposed models, the models have been run ten times for each dataset. Besides, different initialization in terms of epoch has been conducted. The number of epochs was set to 10, 20, 30, 40, 50, 60, 70, 80, 90, and 100.

7.1. Results of Proposed Models

Table 5 illustrates the results achieved by proposed VRNN-SCA model in all four datasets. In all epochs, this model demonstrates good performance in terms of MSE, RMSE, and MAE. For example, in the 41010 dataset, the best MSE, RMSE, and MAE were in epoch 40 with 0.0002, 0.01152, and 0.0110, respectively. Similarly, in the 41040 dataset, the best performance was in epoch 40 with 0.0016 MSE, 0.0396 RMSE, and 0.0230 MAE, whereas in the 41,043 dataset, epoch 10 obtained best performance with 0.0029, 0.0543, and 0.0232 for MSE, RMSE, and MAE, respectively. In the 41,060 dataset, the obtained MSE, RMSE, and MAE best results was on epoch 90 with 0.0010, 0.0313, and 0.0234, respectively.
As can be seen form Table 6, which shows the results of the proposed LSTM-SCA model in all datasets, this model demonstrates a promising performance in terms of MSE, RMSE, and MAE. For example, in dataset from station 41010, the best MSE was in epochs 10, 20, and 30 with values of 0.0001, whereas the best performance in terms of RMSE and MAE was in epoch 30 with 0.0106 and 0.0065, respectively. In dataset 41040, the best obtained MSE, RMSE, and MAE were in epoch 20 with the values of 0.0014, 0.0371, and 0.0210, respectively. In the 4143 dataset, epoch 10 achieved best results in terms of MSE, RMSE, and MAE with the following values: 0.0019, 0.0435, and 0.0168, respectively, whereas epoch 100 achieved the best performance in the 41060 dataset with the values of 0.0010, 0.0320, and 0.0239 for MSE, RMSE, and MAE, respectively.
Table 7 illustrates the results of the proposed GRU-SCA model in all datasets; this model achieves an outstanding performance in terms of all evaluation measures MSE, RMSE, and MAE. For instance, MSE in dataset 41010 achieved the best results in epochs 20, 30, 40, and 50 with the value of 0.0001, while RMSE and MAE obtained the best results in epoch 40 with values of 0.0092 and 0.0063, respectively. Epoch 20 obtained the best results for dataset from station 41040 with the values of 0.0008, 0.0282, and 0.0180 for MSE, RMSE, and MAE, respectively. In dataset 41043, the best results were achieved in epoch 10, whereas in epoch 100, the best performance was achieved for dataset 41060.
Table 8 shows the results of the original model of VRNN. These results demonstrated how effective the integration of the SCA algorithm and the grid search mechanism are in producing better results and overcome the existing work limitations. Table 9 explains the results of the original LSTM. Table 10 explains the results of the original GRU benchmarking over all the datasets.

7.2. Comparison of the Proposed Models with Existing Models

In this subsection, we benchmark our proposed models with three existing recurrent neural networks. Table 11 shows the comparison in terms of average MSE on all dataset for all the three proposed models as well as the three original RNNs models. As can be seen from Table 11, the proposed VRNN-SCA model outperforms the other two proposed models as well as the three original models in terms of MSE in datasets 41040 and 41060. The proposed GRU-SCA outperforms all other models in dataset 41010. However, the original VRNN-SCA clearly outperforms the three original models (VRNN, LSTM, and GRU) in the 41040 dataset and takes the third place in general.
Table 12 shows the comparison in terms of RMSE average for optimized models and original ones on all datasets. The results clearly show that the proposed models outperformed the original RNNs and GRU-SCA shows best results in datasets 41010 and 41060, while LSTM-SCA achieved the best results on dataset 41043. Finally, VRNN-SCA shows the best RMSE results on dataset 41040.
Table 13 explains the average of MAE comparison for all models on all dataset. It is obvious that GRU-SCA came at the first rank with best two results on datasets from stations 41010, and 41060. LSTM-SCA model shows best result value on dataset 41043. On the dataset from station 41040, the best RMSE average results was achieved by VRNN-SCA.
Figures 3a, 6a, 9a and 12a show the best selected prediction vs. actual data for VRNN-SCA on all datasets. The blue color line represents the actual data, while the orange color represents the prediction. Figures 3b, 6b, 9b and 12b show the result that was produced from the non-optimized model.
LSTM-SAC model’s best predicted results are illustrated in prediction graphs as shown in Figures 4a, 7a, 10a and 13a and compared with the best selected results of original LSTM, which are shown in Figures 4b, 7b, 10b and 13b.
The best selected results produced by GRU-SCA can be seen in Figures 5a, 8a, 11a and 14a. On the other hand, Figures 5b, 8b, 11b and 14b show the results for the original GRU model.

7.3. Discussion

All experiments on the three models showed that our proposed technique can be effectively used to forecast wave heights with more prediction accuracy. The simple architecture of all variant of recurrent neural networks, which are VRNN, LSTM, and GRU, can be optimized in terms of weights generation by sine cosine optimization algorithm. The proposed RNN-SCA models have shown an outstanding performance and outperformed the state-of-the-art models in terms of MSE, RMSE, and MAE.
The difference between the graphs in Figure 3 is slightly noticeable. The enhanced model VRNN-SCA’s best prediction for dataset 41010 was in epoch 40 and is shown in Figure 3a. It is more accurate than original model (VRNN) prediction illustrated in Figure 3b. Similarly, Figure 4 shows the comparison over dataset 41010 in terms of prediction between the enhanced LSTM-SCA and original LSTM. As can be seen in Figure 4a, LSTM-SCA in epoch 40 produces the best results for predicating the wave heights. The results outperform and is more precise than that of original LSTM which is shown in Figure 4b. The performance of GRU-SCA is compared with the original GRU in terms of prediction as shown in Figure 5. As can be seen form Figure 5a, the best results among all models over dataset 41010 was for GRU-SCA with the value of 8.54 × 10 5 , which clearly outperform the original GRU shown in Figure 5b.
Figure 6 shows the comparison between the proposed VRNN-SCA and the original VRNN models in term of prediction for dataset form station 41040. Figure 6a shows that the prediction of VRNN-SCA outperforms the original VRNN shown in Figure 6b. This is because the SCA algorithm is effective at producing better prediction accuracy compared to the original one. Similarly, Figure 7 shows the comparison between the proposed model LSTM-SCA and the original LSTM; the proposed model (Figure 7a) clearly outperforms the original one (Figure 7b) in terms of accurately predicating the wave heights. The performance of GRU-SCA is compared with the original GRU in terms of prediction as shown in Figure 8. As can be seen form Figure 8a, the best results among all models over dataset 41040 was for GRU-SCA, which clearly outperform the original GRU shown in Figure 8b.
In Figure 9, there is the comparison between VRNN-SCA model and VRNN in terms of best prediction on dataset 41043. Figure 9a demonstrates the best prediction of the enhanced model VRNN-SCA, while Figure 9b explains the prediction of the original VRNN. The performance of LSTM-SCA on dataset 41043 is demonstrated in Figure 10a, similarly the performance of original LSTM is depicted on Figure 10b. Figure 11a demonstrates the prediction of the enhanced model GRU-SCA, whereas Figure 11b model’s prediction.
Figure 12 shows the comparison between the proposed VRNN-SCA and the original VRNN models in term of prediction for the dataset from station 41060. Figure 12a shows the prediction of VRNN-SCA outperforms the original VRNN showed in Figure 12b. Similarly, Figure 13 shows the comparison between the proposed model LSTM-SCA and the original LSTM; the proposed model (Figure 13a) clearly outperform the original one (Figure 13b) in terms of accurately predicate the wave heights. The performance of GRU-SCA is compared with the original GRU in terms of prediction as shown in Figure 14. As can be seen form Figure 14a, the best results among all models over dataset 41043 was for GRU-SCA, which clearly outperforms the original GRU shown in Figure 14b. This is due to the effective of the SCA algorithm in producing better prediction accuracy comparing to the original one.
The graphs give a clear point of view that experiments of the proposed models for all four datasets compared to the original models have more accurate prediction with minimum loss values. The datasets have different sizes and different variations of numbers in each year. The metrics used to evaluate the performance and prediction accuracy are MSE, RMSE, and MAE. As the experiments have been run in the interval of 10 epochs, we notice that every stop point has different values. These values improve as the epochs number increases. The best values for every metric in the table is highlighted in bold for each dataset.

7.4. Significance Analysis

The one-way analysis of variance (ANOVA) test was used to assess the statistical significance of the differences between the resulting MSE obtained by proposed models versus other models. The findings of this analysis indicate whether the findings of the experiments are independent. No significance difference between the MSE of the proposed models and other models is assumed by the null hypothesis. The null hypothesis is accepted at state level greater than 0.05 and rejected at state level less than 0.05. ANOVA is an effective analysis technique as it accepts more than two groups to find the significance differences, and because we have six groups, the ANOVA test is selected. The procedures of this analysis are adopted from in [80].
In dataset 41010, the obtained p-value, as can be seen in Table 14, is 0.000003, which is less than 0.05, and we can thus reject the null hypothesis and indicate there is a significant difference between the proposed models and the original models. Figure 15 shows the boxplot of the differences between the proposed models and benchmarking models.
Similarly, in dataset 41040, the obtained p-value, as can be seen from Table 15, is 0.000001, which strongly indicates that there is a significance difference between the proposed models and original models. Figure 16 illustrates the boxplot of the differences between the proposed models and original models on this dataset.
For dataset 41043, there is a slight difference, and, as can be seen from Table 16, the obtained p-values is 0.0183, which still less than state level 0.05. Figure 17 demonstrates these difference as a boxplot.
Finally, in dataset 41060, the obtained p-value, as shown in Table 17, is 0.0006, which demonstrates a strong indication of the superiority of the proposed models. Figure 18 illustrates the boxplot differences on this dataset.

8. Conclusions and Future Work

This paper proposed an enhanced weight-optimized recurrent neural network based on the sine cosine algorithm for predicting with high accuracy the wave heights. The proposed models’ structures were first configured with optimal hyperparameters using grid search technique. The grid search is used to find the best values for learning rate. Three models are proposed, namely VRNN-SCA, LSTM-SCA, and GRU-SCA, by integrating the sine cosine algorithm. The results proved that the proposed models have the capability of improving waves prediction and producing better results than original models. The results of the proposed models demonstrate much better results comparing the original ones, for example, the best averages MSE on 41010 datasets were 0.0003, 0.0005, and 0.0008 for GRU-SCA, LSTM-SCA, and VRNN-SCA, respectively, whereas the best average RMSE was 0.0165, 0.0205, and 0.0269 for GRU-SCA, LSTM-SCA, and VRNN-SCA, respectively. Similarly, the best average MAE for GRU-SCA, LSTM-SCA, and VRNN-SCA was 0.0113, 0.0130, and 0.0133, respectively.
The integration of SCA has helped the simple architectures of RNNs to generate weights that are adaptable to the selected data set and models structures. In traditional training of RNN, the initialization of weights happens randomly, ignoring the datasets size. This increases the possibility of the model vanishing, exploding, or becoming trapped in local optima. Therefore, our technique utilized the advantages of SCA by generating adaptive weight values that can adapted with the model’s parameters and the datasets simultaneously. The proposed VRNN-SCA, LSTM-SCA, and GRU-SCA models are effective tools in forecasting wave height and can be recommended to solve other prediction problems. In future work, according to the “No Free Lunch” theorem, other optimization algorithms such as gray wolf optimizer or dragonfly algorithm could be investigated to optimize the weight of recurrent neural networks. Besides, the proposed models could be investigated on other domains such as forecasting air pollution, flood prediction, and wind speed forecasting. Another future direction is to use different evaluation indices to validate the performance of models such as Moving Average, Weighted MA, or Exponential smoothing.

Author Contributions

Conceptualization, A.A.; Methodology, A.A., S.J.A., Q.A.-T.; Software, A.A., Q.A.-T. and M.G.R.; Validation, A.A., S.J.A., Q.A.-T.; Formal analysis, A.A., H.M.R. and H.A.; Investigation, A.A., S.J.A. and Q.A.-T.; Resources, A.A., H.A. and Q.A.-T.; Data curation, A.A. and Q.A.-T.; Writing–original draft preparation, A.A. and Q.A.-T.; Writing–review and editing, S.J.A., H.M.R., H.A., Q.A.-T., and M.G.R.; Visualization, H.A., H.M.R.; Supervision, S.J.A.; Project administration, S.J.A.; Funding acquisition, S.J.A. All authors have read and agreed to the published version of the manuscript.

Funding

Research reported in this publication was supported by Fundamental Research Grant Project (FRGS) from the Ministry of Education Malaysia (FRGS/1/2018/ICT02/UTP/03/1) under UTP grant number 015MA0-013.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

We are grateful to the Editor and two anonymous reviewers for their valuable suggestions and comments, which significantly improved the quality of the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fan, S.; Xiao, N.; Dong, S. A novel model to predict significant wave height based on long short-term memory network. Ocean Eng. 2020, 205, 107298. [Google Scholar] [CrossRef]
  2. Jain, P.; Deo, M.; Latha, G.; Rajendran, V. Real time wave forecasting using wind time history and numerical model. Ocean Model. 2011, 36, 26–39. [Google Scholar] [CrossRef]
  3. Aderinto, T.; Li, H. Ocean wave energy converters: Status and challenges. Energies 2018, 11, 1250. [Google Scholar] [CrossRef] [Green Version]
  4. Abdulkadir, S.J.; Yong, S.P. Unscented kalman filter for noisy multivariate financial time-series data. In International Workshop on Multi-Disciplinary Trends in Artificial Intelligence; Springer: Berlin/Heidelberg, Germany, 2013; pp. 87–96. [Google Scholar]
  5. Aisjah, A.S.; Arifin, S.; Danistha, W.L. Sverdruv Munk Bretschneider Modification (SMB) for Significant Wave Height Prediction in Java Sea. Curr. J. Appl. Sci. Technol. 2016, 1–8. [Google Scholar] [CrossRef]
  6. McCormick, M. Wind-wave power available to a wave energy converter array. Ocean Eng. 1978, 5, 67–74. [Google Scholar] [CrossRef]
  7. Wang, W.; Tang, R.; Li, C.; Liu, P.; Luo, L. A BP neural network model optimized by Mind Evolutionary Algorithm for predicting the ocean wave heights. Ocean Eng. 2018, 162, 98–107. [Google Scholar] [CrossRef]
  8. Abdulkadir, S.J.; Yong, S.P. Scaled UKF–NARX hybrid model for multi-step-ahead forecasting of chaotic time series data. Soft Comput. 2015, 19, 3479–3496. [Google Scholar] [CrossRef]
  9. Alqushaibi, A.; Abdulkadir, S.J.; Rais, H.M.; Al-Tashi, Q. A Review of Weight Optimization Techniques in Recurrent Neural Networks. In Proceedings of the 2020 International Conference on Computational Intelligence (ICCI), Bandar Seri Iskandar, Malaysia, 8–9 October 2020; pp. 196–201. [Google Scholar] [CrossRef]
  10. Abdulkadir, S.J.; Yong, S.P.; Zakaria, N. Hybrid neural network model for metocean data analysis. J. Inform. Math. Sci. 2016, 8, 245–251. [Google Scholar]
  11. Alqushaibi, A.; Abdulkadir, S.J.; Rais, H.M.; Al-Tashi, Q.; Ragab, M.G. An Optimized Recurrent Neural Network for Metocean Forecasting. In Proceedings of the 2020 International Conference on Computational Intelligence (ICCI), Bandar Seri Iskandar, Malaysia, 8–9 October 2020; pp. 190–195. [Google Scholar] [CrossRef]
  12. An, P.; Liu, Q.; Abedi, F.; Yang, Y. Novel calibration method for camera array in spherical arrangement. Signal Process. Image Commun. 2020, 80, 115682. [Google Scholar] [CrossRef]
  13. Pradhan, R.; Aygun, R.S.; Maskey, M.; Ramachandran, R.; Cecil, D.J. Tropical cyclone intensity estimation using a deep convolutional neural network. IEEE Trans. Image Process. 2017, 27, 692–702. [Google Scholar] [CrossRef] [PubMed]
  14. Deo, M.; Naidu, C.S. Real time wave forecasting using neural networks. Ocean Eng. 1998, 26, 191–203. [Google Scholar] [CrossRef]
  15. Mandal, S.; Prabaharan, N. Ocean wave forecasting using recurrent neural networks. Ocean Eng. 2006, 33, 1401–1410. [Google Scholar] [CrossRef]
  16. Mahjoobi, J.; Mosabbeb, E.A. Prediction of significant wave height using regressive support vector machines. Ocean Eng. 2009, 36, 339–347. [Google Scholar] [CrossRef]
  17. Etemad-Shahidi, A.; Mahjoobi, J. Comparison between M5 model tree and neural networks for prediction of significant wave height in Lake Superior. Ocean Eng. 2009, 36, 1175–1181. [Google Scholar] [CrossRef]
  18. Azencott, R.; Muravina, V.; Hekmati, R.; Zhang, W.; Paldino, M. Automatic clustering in large sets of time series. In Contributions to Partial Differential Equations and Applications; Springer: Berlin/Heidelberg, Germany, 2019; pp. 65–75. [Google Scholar]
  19. Lehmann, M.; Karimpour, F.; Goudey, C.A.; Jacobson, P.T.; Alam, M.R. Ocean wave energy in the United States: Current status and future perspectives. Renew. Sustain. Energy Rev. 2017, 74, 1300–1313. [Google Scholar] [CrossRef]
  20. Pirhooshyaran, M.; Snyder, L.V. Forecasting, hindcasting and feature selection of ocean waves via recurrent and sequence-to-sequence networks. Ocean Eng. 2020, 207, 107424. [Google Scholar] [CrossRef]
  21. Tolman, H.L. A third-generation model for wind waves on slowly varying, unsteady, and inhomogeneous depths and currents. J. Phys. Oceanogr. 1991, 21, 782–797. [Google Scholar] [CrossRef]
  22. Kagemoto, H. Forecasting a water-surface wave train with artificial intelligence—A case study. Ocean Eng. 2020, 207, 107380. [Google Scholar] [CrossRef]
  23. Reikard, G.; Rogers, W.E. Forecasting ocean waves: Comparing a physics-based model with statistical models. Coast. Eng. 2011, 58, 409–416. [Google Scholar] [CrossRef]
  24. Abdulkadir, S.J.; Alhussian, H.; Nazmi, M.; Elsheikh, A.A. Long Short Term Memory Recurrent Network for Standard and Poor’s 500 Index Modelling. Int. J. Eng. Technol. 2018, 7, 25–29. [Google Scholar] [CrossRef]
  25. Abraham, B.; Ledolter, J. Statistical Methods for Forecasting; John Wiley & Sons: Hoboken, NJ, USA, 2009; Volume 234. [Google Scholar]
  26. Abdulkadir, S.J.; Yong, S.P.; Marimuthu, M.; Lai, F.W. Hybridization of ensemble Kalman filter and non-linear auto-regressive neural network for financial forecasting. In Mining Intelligence and Knowledge Exploration; Springer: Berlin/Heidelberg, Germany, 2014; pp. 72–81. [Google Scholar]
  27. Rasp, S.; Lerch, S. Neural networks for postprocessing ensemble weather forecasts. Mon. Weather Rev. 2018, 146, 3885–3900. [Google Scholar] [CrossRef] [Green Version]
  28. Campos, R.M.; Krasnopolsky, V.; Alves, J.H.; Penny, S.G. Improving NCEP’s global-scale wave ensemble averages using neural networks. Ocean Model. 2020, 149, 101617. [Google Scholar] [CrossRef]
  29. Harpham, Q.; Tozer, N.; Cleverley, P.; Wyncoll, D.; Cresswell, D. A Bayesian method for improving probabilistic wave forecasts by weighting ensemble members. Environ. Model. Softw. 2016, 84, 482–493. [Google Scholar] [CrossRef]
  30. Campos, R.M.; Krasnopolsky, V.; Alves, J.H.G.; Penny, S.G. Nonlinear wave ensemble averaging in the Gulf of Mexico using neural networks. J. Atmos. Ocean. Technol. 2019, 36, 113–127. [Google Scholar] [CrossRef]
  31. Durrant, T.H.; Woodcock, F.; Greenslade, D.J. Consensus forecasts of modeled wave parameters. Weather Forecast. 2009, 24, 492–503. [Google Scholar] [CrossRef]
  32. Özger, M. Prediction of ocean wave energy from meteorological variables by fuzzy logic modeling. Expert Syst. Appl. 2011, 38, 6269–6274. [Google Scholar] [CrossRef]
  33. Zhang, Y.; Liu, R.; Wang, X.; Chen, H.; Li, C. Boosted binary Harris hawks optimizer and feature selection. Eng. Comput. 2020, 1–30. [Google Scholar] [CrossRef]
  34. Al-Tashi, Q.; Rais, H.M.; Abdulkadir, S.J.; Mirjalili, S.; Alhussian, H. A review of grey wolf optimizer-based feature selection methods for classification. In Evolutionary Machine Learning Techniques; Springer: Berlin/Heidelberg, Germany, 2020; pp. 273–286. [Google Scholar]
  35. Al-Wajih, R.; Abdulkadir, S.J.; Aziz, N.; Al-Tashi, Q.; Talpur, N. Hybrid Binary Grey Wolf With Harris Hawks Optimizer for Feature Selection. IEEE Access 2021, 9, 31662–31677. [Google Scholar] [CrossRef]
  36. Balogun, A.O.; Basri, S.; Mahamad, S.; Abdulkadir, S.J.; Almomani, M.A.; Adeyemo, V.E.; Al-Tashi, Q.; Mojeed, H.A.; Imam, A.A.; Bajeh, A.O. Impact of feature selection methods on the predictive performance of software defect prediction models: An extensive empirical study. Symmetry 2020, 12, 1147. [Google Scholar] [CrossRef]
  37. Al-Tashi, Q.; Abdulkadir, S.J.; Rais, H.M.; Mirjalili, S.; Alhussian, H. Approaches to multi-objective feature selection: A systematic literature review. IEEE Access 2020, 8, 125076–125096. [Google Scholar] [CrossRef]
  38. Deo, M.C.; Jha, A.; Chaphekar, A.; Ravikant, K. Neural networks for wave forecasting. Ocean Eng. 2001, 28, 889–898. [Google Scholar] [CrossRef]
  39. Alexandre, E.; Cuadra, L.; Nieto-Borge, J.; Candil-Garcia, G.; Del Pino, M.; Salcedo-Sanz, S. A hybrid genetic algorithm—Extreme learning machine approach for accurate significant wave height reconstruction. Ocean Model. 2015, 92, 115–123. [Google Scholar] [CrossRef]
  40. Zhang, Y.; Liu, R.; Heidari, A.A.; Wang, X.; Chen, Y.; Wang, M.; Chen, H. Towards augmented kernel extreme learning models for bankruptcy prediction: Algorithmic behavior and comprehensive analysis. Neurocomputing 2021, 430, 185–212. [Google Scholar] [CrossRef]
  41. Zhao, D.; Liu, L.; Yu, F.; Heidari, A.A.; Wang, M.; Liang, G.; Muhammad, K.; Chen, H. Chaotic random spare ant colony optimization for multi-threshold image segmentation of 2D Kapur entropy. Knowl. Based Syst. 2020, 216, 106510. [Google Scholar] [CrossRef]
  42. Tu, J.; Chen, H.; Liu, J.; Heidari, A.A.; Zhang, X.; Wang, M.; Ruby, R.; Pham, Q.V. Evolutionary biogeography-based whale optimization methods with communication structure: Towards measuring the balance. Knowl. Based Syst. 2021, 212, 106642. [Google Scholar] [CrossRef]
  43. Wang, M.; Chen, H. Chaotic multi-swarm whale optimizer boosted support vector machine for medical diagnosis. Appl. Soft Comput. 2020, 88, 105946. [Google Scholar] [CrossRef]
  44. Shan, W.; Qiao, Z.; Heidari, A.A.; Chen, H.; Turabieh, H.; Teng, Y. Double adaptive weights for stabilization of moth flame optimizer: Balance analysis, engineering cases, and medical diagnosis. Knowl. Based Syst. 2020, 214, 106728. [Google Scholar] [CrossRef]
  45. Xu, Y.; Chen, H.; Luo, J.; Zhang, Q.; Jiao, S.; Zhang, X. Enhanced Moth-flame optimizer with mutation strategy for global optimization. Inf. Sci. 2019, 492, 181–203. [Google Scholar] [CrossRef]
  46. Wang, M.; Chen, H.; Yang, B.; Zhao, X.; Hu, L.; Cai, Z.; Huang, H.; Tong, C. Toward an optimal kernel extreme learning machine using a chaotic moth-flame optimization strategy with applications in medical diagnoses. Neurocomputing 2017, 267, 69–84. [Google Scholar] [CrossRef]
  47. Chen, H.; Heidari, A.A.; Chen, H.; Wang, M.; Pan, Z.; Gandomi, A.H. Multi-population differential evolution-assisted Harris hawks optimization: Framework and case studies. Future Gener. Comput. Syst. 2020, 111, 175–198. [Google Scholar] [CrossRef]
  48. Savitha, R.; Al Mamun, A. Regional ocean wave height prediction using sequential learning neural networks. Ocean Eng. 2017, 129, 605–612. [Google Scholar]
  49. Akbarifard, S.; Radmanesh, F. Predicting sea wave height using Symbiotic Organisms Search (SOS) algorithm. Ocean Eng. 2018, 167, 348–356. [Google Scholar] [CrossRef]
  50. Mnasri, S.; Nasri, N.; Van den Bossche, A.; Val, T. A new multi-agent particle swarm algorithm based on birds accents for the 3D indoor deployment problem. ISA Trans. 2019, 91, 262–280. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  51. Cerqueira, V.; Torgo, L.; Smailović, J.; Mozetič, I. A comparative study of performance estimation methods for time series forecasting. In Proceedings of the 2017 IEEE International Conference on Data Science and Advanced Analytics (DSAA), Tokyo, Japan, 19–21 October 2017; pp. 529–538. [Google Scholar]
  52. Bergmeir, C.; Hyndman, R.J.; Koo, B. A note on the validity of cross-validation for evaluating autoregressive time series prediction. Comput. Stat. Data Anal. 2018, 120, 70–83. [Google Scholar] [CrossRef]
  53. Rashid, T.A.; Fattah, P.; Awla, D.K. Using accuracy measure for improving the training of LSTM with metaheuristic algorithms. Procedia Comput. Sci. 2018, 140, 324–333. [Google Scholar] [CrossRef]
  54. Somu, N.; MR, G.R.; Ramamritham, K. A hybrid model for building energy consumption forecasting using long short term memory networks. Appl. Energy 2020, 261, 114131. [Google Scholar] [CrossRef]
  55. Rosli, S.J.; Rahim, H.A.; Abdul Rani, K.N.; Ngadiran, R.; Ahmad, R.B.; Yahaya, N.Z.; Abdulmalek, M.; Jusoh, M.; Yasin, M.N.M.; Sabapathy, T.; et al. A Hybrid Modified Method of the Sine Cosine Algorithm Using Latin Hypercube Sampling with the Cuckoo Search Algorithm for Optimization Problems. Electronics 2020, 9, 1786. [Google Scholar] [CrossRef]
  56. Al-Tashi, Q.; Abdulkadir, S.J.; Rais, H.M.; Mirjalili, S.; Alhussian, H.; Ragab, M.G.; Alqushaibi, A. Binary Multi-Objective Grey Wolf Optimizer for Feature Selection in Classification. IEEE Access 2020, 8, 106247–106263. [Google Scholar] [CrossRef]
  57. Al-Tashi, Q.; Kadir, S.J.A.; Rais, H.M.; Mirjalili, S.; Alhussian, H. Binary optimization using hybrid grey wolf optimization for feature selection. IEEE Access 2019, 7, 39496–39508. [Google Scholar] [CrossRef]
  58. Steele, K.; Teng, C.C.; Wang, D. Wave direction measurements using pitch-roll buoys. Ocean Eng. 1992, 19, 349–375. [Google Scholar] [CrossRef]
  59. Fernández, J.C.; Salcedo-Sanz, S.; Gutiérrez, P.A.; Alexandre, E.; Hervás-Martínez, C. Significant wave height and energy flux range forecast with machine learning classifiers. Eng. Appl. Artif. Intell. 2015, 43, 44–53. [Google Scholar] [CrossRef]
  60. Hashim, R.; Roy, C.; Motamedi, S.; Shamshirband, S.; Petković, D. Selection of climatic parameters affecting wave height prediction using an enhanced Takagi-Sugeno-based fuzzy methodology. Renew. Sustain. Energy Rev. 2016, 60, 246–257. [Google Scholar] [CrossRef]
  61. Salehinejad, H.; Sankar, S.; Barfett, J.; Colak, E.; Valaee, S. Recent advances in recurrent neural networks. arXiv 2017, arXiv:1801.01078. [Google Scholar]
  62. Alom, M.Z.; Taha, T.M.; Yakopcic, C.; Westberg, S.; Sidike, P.; Nasrin, M.S.; Hasan, M.; Van Essen, B.C.; Awwal, A.A.; Asari, V.K. A state-of-the-art survey on deep learning theory and architectures. Electronics 2019, 8, 292. [Google Scholar] [CrossRef] [Green Version]
  63. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  64. Lipton, Z.C.; Berkowitz, J.; Elkan, C. A critical review of recurrent neural networks for sequence learning. arXiv 2015, arXiv:1506.00019. [Google Scholar]
  65. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  66. Werbos, P.J. Backpropagation through time: What it does and how to do it. Proc. IEEE 1990, 78, 1550–1560. [Google Scholar] [CrossRef] [Green Version]
  67. Abdulkadir, S.J.; Alhussian, H.; Alzahrani, A.I. Analysis of recurrent neural networks for henon simulated time-series forecasting. J. Telecommun. Electron. Comput. Eng. 2018, 10, 155–159. [Google Scholar]
  68. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  69. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  70. Mirjalili, S. SCA: A sine cosine algorithm for solving optimization problems. Knowl. Based Syst. 2016, 96, 120–133. [Google Scholar] [CrossRef]
  71. Attia, A.F.; El Sehiemy, R.A.; Hasanien, H.M. Optimal power flow solution in power systems using a novel Sine-Cosine algorithm. Int. J. Electr. Power Energy Syst. 2018, 99, 331–343. [Google Scholar] [CrossRef]
  72. Cherkassky, V.; Ma, Y. Selection of meta-parameters for support vector regression. In International Conference on Artificial Neural Networks; Springer: Berlin/Heidelberg, Germany, 2002; pp. 687–693. [Google Scholar]
  73. Hsu, C.W.; Chang, C.C.; Lin, C.J. A practical guide to support vector classification. Precis. Agric. 2003. [Google Scholar] [CrossRef]
  74. Feng, G.H. Support Vector Machine parameter selection method. Comput. Eng. Ring Appl. 2011, 47, 123. [Google Scholar]
  75. Hinz, T.; Navarro-Guerrero, N.; Magg, S.; Wermter, S. Speeding up the hyperparameter optimization of deep convolutional neural networks. Int. J. Comput. Intell. Appl. 2018, 17, 1850008. [Google Scholar] [CrossRef] [Green Version]
  76. Shuai, Y.; Zheng, Y.; Huang, H. Hybrid Software Obsolescence Evaluation Model Based on PCA-SVM-GridSearchCV. In Proceedings of the 2018 IEEE 9th International Conference on Software Engineering and Service Science (ICSESS), Beijing, China, 23–25 November 2018; pp. 449–453. [Google Scholar]
  77. Ragab, M.G.; Abdulkadir, S.J.; Aziz, N.; Al-Tashi, Q.; Alyousifi, Y.; Alhussian, H.; Alqushaibi, A. A Novel One-Dimensional CNN with Exponential Adaptive Gradients for Air Pollution Index Prediction. Sustainability 2020, 12, 10090. [Google Scholar] [CrossRef]
  78. Ragab, M.G.; Abdulkadir, S.J.; Aziz, N. Random Search One Dimensional CNN for Human Activity Recognition. In Proceedings of the 2020 International Conference on Computational Intelligence (ICCI), Bandar Seri Iskandar, Malaysia, 8–9 October 2020; pp. 86–91. [Google Scholar]
  79. Cao, W.; Wang, X.; Ming, Z.; Gao, J. A review on neural networks with random weights. Neurocomputing 2018, 275, 278–287. [Google Scholar] [CrossRef]
  80. Al-Tashi, Q.; Rais, H.M.; Abdulkadir, S.J.; Mirjalili, S. Feature Selection Based on Grey Wolf Optimizer for Oil & Gas Reservoir Classification. In Proceedings of the 2020 International Conference on Computational Intelligence (ICCI), Bandar Seri Iskandar, Malaysia, 8–9 October 2020; pp. 211–216. [Google Scholar]
Figure 1. Stations’ location in the map.
Figure 1. Stations’ location in the map.
Jmse 09 00524 g001
Figure 2. General model of the proposed RNN-SCA method.
Figure 2. General model of the proposed RNN-SCA method.
Jmse 09 00524 g002
Figure 3. Comparison of VRNN models on dataset 41010.
Figure 3. Comparison of VRNN models on dataset 41010.
Jmse 09 00524 g003
Figure 4. Comparison of LSTM models on dataset 41010.
Figure 4. Comparison of LSTM models on dataset 41010.
Jmse 09 00524 g004
Figure 5. Comparison of GRU models on dataset 41010.
Figure 5. Comparison of GRU models on dataset 41010.
Jmse 09 00524 g005
Figure 6. Comparsion of VRNN models on dataset 41040.
Figure 6. Comparsion of VRNN models on dataset 41040.
Jmse 09 00524 g006
Figure 7. Comparison of LSTM models on dataset 41040.
Figure 7. Comparison of LSTM models on dataset 41040.
Jmse 09 00524 g007
Figure 8. Comparison of GRU models on dataset 41040.
Figure 8. Comparison of GRU models on dataset 41040.
Jmse 09 00524 g008
Figure 9. Comparison of VRNN models on dataset 41043.
Figure 9. Comparison of VRNN models on dataset 41043.
Jmse 09 00524 g009
Figure 10. Comparison of LSTM models on dataset 41043.
Figure 10. Comparison of LSTM models on dataset 41043.
Jmse 09 00524 g010
Figure 11. Comparison of GRU models on dataset 41043.
Figure 11. Comparison of GRU models on dataset 41043.
Jmse 09 00524 g011
Figure 12. Comparison of VRNN models on dataset 41060.
Figure 12. Comparison of VRNN models on dataset 41060.
Jmse 09 00524 g012
Figure 13. Comparison of LSTM models on dataset 41060.
Figure 13. Comparison of LSTM models on dataset 41060.
Jmse 09 00524 g013
Figure 14. Comparison of GRU models on dataset 41060.
Figure 14. Comparison of GRU models on dataset 41060.
Jmse 09 00524 g014
Figure 15. Boxplot differences on the 41010 dataset.
Figure 15. Boxplot differences on the 41010 dataset.
Jmse 09 00524 g015
Figure 16. Boxplot differences on the 41040 dataset.
Figure 16. Boxplot differences on the 41040 dataset.
Jmse 09 00524 g016
Figure 17. Boxplot differences on the 41043 dataset.
Figure 17. Boxplot differences on the 41043 dataset.
Jmse 09 00524 g017
Figure 18. Boxplot differences on the 41060 dataset.
Figure 18. Boxplot differences on the 41060 dataset.
Jmse 09 00524 g018
Table 1. Stations’ dataset.
Table 1. Stations’ dataset.
Station IDLon (W)Lat (N)Water Depth (M)SWH (M)Range Years (Interval)Training PointsTest Points
4101028.87878.485890[0.00, 4.6][2005,2017]8635637011
4104053.04514.5545112[0.00, 8.09][2005, 2017]5479418617
4104364.83021.1245271[0.00, 13.42][2007, 2017]11499049282
4106051.01714.8245021[0.72, 4.89][2012, 2017]2895212409
Table 2. Grid search space parameters.
Table 2. Grid search space parameters.
Grid Search ParametersInitialization Values
learn_rate[0.001, 0.01, 0.1, 0.2, 0.3]
dropout_rate[0.0, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9]
batch_size[32, 50, 64]
momentum[0.0, 0.2, 0.4, 0.6, 0.8, 0.9]
Table 3. Sine Cosine Algorithm (SCA) setup.
Table 3. Sine Cosine Algorithm (SCA) setup.
SCA ParametersInitialization Values
SCA Upper Bound    5
SCA Lower Bound  −5
SCA Search Agents  25
SCA Max Iteration200
Table 4. Parameter settings and model configuration.
Table 4. Parameter settings and model configuration.
Models ParametersVRNN-SCALSTM-SCAGRU-SCA
No of iterations[10–100][10–100][10–100]
Input Features121212
Hidden units323232
Activation FunctionReluReluRelu
Training Size70%70%70%
Testing Size30%30%30%
Table 5. Results of the proposed VRNN-SCA model in all datasets.
Table 5. Results of the proposed VRNN-SCA model in all datasets.
Epoch41010410404104341060
MSERMSEMAEMSERMSEMAEMSERMSEMAEMSERMSEMAE
100.00030.01850.01310.00610.07790.04430.00290.05430.02320.00100.03180.0238
200.00040.01920.01470.00310.05540.03900.00790.08910.03290.00100.03210.0243
300.00030.01800.01370.00220.04730.02800.01200.10960.03990.00110.03260.0249
400.00020.01520.01100.00160.03960.02300.01670.12920.04530.00100.03170.0237
500.00030.01780.01300.00200.04470.03140.03220.17940.05730.00100.03160.0235
600.00090.02990.01350.00270.05190.04270.04430.21040.06250.00100.03150.0235
700.00100.03150.01490.00410.06400.05610.06080.24660.07020.00100.03150.0234
800.00140.03790.01300.00360.06000.05460.06880.26240.07440.00100.03140.0234
900.00170.04120.01290.00530.07290.06560.08670.29450.08840.00100.03130.0234
1000.00160.03980.01300.00590.07680.06980.06100.24700.06680.00100.03130.0234
Table 6. Results of the proposed LSTM-SCA model in all datasets.
Table 6. Results of the proposed LSTM-SCA model in all datasets.
Epoch41010410404104341060
MSERMSEMAEMSERMSEMAEMSERMSEMAEMSERMSEMAE
100.00010.01170.00810.00240.04940.02690.00190.04350.01680.00110.03240.0243
200.00010.01110.00790.00140.03710.02100.01340.11580.03890.00110.03240.0246
300.00010.01060.00650.00200.04440.02930.03250.18030.05770.00100.03210.0242
400.00020.01230.00690.00220.04720.03740.04840.22000.06430.00100.03230.0242
500.00060.02370.01100.00390.06280.05290.08780.29640.08570.00100.03220.0241
600.00040.02090.01470.00790.08900.07060.13440.36660.10120.00100.03220.0240
700.00050.02130.01640.01180.10850.08530.19860.44560.11440.00100.03220.0239
800.00050.02180.01660.01290.11340.08880.30380.55120.14390.00100.03220.0240
900.00070.02650.01820.01230.11080.08660.37140.60940.14630.00100.03210.0239
1000.00210.04540.02240.01440.12000.08970.45760.67640.16610.00100.03200.0239
Table 7. Results of the proposed GRU-SCA model in all datasets.
Table 7. Results of the proposed GRU-SCA model in all datasets.
Epoch41010410404104341060
MSERMSEMAEMSERMSEMAEMSERMSEMAEMSERMSEMAE
100.00020.01300.00840.00230.04750.03210.00350.05880.02310.00100.03230.0248
200.00010.01130.00710.00080.02820.01800.01200.10950.03520.00110.03300.0253
300.00010.00980.00630.00310.05580.05070.02120.14560.04270.00110.03260.0249
400.00010.00920.00630.00930.09640.09060.02940.17160.04900.00100.03200.0240
500.00010.01130.00820.01650.12860.12250.03640.19080.05590.00100.03190.0238
600.00020.01370.01020.02640.16260.15330.06050.24590.07560.00100.03190.0238
700.00050.02270.01710.03250.18030.17060.09240.30400.09240.00100.03200.0239
800.00070.02610.01930.03610.19010.18060.11570.34010.10240.00100.03200.0239
900.00060.02380.01630.03810.19520.18510.12590.35480.10640.00100.03190.0239
1000.00060.02380.01380.03890.19730.18710.12770.35740.10680.00100.03170.0238
Table 8. Results of the original VRNN model in all datasets.
Table 8. Results of the original VRNN model in all datasets.
Epoch41010410404104341060
MSERMSEMAEMSERMSEMAEMSERMSEMAEMSERMSEMAE
100.19520.44180.43360.01890.13740.10210.45430.67400.20820.02560.15990.1143
200.57500.75830.75260.01890.13740.10200.11210.33490.14860.02220.14890.1101
300.14440.37990.37080.01860.13630.10140.78710.88720.31770.02210.14860.1101
400.04580.21410.19890.01740.13180.09890.01120.10570.03970.02190.14800.1100
500.08540.29230.21640.01820.13470.10050.02270.15070.05310.02190.14800.1100
600.93930.96920.92140.01910.13810.10250.03980.19950.06680.02220.14890.1101
700.87050.93300.81640.01850.13590.10110.04800.21900.06910.02240.14960.1102
800.95740.97850.87240.01680.12960.09800.06940.26340.07960.02190.14790.1100
901.50221.22561.17720.01730.13150.09880.07650.27660.08180.02190.14790.1100
1000.68310.82650.79060.01900.13790.10230.08990.29980.08750.02190.14800.1100
Table 9. Results of the original long short-term memory (LSTM) model in all datasets.
Table 9. Results of the original long short-term memory (LSTM) model in all datasets.
Epoch41010410404104341060
MSERMSEMAEMSERMSEMAEMSERMSEMAEMSERMSEMAE
100.62450.79020.78580.44220.66500.64030.01070.10370.07180.02160.14700.1068
200.66540.81570.81260.37490.61230.56790.01050.10230.07350.01550.12430.0968
300.00670.08200.05360.07320.27060.20760.10790.32850.31730.01740.13190.1017
400.12450.35290.32710.39490.62840.61430.00840.09150.07350.14810.38480.3646
500.32130.56690.56410.11360.33710.31930.01030.10150.08470.01950.13970.1067
600.31380.56020.55610.06100.24690.21460.00550.07400.05880.01580.12560.0934
700.53170.72920.72660.03780.19430.15270.01820.13490.11480.02140.14640.1156
800.65410.80870.80320.06920.26300.22710.02650.16270.09260.01760.13270.1010
900.71020.84270.83780.10750.32790.29700.00550.07400.05720.01510.12280.0919
1000.35990.59990.59420.05700.23880.20370.00610.07790.06370.01940.13940.1058
Table 10. Results of the original GTU model in all datasets.
Table 10. Results of the original GTU model in all datasets.
Epoch41010410404104341060
MSERMSEMAEMSERMSEMAEMSERMSEMAEMSERMSEMAE
100.26780.51750.51110.00230.04750.03210.03040.17440.16280.09900.31470.2932
200.78780.88760.88300.00080.02820.01800.13780.37120.35890.02170.14730.1144
301.31721.14771.14580.00310.05580.05070.01150.10730.08870.01820.13480.1031
400.58310.76360.75780.00930.09640.09060.00460.06750.04760.01630.12760.0966
502.82681.68131.68000.01650.12860.12250.02080.14430.13120.01930.13890.1052
602.49971.58101.57800.02640.16260.15330.01610.12690.10690.01720.13130.1010
703.96381.99091.63400.03250.18030.17060.00530.07290.05380.01650.12840.0989
801.07731.03801.03370.03610.19010.18060.02370.15390.14000.01810.13440.1023
900.26980.51940.51160.03810.19520.18510.00510.07170.05220.01950.13970.1050
1000.12930.35950.35010.03890.19730.18710.00650.08060.06000.01560.12500.0931
Table 11. Comparison in terms of average mean squared error (MSE) on all dataset.
Table 11. Comparison in terms of average mean squared error (MSE) on all dataset.
DatasetVRNN-SCALSTM-SCAGRU-SCAVRNNLSTMGRU
410100.00080.00050.00030.59980.43121.3723
410400.00370.00710.02040.01820.17310.0204
410430.03930.16500.06250.17110.02090.0262
410600.00100.00100.00100.02240.03110.0261
Table 12. Comparison in terms of average root mean square error (RMSE) on all dataset.
Table 12. Comparison in terms of average root mean square error (RMSE) on all dataset.
DatasetVRNN-SCALSTM-SCAGRU-SCAVRNNLSTMGRU
410100.02690.02050.01650.70190.61481.0487
410400.05910.07830.12820.13510.37840.1282
410430.18220.35050.22790.34110.12510.1371
410600.03170.03220.03210.14960.15940.1522
Table 13. Comparison in terms of average mean absolute error (MAE) on all dataset.
Table 13. Comparison in terms of average mean absolute error (MAE) on all dataset.
DatasetVRNN-SCALSTM-SCAGRU-SCAVRNNLSTMGRU
410100.01330.01300.01130.65500.60611.0085
410400.04550.05890.11910.10080.34440.1191
410430.05610.09350.06900.11520.10080.1202
410600.02370.02410.02420.11050.12840.1213
Table 14. ANOVA table results on dataset 41010.
Table 14. ANOVA table results on dataset 41010.
SourceSSdfMSFp-Value
Groups14.648852.929758.910.000003
Error17.7537540.32877
Total32.402559
Table 15. ANOVA table results on the dataset 41040.
Table 15. ANOVA table results on the dataset 41040.
SourceSSdfMSFp-Value
Groups0.2136450.042738.910.0000001
Error0.24003540.00445
Total0.4536859
Table 16. ANOVA table results on dataset 41043.
Table 16. ANOVA table results on dataset 41043.
SourceSSdfMSFp-Value
Groups0.2386150.0477230.0183
Error0.85785540.01589
Total1.0964659
Table 17. ANOVA table results on dataset 41060.
Table 17. ANOVA table results on dataset 41060.
SourceSSdfMSFp-Value
Groups0.0101750.002035.180.0006
Error0.02118540.00039
Total0.0313559
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Alqushaibi, A.; Abdulkadir, S.J.; Rais, H.M.; Al-Tashi, Q.; Ragab, M.G.; Alhussian, H. Enhanced Weight-Optimized Recurrent Neural Networks Based on Sine Cosine Algorithm for Wave Height Prediction. J. Mar. Sci. Eng. 2021, 9, 524. https://doi.org/10.3390/jmse9050524

AMA Style

Alqushaibi A, Abdulkadir SJ, Rais HM, Al-Tashi Q, Ragab MG, Alhussian H. Enhanced Weight-Optimized Recurrent Neural Networks Based on Sine Cosine Algorithm for Wave Height Prediction. Journal of Marine Science and Engineering. 2021; 9(5):524. https://doi.org/10.3390/jmse9050524

Chicago/Turabian Style

Alqushaibi, Alawi, Said Jadid Abdulkadir, Helmi Md Rais, Qasem Al-Tashi, Mohammed G. Ragab, and Hitham Alhussian. 2021. "Enhanced Weight-Optimized Recurrent Neural Networks Based on Sine Cosine Algorithm for Wave Height Prediction" Journal of Marine Science and Engineering 9, no. 5: 524. https://doi.org/10.3390/jmse9050524

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop