Next Article in Journal
Full-Scale Numerical Simulation of a Free-Running Cruise Ship in Heavy Head Sea Conditions
Previous Article in Journal
Classification of Underwater Sediments in Lab Based on LiDAR Full-Waveform Data
Previous Article in Special Issue
Experimental and Numerical Study of Suspended Inter-Array Cable Configurations for Floating Offshore Wind Farm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Maximum Individual Wave Height Prediction Using Different Machine Learning Techniques with Data Collected from a Buoy Located in Bilbao (Bay of Biscay)

by
Lucia Porlan-Ferrando
1,
J. David Nuñez-Gonzalez
1,
Alain Ulazia Manterola
2,*,
Nahia Martinez-Iturricastillo
3 and
John V. Ringwood
3
1
Applied Mathematics Department, University of the Basque Country (UPV/EHU), Otaola 29, 20600 Eibar, Spain
2
Energy Engineering Department, University of the Basque Country (UPV/EHU), Otaola 29, 20600 Eibar, Spain
3
Centre for Ocean Energy Research, Maynooth University, W23 F2H6 Maynooth, Ireland
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2025, 13(4), 625; https://doi.org/10.3390/jmse13040625
Submission received: 18 February 2025 / Revised: 17 March 2025 / Accepted: 19 March 2025 / Published: 21 March 2025
(This article belongs to the Special Issue New Era in Offshore Wind Energy)

Abstract

:
Accurate prediction of extreme waves, particularly the maximum wave height and the ratio between the maximum and significant wave heights of individual waves, is crucial for maritime safety and the resilience of offshore infrastructure. This study employs machine learning (ML) techniques such as linear regression modeling (LM), support vector regression (SVR), long short-term memory (LSTM), and gated recurrent units (GRU) to develop predictive models based on historical data (1990–2024) obtained from a buoy at a specific oceanic location. The results show that the SVR model provides the highest accuracy in predicting the maximum wave height ( H m a x ), achieving a coefficient of determination ( R 2 ) of 0.9006 and mean squared error (MSE) of 0.0185. For estimation of the ratio between maximum and significant wave heights ( H m a x / H s ), the SVR and LM models exhibit comparable performance, with MSE values of 0.0229. These findings have significant implications for improving early warning systems, optimizing the structural design of offshore infrastructure, and enhancing the efficiency of energy extraction under changing climate conditions.

1. Introduction

Climate change, primarily driven by anthropogenic activities, is inducing rapid and significant alterations in global temperature and other climatic variables, with potentially catastrophic consequences. The predominant driver of this phenomenon is the emission of greenhouse gases resulting from the combustion of fossil fuels. However, the integration of renewable energy sources, advances in energy efficiency, electrification, demand-side management, and the deployment of intelligent solutions present viable pathways to implementing transformative reforms in electrical and transportation systems. These interventions not only mitigate the impacts of climate change but also promote job creation and reduce energy costs [1].
Renewable energy sources are characterized by their cleanliness and inexhaustibility, and constitute a competitive alternative to fossil fuels because of their diversity, abundance, and global availability. These energy sources do not produce greenhouse gas emissions or pollutants during operation [2]. Among renewable energy sources, oceans represent a substantial energy reservoir, offering opportunities through wave energy, tidal forces, ocean currents, and thermal gradients between surface and deep waters 1. Covering approximately 70% of the Earth’s surface, and with 97% of this area represented by seas and oceans, marine energy demonstrates several advantages, including low environmental impact, high predictability, and continuous energy generation, making it a compelling option compared to intermittent sources such as solar and wind power [3]. Wave energy in particular exhibits high energy density and reliability, especially in coastal regions, and holds the potential to address a significant portion of global energy demand. However, climate change may alter oceanographic conditions such as wave height and wind speed, potentially affecting the viability of marine energy generation [4]. Moreover, the influence of climate change in the past and future is globally affecting the long-term behavior of wave loads and energy, as has been shown by several previous authors [5,6,7,8]. This lends more importance to the study of extreme oceanic events in terms of maximum individual wave height.
Accurate prediction of extreme wave events, specifically the maximum wave height ( H m a x ) and ratio of maximum height to significant height ( H m a x / H s ), is critical to ensuring the safety and sustainability of essential maritime operations such as navigation, positioning of offshore platforms, and the design of wave energy converters [9]. The significant wave height ( H s ), defined as the average height of the highest third of the waves recorded during a given period, is widely used in oceanography to describe general sea conditions. However, in scenarios that involve extreme wave activity, H m a x and its ratio to H s are more relevant metrics. For wave height distributions adhering to a Rayleigh distribution, theoretical models indicate that the maximum wave height can reach approximately twice the significant wave height. Consequently, reliable forecasting of these extreme events is vital not only to ensuring maritime safety but also to optimizing the design and placement of offshore infrastructure, reducing risks to personnel, and enhancing navigation route planning [10].
Two principal approaches are used for predicting meteorological and geographic parameters such as temperature, wind speed, and humidity. The first approach relies on physical models that use numerical simulations based on geographic and meteorological input, while the second employs machine learning (ML) methodologies, which can enable more flexible and long-term predictive capabilities [11]. However, traditional ML models exhibit limitations in capturing large-scale extreme phenomena, underscoring the need for development of robust ML-based techniques to enhance the precision of maximum wave height estimation [12]. The integration of meteorological and oceanographic data with ML approaches has demonstrated significant improvements in the accuracy of extreme wave event predictions [13].
This research focuses on estimation of the maximum wave height H m a x and H m a x / H s ratio using observational data collected from a buoy located in Bilbao. By employing machine learning algorithms, including linear regression (LM), support vector regression (SVR), long short-term memory (LSTM), and gated recurrent units (GRU), this study seeks to identify the most suitable methodology for modeling these specific phenomena and improving predictive accuracy on a global scale. We hypothesize that the SVR model will provide the highest accuracy in predicting H m a x , while the LM and SVR algorithms will yield similar results when estimating the H m a x / H s ratio. These findings are expected to contribute to the design of resilient marine structures and the optimization of maritime operations.

2. Related Works

Climate change has significantly altered wave characteristics on a global scale. In [14], the authors identified increasing trends in significant wave height across various oceanic regions, attributing these changes to wind pattern variations driven by global warming. In [15], the authors projected an increase in wave energy due to ocean warming, with potential implications for coastal erosion and maritime infrastructure.
Extreme waves pose a considerable challenge in the design of marine structures. In [16], the authors analyzed impact loads from solitary waves on vertical structures, emphasizing the importance of breaking wave forces, while ref. [17] observed a 30 cm increase in the mean height of winter waves off the California coast since 1970 along with a doubling in the frequency of waves exceeding 4 m between 1996–2016 compared to 1949–1969. In [18], the authors reported significant positive trends in extreme significant wave height and wind speed, particularly in the Beaufort and East Siberian Seas, which they attributed to sea ice retreat and the expansion of open water areas. In [19], the authors utilized satellite observations to detect increasing trends in significant wave height and wind speed, particularly in the high latitudes of the Southern Hemisphere. In [20], the authors developed a theoretical framework for evaluating wave–structure interactions, enabling assessment of wave-induced loads under varying maritime conditions. Accurate prediction of maximum wave height is critical for the design of maritime infrastructure. In [21], the authors highlighted the importance of estimating the relationship between maximum and significant wave heights for establishing reliable design criteria, while ref. [22] introduced a statistical model for predicting extreme wave heights.
Recent advancements have improved wave height forecasting. In [23], the authors demonstrated that autoregressive models could predict wave heights up to 5 s ahead, enhancing the stability of marine energy structures. In [24], the authors developed a vision transformer-based model for regional significant wave height prediction, leveraging attention mechanisms to capture wind–wave relationships while preserving positional information via transposed convolutions. Sargol et al. applied adaptive neuro-fuzzy inference systems (ANFIS) and support vector regression (SVR) to forecast seasonal maximum wave heights using buoy data from the southern Baltic Sea. In [12,25], the authors analyzed critical ocean wave parameters such as the maximum crest height, crest-to-trough height, and envelope height, validating their estimates with observational data from the North Pacific.
Reguero et al. [9] investigated the impact of climate change on the variability of the H m a x / H s ratio, a key parameter in marine structure design and risk assessment. Their findings indicated significant regional variations in this ratio driven by shifts in wind intensity and direction, which in turn could affect the prediction of extreme wave events. An increase in H m a x / H s suggests an elevated risk of extreme waves in specific regions, potentially compromising the safety of maritime infrastructure and vessels. Their study highlights the necessity of integrating updated climate projections into wave forecasting models and coastal engineering practices to enhance resilience against increasingly variable and intense oceanic conditions.
In conclusion, climate change is intensifying wave conditions worldwide, posing significant challenges for coastal and offshore infrastructure. Advances in maximum wave height prediction and better understanding of the H m a x / H s ratio are crucial for improving the resilience of marine structures in response to an increasingly extreme ocean climate.
While previous studies have used certain machine learning techniques, they have either not compared a comprehensive range of models or not applied these techniques to long-term datasets. The goal of this paper is to cover this gap, providing a predictive experimental study covering different phases and techniques of machine learning that can help to predict maximum wave heights in the future from time series data.
The rest of this paper is structured as follows: Section 3 presents the experimental data, including characteristics and preprocessing; Section 4 describes the methodology; Section 5 shows the experimental pipeline; Section 6 presents the results; and Section 7 concludes the paper.

3. Data

3.1. Location and Attributes

The dataset used in this paper was acquired from the Bilbao buoy, provided by Puertos del Estado 2. The geographical location of the buoy is indicated in Figure 1, represented by a red marker. The primary technical specifications of the buoy are summarized in Table 1.
The dataset from Puertos del Estado comprised 17 attributes, categorized as temporal data, wave characteristics, and meteorological-oceanographic data, as detailed below:
  •  Time:
    Date (GMT), date and hour.
  •  Waves:
    Significant height, H s (m).
    Mean period, T m  (s).
    Peak period, T p  (s).
    Maximum height, H m a x  (m).
    Period associated with maximum height, T m a x (s).
    Mean direction, D i r . m (degrees).
    Mean direction at peak energy, D i r . p (degrees).
    Directional spread at peak energy, D i s p (degrees).
  • Meteorology (data recorded at 3 m above the surface):
    Average wind speed, U v (m/s).
    Average wind direction, D i r . v (degrees).
    Air temperature, T a i r  ( ° C ).
    Atmospheric pressure, p (HPa).
  • Oceanography (data recorded at 3 m below the surface):
    Average current velocity U c (cm/s).
    Average current direction, D i r . o f . p r o p (degrees).
    Salinity, a l (psu).

3.2. Data Preprocessing

The main goal of this research is to predict the maximum wave height ( H m a x ) and ratio of the maximum to the significant wave height ( H m a x / H s ) based on buoy-derived variables. Initially, a correlation matrix (Figure 2) was computed to examine the relationships between predictor variables and target variables ( H m a x and H m a x / H s ). This analysis facilitated our selection of the most pertinent variables for each prediction task.
For prediction of H m a x , the following five variables exhibited the strongest correlations:
  • Significant wave height ( H s ).
  • Mean wave period ( T m ).
  • Peak wave period ( T p ).
  • Period associated with the maximum wave height ( T _ H m a x ).
  • Wind speed (U).
Similarly, for the prediction of H m a x / H s , the selected variables included the following:
  • Mean wave period ( T m ).
  • Directional spread at the energy peak ( D i s p ).
  • Period associated with the maximum wave height ( T _ H m a x ).
  • Mean wave direction ( D i r . m ).
  • Peak wave period ( T p ).
The preprocessing pipeline commenced with variable selection and an assessment of missing data. To address missing values, three distinct imputation methods were implemented:
  • Predictive Mean Matching (PMM): This method identifies the observed values closest to the predictive mean, randomly selects one, and assigns it to the missing entry.
  • Classification and Regression Trees (CART):
    Constructs decision trees via recursive partitioning.
    For each missing value, determines the terminal node of the fitted tree.
    Assigns the observed value through random sampling within the terminal node.
  • LASSO Regression with Normalization (LASSO.NORM): This method employs LASSO linear regression combined with bootstrapping to handle univariate normal missing data.
Following imputation, the dataset was normalized to ensure uniform feature scaling in order to enhance the performance and stability of the machine learning models.

4. Methodology

To address the prediction tasks, three regression algorithms were employed:
  • Linear Regression (LM): Linear regression can model the relationship between dependent (target) variables and independent (predictor) variables as a linear equation:
    Y = β 0 + β 1 X 1 + + β n X n + ϵ
    where β 0 is the intercept, β i are the coefficients, and  ϵ represents the error term.
  • Support Vector Regression (SVR): SVR is an extension of support vector machine (SVM) that constructs hyperplanes in high-dimensional space to minimize the prediction error within a defined tolerance margin. It is particularly effective for complex nonlinear datasets.
  • Long Short-Term Memory (LSTM): LSTM is a recurrent neural network (RNN) architecture designed to model data sequences while maintaining long-term information through a memory cell structure. It uses different types of gates (input, forget, and output) to regulate the flow of information, allowing it to learn long-term dependencies without encountering the problem of gradient vanishing.
  • Gated Recurrent Units (GRU): GRU is another variant of the recurrent neural network architecture. Similar to LSTM, it is designed to handle long-term dependencies in data sequences. Unlike LSTM, however, GRU has a simpler structure, using only two gates (reset and update) to control the flow of information, which makes it more computationally efficient without sacrificing the ability to model temporal dependencies.
Regarding the hyperparameter configuration of each model, linear regression (LM) does not require predefined hyperparameters, making it a simple and computationally efficient baseline approach. The support vector regression (SVR) model utilizing a linear kernel (kernel = “linear”) adopts the following default hyperparameter values: cost = 1, gamma = 1, epsilon = 0.1, and type = “eps regression”, which govern its optimization process by defining the error margin constraints.
For the LSTM and GRU models, both of which are based on recurrent neural networks, an identical hyperparameter configuration was employed. The architecture consisted of an initial layer with 128 units followed by a second layer with 64 units. The input dimensionality was defined as input_shape = (1, 5), indicating a single time step with five input features. The output layer comprised a single neuron with linear activation, ensuring the production of continuous numerical outputs. Additionally, L2 regularization with a coefficient of 0.01 was applied to the final layer in order to mitigate overfitting by controlling model complexity.
For this study, we selected models specifically designed for regression tasks in order to facilitate a rigorous comparative evaluation of their predictive performance. The LM and SVR models were chosen due to their capacity to effectively model both linear and nonlinear relationships. Conversely, the LSTM and GRU models were incorporated to assess their performance in capturing temporal dependencies, enabling a comparative analysis between conventional and advanced approaches with an appropriate balance between predictive accuracy and computational efficiency.
The integration of these diverse machine learning models allows for a comprehensive assessment of their predictive capabilities in estimating maximum individual wave height ( H m a x ) and the ratio between maximum and significant wave height ( H m a x / H s ) using meteorological and oceanographic data. This comparative framework can provide valuable insights into the relative strengths and limitations of each approach in the context of wave prediction.

5. Proposed Experiment

As detailed above, the experiment described in this study was conducted using databases obtained from the Spanish State Ports website referenced previously, where data for buoys are available.
To achieve robust and accurate predictions, a structured machine learning (ML) lifecycle was implemented. This lifecycle comprised multiple phases, including planning, data preprocessing, model engineering, evaluation, deployment, monitoring, maintenance, and knowledge extraction.
A rigorous variable selection process was carried out during the data preprocessing stage, identifying five explanatory variables in addition to the target variable. Subsequently, missing values were imputed and various transformations were applied, including standardization of variables to ensure consistency in the input features.
Additionally, an autoregressive modeling approach was integrated, expressed mathematically as
y ( t ) = f ( y ( t 1 ) , , y ( t n ) , x ( t 1 ) , , x ( t m ) ) ,
where past instances of both the independent and dependent variables were utilized. This methodological framework enabled the models to capture the temporal dependencies inherent in time series data, thereby enhancing predictive performance by incorporating relevant historical information during training.
Following data preprocessing, different regression-based predictive models were trained and evaluated, including LM, SVR, LSTM, and GRU models. A comparative analysis of these models was then performed using a range of performance metrics to assess their predictive accuracy and reliability.
A fundamental aspect of our methodological framework involved systematically testing different database configurations and preprocessing strategies in order to identify the most effective approach for training the regressors and optimizing their predictive performance.
The specific database configurations considered in this study are detailed in the following section.
For the experiments detailed herein, the database spanned the years 1990 to 2024. For training and testing purposes, individual years were selected as subsets to create training and test datasets. This procedure was repeated with multiple randomly selected years from the complete database, ensuring robust and representative results.
The computational experiments were conducted on a machine equipped with an 11th-Gen Intel® Core™ i7-1165G7 processor operating at a base frequency of 2.80 GHz and with a maximum clock speed of 2.803 GHz. Due to system resource limitations, the data analyses were performed using one-year temporal windows.

5.1. Hmax Predictions

The primary objective of this experiment was to systematically assess the efficacy of different validation methodologies in training predictive models. These methods vary in their partitioning strategies for training and test sets as well as in the configuration of prediction windows within the test set. Although conceptually similar, they differ in terms of the dimensions of the prediction windows and the allocation of data between training and testing subsets. The experiment was conducted using the preprocessed database, where each subset (training and testing) corresponded to a full year, and predictions were performed for multiple years randomly selected from the dataset. This approach ensured a robust evaluation of the models’ generalization ability across different temporal periods.
  • Method 1: The training set consisted of the first three months of the year, while the test set comprised the remaining nine months. Predictions were generated iteratively; the model was initially trained on the first three months, after which it predicted the following month. After each prediction, the actual values of the predicted month were added to the training set and the process was repeated until predictions for all months in the test set were completed.
  • Method 2: The training set was defined as the first 1000 observations of the year, while the remaining observations were assigned to the test set. Predictions were generated for the 50 observations immediately following the training set. After obtaining these predictions, the actual values corresponding to the predictions were incorporated into the training set and iteration continued until predictions were generated for the entire test set.
  • Method 3: The training set comprised the first two weeks of the year, while the remaining data were allocated to the test set. Predictions were made for three consecutive days following the initial two-week training period. After obtaining these predictions, the actual values corresponding to the predictions were incorporated into the training set and iteration continued until all test set predictions were completed.
  • Method 4: This method was similar to Method 3 but introduced an additional adjustment. After generating predictions for three consecutive days, the actual values corresponding to the predictions were added to the training set, while the three oldest days were removed. This iterative process continued until all predictions for the test set were completed.
All four methods shared a common iterative framework, with specific differences in the size and configuration of the training and test sets. Below, Algorithm 1 describes the shared approach for the methods described above.
The average results for all the trained years are presented in Figure 3, graphically represented through bar diagrams.
Validation of the proposed methods was conducted to assess their performance across both short-term and long-term prediction horizons, as evaluated through the designated prediction windows applied to the test set. Among the analyzed methods, Method 1 yielded the most notable results. Additionally, a comparative analysis of different predictive models indicated that the linear regression (LM) and support vector regression (SVR) models exhibited superior performance, demonstrating higher levels of accuracy and robustness.
Algorithm 1 General algorithm
Input: data
Goal: predict the class variable with the cleaned and normalized database data.
data_train← each method has its own training set.
data_test← each method has its own test set.
n_rows← number of rows of data_test
increment← data in the prediction window
for i in seq(1, n_rows, by = increment)
     for (j in 1:increment)
          model ← regressor model with data_train
          prediction ← prediction of the regressor with model and data_test
          data_train ← rbind(data_train, prediction)
     end for
end for

5.2. Hmax/Hs Predictions

The objective of the second experiment was to assess the effectiveness of the same validation methodologies in predicting the H m a x / H s ratio, as opposed to the individual maximum wave height H m a x examined in the first experiment. The methodology remained identical, with the primary distinction being the target variable, which in this case corresponded to the ratio between the maximum wave height H m a x and significant wave height H s . In this experiment, predictions were conducted exclusively using the LM and SVR models, as they provided the best results in the previous experiment.
Despite using the same approaches and same preprocessed dataset, the obtained results showed significantly lower performance compared to the prediction of H m a x , as shown in Figure 4, which contains the average values of the results for all the trained years. Both methods faced significant difficulties in accurately predicting the relationship, as reflected in the low R 2 values. This coefficient was notably lower compared to the values obtained for H m a x , indicating that the models were limited in their ability to explain the variability in the relationship data.
However, improvements were observed in evaluation metrics such as the mean absolute error (MAE), mean squared error (MSE), and root mean squared error (RMSE), which remained comparable to those obtained in the H m a x prediction task. This finding suggests that although predicting the H m a x / H s ratio is more challenging, the models still generate reasonable estimates in terms of the absolute error magnitudes.
One potential explanation for the observed decline in predictive performance is the low correlation between the target variable ( H m a x / H s ratio) and the predictor variables, which does not exceed 0.16. This weak correlation indicates a limited relationship between the input features and the target variable, potentially constraining the model’s ability to learn robust patterns.

6. Results

The following performance metrics were used to determine the effectiveness of the chosen regressors:
  • R-Squared ( R 2 ): The coefficient of determination, often referred to as R-squared, is the proportion of the total variance in the dependent variable that can be explained by the estimated regression model. A result close to 1 means that the model is able to explain much of the variation in the data, while lower values mean that it lacks this ability and does not fit the data well. The R-squared value is calculated using the following formula:
    R 2 = i = 1 N ( y i ^ y i ) 2 ( y i y ¯ ) 2 .
  • Mean Absolute Error (MAE): The MAE is used to quantify the prediction accuracy. It measures the average magnitude of the errors in a set of predictions, providing a numerical value that represents how far the predictions are from the actual values. An MAE value close to 0 indicates that the model is more accurate, as there is not much difference between the actual and predicted values.
    MAE = i = 1 N | y i y i ^ | N
  • Mean Square Error (MSE): The MSE measures the average of the squares of the errors, i.e., the average squared difference between the estimated values and the actual value:
    MSE = i = 1 N ( y i y i ^ ) 2 N .
  • Root Mean Square Error (RMSE): The RMSE is another metric used to measure the accuracy of predictions. In this case, the square of the errors is taken into account, which assigns more weight to larger errors. The ideal value is 0. The formula for RMSE is
R M S E = i = 1 N ( y i y i ^ ) 2 N ,
where y i ^ represents the predicted values, y i the observed values, y ¯ the mean value, and N the number of observations.
As previously discussed, model prediction of both the maximum wave height ( H m a x ) and the ratio ( H m a x / H s ) were evaluated using these four key performance metrics: MAE, MSE, RMSE, and R 2 , for which definitions and detailed explanations can be found in Section 4. In the following, we present the best results accompanied by the standard deviation of the errors, allowing for assessment of the statistical significance of the predictions.

6.1. H m a x Predictions

For the initial experiment, the best results were obtained from the H m a x predictions for the year 2014, which exhibited the highest performance among all the years we evaluated.
Based on the results presented in Table 2, it can be observed that Method 1 exhibits the highest performance in terms of the considered evaluation metrics of R 2 , MAE, MSE, and RMSE, which has been highlighted in red for easy identification. In particular, the model that demonstrated the best performance on this dataset was SVR, suggesting its superior ability to capture the underlying patterns when predicting the target variable.
A key aspect of these results is that the standard deviation of the errors obtained with Method 1 is significantly lower compared to the other evaluated approaches. This indicates greater stability in predictions and reduced variability in errors, which in turn suggests higher consistency and accuracy across the entire dataset. The reduction in error dispersion is a critical indicator of the model’s robustness, further reinforcing its reliability compared to alternative methods and highlighting its potential as an optimal strategy for this type of predictive analysis.
To ensure the statistical validity of the results, normality tests were conducted. These tests are essential as they assess whether the data follow a normal distribution, which is a key requirement for many parametric statistical tests. The results indicated that the data do not exhibit a normal distribution, which has significant implications for selecting the most appropriate statistical tests for analysis.
Because many parametric tests assume data normality, applying them in cases where the assumption of normality does not hold could lead to misleading or biased conclusions. Therefore, the non-parametric Kruskal–Wallis test was chosen as a robust alternative that does not assume normality in order to effectively compare multiple groups. This test evaluates whether significant differences exist among the groups considered in the analysis and provides p-values that determine the statistical significance of these differences.
The p-values obtained from the Kruskal–Wallis test are presented in Table 3, and serve as a key criterion for interpreting the significance of the results. The application of these methods ensures a rigorous and appropriate analysis, minimizing the risk of erroneous inferences and strengthening the reliability of the conclusions drawn. This approach enhances the robustness of the study and its applicability to similar contexts where the data may not conform to a normal distribution.
Because the obtained p-values are all below 0.05, it can be concluded that the observed differences are statistically significant.

6.2. H m a x / H s Predictions

In the second experiment, the most accurate results in predicting the ratio ( H m a x / H s ) were obtained for the year 2011, as shown in Table 4, highlighting it as the period with the best performance among all the years we evaluated. It is important to note that we used only the models that yielded the best results in the previous experiment, specifically, the linear regression and SVR models.
Based on the results presented in Table 4, it can be observed that the SVR model exhibits superior performance in terms of R 2 and MAE, while the linear regression model demonstrates better performance for MSE and RMSE. However, in the context of this experiment, it is not possible to definitively determine the most suitable model, as the results obtained by all methods are comparatively similar. Nonetheless, it can be stated that Method 1 shows significantly smaller standard deviations than the other methods, suggesting greater consistency and accuracy in predictions across the dataset. This behavior underscores the robustness and reliability of this method in comparison to the evaluated alternatives.
The performance in term of R 2 exhibits notably low values across all predictions. However, it is important to highlight that the trend line between the actual observations and the predictions demonstrates a high degree of congruence, as shown in Figure 5. In this figure, the x-axis represents the indices of the test set values, while the y-axis displays the predicted and actual values of H m a x . This finding suggests that, despite the imprecision of individual predictions, the models are capable of adequately capturing the overall trend of the data. This behavior indicates that with further adjustments, such as the inclusion of variables with higher correlation, the predictive capacity of the model could be significantly improved in future studies.
In addition, normality tests were conducted to analyze the prediction results for the ( H m a x / H s ) ratio. The p-value results indicated that the data do not follow a normal distribution; consequently, the non-parametric Kruskal–Wallis test was applied to assess whether there were significant differences between the groups. Upon examining the results presented in Table 4, it can be observed that the metrics obtained for the two models are notably similar. Therefore, the test is expected to reflect the absence of statistical differences between the two models. The results of the Kruskal–Wallis test are detailed in Table 5.
As shown in Table 5, the results align with expectations. The p-values are all greater than 0.05, allowing us to conclude that there are no statistically significant differences between the models used for predicting the variable in question.

7. Conclusions

The analysis conducted in this study demonstrates that long-term predictions are significantly more accurate than short-term ones. This phenomenon can be attributed to the capacity of Method 1 to more effectively identify and model stable trends in the data, which is thanks to this method utilizing a larger data window in the test set. In the context of ocean engineering, this improvement is particularly relevant for estimating extreme wave conditions such as swells, which critically impact both the safety and operational efficiency of maritime activities. The ability of models to capture general trends over extended intervals provides a robust tool for risk planning and mitigation in harsh ocean environments, where the predictability of extreme conditions is essential for decision-making.
Regarding prediction of the H m a x variable, the models that best fit the data and achieved the most accurate results were those based on linear regression (LM) and support vector regression (SVR), which is further supported by statistical significance analysis.
This study also examined prediction of the H m a x / H s ratio. This ratio shows a low correlation with other variables, affecting the models’ accuracy. Nevertheless, error metrics such as MAE, MSE, and RMSE indicate that the models were able to successfully reduce prediction errors. Although the models did not fully capture the variability, the observed and predicted trends are similar, suggesting that they were able to adequately capture the global dynamics.

Author Contributions

All authors contributed equally to conceptualization, methodology, software, validation, formal analysis, investigation, resources, data curation, writing—original draft preparation, writing—review and editing, visualization, supervision, project administration, and funding acquisition. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Raw data are available upon request from the website www.puertos.es, accessed on 15 Septembre 2024.

Acknowledgments

We thank Puertos de España (Ports of Spain) for providing the requested data for this work.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ANFISAdaptative Neuro-Fuzzy Interference System
CARTClassification and Regression Trees
GHGGreenhouse Gases
LASSO.NORMLasso Linear Regression
LMLinear Regression Modeling
MLMachine Learning
MAEMean Average Error
MICEMultiple Imputation using Chained Equations
MREMarine Renewable Energy
MSEMean Square Error
OTECOcean Thermal Energy Conversion
PMMPredictive Mean Matching
PPCAProbabilistic Principal Component Analysis
R 2 R-Squared
RFRandom Forest
RMSERoot Mean Square Error
SVRSupport Vector Regression

Notes

1
2
www.puertos.es, 15 Septembre 2024

References

  1. Yang, Y.; Zhao, L.; Chen, L.; Wang, C.; Wang, G. The spillover effects between renewable energy tokens and energy assets. Res. Int. Bus. Financ. 2024, 74, 102672. [Google Scholar] [CrossRef]
  2. Zhang, X.; Xu, W.; Rauf, A.; Ozturk, I. Transitioning from conventional energy to clean renewable energy in G7 countries: A signed network approach. Energy 2024, 307, 132655. [Google Scholar] [CrossRef]
  3. Khurshid, H.; Mohammed, B.S.; Al-Yacoubya, A.M.; Liew, M.; Zawawi, N.A.W.A. Analysis of hybrid offshore renewable energy sources for power generation: A literature review of hybrid solar, wind, and waves energy systems. Dev. Built Environ. 2024, 19, 100497. [Google Scholar] [CrossRef]
  4. Ibarra-Berastegui, G.; Sáenz, J.; Ulazia, A.; Sáenz-Aguirre, A.; Esnaola, G. CMIP6 projections for global offshore wind and wave energy production (2015–2100). Sci. Rep. 2023, 13, 18046. [Google Scholar] [CrossRef]
  5. Ulazia, A.; Esnaola, G.; Serras, P.; Penalba, M. On the impact of long-term wave trends on the geometry optimisation of oscillating water column wave energy converters. Energy 2020, 206, 118146. [Google Scholar] [CrossRef]
  6. Ulazia, A.; Sáenz, J.; Saenz-Aguirre, A.; Ibarra-Berastegui, G.; Carreno-Madinabeitia, S. Paradigmatic case of long-term colocated wind–wave energy index trend in Canary Islands. Energy Convers. Manag. 2023, 283, 116890. [Google Scholar] [CrossRef]
  7. Ulazia, A.; Saenz-Aguirre, A.; Ibarra-Berastegui, G.; Sáenz, J.; Carreno-Madinabeitia, S.; Esnaola, G. Performance variations of wave energy converters due to global long-term wave period change (1900–2010). Energy 2023, 268, 126632. [Google Scholar] [CrossRef]
  8. Martinez-Iturricastillo, N.; Ulazia, A.; Ringwood, J. Long term wave load trends against offshore monopile structures: A case study in the Bay of Biscay. Proc. EWTEC 2023, 15, 1–7. [Google Scholar] [CrossRef]
  9. Sierra, J.P.; Castrillo, R.; Mestres, M.; Mösso, C.; Lionello, P.; Marzo, L. Impact of climate change on wave energy resource in the Mediterranean coast of Morocco. Energies 2020, 13, 2993. [Google Scholar] [CrossRef]
  10. Stansell, P. Distributions of extreme wave, crest and trough heights measured in the North Sea. Ocean Eng. 2005, 32, 1015–1036. [Google Scholar] [CrossRef]
  11. Neshat, M.; Sergiienko, N.Y.; Rafiee, A.; Mirjalili, S.; Gandomi, A.H.; Boland, J. Meta Wave Learner: Predicting wave farms power output using effective meta-learner deep gradient boosting model: A case study from Australian coasts. Energy 2024, 304, 132122. [Google Scholar] [CrossRef]
  12. Barbariol, F.; Bidlot, J.R.; Cavaleri, L.; Sclavo, M.; Thomson, J.; Benetazzo, A. Maximum wave heights from global model reanalysis. Prog. Oceanogr. 2019, 175, 139–160. [Google Scholar] [CrossRef]
  13. White, C.; Carrelhas, A.; Gato, L.; Portillo, J.; Cândido, J. Floating wind and wave energy technologies: Applications, synergies and role in decarbonization in Portugal. Proc. EWTEC 2023, 15. [Google Scholar] [CrossRef]
  14. Young, I.R.; Zieger, S.; Babanin, A.V. Global trends in wind speed and wave height. Science 2011, 332, 451–455. [Google Scholar] [CrossRef]
  15. Reguero, B.G.; Losada, I.J.; Méndez, F.J. A recent increase in global wave power as a consequence of oceanic warming. Nat. Commun. 2019, 10, 205. [Google Scholar] [CrossRef] [PubMed]
  16. Stansby, P.K. Solitary wave run up and overtopping by a semi-analytical model. Coast. Eng. 2003, 47, 159–179. [Google Scholar] [CrossRef]
  17. Bromirski, P.D.; Miller, A.J.; Flick, R.E.; Auad, G. Wave power variability and trends across the North Pacific. J. Geophys. Res. Ocean. 2020, 125, e2019JC015419. [Google Scholar] [CrossRef]
  18. Cabral, I.S.; Young, I.R.; Toffoli, A. Long-term and seasonal variability of wind and wave extremes in the Arctic Ocean. J. Geophys. Res. Ocean. 2020, 125, e2020JC016708. [Google Scholar] [CrossRef]
  19. Young, I.R.; Ribal, A. Multiplatform evaluation of global trends in wind speed and wave height. Science 2019, 364, 548–552. [Google Scholar] [CrossRef]
  20. Boccotti, P. Wave Mechanics for Ocean Engineering; Elsevier Oceanography Series; Elsevier: Amsterdam, The Netherlands, 2000; Volume 64, Chapter 11. [Google Scholar] [CrossRef]
  21. Goda, Y. Random Seas and Design of Maritime Structures; World Scientific Publishing Company: Singapore, 2000. [Google Scholar] [CrossRef]
  22. Tayfun, M.A. Narrow-band nonlinear sea waves. J. Geophys. Res. Ocean. 1980, 85, 1548–1552. [Google Scholar] [CrossRef]
  23. Liu, Y.; Zhang, X.; Dong, Q.; Chen, G.; Li, X. Phase-resolved wave prediction with linear wave theory and physics-informed neural networks. Appl. Energy 2024, 355, 121602. [Google Scholar] [CrossRef]
  24. Liu, Y.; Huang, L.; Ma, X.; Zhang, L.; Fan, J.; Jing, Y. A fast, high-precision deep learning model for regional wave prediction. Ocean Eng. 2023, 288, 115949. [Google Scholar] [CrossRef]
  25. Memar, S.; Mahdavi-Meymand, A.; Sulisz, W. Prediction of seasonal maximum wave height for unevenly spaced time series by Black Widow Optimization algorithm. Mar. Struct. 2021, 78, 103005. [Google Scholar] [CrossRef]
Figure 1. Position of the buoy.
Figure 1. Position of the buoy.
Jmse 13 00625 g001
Figure 2. Heatmap of correlations.
Figure 2. Heatmap of correlations.
Jmse 13 00625 g002
Figure 3. Average results of H m a x predictions.
Figure 3. Average results of H m a x predictions.
Jmse 13 00625 g003
Figure 4. Average results of H m a x / H s prediction.
Figure 4. Average results of H m a x / H s prediction.
Jmse 13 00625 g004
Figure 5. Observations, predictions, and trend lines.
Figure 5. Observations, predictions, and trend lines.
Jmse 13 00625 g005
Table 1. Characteristics of buoy.
Table 1. Characteristics of buoy.
BuoyLongitudeLatitudeDepthCadenceModelYears
Bilbao-Biscay 5.100 W43° 37.940 N870 m60 minSeaWatch1990–2024
Table 2. Results of H m a x predictions for 2014.
Table 2. Results of H m a x predictions for 2014.
Method 1Method 2
R2MAEMSERMSER2MAEMSERMSE
LM x ¯ 0.89320.09670.01960.13350.62020.12930.05360.1702
sd0.02150.03380.01310.04530.24740.11560.12000.1578
SVR x ¯ 0.90060.09160.01850.12860.63390.12460.05340.1659
sd0.02220.03480.01330.04730.24630.11680.12390.1617
LSTM x ¯ 0.81620.14330.03890.18840.38580.18040.07860.2374
sd0.07900.05180.02340.06250.27640.11300.11360.1494
GRU x ¯ 0.81240.14350.03970.18920.39060.18410.07960.2397
sd0.07920.05530.02560.06620.27770.11380.11310.1493
Method 3Method 4
R2MAEMSERMSER2MAEMSERMSE
LM x ¯ 0.67480.15390.07280.20460.68010.14970.06470.1966
sd0.23670.12900.14310.17660.24710.12160.11820.1619
SVR x ¯ 0.68840.14840.07250.19970.68420.14550.06460.1947
sd0.23760.13090.14760.18130.24270.11890.12270.1642
LSTM x ¯ 0.70470.30780.18620.40990.68960.35470.26870.4699
sd0.17780.10660.12020.14290.17970.19290.27280.2323
GRU x ¯ 0.46400.20660.10280.27460.69600.36640.29560.4802
sd0.29020.12420.13660.16520.17290.22910.35690.2703
Table 3. The p values for the H m a x predictions.
Table 3. The p values for the H m a x predictions.
Method 1Method 2Method 3Method 4
R21.1 × 10 3 2.2 × 10 16 7.8 × 10 11 6.8 × 10 7
MAE1.5 × 10 2 2.2 × 10 16 8.9 × 10 11 9.4 × 10 7
MSE3.0 × 10 2 2.2 × 10 16 6.3 × 10 11 3.5 × 10 7
RMSE3.0 × 10 2 2.2 × 10 16 6.3 × 10 11 9.07 × 10 7
Table 4. Results of H m a x / H s predictions for 2011.
Table 4. Results of H m a x / H s predictions for 2011.
Method 1Method 2
R2MAEMSERMSER2MAEMSERMSE
LM x ¯ 0.01080.11700.02290.15160.02780.11880.02590.1532
sd0.00990.00290.00130.00430.03670.02620.03800.0489
SVR x ¯ 0.01140.11660.02350.15330.02830.11820.02640.1547
sd0.00990.00300.00130.00430.03820.02620.03830.0492
Method 3Method 4
R2MAEMSERMSER2MAEMSERMSE
LM x ¯ 0.02210.11840.02550.15350.02350.12050.02600.1552
sd0.03030.02190.03050.04440.03080.02150.03070.0440
SVR x ¯ 0.02210.11790.02600.15490.02290.11890.02620.1557
sd0.03060.02190.03070.04470.02190.02160.03040.0439
Table 5. The p values for H m a x / H s predictions.
Table 5. The p values for H m a x / H s predictions.
Method 1Method 2Method 3Method 4
R20.82530.89460.75150.9070
MAE0.69110.67260.63410.3464
MSE0.30990.55350.41360.7876
RMSE0.30990.55350.53910.7876
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Porlan-Ferrando, L.; Nuñez-Gonzalez, J.D.; Ulazia Manterola, A.; Martinez-Iturricastillo, N.; Ringwood, J.V. Maximum Individual Wave Height Prediction Using Different Machine Learning Techniques with Data Collected from a Buoy Located in Bilbao (Bay of Biscay). J. Mar. Sci. Eng. 2025, 13, 625. https://doi.org/10.3390/jmse13040625

AMA Style

Porlan-Ferrando L, Nuñez-Gonzalez JD, Ulazia Manterola A, Martinez-Iturricastillo N, Ringwood JV. Maximum Individual Wave Height Prediction Using Different Machine Learning Techniques with Data Collected from a Buoy Located in Bilbao (Bay of Biscay). Journal of Marine Science and Engineering. 2025; 13(4):625. https://doi.org/10.3390/jmse13040625

Chicago/Turabian Style

Porlan-Ferrando, Lucia, J. David Nuñez-Gonzalez, Alain Ulazia Manterola, Nahia Martinez-Iturricastillo, and John V. Ringwood. 2025. "Maximum Individual Wave Height Prediction Using Different Machine Learning Techniques with Data Collected from a Buoy Located in Bilbao (Bay of Biscay)" Journal of Marine Science and Engineering 13, no. 4: 625. https://doi.org/10.3390/jmse13040625

APA Style

Porlan-Ferrando, L., Nuñez-Gonzalez, J. D., Ulazia Manterola, A., Martinez-Iturricastillo, N., & Ringwood, J. V. (2025). Maximum Individual Wave Height Prediction Using Different Machine Learning Techniques with Data Collected from a Buoy Located in Bilbao (Bay of Biscay). Journal of Marine Science and Engineering, 13(4), 625. https://doi.org/10.3390/jmse13040625

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop