Next Article in Journal
Biometry, Growth, and Recruitment Pattern of a Commercially Important Nereid polychaete, Namalycastis fauveli, from the East Coast of Bangladesh
Previous Article in Journal
Soft-Shell Production of the Invasive Atlantic Blue Crab Callinectes sapidus in the Lesina Lagoon (SE Italy): A First Assessment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Wave Height and Period Estimation from X-Band Marine Radar Images Using Convolutional Neural Network

1
School of Mechanical Engineering, Dalian University of Technology, Dalian 116081, China
2
State Key Laboratory of High-Performance Precision Manufacturing, Dalian University of Technology, Dalian 116024, China
3
Ningbo Institute, Dalian University of Technology, Ningbo 315000, China
4
Department of Navigation, Dalian Naval Academy, Dalian 116081, China
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2024, 12(2), 311; https://doi.org/10.3390/jmse12020311
Submission received: 29 December 2023 / Revised: 30 January 2024 / Accepted: 7 February 2024 / Published: 9 February 2024
(This article belongs to the Section Physical Oceanography)

Abstract

:
In this study, a deep learning network for extracting spatial-temporal features is proposed to estimate significant wave height ( H s ) and wave period ( T s ) from X-band marine radar images. Since the shore-based radar image in this study is interfered with by other radar radial noise lines and solid target objects, to ensure that the proposed convolutional neural network (CNN) extracts the image features accurately, it is necessary to pre-process the radar image to eliminate interference. Firstly, a pre-trained GoogLeNet is used to extract multi-scale depth space features from the radar images to estimate H s and T s . Since CNN-based models cannot analyze the temporal behavior of wave features in radar image sequences, self-attention is connected after the deep convolutional layer of the CNN to construct a convolutional self-attention (CNNSA)-based model that generates spatial-temporal features for H s and T s estimation. Simultaneously, H s and T s measured by nearby buoys are used for model training and reference. The experimental results show that the proposed CNNSA model reduces the RMSD by 0.24 m and 0.11 m, respectively, in H s estimation compared to the traditional SNR-based and CNN-based methods. In T s estimation, the RMSD is reduced by 0.3 s and 0.08 s, respectively.

1. Introduction

X-band radar has become a valuable tool for oceanographic studies due to its high spatial and temporal resolution [1]. The Bragg resonance interaction between X-band electromagnetic waves and the centimeter-scale ripple waves induced by local winds generate radar backscatter from the sea surface. The sea surface changes are imaged by the backscattering of electromagnetic waves [2]. In addition, marine radar images are used to effectively estimate sea surface features, such as wind [3,4], wave parameters [2,5,6,7], and current [8,9,10]. Different sea surface parameters, such as significant wave height ( H s ), wave period ( T s ), and wave direction, are essential for the safety of various marine activities and the development of coastal areas [11]. In situ sensors, such as wave buoys, have traditionally been used for wave measurements. In contrast, X-band marine radars can detect and measure wave parameters over a broader range and have relatively low maintenance costs.
The echo signal produced by the X-band radar, presented in a light and dark striped image, is called sea clutter, formed by the backscattering of electromagnetic waves emitted by radar across the sea. The base signal (short waves) is modulated by longer waves with several mechanisms, such as hydrodynamic modulation, tilt modulation, and shadow modulation, resulting in longer waves visible in radar images [12]. The conventional spectral analysis method obtains the image spectrum via 3-D Fast Fourier transformation (FFT) of the radar image sequence [2], using the dispersion relation as a filter to extract the image spectrum from the background noise. Then, the wave period, wave direction, and wavelength can be deduced from the filtered spectrum [2,6]. Based on the estimation relation of the synthetic aperture radar [13], the signal-to-noise ratio (SNR) of the wave is calculated from the wave spectrum, and the estimation is completed by the linear relationship between the H s and the square root of the SNR. The coefficients can be determined using in situ measurements. Later studies have found that H s is not exactly linearly proportional to the square root of the SNR due to variations in sea states, different methods of calculating SNR, and differences in radar systems. In addition to SNR-based methods, there are some alternative methods that have been proposed to estimate wave parameters, such as empirical orthogonal function-based methods [14], iterative least squares-based methods [1], 2D continuous wavelet transform-based methods [15], array-beamforming-based methods [16], shadowing mitigation-based methods [17], and synchrosqueezed wavelet transform-based methods [18]. Additionally, there are some other methods for estimating H s that have been proposed, such as shadowing-based methods [19,20], ensemble empirical mode decomposition-based methods [21], correlation analysis-based methods [22], and variational mode decomposition-based methods [23]. It is worth noting that machine learning algorithms have been applied to H s estimation, which can simplify the cumbersome steps of previous algorithms and improve computational efficiency. It is also possible to estimate more accurate results using methods including a support vector regression (SVR)-based method [24], artificial neural network (ANN)-based methods [25,26], a convolutional neural network (CNN)-based method [27], a convolutional gated recurrent unit network (CGRU)-based method [14], or a temporal convolutional network (TCN)-based method [28]. In addition, random forest (RF)-based machine learning methods have been used to estimate wave directions and periods [29].
CNNs can extract spatial features of image sequences but are incapable of temporal analysis. In this study, a novel H s and T s estimation model combining CNN with self-attention (abbreviated as CNNSA) is proposed. This method can extract the spatial features of radar image sequences via CNN, and introduces a self-attention layer in the feature vector of the CNN to capture dependencies in the time series and achieve spatial-temporal feature extraction of radar images. Section 2 introduces the data pre-processing methods used, including median filtering based on the two-layer decision and the adaptive region growing repair method. The structure and components of the proposed CNN and convolutional self-attention (CNNSA)-based models are illustrated in Section 3. The training and testing results obtained from shore-based marine radar data using SNR, CNN, and CNNSA-based models are presented and compared with each other in Section 4. The results are discussed in Section 5. Finally, the conclusion and outlook of this work appear in Section 6.

2. Data Pre-Processing

Interference from other marine radars produces radial noise lines in radar images, as shown by the dense radial pixel lines in Figure 1. In this study, the radar images collected are from shore-based X-band radar, which also is affected by interference from targets like ships, which appear as bright spots in the radar images. The pre-processing process of the radar images is shown in Figure 2.

2.1. Median Filtering Based on the Two-Layer Decision

According to the characteristics of the wave texture in the radar image, an improved median filtering method is used to screen the pixel points for noise. The region involved in H s and T s estimation is intercepted in a radar polar coordinates image. In this study, a sub-image with azimuth ±30° and range in 600–2500 m is selected and converted into a grey image of 0–255, denoted as I ( x , y ) , which is of size n × n , n = 256. First, the size of the sliding window L 1 is set to m × 1 , and in this study, m = 3. The I ( x , y ) is expanded according to the sliding window size; the pixel image size after the expansion is ( n + m 1 ) , and the starting point is the first position of the expanded image in order to find the noise point to be processed. Then, the center point L 1 is set to determine whether the median of the grey value of each pixel point is the L 1 . If it is the median, the grey value of the center point will not be processed. If it is not the median, L 1 is set to determine whether it is a noise point. Next, a threshold C 1 is set to determine whether the pixel point is a noise point. C 1 = 1 5 A a v g calculates the difference between the grayscale value of the maximum pixel point in each sliding window ( B m a x ( i , j ) ) and the grayscale value of the minimum pixel point ( B m i n ( i , j ) ), expressed as B. If B > C 1 , the noise point is determined. Finally, the noise point is replaced by the median value calculated in L 1 . A a v g is represented as follows:
A a v g = 1 x 1 y j = 1 y i = 1 x I ( i , j )

2.2. Adaptive Region Growing Repair Method

Targets on the sea surface can cause a lack of wave texture, leading to errors when estimating H s and T s from radar images using deep learning-based methods. To avoid this problem, this study uses an adaptive region-growing repair method to recover the image region disturbed by targets. The process is divided into four parts. The first part uses the adaptive threshold to determine whether the target is produced. The initial parameter C 2 is set with threshold D 1 , where A m a x is the maximum value in the grayscale value of all pixel points in I ( x , y ) . It is then determined whether A m a x is greater than D 1 , and if the output is based on the original grayscale value, the following calculation is required.
D 1 = A a v g + C 2
C 2 = A m a x + 1 2
The second part uses gradient descent to find the initial growth point of the target and identifies the target area in I ( x , y ) . The grayscale values of all the pixel points in I ( x , y ) are sorted from smallest to largest, and the location of a specific grayscale value is selected as the growth point of the target. The sliding window is set to L 2 = m × m , so that A m a x is used as the center point and initial growth point of the sliding window L 2 . This point is used as the starting point to search for pixel points in the sliding window that have similar features to the centre point of the sliding window. It is also used as the noise starting point until the traversal stops when there are no pixel points in the sliding window that have similar features, in order to determine the area where the target is located. The pixel point screening method for similar features is C ( i , j ) C c e n < D 2 , where C ( i , j ) denotes the grayscale value of the pixel point inside L 2 , C c e n denotes the grayscale value of the center point of L 2 , and D 2 is the filtering threshold. If the equation is valid, pixel points with similar features are searched for using C ( i , j ) as a new starting point until the equation is not valid inside the sliding window. These steps are repeated to find the remaining targets in the image. D 2 is defined as follows:
D 2 = 1 3 A a v g
The third part determines whether the target is a real target at the time. The number of pixel locations occupied by each target object N i is calculated, and it is determined whether the total number of pixel locations occupied by each target object is less than the maximum upper limit of single target object recognition where a r e a = m × m   ( m = N i / 6 ) . The integer is retained, and if it is valid, it is considered to be a target object. The average grey value B a v g of the pixel points in the target area and the target threshold D 3 are calculated. If B a v g > D 3 , then it is considered a real target. D 3 is defined as follows:
D 3 = 5 4 A a v g
The fourth part is based on a mean-filling transition algorithm to restore the target and real wave recovery. The grayscale value of the pixel point where the target is located is set to 0. According to the size of the sliding window L 2 , I ( x , y ) is expanded to ( n + 2 m ) × ( n + 2 m ) , and the mean value of the four-pixel points that are m points away from the centre of the noise point in the distance direction and orientation direction is selected to fill the image, instead of the noise point. Figure 3 shows the radar sub-image, processed radial noise line, and the image of the target object.

3. The CNNSA-Based Estimation Model

3.1. CNN

CNNs are a deep learning model designed to process and analyze data with a grid structure, such as images and videos. They are unique in their convolutional layers, which are locally aware of the input data through the utilization of sliding windows, thus effectively capturing the spatial relationships of image or video features [30]. The basic building blocks of CNNs include convolutional operations, pooling operations, and fully connected layers. Convolutional operations detect various features in images, such as edges, textures, and shapes. In contrast, pooling operations are used to reduce the size of the feature map, reduce computational complexity, and extract higher-level features. In this study, GoogLeNet [31] is selected as the pre-training model among many convolutional network models. Compared with other classical CNNs, such as AlexNet [32] and VGGNet [33], GoogLeNet has a lighter weight and more efficient structure, adopting the “Inception” module structure to reduce the number of parameters by using multi-scale convolutional kernels. There are also deeper networks that can learn more complex feature representations.
GoogLeNet consists of several “Inception” modules, each containing a series of convolutional kernels of different scales to capture image features in parallel. Its overall structure consists of the following key components:
(1)
Convolutional and pooling layers. These are used to extract the basic features of the image, such as edges and texture;
(2)
Inception modules. Each “Inception” contains multiple parallel convolutional kernels and pooling operations to capture features at different scales and levels. The results of these parallel operations are cascaded together to form the module outputs, with the primary goal of improving the feature representation of the model without adding too many parameters;
(3)
Global average pooling layer. This layer averages the values of each channel of the feature map to generate a fixed-size feature vector. This reduces the fully connected layer’s dimensionality and helps reduce overfitting.
(4)
Fully connected layer. This layer integrates information from different features to capture complex relationships in the data. In this study, the regression task of the fully connected layer is utilized to estimate the HS and TS of the radar image.
The structure of the GoogLeNet-based estimation model is shown in Figure 4.

3.2. Self-Attention

The CNN-based model performs well in capturing the depth features of the radar image, but there are no features to capture the temporal correlation between the entire image sequence. Self-attention allows the model to dynamically capture long-distance dependencies throughout the input sequence, which can relate to the context of the image sequence. The computation process of self-attention, which includes the computation of Query, Key, Value, Attention Score, and the final weighted summation, is shown in Figure 5. Firstly, the mapping map x obtained by deep convolution is multiplied by the weight matrices W Q , W K , and W K to get the feature spaces f ( x ) , g ( x ) , and h x , respectively:
f x = W Q · x f x = W K · x f x = W V · x
Then, after multiplying the transpose of the feature space f ( x i ) with the feature space g ( x j ) , the product is normalized with a soft classifier to obtain the attention map β j , i :
β j , i = e x p ( s i j ) i = 1 N e x p ( s i j ) ,   w h e r e   s i j = f ( x i ) T · g ( x j )
Finally, the feature space h x is multiplied by the attention map and then with a 1 × 1 convolution to obtain the self-attention feature map, o:
o j = v j = 1 N β j , i · h x i ,   w h e r e   h x i = W V · x i

3.3. CNN-SA Model

In this study, a novel deep learning model with temporal and spatial feature extraction is proposed by embedding self-attention as a feature extraction module in the CNN model. The model captures the spatial information of radar images through the convolutional layer, pooling layer, and Inception structure of GoogLeNet. Adding self-attention after the “Inception” structure can dynamically capture long-range dependencies in image sequences. This makes capturing spatial and temporal relationships in radar image sequences more conducive to estimating H s and T s from the model. The flowchart of the CNNSA-based estimation model for H s and T s is shown in Figure 6, and detailed parameter information is shown in Table 1.

4. Results

4.1. Data Overview

The X-band marine radar data used in this study were collected in Pingtan, Fujian Province, from 16 December 2010 to 10 January 2011. The detailed parameter information of the radar is shown in Table 2. The rotation period of the radar antenna was about 2.5 s, and a radar image sequence contained 32 images, that is, 80 s of each image sequence interval. The wave buoy that measured the reference data was deployed approximately 0.85 km from the radar. Since the references H s and T s were measured every 20 min, temporal interpolation is required to provide synchronized references for each radar image sequence. Figure 1 shows the position relationship between a radar image under polar coordinates and wave buoy placement. During data collection, H s was between 1.3 and 3.5 m and T s was between 6.5 and 10 s. At the same time, the synchronized local wind speeds ranged from 5.8 to 18.5 m/s, as shown in Figure 7, making the proposed method estimated under typical wind and wave conditions. The synchronized buoy references and radar images used in the study were not present at every moment, and the corresponding data are shown in Figure 7.

4.2. Model Train

In this study, the first image of each sequence was input into the pre-processing process in Figure 2, which selected the radar image sub-area and filtered out the radial noise lines and targets so that high-quality sub-images could be provided for the estimation of H s and T s , based on deep learning for better learning of feature information in the images. The significant wave heights used in this study were 1.3–3.5 m, and the training sample contained images from all ranges. Since the reference H s and T s were measured around every 20 min before training, all the image sequences (excluding the rainy day) were first sorted according to the simultaneous H s and T s obtained from interpolated buoy measurements. Then, at each 0.5 m interval between 1.3–3.5 m, 70% and 30% of the samples were taken for training and testing of the model. The proposed CNNSA-based model was trained using the Pytorch framework, installed on a Windows 10 PC with two 2.10 GHz Xeon (R) Gold 6230R CPUs. The initial learning rate was 0.0003 using the Adam optimizer and the MSE loss function. Batch size was set to 36, and the number of epochs was 50.

4.3. Result Analysis

The SNR-based method [7] and the CNN-based method were also applied to the same testing set for comparison, to demonstrate the effectiveness of the CNNSA-based H s and T s estimation model. In this study, the H s and T s obtained in two continuous buoy measurement intervals were averaged using a time-moving average with the same moving time as the buoy measurement interval. The root-mean-square differences (RMSDs), correlation coefficients (CCs), and biases between the reference value and the estimated value using the SNR-based, CNN-based, and CNNSA-based methods are summarized in Table 3 and Table 4, respectively. The estimation results of the three methods for the test sample before and after the moving average are shown in Figure 8. In addition, Figure 9 shows the estimation results in the time series.
From Table 3, it can be seen that the RMSD of the proposed method is reduced by 0.21 m and 0.1 m, and the CC is improved to 0.85 when compared with the SNR-based and CNN-based methods, which is the H s estimation without averaging. Further, when the H s estimation result is processed by sliding for an average of 30 min, it can be seen from Table 3 that the H s estimation result is further improved, with an RMSD and CC of 0.3 m and 0.86, respectively, based on the CNNSA method. Similarly, Table 4 shows the T s estimation. The RMSD of the proposed method is reduced by 0.23 s and 0.09 s, and the CC is improved to 0.89 compared to the SNR-based and CNN-based methods, which is the T s estimation without averaging. The estimated results of T s were processed by sliding for an average of 30 min. As can be seen from Table 4, the estimated T s is also improved, and the RMSD and CC based on the proposed method are 0.27 s and 0.91. To visualize the estimation results of the three methods regarding H s and T s , Figure 8 shows the estimation results of H s and T s under average sliding and without average sliding of the three methods.

5. Discussion

As can be seen from Figure 8a, when Hs is 2–3 m, the estimated result of H s is small, while when H s is greater than 3 m, the estimated result of H s is larger. In Figure 8b, the SNR-based T s estimation results show a large dispersion as a whole, which is due to the sensitivity of SNR-based methods to radar image noise, resulting in a large deviation in the estimation results. It can be seen from Figure 8c,e that the CNNSA-based method has a better linear correlation in H s estimation than the CNN-based method. Similarly, in Figure 8d,f, the CNNSA-based method also produces a better linear correlation in the estimation of T s . It can be seen from Table 3 and Table 4 that the correlation coefficients of H s and T s estimation based on the CNNSA method are 0.86 and 0.91, respectively, after the 30-min moving average, both of which are higher than those of the CNN-based method and SNR-based method. Further, Figure 9a and b, respectively, show the variation trend of H s and T s estimated by the three methods over time, and it can be seen that the results estimated by the CNNSA-based method are closer to the variation trend of buoy references.
In summary, the proposed CNNSA-based regression estimation model has different degrees of improvement in the estimation of H s and T s compared to the other two methods. Figure 8 shows the degree of fit of the three methods; the proposed CNNSA-based model estimates of H s and T s are closer to the black line after applying the time-moving average. Further, the closeness of the three methods to the buoy reference is demonstrated in Figure 9, with H s and T s estimation of the proposed method being closer to the buoy references. All these results show that the proposed method improves the accuracy of H s and T s estimation using X-band marine radar images and simultaneously validates the reasonableness and effectiveness of the proposed CNNSA-based estimation model.

6. Conclusions

In this study, a deep neural network was used to estimate H s and T s in an X-band marine radar backscatter image sequence. Since the shore-based radar images in this study were interfered with by other radar radial noise lines and solid target objects, to make the convolutional neural network (CNN) extract the image features more accurately, it was necessary to pre-process the radar images to eliminate the interference. Firstly, a pre-trained GoogLeNet was used to extract multi-scale depth space features from the radar images to estimate H s and T s . Since CNN-based models cannot analyse the temporal features of wave features in radar image sequences, self-attention was connected after the deep convolutional layer of the CNN in order to construct a convolutional self-attention (CNNSA)-based model that generated spatial-temporal features for H s and T s estimation. Both models were trained and tested using collected shore-based radar data. The buoy measurements were also interpolated and used as baseline values for performance evaluation. The experimental results showed that both CNN-based and CNNSA-based models can improve the estimation results of H s and T s . The RSMD of H s estimation is reduced by 0.13 m and 0.24 m, respectively, and the RSMD of T s estimation is reduced by 0.22 s and 0.3 s, respectively, compared with the traditional SNR-based method. Overall, the CNNSA-based model showed better results for estimating H s and T s simultaneously. Therefore, the estimation accuracy is further improved by using temporal information while extracting multi-scale depth space features.
Due to the data limitations of this experiment, in future work, a collection of sea state data and radar data containing a larger number of images and a wider range for model training and validation is required, and radar data and reference data should also be collected at different locations to validate the proposed model further.

Author Contributions

Conceptualization, S.Z., D.W., X.W. and L.S.; methodology, S.Z.; software, S.Z.; validation, S.Z., S.L., Y.Z. and D.L.; formal analysis, S.Z., D.W., X.W. and L.S.; investigation, S.Z.; resources, D.W. and X.W.; data curation, S.Z., S.L., Y.Z. and D.L.; writing—original draft preparation, S.Z.; writing—review and editing, S.Z.; visualization, S.Z.; supervision, S.Z., D.W., X.W. and L.S.; project administration, S.Z., D.W., X.W. and L.S.; funding acquisition, D.W. and X.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant Nos. 62074138 and 51975104), the Natural Science Foundation of Ningbo (2022J008), the Fundamental Research Funds for the Central Universities (DUT22LAB405 and DUT22QN227) and Ningbo Institute of Dalian University of Technology.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw/processed data required to reproduce these findings cannot be shared at this time as the data are also part of an ongoing study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Huang, W.M.; Gill, E.; An, J.Q. Iterative least-squares-based wave measurement using X-band nautical radar. IET Radar Sonar Nav. 2014, 8, 853–863. [Google Scholar] [CrossRef]
  2. Young, I.R.; Rosenthal, W.; Ziemer, F. A three-dimensional analysis of marine radar images for the determination of ocean wave directionality and surface currents. J. Geophys. Res. Ocean. 1985, 90, 1049–1059. [Google Scholar] [CrossRef]
  3. Lund, B.; Graber, H.C.; Romeiser, R. Wind Retrieval From Shipborne Nautical X-Band Radar Data. IEEE Trans. Geosci. Remote Sens. 2012, 50, 3800–3811. [Google Scholar] [CrossRef]
  4. Dankert, H.; Horstmann, J.; Rosenthal, W. Ocean wind fields retrieved from radar-image sequences. J. Geophys. Res Ocean. 2003, 108, 3352. [Google Scholar] [CrossRef]
  5. Lund, B.; Graber, H.C.; Xue, J.; Romeiser, R. Analysis of Internal Wave Signatures in Marine Radar Data. IEEE Trans. Geosci. Remote Sens. 2013, 51, 4840–4852. [Google Scholar] [CrossRef]
  6. Nieto-Borge, J.C.; Hessner, K.; Jarabo-Amores, P.; de la Mata-Moya, D. Signal-to-noise ratio analysis heights from X-band marine to estimate ocean wave radar image time series. IET Radar Sonar Nav. 2008, 2, 35–41. [Google Scholar] [CrossRef]
  7. Borge, J.C.N.; Soares, C.G. Analysis of directional wave fields using X-band navigation radar. Coast. Eng. 2000, 40, 375–391. [Google Scholar] [CrossRef]
  8. Gangeskar, R. Ocean current estimated from X-band radar sea surface images. IEEE Trans. Geosci. Remote Sens. 2002, 40, 783–792. [Google Scholar] [CrossRef]
  9. Senet, C.M.; Seemann, J.; Ziemer, F. The near-surface current velocity determined from image sequences of the sea surface. IEEE Trans. Geosci. Remote Sens. 2001, 39, 492–505. [Google Scholar] [CrossRef]
  10. Shen, C.; Huang, W.; Gill, E.W.; Carrasco, R.; Horstmann, J. An Algorithm for Surface Current Retrieval from X-band Marine Radar Images. Remote Sens. 2015, 7, 7753–7767. [Google Scholar] [CrossRef]
  11. Greenwood, C.; Vogler, A.; Morrison, J.; Murray, A. The approximation of a sea surface using a shore mounted X-band radar with low grazing angle. Remote Sens. Environ. 2018, 204, 439–447. [Google Scholar] [CrossRef]
  12. Huang, W.M.; Liu, X.L.; Gill, E.W. Ocean Wind and Wave Measurements Using X-Band Marine Radar: A Comprehensive Review. Remote Sens. 2017, 9, 39. [Google Scholar] [CrossRef]
  13. Alpers, W.; Hasselmann, K. Spectral signal to clutter and thermal noise properties of ocean wave imaging synthetic aperture radars. Int. J. Remote Sens. 1982, 3, 423–446. [Google Scholar] [CrossRef]
  14. Chen, X.W.; Huang, W.M. Spatial-Temporal Convolutional Gated Recurrent Unit Network for Significant Wave Height Estimation From Shipborne Marine Radar Data. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4201711. [Google Scholar] [CrossRef]
  15. An, J.Q.; Huang, W.M.; Gill, E.W. A Self-Adaptive Wavelet-Based Algorithm for Wave Measurement Using Nautical Radar. IEEE Trans. Geosci. Remote Sens. 2015, 53, 567–577. [Google Scholar]
  16. Ma, K.; Wu, X.; Yue, X.; Wang, L.; Liu, J. Array Beamforming Algorithm for Estimating Waves and Currents From Marine X-Band Radar Image Sequences. IEEE Trans. Geosci. Remote Sens. 2017, 55, 1262–1272. [Google Scholar] [CrossRef]
  17. Navarro, W.; Velez, J.C.; Orfila, A.; Lonin, S. A Shadowing Mitigation Approach for Sea State Parameters Estimation Using X-Band Remotely Sensing Radar Data in Coastal Areas. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6292–6310. [Google Scholar] [CrossRef]
  18. Wei, Y.B.; Zheng, Y.; Lu, Z.Z. A Method for Retrieving Wave Parameters From Synthetic X-Band Marine Radar Images. IEEE Access 2020, 8, 204880–204890. [Google Scholar] [CrossRef]
  19. Gangeskar, R. An Algorithm for Estimation of Wave Height From Shadowing in X-Band Radar Sea Surface Images. IEEE Trans. Geosci. Remote Sens. 2014, 52, 3373–3381. [Google Scholar] [CrossRef]
  20. Liu, X.L.; Huang, W.M.; Gill, E.W. Wave Height Estimation from Shipborne X-Band Nautical Radar Images. J. Sens. 2016, 2016, 1078053. [Google Scholar] [CrossRef]
  21. Liu, X.L.; Huang, W.M.; Gill, E.W. Estimation of Significant Wave Height From X-Band Marine Radar Images Based on Ensemble Empirical Mode Decomposition. IEEE Geosci. Remote Sens. 2017, 14, 1740–1744. [Google Scholar] [CrossRef]
  22. Ludeno, G.; Serafino, F. Estimation of the Significant Wave Height from Marine Radar Images without External Reference. J. Mar. Sci. Eng. 2019, 7, 432. [Google Scholar] [CrossRef]
  23. Yang, Z.D.; Huang, W.M. Wave Height Estimation From X-Band Radar Data Using Variational Mode Decomposition. IEEE Geosci. Remote Sens. 2022, 19, 1505405. [Google Scholar] [CrossRef]
  24. Cornejo-Bueno, L.; Borge, J.N.; Alexandre, E.; Hessner, K.; Salcedo-Sanz, S. Accurate estimation of significant wave height with Support Vector Regression algorithms and marine radar images. Coast. Eng. 2016, 114, 233–243. [Google Scholar] [CrossRef]
  25. Park, J.; Ahn, K.; Oh, C.; Chang, Y.S. Estimation of Significant Wave Heights from X-Band Radar Using Artificial Neural Network. J. Korean Soc. Coast. Ocean Eng. 2020, 32, 561–568. [Google Scholar] [CrossRef]
  26. Kim, H.; Ahn, K.; Oh, C. Estimation of Significant Wave Heights from X-Band Radar Based on ANN Using CNN Rainfall Classifier. J. Korean Soc. Coast. Ocean Eng. 2021, 33, 101–109. [Google Scholar] [CrossRef]
  27. Duan, W.; Yang, K.; Huang, L.; Ma, X. Numerical Investigations on Wave Remote Sensing from Synthetic X-Band Radar Sea Clutter Images by Using Deep Convolutional Neural Networks. Remote Sens. 2020, 12, 1117. [Google Scholar] [CrossRef]
  28. Huang, W.M.; Yang, Z.D.; Chen, X.W. Wave Height Estimation From X-Band Nautical Radar Images Using Temporal Convolutional Network. IEEE J. Stars 2021, 14, 11395–11405. [Google Scholar] [CrossRef]
  29. Yang, Z.D.; Huang, W.M.; Chen, X.W. Evaluation and Mitigation of Rain Effect on Wave Direction and Period Estimation From X-Band Marine Radar Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 5207–5219. [Google Scholar] [CrossRef]
  30. Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; et al. Recent Advances in Convolutional Neural Networks. Pattern Recognit. 2018, 77, 354–377. [Google Scholar] [CrossRef]
  31. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  32. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  33. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
Figure 1. For calculation, Radar images under polar coordinates, buoy position (black dot), and sub-image areas (red box).
Figure 1. For calculation, Radar images under polar coordinates, buoy position (black dot), and sub-image areas (red box).
Jmse 12 00311 g001
Figure 2. Pre-processing of radar images.
Figure 2. Pre-processing of radar images.
Jmse 12 00311 g002
Figure 3. Pre-processing of radar image sub-area. (a) Contains radial noise lines and target. (b) Removed radial noise lines. (c) Removed target.
Figure 3. Pre-processing of radar image sub-area. (a) Contains radial noise lines and target. (b) Removed radial noise lines. (c) Removed target.
Jmse 12 00311 g003
Figure 4. Flowchart of the GoogLeNet-based HS and TS estimation model.
Figure 4. Flowchart of the GoogLeNet-based HS and TS estimation model.
Jmse 12 00311 g004
Figure 5. The structure of self-attention.
Figure 5. The structure of self-attention.
Jmse 12 00311 g005
Figure 6. Flowchart of the CNNSA-based estimation model for HS and TS.
Figure 6. Flowchart of the CNNSA-based estimation model for HS and TS.
Jmse 12 00311 g006
Figure 7. Synchronized anemometer-measured wind speed and buoy-measured data.
Figure 7. Synchronized anemometer-measured wind speed and buoy-measured data.
Jmse 12 00311 g007
Figure 8. Scatter plots of the buoy-measured interpolated references and the radar-derived results. Blue and red dots correspond to HS and TS, respectively. (a,b) Testing results obtained from the SNR-based method. (c,d) Testing results obtained from the CNN-based method. (e,f) Testing results obtained from the proposed CNNSA-based method.
Figure 8. Scatter plots of the buoy-measured interpolated references and the radar-derived results. Blue and red dots correspond to HS and TS, respectively. (a,b) Testing results obtained from the SNR-based method. (c,d) Testing results obtained from the CNN-based method. (e,f) Testing results obtained from the proposed CNNSA-based method.
Jmse 12 00311 g008
Figure 9. Radar-derived results in time series using three different methods, where (a) HS estimation results and (b) TS estimation results.
Figure 9. Radar-derived results in time series using three different methods, where (a) HS estimation results and (b) TS estimation results.
Jmse 12 00311 g009
Table 1. Detailed parameter information for CNNSA-based estimation models.
Table 1. Detailed parameter information for CNNSA-based estimation models.
TypeInput SizeOutput SizePatch Size/StrideFilters
Input256 × 256256 × 256 × 1
Convolution256 × 256 × 1128 × 128 × 647 × 7/264
Max pool128 × 128 × 6464 × 64 × 643 × 3/2
Convolution64 × 64 × 6464 × 64 × 641 × 1/164
Convolution64 × 64 × 6464 × 64 × 1923 × 3/1192
9 × Inception64 × 64 × 1928 × 8 × 1024
Self-attention8 × 8 × 10248 × 8 × 1024
Average pool8 × 8 × 10241 × 1 × 10247 × 7/1
linear 1 × 1 × 2
Table 2. Radar Information.
Table 2. Radar Information.
ParametersValue
Transmit frequency9.41 GHz
PolarizationHorizontal
Antenna rotation speed 22 r/min
Range resolution7.5 m
Antenna height45 m
Horizontal beam width
Azimuth coverage360°
Table 3. Comparisons of results using different methods for HS estimation.
Table 3. Comparisons of results using different methods for HS estimation.
MethodTesting
(Without Averaging)
Testing
(With Averaging)
RMSD (m)CCBiasRMSD (m)CCBias
SNR0.560.64−0.200.540.65−0.20
CNN0.450.76−0.040.410.77−0.04
CNNSA0.350.85−0.030.300.860.04
Table 4. Comparisons of results using different methods for TS estimation.
Table 4. Comparisons of results using different methods for TS estimation.
MethodTesting
(Without Averaging)
Testing
(With Averaging)
RMSD (m)CCBiasRMSD (m)CCBias
SNR0.600.650.120.570.650.11
CNN0.460.74−0.100.350.77−0.1
CNNSA0.370.890.030.270.910.03
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zuo, S.; Wang, D.; Wang, X.; Suo, L.; Liu, S.; Zhao, Y.; Liu, D. Wave Height and Period Estimation from X-Band Marine Radar Images Using Convolutional Neural Network. J. Mar. Sci. Eng. 2024, 12, 311. https://doi.org/10.3390/jmse12020311

AMA Style

Zuo S, Wang D, Wang X, Suo L, Liu S, Zhao Y, Liu D. Wave Height and Period Estimation from X-Band Marine Radar Images Using Convolutional Neural Network. Journal of Marine Science and Engineering. 2024; 12(2):311. https://doi.org/10.3390/jmse12020311

Chicago/Turabian Style

Zuo, Shaoyan, Dazhi Wang, Xiao Wang, Liujia Suo, Shuaiwu Liu, Yongqing Zhao, and Dewang Liu. 2024. "Wave Height and Period Estimation from X-Band Marine Radar Images Using Convolutional Neural Network" Journal of Marine Science and Engineering 12, no. 2: 311. https://doi.org/10.3390/jmse12020311

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop