1. Introduction
The Normalized Radar Cross Section (NRCS) is crucial in marine science and has been used in numerous studies, including oil pollution monitoring [
1], wind speed retrieval [
2], wave height estimation [
3], rainfall monitoring [
4], marine target identification and classification [
5], and image interpretation [
6]. Enhancing the NRCS resolution aids in improving spatial feature expressiveness, enabling more accurate detection and identification of surface features like water bodies and vegetation [
7], and enhancing structural characteristics of marine targets, thereby improving identification abilities. The factors constraining the resolution of radar systems are multifaceted, including constraints of radar system architecture (signal bandwidth, sampling rate, synthetic aperture length, antenna beamwidth, and receiver performance), external environmental factors (celestial interference, atmospheric interference, and sea surface conditions) and constraints of data processing techniques. Enhancing the performance of radar systems, such as sharpening antenna beams or employing arrays with tens of antenna elements, is an effective way to address low-resolution issues. However, these methods incur substantial economic costs [
8]. Therefore, it is necessary to develop more practical and cost-effective super-resolution algorithms to improve the resolution of scattering data.
Previous studies have significantly enhanced the spatial resolution of scatterometer and radiometer data through traditional numerical optimization algorithms. For instance, Lindsley utilized the Weighted Average (AVE) algorithm and the Scatterometer Image Reconstruction (SIR) algorithm to reconstruct the backscatter data measured by the MetOp satellite, achieving a resolution of 15 to 20 km, which significantly improves upon the nominal spatial resolutions of 25 and 50 km [
9]. D. G. Long employed the Backus–Gilbert Inversion (BGI) technique in combination with the SIR algorithm to enhance the spatial resolution of land and vegetation areas in data collected by the Special Sensor Microwave/Imager (SSM/I) [
10]. D. G. Long improved the resolution of the Cassini Titan Radar Mapper data by applying the AVE and SIR algorithms [
11], and further optimized the spatial resolution of SMAP data using the rSIR and BG algorithms [
12]. Additionally, D. G. Long combined the Drop-In-The-Bucket (DIB) technique with the SIR algorithm to enhance the resolution of SMAP NRCS data [
13]. Santi, E. applied the Smoothing Filter-Based Intensity Modulation (SFIM) technique to improve the spatial resolution of land-, forest-, and snow-covered area data measured by AMSR-E [
14]. J. Z. Miller achieved spatial resolution improvement for SMAP NRCS data using the SIR algorithm [
15]. Although various algorithms have been proposed for enhancing the spatial resolution of NRCS data, such as BG and SIR algorithms, they still face numerous limitations in practical applications. These algorithms are computationally complex, and require additional parameter information (e.g., antenna gain functions and acquisition geometry), which not only increases operational complexity but also restricts their broader applicability [
14]. Most importantly, these algorithms often trade off other performance aspects—such as noise levels, NRCS accuracy, and temporal resolution—when improving spatial resolution [
16].
In recent years, with the advancement of deep learning techniques, neural network-based super-resolution methods have demonstrated remarkable advantages in the field of synthetic aperture radar (SAR) image enhancement. For instance, L. G. Wang et al. proposed a generative adversarial network (GAN)-based super-resolution reconstruction method for SAR images, achieving resolution enhancement from low to high resolution [
17]. Y. H. Li et al. introduced a residual attention module (ERAM), integrating channel attention (CA) and spatial attention (SA), and implemented a SAR image super-resolution U-Net architecture leveraging this module [
18]. L. J. Bu et al. proposed a deep learning network for integrated speckle reduction and super-resolution in multi-temporal SAR (ISSMSAR), which employs two parallel subnetworks to simultaneously process dual-temporal SAR image inputs [
19]. Y. Y. Kong et al. introduced a feature extraction module combining convolutional operations with deformable multi-head self-attention (DMSA), forming a SAR image super-resolution network named DMSC-GAN [
20]. These methods primarily focus on resolution enhancement networks designed for natural images, whose pixel values are typically non-negative. Since NRCS data far exceed the range of natural image pixel values, directly applying these approaches to NRCS resolution enhancement would result in partial numerical information loss.
On the other hand, combining resolution reconstruction algorithms with data down-sampling can serve as an effective approach for data compression and reconstruction. In synthetic aperture radar (SAR) applications, while hardware advancements have enabled high-resolution data acquisition, they have concurrently exacerbated data storage and transmission challenges. To address this issue, previous researchers have proposed various methods for SAR data compression and reconstruction. For instance, Yang, D. et al. applied the Matrix Completion (MC) theory to the SAR imaging process, leveraging the low-rank property of the radar echo data matrix to reconstruct the full aperture data from undersampled measurements [
21]. Lee, S. et al. employed Compressive Sensing (CS) techniques to reconstruct SAR images [
22]. Alaa M. El-Ashkar et al. utilized CS techniques for SAR image reconstruction, successfully addressing the challenges of data storage and transmission while maintaining image quality [
23]. Slim Rouabah et al. adopted a Fast Fourier Transform (FFT)-based recovery algorithm for SAR compression and reconstruction. Their experimental results demonstrated that high-quality image recovery could be achieved with only 30% of the data volume [
24]. While these conventional approaches have been extensively applied in terrestrial monitoring, deep learning architectures are increasingly emerging as a transformative paradigm for SAR image reconstruction. Shaoman Fu et al. proposed a content-adaptive transform network and a Context-Aware Entropy Model (CAEM) to achieve SAR image compression and reconstruction [
25]. Zhixiong Di et al. introduced a compression framework based on Variational Autoencoders (VAE), which integrates pyramid features and quality enhancement techniques. This framework significantly improves the compression ratio and reconstruction quality for SAR images [
26].
Ocean remote sensing is more challenging than land observation, owing to the inherent stochasticity and hydrodynamic complexity of sea waves, for which deep learning architectures have demonstrated superior efficacy. Ma proposed a multi-task one-dimensional convolutional neural network (MT1DCNN) for the joint prediction of the type and parameters of sea clutter amplitude distribution [
27]. An improved H-YOLOv4 model is proposed for effective sea clutter region extraction [
28]. Ji established an over-the-horizon propagation loss (OHPL) prediction model by incorporating prior information into the long short-term memory (LSTM)–transformer structure (IPILT–OHPL) for efficiently predicting the OHPL in nonuniform evaporation ducts [
29]. He, H. et al. managed to forecast sea surface temperature (SST) values using the Attention-based Context Fusion Network (ACFN) [
30]. As deep learning has achieved significant advancements in enhancing image pixel resolution, some researchers have proposed high-resolution reconstruction methods based on deep learning for non-image data, such as sea surface temperature and ocean salinity. Wang applied the LightGBM algorithm to establish a high-resolution reconstruction model for sea surface salinity data [
31]. Wang proposed an implicit neural representation-based interpolation method with temporal information (T_INRI) to reconstruct high spatial resolution sea surface temperature data [
32].
Through a comprehensive review of deep learning literature, attention mechanisms have emerged as a pivotal technique for obtaining discriminative feature representations, thereby enhancing model performance. Consequently, numerous studies have adopted attention mechanisms in their implementations. Specifically, spatial attention [
33] achieves selective feature processing through spatial position focusing, while the processing strategy of spatial masking is coarse and has no intrinsic effect on modulating fine-grained channel knowledge. Channel attention [
34] can model inter-channel dependencies but is constrained by first-order statistics, failing to capture high-order feature interactions. Building upon this, researchers have proposed a dual-attention module of spatial and channel attention in both serial [
35] and parallel [
36], yet these still suffer from insufficient cross-dimensional interactions. To address this limitation, the Global Attention Mechanism (GAM) innovatively integrates an MLP-based 3D permutation channel submodule with a convolutional spatial submodule, effectively suppressing information decay while enhancing global representation capability [
37]. Meanwhile, High-Order Attention (HOA) employs a high-order polynomial predictor to model the higher-order statistics of convolutional activations, significantly improving the capture of complex feature correlations [
38].
Inspired by the successful application of deep learning techniques to oceanic phenomena by previous researchers, this paper explores the potential of a deep learning model for the enhanced-resolution reconstruction of oceanic scattering data. We propose a deep neural network that innovatively introduces a Self-Attention Feature Fusion based on the Weighted Channel Concatenation (SAFF-WCC) module, combining the Global Attention Mechanism (GAM) module and the High-Order Attention (HOA) module. The feature fusion module can automatically allocate weight proportions according to the importance of input features, effectively avoiding potential issues of feature loss or over-fusion during the feature fusion process, thereby achieving better feature fusion results and higher reconstruction quality. Furthermore, the proposed method eliminates the need for additional parameter requirements and high computational complexity inherent in traditional algorithms while maintaining temporal resolution. Moreover, the resolution reconstruction technique proposed in this study offers an alternative innovative approach to alleviating data storage and transmission pressures. Experiments were conducted using SAR data with alternating down-sampling scaling factors of 2 and 4, resulting in data sizes reduced to 1/4 and 1/16 of the original size, respectively, and significantly lowering storage and transmission demands. Additionally, we conducted comparative analyses between the proposed technique and traditional Compressed Sensing (CS) methods to further validate its effectiveness and advantages.
This paper is organized as follows:
Section 2 outlines the acquisition and processing of the scattering data from the Sentinel-1 satellite.
Section 3 elaborates on the deep learning model for high-resolution reconstruction.
Section 4 shows the ablation experiments and the comparison experiments for the proposed SAFF-WCC module, along with other related feature fusion modules. The experimental results for the enhancement and reconstruction of the NRCS data, as well as comparing the results with those from traditional methods, are also presented in this section.
Section 5 discusses the advantages and the limitations of the proposed approach and briefly describes future work.
Section 6 summarizes this paper.
2. Datasets
The satellites capable of offering NRCS products include JERS-1, ERS-1/2, RADARSAT-1/2, Envisat-1, Sentinel-1, Terra SAR-X, ALOS-2, and China’s GaoFen-3, among others. The RADARSAT Constellation Mission (RCM) consists of three identical satellites equipped with C-band SAR, covering HH, VV, HV, VH, and compact polarization modes, with spatial resolutions ranging from 5 m to 100 m [
39]. The Gaofen-3 (GF-3) satellite is China’s first civilian quad-polarized C-band imaging microwave satellite that can provide high-resolution marine and land observations. It operates in 12 different imaging modes. Among these, there are two full-polarization measurement modes: the QPSI mode with a resolution of 8 m, a swath width of 30 km, and an incidence angle range of 20 to 50 degrees; and the QPSII mode with a resolution of 25 m, a swath width of 45 km, and an incidence angle range of 19 to 50 degrees [
40]. Sentinel-1 consists of two polar-orbiting satellites, each carrying an imaging C-band SAR to provide long-term radar backscatter data [
41]. Equipped with active phased array antennas, it achieves high-precision beam pointing in both elevation and azimuth angles, enabling high flexibility in data acquisition [
42]. Sentinel-1′s SAR has four imaging modes: strip map, interferometric wide swath (IW), extra-wide swath, and wave mode (WV). Core products are provided at Levels 0, 1, and 2, respectively. In this paper, VV- and VH-polarized data from the high-resolution Level 1 ground range detected product of the IW mode are used. The resolution is 20 m × 22 m, with a swath width of 250 km [
43].
2.1. Sentinel-1 Data Extraction
The Sentinel Application Platform (SNAP) version 9.0.0 is software provided by the European Space Agency (ESA) for preprocessing Sentinel-1 data [
44]. SNAP offers two primary operational modes: the Graphical User Interface (GUI) and the Graph Processing Tool (GPT). While the GUI is user-friendly and convenient, it is less efficient compared to the GPT tool. Therefore, this study utilizes the GPT mode for data processing. As illustrated in
Figure 1, the entire processing workflow consists of eight steps: data input, thermal noise removal, orbit file correction, radiometric calibration, noise filtering, Doppler terrain correction, conversion to decibel scale, and scattering data output.
After preprocessing the data using SNAP, an output file in the NetCDF format can be obtained. Since the size of the NRCS data in a single file is approximately 24,000 × 35,000, which is not suitable for direct input into deep learning models, the data requires further processing. As shown in
Figure 2, the scattering data is gridded and divided into sections of 1024 × 1024.
In this paper, the focus of the high-resolution reconstruction is primarily on oceanic areas. Due to the presence of uncertain factors, such as cargo and cruise ships, it is necessary to remove data from large land areas to avoid negatively impacting the reconstruction of objects like ships within the ocean. Simultaneously, it is important to retain parts of small islands within the ocean to enhance the number and feature diversity of non-sea-surface targets.
Figure 3a shows the scattering data image of a specific area in the Yellow Sea at 09:47:42 on 14 January 2021.
Figure 3b displays the scattering data image of the ocean area after removing large land areas, using the gridding method, and reassembling the remaining sections.
China’s offshore seas can be divided into five main parts, from north to south: the Bohai Sea, Yellow Sea, East China Sea, South China Sea, and the Taiwan Strait, along with the eastern Pacific Ocean. In this section, based on their specific locations, these five representative marine areas are selected as the subjects for high-resolution reconstruction studies of the ocean scattering data. Specific longitude and latitude data can be found in reference [
45].
Table 1 presents the longitude and latitude information for each marine area, as well as the division of the training and validation sets for the corresponding Sentinel-1 scattering data based on the marine area information.
2.2. Data Processing for Deep Learning Model
Supervised learning is used in this paper to train the deep learning models. The input data, which has a lower resolution than the output, is processed by alternating downsampling to reduce its resolution. The loss value is then calculated by comparing the prediction results with the original data, which is not subject to resolution reduction, to train the deep learning models.
Due to the relatively small proportion of scattering data from individual objects (such as ships) in the overall sea surface scattering data, the model struggles to effectively extract features from this subset, resulting in the loss of localized details. To improve the prediction of high-resolution data from low-resolution data, while avoiding the loss of edge data during feature extraction, preprocessing is applied based on the significant differences in scattering characteristics between the sea surface and targets. The Sentinel-1 scattering data is processed using SNAP software version 9.0.0, then gridded and divided into 1024 × 1024-pixel blocks. Typically, the backscattered signal energy from the ocean surface is significantly lower than that reflected from target objects. Based on this characteristic, the backscattering data can be binarized to separate edge data from non-edge data, thereby further distinguishing small targets from sea surface features. For example, the structural contour characteristics of small targets are preserved in the edge data, enabling the model to focus more on learning the feature information of small targets during edge data branch training. Additionally, to enhance the distinguishability between targets and the sea surface, the preprocessing stage normalizes the backscattering data to a range of 0–255, thereby amplifying the contrast between the target scattering data (higher values) and the sea surface scattering data (lower values), as shown in
Figure 4. (Note: the 1024 × 1024 resolution data is used here as an example, and downsampled resolutions should be used for actual model training.)
6. Conclusions
This paper proposes a deep learning model for enhancing and reconstructing the spatial resolution of Sentinel-1 backscattering data from the Bohai Sea, Yellow Sea, East China Sea, Taiwan Strait, and South China Sea. The SNAP GPT tool is employed to preprocess the Sentinel-1 data, and the decibel-processed data undergo gridding segmentation to extract effective data for specific sea areas. The extracted data is then downsampled according to different scale factors and binarized to obtain edge and non-edge data, which serve as inputs for the proposed deep learning model. The deep learning model innovatively incorporates a self-attention feature fusion based on the WCC (SAFF-WCC) module, combined with a Global Attention Mechanism (GAM) and a High-Order Attention (HOA) module, which significantly improves the efficiency of multi-feature integration. To evaluate the effectiveness of the proposed model, four key metrics were employed: Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index (SSIM), Multi-Scale Structural Similarity Index (MS-SSIM), and Mean Absolute Percentage Error (MAPE). Specifically, we conducted experiments on the scattering data from a marine target in the Bohai Sea to enhance its spatial resolution by factors of 2 and 4. Additionally, we reconstructed the NRCS data for the East China Sea under sea states ranging from levels 2 to 7 and used the NRCS data under sea state level 3 as an example for comparative analysis with traditional reconstruction algorithms. The results demonstrate that for the scattering data on ocean surface targets, the proposed model effectively enhances the fine structural features of the targets when the resolution is doubled. For high-resolution NRCS reconstruction under different sea states, the proposed method outperforms the traditional approaches across all four evaluation metrics under a scaling factor of 2. It should be noted that China’s coastal areas exhibit three distinct climate types: temperate (covering the Bohai Sea, the Yellow Sea, and northern portions of the East China Sea), subtropical (the southern East China Sea), and tropical (the South China Sea). Since our dataset exclusively focuses on these mid- to low-latitude regions, polar zones are not included, and any generalization to polar regions requires further validation.
In future work, we aim to further optimize the model by integrating the scattering model, adjusting the network architecture, and balancing the proportion of the NRCS data for different sea states in the training dataset. These efforts will further improve the model’s ability to enhance the structural features of marine targets and increase the reconstruction accuracy of the NRCS data under high sea state conditions.