Next Article in Journal
Towards a New MAX-DOAS Measurement Site in the Po Valley: Aerosol Optical Depth and NO2 Tropospheric VCDs
Previous Article in Journal
Spatial Resolution Enhancement of Microwave Radiation Imager (MWRI) Data
Previous Article in Special Issue
Parcel-Based Sugarcane Mapping Using Smoothed Sentinel-1 Time Series Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Paddy Rice Mapping in Hainan Island Using Time-Series Sentinel-1 SAR Data and Deep Learning

by
Guozhuang Shen
1,2,3,* and
Jingjuan Liao
1,2,3
1
Key Laboratory of Earth Observation of Hainan Province, Hainan Aerospace Information Research Institute, Sanya 572029, China
2
International Research Center of Big Data for Sustainable Development Goals, Beijing 100094, China
3
Key Laboratory of Digital Earth Science, Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100094, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(6), 1033; https://doi.org/10.3390/rs17061033
Submission received: 8 February 2025 / Revised: 11 March 2025 / Accepted: 13 March 2025 / Published: 15 March 2025
(This article belongs to the Special Issue Radar Remote Sensing for Monitoring Agricultural Management)

Abstract

:
Rice serves as a fundamental staple food for a significant portion of the global population, and accurate monitoring of paddy rice cultivation is essential for achieving Sustainable Development Goal (SDG) 2–Zero Hunger. This study proposed two models, RiceLSTM and RiceTS, designed for the precise extraction of paddy rice areas in Hainan Island using time-series Synthetic Aperture Radar (SAR) data. The RiceLSTM model leverages a Bidirectional Long Short-Term Memory (BiLSTM) network to capture temporal variations in SAR backscatter and integrates an attention mechanism to enhance sensitivity to paddy rice phenological changes. This model achieves classification accuracies of 0.9182 and 0.9245 for early and late paddy rice, respectively. The RiceTS model extends RiceLSTM by incorporating a U-Net architecture with MobileNetV2 as its backbone, further improving the classification performance, with accuracies of 0.9656 and 0.9808 for early and late paddy rice, respectively. This enhancement highlights the model’s capability to effectively integrate both spatial and temporal features, leading to more precise paddy rice mapping. To assess the model’s generalizability, the RiceTS model was applied to map paddy rice distributions for the years 2020 and 2023. The results demonstrate strong spatial and temporal transferability, confirming the model’s adaptability across varying environmental conditions. Additionally, the extracted rice distribution patterns exhibit high consistency with statistical data, further validating the model’s effectiveness in accurately delineating paddy rice areas. This study provides a robust and reliable approach for paddy rice mapping, particularly in regions that are characterized by frequent cloud cover and heavy rainfall, where optical remote sensing is often limited.

1. Introduction

Ensuring a sufficient food supply is a fundamental objective of the United Nations’ 2030 Agenda for Sustainable Development Goals (SDGs). As a major crop, rice plays a key role in global food security, serving as the primary staple food for more than 50% of the world’s population [1], particularly in Asia [2]. Over 90% of the world’s rice is produced in the Asia–Pacific Region, with major production centers in East Asia, South Asia, and Southeast Asia [3]. However, rapid urbanization and industrialization introduce uncertainty in paddy rice cultivation, alongside other influential factors such as weather conditions, water availability, paddy rice diseases and pests, agricultural practices, and economic policies [4]. Concurrently, the International Food Policy Research Institute (IFPRI) reports that the rice demand is growing at an annual rate of 1.8% [5]. Given these challenges, the timely and accurate monitoring of paddy rice planting areas, growth dynamics, and yield estimation are essential for informed agricultural policymaking and the realization of SDG 2–Zero Hunger.
The traditional methods for large-scale paddy rice monitoring are time-consuming and labor-intensive. Remote sensing technology has become a key tool for regional and global crop mapping, offering broad spatial coverage, frequent data acquisition, and reduced observation costs. It provides precise and timely insights into crop phenology and development [3,6]. Several studies have been conducted on the use of coarser- to finer-resolution optical remote data for crop mapping [7] and paddy rice mapping [8,9], mainly including MODIS, Landsat, and Sentinel-2 data. Time-series Landsat-8 OLI data, coupled with a deep learning (DL) model, FR-Net, successfully mapped paddy rice in the main rice-producing area of northern China, demonstrating the significant potential of Landsat data in agricultural remote sensing [10]. Similarly, 30 m harmonized Landsat and Sentinel-2 surface reflectance data, processed using the Feature Selection and Hierarchical Classification (FSHC) method, were applied for paddy rice mapping in South China [11].
However, optical remote sensing faces challenges in paddy rice regions, which are frequently cloudy and rainy, limiting effective data acquisition during critical growth stages such as transplanting and flooding. Synthetic Aperture Radar (SAR) overcomes these limitations by offering all-weather, day-and-night imaging [12]. SAR data are highly sensitive to the crop structure and dielectric properties [13], making them ideal for monitoring growth stages and crop development. By analyzing backscatter variations, SAR time-series data can track the entire rice growth cycle [14], distinguishing flooded fields (low backscatter during transplanting) from dry fields (higher backscatter) [15]. SAR-based paddy rice mapping has been extensively studied using sensors such as ENVISAT/ASAR [16,17,18,19,20,21], ALOS/PALSAR [22], RADARSAT/RADARSAT-2 [13,23,24], and TerraSAR-X [25,26,27]. The launch of Sentinel-1, a C-band SAR satellite constellation, has further enhanced the capability for large-scale paddy rice monitoring, providing high-temporal-resolution data every six days (Sentinel-1A and 1B) [28], making large-area paddy rice plant area mapping feasible [5,29]. Consequently, long-term Sentinel-1 SAR time-series data have been widely adopted for paddy rice mapping [5,30,31,32] and demonstrated as a stable SAR data source.
Accurate crop mapping, monitoring, and management require the extraction of spatial–temporal phenological information [6,33]. SAR-based paddy rice mapping typically utilizes backscatter characteristics, which capture both spatial and temporal features as key indicators [4,31,32,34]. Spatial features include texture, shape, and contextual information [35], while temporal features reflect backscatter variations associated with growth-stage-specific physiological changes [36]. The distinct backscatter variation patterns of rice fields differentiate them from other land cover types [35]. For high-temporal-resolution SAR data, features of paddy rice backscatter variation curves—such as maximum, minimum, mean, variance, curvature, and intervals between inflection points—can be extracted and used to construct phenological indicators [15]. These indicators facilitate paddy rice identification through thresholding, curve similarity methods [15,37], or the derived robust SAR-based Paddy Rice Index (SPRI) [38].
Traditional machine learning methods such as K-Nearest Neighbor (KNN), Decision Tree (DT), Support Vector Machine (SVM), and Random Forest (RF) have been applied in paddy rice mapping [12]. However, these approaches often struggle to fully exploit spatial–temporal information. Since the introduction of DL by Hinton et al. [39] in 2006, DL-based models have demonstrated superior performance in remote sensing crop identification. Convolutional Neural Networks (CNNs), particularly Fully Convolutional Neural Network (FCN), U-Net [40], DeepLabv3+, PSPNet, and SegNet, have been widely employed for paddy rice segmentation, achieving higher classification accuracy [36,41]. However, CNN-based models primarily capture spatial patterns and are not optimized for time-series data [42]. To address this limitation, models such as 1D-CNN, Recurrent Neural Network (RNN) [43,44], Long Short-Term Memory (LSTM) [45,46,47], and BiLSTM [48,49] have been explored for time-series SAR-based mapping, with BiLSTM achieving the highest accuracy [12]. However, these methods do not fully exploit spatial information. Recently, 3D-CNN, ConvLSTM, and Transformer-based models have been developed to leverage both spatial and temporal variations for improved remote sensing image semantic segmentation [50,51], especially paddy rice mapping [52,53,54].
In recent years, some high-resolution paddy rice distribution maps have also been constructed, leveraging optical and SAR data for different regions, including Indonesia [46], Thailand [55], Nigeria [56], China [3,57,58], especially Northeast China [59], South China [60], and the Asian monsoon region [61]. Several studies have already focused on paddy rice mapping in Hainan Island. These studies utilized machine learning and DL methods to identify and delineate paddy rice fields [51,62,63]. However, these products can be further improved to fulfill the monitoring requirements for small and irregular ploughlands due to their limited spatial coverage and low spatial resolution.
Here, we proposed a paddy rice extraction and mapping framework using long-time-series SAR data for Hainan Island. BiLSTM- and LSTM-UNet-based models (named RiceLSTM and RiceTS) were proposed. The RiceLSTM model only uses the temporal information of SAR data, while the RiceTS model uses both the temporal and spatial information of SAR data. The remainder of this paper is organized as follows. Section 2 describes the study area and the dataset used in this study. Section 3 outlines the data processing workflow, as well as the proposed RiceLSTM and RiceTS models with their implementation. Section 4 presents the results. Section 5 discusses the influencing factors. Section 6 concludes the paper.

2. Study Area and Materials

2.1. Study Area

Hainan Island, located in the southernmost part of China (18°8′36.79″~20°9′48.71″N, 108°35′27.90″~111°4′16.78″E), covers an area of approximately 33,920 km 2 (Figure 1). The island’s terrain exhibits a distinct elevational gradient, with higher elevations concentrated in the central region and progressively decreasing towards the periphery. The Five-Finger Mountain (1867 m) and Parrot Mountain (1811 m), both located near the island’s center, represent the highest peaks. Moving outward from the central highlands, the landscape transitions through a series of descending mountains, hills, terraces, and plains, forming a concentric stratified landform (Figure 1c). Due to this topographic pattern, paddy rice cultivation is predominantly concentrated in the low-elevation peripheral regions of the island. Hainan Island encompasses diverse land cover types, including agricultural fields, urban areas, wetlands, grasslands, water bodies, and forests. The island’s unique geographical and ecological characteristics provide an optimal environment for agricultural production, particularly for paddy rice cultivation.
Hainan Island experiences a tropical monsoon climate, characterized by high temperatures, abundant sunshine, and ample rainfall throughout the year. These climatic conditions create a favorable environment for rice cultivation, supporting multiple cropping cycles. Additionally, the island’s fertile and diverse soil types supply essential nutrients for rice growth, further enhancing its agricultural productivity. Due to these advantages, Hainan Island has become one of China’s key rice-producing regions, capable of sustaining two or even three rice-growing seasons annually. The early paddy rice season extends from March to early July, while the late paddy rice season lasts from late July to late November. There is a slight difference in planting areas between the two seasons, with the late rice area typically exceeding that of the early crop [64].

2.2. Data

2.2.1. SAR Data

The frequent rainy and cloudy weather in Hainan Island limits the availability of high-quality cloud-free optical images from non-commercial platforms such as Landsat or Sentinel-2, making it challenging to obtain sufficient data for paddy rice mapping. To address this, 150 scenes of Level-1 Ground-Range-Detected (GRD) Sentinel-1A C-band (5.405 GHz) SAR data from the Interferometric Wide (IW) swath mode and VV/VH polarization were collected in 2020, along with 134 scenes in 2023, from Copernicus Open Access Hub (https://dataspace.copernicus.eu/, accessed on 18 April 2024). After Sentinel-1 slice assembly, a total of 90 effective observations were obtained for 2020 due to the assembly of two consecutive slices, resulting in 30 observations per orbit for relative orbits 157, 55, and 84 (Table A1). For 2023, 84 effective observations were obtained, with 30, 25, and 29 observations corresponding to relative orbits 157, 55, and 84, respectively (Table A1). All 174 observations were acquired in ascending mode. From Figure 2, we can see that the SAR data are primarily concentrated in the rainy season, aligning with the paddy rice planting period. Hainan Island is predominantly covered by Sentinel-1 SAR data from relative orbit 157 (~90%), while the eastern and western regions are covered by relative orbits 55 and 84, respectively (Figure 1). In 2020, data from all three relative orbits were relatively complete, with only one missing acquisition from orbit 84 (Figure 2). However, in 2023, data gaps were more prominent, with five missing acquisitions from orbit 55, one from orbit 84, and one from orbit 157 (Figure 2). These data gaps pose certain challenges for time-series SAR-based paddy rice mapping. The next section will describe the methods used to address these gaps.

2.2.2. Auxiliary Data

In this study, high-resolution Google Earth satellite images (Level 17, ~2 m ) were also collected (Figure 1d) and used to identify the land cover types and the paddy rice fields for training. Meanwhile, the 10-meter-resolution paddy rice planting distribution map for Hainan Province in 2020, obtained from the Geographic Remote Sensing Ecological Network platform, was collected and utilized for training sample generation. The paddy rice mapping process followed a pixel classification strategy based on the planting and growth patterns of paddy rice in Hainan Province, achieving an overall accuracy of 85% ( ± 2 % ) in 2020 (Figure 1e). Furthermore, a 10-meter-resolution land cover dataset (AIEC, https://engine-aiearth.aliyun.com/#/dataset/DAMO_AIE_CHINA_LC (accessed on 28 May 2024)) developed by DAMO AI Earth team was collected and used for training sample generation. This dataset integrates Sentinel-2 optical imagery and DL methods for detailed land cover classification.

3. Methodology

The flowchart of this study is presented in Figure 3. The workflow contains three main steps: SAR image preprocessing, paddy rice extraction using RiceLSTM and RiceTS models, and accuracy assessment. In the first step, long-time-series Sentinel-1 SAR σ 0 images are generated using the SARProcMod module. In the second step, two paddy rice mapping models, RiceLSTM and RiceTS, are proposed. To assess the performance of these models, artificial samples are selected for the confusion matrix and F1 score.

3.1. Data Preprocessing

Although VH polarization is more sensitive to rice growth stages, particularly during the flooding period [47], both VV and VH polarization data were utilized in this study to capture additional backscatter variations across non-rice land covers. The SARProcMod module proposed in [28] was used to preprocess Sentinel-1 SAR dataset (Figure 1c). The workflow included nine steps:
  • Thermal noise removal was applied to all Sentinel-1 SAR data.
  • Slice assembly was performed on consecutive slices when possible.
  • Orbit files were applied to update the orbit state vectors in the abstract metadata.
  • Radiometric calibration was conducted to generate radar backscatter bands ( σ 0 ).
  • A Refined Lee filter was used to suppress speckle noise.
  • Multi-looking was applied to ensure square pixels.
  • Orthorectification of the radar backscatter bands was performed using the Range Doppler Terrain Correction algorithm with SRTM DEM data (spatial resolution: 3 )
  • The backscatter coefficient (in dB) was computed from the orthorectified radar backscatter band using the equation 10 × log 10 ( σ 0 ) .
  • All Sentinel-1 SAR data were clipped and reprojected to a uniform UTM49N coordinate system, ensuring alignment with the study area extent.
Finally, the 10 m resolution Sentinel-1 SAR dataset was generated for paddy rice mapping. Additionally, Google Earth imagery, paddy rice planting distribution products, and land cover data underwent cropping, reprojection, and resampling to ensure spatial alignment with the Sentinel-1 SAR dataset.

3.2. RiceLSTM Model

The complex rice planting patterns and the heterogeneous paddy field environment pose significant challenges to accurately mapping paddy rice distribution using traditional methods. LSTM, known as a type of RNN, can effectively capture the temporal dependencies in time-series data [65]. BiLSTM, an extension of LSTM, processes sequences in both forward and backward directions, allowing it to capture contextual information more effectively [66]. Unlike Transformer, which relies on self-attention mechanisms and requires large datasets and high computational resources for effective training, LSTM and BiLSTM offer a more efficient and robust solution for capturing temporal dependencies in time-series SAR data [67], making them suitable for paddy rice mapping. The proposed RiceLSTM model, illustrated in Figure 4, consists of two BiLSTM layers, each containing forward and backward layers (with two layers per direction) and a hidden size of 64, followed by an attention layer and a fully connected layer. The attention layer was introduced to make the model focus on key features and critical time steps, thereby enhancing its accuracy, interpretability, and robustness. This RiceLSTM model processes an input tensor with dimensions 256 × 256 × 15 × 2 , where 256 × 256 represents the spatial dimensions of the image patches, 15 represents the number of observations, and 2 corresponds to the backscatter coefficient values from VV and VH polarization. The tensor was first reshaped into 65 , 536 × 15 × 2 and fed into the BiLSTM layers, producing an output of size 128 × 15 . The output is processed through an attention layer with 128 ( 64 × 2 ) neurons, followed by a final dense layer, which determines the paddy rice probability for each pixel. Finally, the output is reshaped back to 256 × 256 to match the input image dimensions. By analyzing the temporal variation of the backscatter coefficient of paddy rice and other land cover types, the RiceLSTM model effectively extracts key temporal features that differentiate paddy rice from non-rice areas.

3.3. RiceTS Model

Although the proposed RiceLSTM model effectively captures the temporal dependencies in time-series data, it has inherent limitations. Specifically, RiceLSTM operates at the pixel level, thereby neglecting the spatial features present in remote sensing imagery. To address this limitation and fully exploit the spatial features of time-series SAR data, we introduce RiceMU (black dashed box in Figure 5), a Deep Convolutional Neural Network (DCNN) model designed for spatial feature extraction. RiceMU adopts the U-Net architecture [68], which features an encoder–decoder structure with skip connections to preserve fine-grained spatial details. The backbone of the U-Net model in RiceMU is MobileNetV2 [69], a lightweight CNN designed for computational efficiency. MobileNetV2 leverages depthwise separable convolutions to significantly reduce parameters and computational costs while maintaining strong feature extraction capabilities. Additionally, it introduces inverted residuals and linear bottlenecks, which enhance feature expressiveness and prevent information loss during non-linear transformations. The combination of U-Net and MobileNetV2 enables RiceMU to efficiently process SAR image patches and extract fine-grained spatial features for paddy rice mapping. RiceMU processes an input tensor of dimensions 256 × 256 × d , where 256 × 256 represents the spatial dimensions of the image patches, and d (in this study of 30 ( 15 × 2 )) denotes the number of feature channels. The model undergoes five levels of downsampling through convolutional layers to capture hierarchical spatial features, followed by upsampling operations to restore spatial resolution while retaining essential details. The final output is a 256 × 256 × 2 probability map, where each pixel is assigned a likelihood of being classified as paddy rice.
To fully exploit and utilize the temporal and spatial features embedded in time-series SAR data, many DL models integrating LSTM and CNN have been proposed, employing different sequences for extracting spatial and temporal features. Here, we proposed a RiceTS model, which first extracts temporal features using RiceLSTM layer and then spatial features afterward using RiceMU layer. The proposed RiceTS model, illustrated in Figure 5, consists of two main components:
  • RiceLSTM layer (temporal feature extraction): The first stage of RiceTS employs RiceLSTM model as a layer to capture temporal dependencies in time-series SAR data. The workflow of the RiceLSTM layer follows that of the RiceLSTM model but generates a feature representation of size 65,536 ×   64 .
  • RiceMU layer (spatial feature extraction): The temporal feature output from the RiceLSTM layer is reshaped into a spatial format of 256 × 256 × 64 and passed into the RiceMU layer. This step allows the spatial encoder–decoder network to further refine the feature representation and perform detailed segmentation of paddy rice regions.
The final output of RiceTS model has a shape of 256 × 256 × 2 , which represents the probability distribution of paddy rice presence at each pixel. By separating spatial and temporal feature extraction into distinct components, RiceTS allows for a comprehensive analysis of their relative importance in paddy rice mapping. Furthermore, RiceMU can be independently applied to assess the effectiveness of spatial feature extraction alone, enabling a comparative evaluation of spatial and temporal feature contributions in SAR-based rice area extraction.

3.4. Training Sample Generation

The training samples for the RiceLSTM and RiceTS models were labeled at the pixel level, with attribute values assigned as 1 for paddy rice and 0 for non-rice. Firstly, all Sentinel-1 SAR images from 2020 with the relative orbit of 155, along with Google Earth image (Figure 1d), and the paddy rice distribution map (Figure 1e) were stacked together to aid in generating training and validation samples. Through visual analysis of high-resolution optical images and the paddy rice distribution map, 490 paddy rice field polygons were delineated within the study area. Considering the 10 m resolution of the imagery, patches were cropped to a size of 256 × 256 using a sliding window to better capture spatial features of paddy rice and non-rice for model training. However, the total paddy rice planting area in Hainan Island is approximately 1.10 × 105 h a , accounting for only 3.3% of the island’s total area (3.392 × 106 h a ). Moreover, not all paddy rice fields were included during sample generation, further exacerbating the issue of severe class imbalance. To address this, patches with fewer than 100 paddy rice pixels were discarded, resulting in a final set of 896 sample patches. For the RiceLSTM model, 973,124 pixels for paddy rice and 2,000,000 non-rice pixels were randomly selected from the 896 patches, yielding a total of 2,973,124 samples. These samples were split into 80% for training and 20% for validation. For the RiceTS model, all 896 patches were used, maintaining the same 80%–20% split for training and validation. To enhance sample diversity and variability, data augmentation techniques such as random vertical and horizontal flip were applied during training. To mitigate class imbalance, a higher weight was assigned to the paddy rice class in the loss function, penalizing the model more for misclassifications of minority samples. The weights for the non-rice and paddy rice were defined as C R i c e / ( C N o n + C R i c e ) , C N o n / ( C N o n + C R i c e ) , where C R i c e and C N o n represent the number of rice and non-rice pixels in the patches, respectively.

3.5. Validation Metrics

The evaluation of a paddy rice extraction model involves the application of various statistical metrics. In this study, Accuracy, Precision, Recall, F1 score, and Cohen’s Kappa coefficient (k) were utilized to assess model performance (Equations (1)–(6)). Here, T P , F P , T N , and F N represent true positive, false positive, true negative, and false negative, respectively.
A c c u r a c y = T N + T P T N + F P + F N + T P
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
k = A c c u r a c y P e 1 P e
where P e = ( T P + F N ) ( T P + F P ) + ( T N + F N ) ( T N + F P ) ( T P + T N + F P + F N ) 2

3.6. Experimental Settings

These two models were implemented using Python 3.10, Pytorch 2.2, and CUDA 12.1. Training and evaluation were conducted on a workstation equipped with NVIDIA RTX 8000 GPU (48 GB memory), two Intel(R) Xeon(R) Gold 6134 CPUs, 512 GB memory, a 512 GB SSD, and an 8 TB HDD. Through trial and error in the training phase, batch sizes of 65,536, 12, and 4 were determined for RiceLSTM, RiceMU, and RiceTS models, respectively. The models were trained for 200 epochs with an early stopping mechanism set at a patience of 20 epochs to prevent overfitting. We employed the CrossEntropy loss function, the Adam optimizer with a constant learning rate of 0.001, and used accuracy as the evaluation function. The most optimal model was ultimately preserved. For post-processing, a threshold value of 0.5 was applied to convert probability values into binary classifications.

4. Results

4.1. The Backscatter Variations for Different Land Cover Types

After preprocessing all the Sentinel-1 SAR data, the SAR images from 2020 were utilized to analyze the temporal variations in backscatter for paddy rice and non-rice crop types. Five land cover types—paddy rice, buildings, water, dryland, and forest—were selected for time-series curve analysis, with a total of 20 polygons chosen as representative samples. The VV and VH backscatter coefficients for each land cover type were averaged at each time step, yielding 30 pairs of values, as shown in Figure 6 and Figure 7. These figures indicate that the temporal backscatter distribution patterns differ significantly among the land cover types. Regardless of the land cover type, the backscatter in VH polarization is consistently lower than that in VV polarization (Figure 6 and Figure 7). Water bodies exhibit lower backscatter due to specular reflection of radar signals (Figure 6), while buildings demonstrate stronger backscatter as a result of double scattering, and forests exhibit high backscatter due to volume scattering (Figure 6). These three land cover types display relatively stable backscatter over time, indicating their temporal consistency (Figure 7). In contrast, paddy rice fields and drylands exhibit lower backscatter with greater temporal variation, indicating their dynamic seasonal changes (Figure 7).
In Figure 6, the backscatter signatures of paddy rice and dryland vegetation exhibit a significant degree of overlap. However, Figure 7 clearly reveals distinct differences in their temporal variations. For paddy rice, two distinct planting cycles are observed, separated by early July, whereas dryland vegetation does not exhibit such cyclic behavior. As paddy rice grows, backscatter increases, reaching its peak between late May and early June. During the maturity stage, backscatter gradually decreases, followed by a sharp decline during the harvesting period in early July. Subsequently, the growth cycle of late paddy rice begins with an agricultural flooding period in mid-July, during which the backscatter of paddy rice closely resembles that of water. Figure 6 and Figure 7 demonstrate that time-series Sentinel-1 SAR data are effective for paddy rice mapping. To better capture the distinct phenological patterns, the 30 observation datasets were divided into two groups, each containing 15 observations, corresponding to the growth cycles of early and late paddy rice.

4.2. Comparison Between RiceLSTM and RiceTS

To evaluate and compare the RiceLSTM (with and without an attention layer) and RiceTS models, the training and validation samples in 2020 were used. The five evaluation metrics outlined in Section 3.5 were employed for quantitative assessment. The RiceLSTM-Attention model was configured with the same BiLSTM layers and parameter settings as the RiceLSTM (without attention) model. The RiceMU model was also configured to evaluate the significance of spatial features. By utilizing stacked training samples as an input (Table A2), it inherently incorporates temporal variations in SAR imagery. This enables RiceMU to exploit both spatial and temporal characteristics, offering a complementary perspective for assessing paddy rice classification performance. For the RiceTS model, the temporal feature extraction layer (BiLSTM layer) was configured identically to that of the RiceLSTM-Attention model, while the spatiotemporal feature extraction layer (RiceMU layer) was designed following the architecture of the RiceMU model.
The quantitative results are presented in Table 1 and Figure 8. The RiceLSTM-Attention model achieved accuracies of 0.9182 ( + 2.1 % ) and 0.9245 ( + 1.8 % ) for early and late paddy rice, respectively, demonstrating a notable improvement over the RiceLSTM model (0.8971 and 0.9067). The results indicate that the attention mechanism significantly enhances the performance of RiceLSTM across all the metrics, particularly in capturing subtle variations in paddy rice field dynamics. The RiceMU model demonstrated competitive performance, achieving high classification accuracy (0.9642 for early rice and 0.9604 for late rice), with recall values exceeding 0.9500. However, its precision was slightly lower than that of RiceTS, particularly for late rice (0.7474), indicating a tendency to classify non-rice areas as rice. The RiceTS model further improves in classification accuracy, achieving 0.9656 ( + 4.74 % ) and 0.9808 ( + 5.63 % ) for early and late paddy rice, respectively, surpassing the RiceLSTM-Attention and RiceMU models. However, the RiceTS model exhibits lower precision (0.7512 for early and 0.8465 for late paddy rice) compared to the RiceLSTM-Attention model, indicating a higher rate of false positives. This overestimation may result from similarities in backscatter patterns between paddy rice fields and non-rice land cover types, leading to the misclassification of non-rice regions as paddy rice fields.
Figure 9 presents the classification results of the proposed models for the test area. For test area (1), the RiceLSTM-Attention, RiceMU, and RiceTS models achieved favorable classification performance, effectively capturing the distribution of paddy rice fields. However, a significant omission of paddy rice is observed in the B1 region (blue box B1) in the paddy rice product (Figure 9b), whereas the RiceLSTM, RiceMU, and RiceTS models successfully identify paddy rice in this area. Both the paddy rice distribution product (Figure 9b) and the RiceLSTM model result (Figure 9c) exhibit salt-and-pepper noise due to pixel-based classifications. The paddy rice distribution product, generated using Sentinel-2 optical imagery, contains salt-and-pepper noise, while the RiceLSTM-Attention model result, derived from SAR data, exhibits even more pronounced noise (blue box B2). For test area (2), the RiceLSTM-Attention model struggles to effectively extract the spatial distribution of paddy rice, whereas the RiceMU and RiceTS models achieve more accurate extraction, highlighting the importance of spatial features in paddy rice mapping. Additionally, the RiceTS model demonstrates superior continuity in the extracted results compared to RiceMU, further highlighting the significance of temporal features. Across both test areas, the RiceTS model effectively suppresses speckle noise, improving overall classification accuracy. However, this suppression comes at the cost of losing some fine details as certain roads or ridges are misclassified as rice (blue box), leading to a reduction in precision. Nevertheless, the extracted paddy rice boundaries from the RiceLSTM-Attention and RiceTS models closely align with the actual remote sensing imagery.
These results indicate that the RiceTS model excels in capturing the temporal and spatial characteristics of rice growth, resulting in improved accuracy in paddy rice mapping. Therefore, the RiceTS model will be adopted for paddy rice mapping and analysis in the subsequent sections.

4.3. Paddy Rice Distribution Map

Figure 10 illustrates the early and late paddy rice distribution map in 2020, as derived from the proposed RiceTS model. According to the model, the early and late paddy rice areas were 1.123 × 105 h a and 9.037 × 104 h a , respectively, resulting in a total of 2.027 × 105 h a in 2020. In comparison, statistical data from the Hainan Statistic Yearbook (2021) [70], published by China Statistics Press, reported early and late paddy rice areas of 1.100 × 105 h a and 1.284 × 105 h a , respectively, with a total of 2.384 × 105 h a . The extracted annual paddy rice planting area achieved an accuracy of 90.6%, demonstrating good consistency with the official statistics.
The differences between the classified paddy rice area and the statistical data were 2.3 × 103 h a (2.1%) for early paddy rice and 3.80 × 104 h a (29.6%) for late paddy rice. The relatively lower extraction accuracy for late rice is primarily due to the training samples derived from early rice. Conversely, this also indicates that the model exhibits a certain degree of temporal transferability.
As shown in Figure 10, paddy rice fields are rarely distributed in the central mountainous region of Hainan Island due to significant terrain fluctuations. In contrast, large-scale rice cultivation is primarily concentrated in the northern and northeastern regions, as well as along the southern coastal areas. Additionally, double-cropping paddy rice is widely distributed in the eastern part of Hainan Island, accounting for approximately 32.5% of the total paddy rice area. Early-season single-crop rice is also prevalent in this region, comprising approximately 26.9% of the total paddy rice area. Meanwhile, in the western part of the island, large-scale single-crop late rice dominates, accounting for approximately 40.6% of the total paddy rice area.
For paddy rice mapping in 2023, the RiceTS model was employed. However, this model requires continuous time-series data as the input, and the dataset for 2023 was incomplete (Figure 2, Table A1), making the direct application of the model unfeasible. To address the issue of missing data, a data imputation approach was implemented. For data from relative orbit 157, five missing acquisitions between 19 June and 6 August 2023 contained only slice 1. The gap was filled using SAR data from the corresponding period in 2020. Similarly, for data from relative orbit 55, the five missing acquisitions (12 June, 24 June, 18 July, 30 July, and 11 August) were also supplemented with SAR data from the same periods in 2020. For relative orbit 84, the missing acquisition on 12 February was reconstructed by averaging the data from 21 January and 14 February 2023. This imputation approach resulted in a complete time-series dataset (30 observations) covering the entire early and late paddy rice growth cycles for 2023.
Using the model trained in 2020, the paddy rice was mapped for 2023, as shown in Figure 11. Due to the unavailability of statistical data and paddy rice distribution products for validation, the classification accuracy could not be directly assessed. According to the classification results, the estimated early paddy rice area is 1.284 × 105 h a , while the late paddy rice area is 9.037 × 104 h a , showing no significant change compared to 2020. The spatial distribution patterns of early rice, late rice, and double-cropping paddy rice in 2023 are consistent with those observed in 2020, further demonstrating the robust spatiotemporal transferability of the proposed model.

5. Discussion

This study aims to investigate the effective use of time-series SAR data for extracting paddy rice areas in Hainan Island. Previous studies have demonstrated the efficacy of integrating temporal, spatiotemporal, and spatiotemporal–spectral feature fusion for paddy rice mapping [12,45,52,53,65,66]. Two models were developed for paddy rice mapping: RiceLSTM and RiceTS. RiceLSTM utilizes a BiLSTM network with an attention mechanism, achieving classification accuracies of 0.9182 (early rice) and 0.9245 (late rice). RiceTS, which integrates a U-Net with MobileNetV2, further improves the classification accuracy to 0.9656 (early rice) and 0.9808 (late rice). The extraction accuracy in this study exceeds 90%, aligning with the performance levels reported in the existing research [5,12,48]. Compared to the 2020 statistics, the estimated areas differed by 2.1% for early rice and 29.6% for late rice, highlighting both the model’s accuracy and robustness. Using the RiceTS model, the paddy rice areas in Hainan were successfully extracted from time-series Sentinel-1A SAR data for both 2020 and 2023. When applied to the 2023 SAR data, the classified paddy rice distribution exhibited spatial patterns that are consistent with 2020, demonstrating strong spatiotemporal transferability. These findings confirm that integrating time-series SAR imagery with DL effectively captures both spatial and temporal features, significantly enhancing paddy rice mapping accuracy. This underscores the model’s potential for large-scale paddy rice monitoring in regions with similar agricultural practices. Hainan Island’s warm and humid climate supports paddy rice cultivation, including single-, double-, and, in some areas, triple-season cropping. Double-season rice, which constitutes approximately 30% of the total paddy rice area, is predominant in the eastern region, whereas single-season late rice is primarily cultivated in the western region.
However, some challenges remain in paddy rice mapping using time-series SAR data. The completeness of time-series data is crucial for accurately mapping paddy rice cultivation areas, particularly for models that rely on temporal backscatter features. Missing data disrupt temporal continuity, reducing classification reliability and accuracy. Accurate paddy rice mapping depends on capturing key phenological stages, such as transplanting, tillering, and maturity [15]. Data gaps during these critical periods hinder the model’s ability to fully capture paddy rice growth dynamics, often leading to misclassification—particularly in distinguishing paddy rice from other vegetation with similar backscatter characteristics, such as dryland crops or wetland vegetation [71]. Most studies have not directly addressed data gaps in time-series SAR data for paddy rice mapping [12,48]. Shen and Nie attempted to mitigate this issue by reconstructing the time-series SAR dataset using a grouping strategy, but this approach also introduced challenges in accurately capturing paddy rice backscatter variations [14]. In this study, data interpolation and reuse methods were applied to address missing data, although these techniques introduced a certain level of uncertainty in paddy rice mapping.
Compared to optical data, SAR data are more susceptible to salt-and-pepper noise [12], which can negatively impact the accuracy of paddy rice mapping. The precise identification of paddy rice fields depends on detecting the subtle variations in backscatter associated with different growth stages [12]. However, speckle noise can obscure these variations, leading to the misclassification of paddy rice fields or their misinterpretation as other land cover types with similar backscatter properties. Additionally, speckle noise complicates feature extraction, particularly for small or heterogeneous paddy rice fields. To mitigate these effects, a 3D-Gamma adaptive filter and the Savitzky and Golay (S–G) smoothing method were applied to process time-series SAR data in both the spatial and temporal domains, facilitating the identification of long-term change trends [12].
DL models require a sufficient number of high-quality training samples to achieve accurate and generalizable results. Sample generation plays a crucial role in the accuracy and reliability of paddy rice mapping using remote sensing data and DL models [63]. The relatively lower extraction accuracy for late rice underscores the importance of complete and representative training samples. While the existing paddy rice distribution products [62,63] covering Hainan Island can be used for training sample generation, their limited temporal and spatial coverage, along with inherent accuracy constraints, may introduce biases into the training data [63]. Field surveys and manual interpretation remain essential but are time-consuming and labor-intensive [48]. Furthermore, the relatively small proportion of paddy rice cultivation areas can lead to sample imbalance, causing the models to favor majority classes and reducing sensitivity to paddy rice fields. Addressing sample generation challenges is thus a critical aspect of improving DL-based paddy rice mapping.
Recent changes in the paddy rice transplantation methods in Hainan [64] have also influenced the backscatter characteristics observed in time-series Sentinel-1 SAR data, potentially impacting mapping accuracy. Additionally, variations in transplanting times across regions and seasons further complicate classification accuracy, particularly when using time-series SAR data. Another persistent challenge is the misclassification of wetland vegetation as paddy rice due to their similar backscatter characteristics, temporal backscatter variations, vegetation structures, and geographical overlap. The similarity in backscatter signals makes it difficult to accurately distinguish between paddy rice and wetland vegetation, increasing the likelihood of misclassification in paddy rice mapping [71].
The long-term SAR paddy rice mapping process requires substantial computational power for data preprocessing. In this study, we collected a total of 284 scenes of Sentinel-1 SAR data. Using the SARProcMod module, the download time for each scene of Sentinel-1 SAR (approximately 1 GB) is approximately 2 min. The SARProcMod module was also used for processing SAR data, with each scene requiring approximately 20 min (preprocessing and clip). The total preprocessing time amounted to around 95 h, resulting in a 10 m resolution dataset with a total disk volume of around 500 GB. Many studies have leveraged the Google Earth Engine (GEE) platform, which offers direct access to preprocessed Sentinel-1 SAR data [72], significantly reducing the time required for data preprocessing [8,9,62]. However, the GEE platform does not support end-to-end processing of SAR data using deep learning frameworks, and it imposes strict limitations on downloading large-scale Sentinel-1 datasets. DL training also demands significant GPU resources (Table A2), which impacts the model’s scalability and practical deployment. The total parameters for the two proposed models reach up to 151,042 and 3,133,066 separately, requiring 24 GB and 48 GB of GPU memory during training, respectively. This high computational demand makes it challenging to deploy these models on edge devices or in resource-constrained environments, limiting their practical application in real-time paddy rice mapping. Recently, Transformer [67], particularly Swin Transformer (Shifted Window Transformer) [50], has garnered significant attention for its effectiveness in remote sensing image analysis. Swin Transformer employs a window-based self-attention mechanism, and its hierarchical structure enables it to capture multi-scale features. This capability is particularly beneficial for extracting both spatial and temporal features at different scales, making it highly suitable for rice field identification across various growth stages. Swin Transformer and its derivatives are increasingly being applied to paddy rice and other crop information extraction [51,73] due to their high efficiency and low system requirements.

6. Conclusions

In this study, we proposed two DL models, RiceLSTM and RiceTS, for accurate paddy rice field extraction from time-series SAR data in Hainan Island. Both models demonstrate the potential of leveraging temporal features in SAR data for effective paddy rice mapping. RiceLSTM, particularly with an attention mechanism, significantly outperforms the traditional LSTM model in terms of accuracy. RiceMU leverages spatiotemporal information by using stacked inputs, leading to improved extraction accuracy compared to RiceLSTM. RiceTS, on the other hand, highlights the effectiveness of temporal convolutional networks in capturing spatial dynamics, achieving a higher F1 score than RiceLSTM. The integration of multi-temporal SAR data effectively captures paddy rice growth dynamics, enhancing the detection of both early- and late-season paddy rice. This approach provides a reliable solution for paddy rice mapping in regions with frequent cloud cover and heavy rainfall.
However, the results also emphasize the dependence of these models on the completeness and accuracy of the training data. Issues such as sample mislabeling, incomplete representation of paddy rice field variability, and temporal data gaps may limit the models’ generalization ability and overall accuracy. Despite these challenges, the models demonstrate strong potential for spatiotemporal generalization across different years, underscoring their adaptability for large-scale crop mapping applications.
Future research will focus on improving the sample generation strategies, optimizing the model efficiency, effectively addressing data gaps, mitigating sample imbalance, and preserving edge details in DL-based paddy rice mapping. These efforts will contribute to improving the robustness and applicability of DL models for agricultural monitoring using SAR data.

Author Contributions

G.S. proposed and designed the technique road map, contributed to the creation of the algorithmic methods, and collected and processed the SAR data. J.L. helped with the results analysis and reviewed the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Hainan Provincial Natural Science Foundation of China (Grant No. 323MS111), Key Research and Development Program of Guangxi (GuikeAB22035060), and Hainan Province Science and Technology Special Fund (Grant No. ATIC-2023010004-06).

Data Availability Statement

The Sentinel-1 SAR data listed in Table A1 are available in https://dataspace.copernicus.eu/ (accessed on 18 April 2024). The other data presented in this study are available on request from the corresponding author.

Acknowledgments

The authors are grateful to ESA for providing the Sentinel-1 SAR data and would like to express thanks to the anonymous reviewers for their voluntary work and constructive comments to improve this manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. SAR data used in this study.
Table A1. SAR data used in this study.
Year: 2020, Relative Orbit: 157
No.Acquisition DateSliceNo.Acquisition DateSliceNo.Acquisition DateSlice
12020/1/61, 2112020/5/51, 2212020/9/21, 2
22020/1/181, 2122020/5/171, 2222020/9/141, 2
32020/1/301, 2132020/5/291, 2232020/9/261, 2
42020/2/111, 2142020/6/101, 2242020/10/81, 2
52020/2/231, 2152020/6/221, 2252020/10/201, 2
62020/3/61, 2162020/7/41, 2262020/11/11, 2
72020/3/181, 2172020/7/161, 2272020/11/131, 2
82020/3/301, 2182020/7/281, 2282020/11/251, 2
92020/4/111, 2192020/8/91, 2292020/12/71, 2
102020/4/231, 2202020/8/211, 2302020/12/191, 2
Year: 2020, Relative Orbit: 55
No.Acquisition DateSliceNo.Acquisition DateSliceNo.Acquisition DateSlice
312020/1/115, 6412020/5/105, 6512020/9/75, 6
322020/1/235, 6422020/5/225, 6522020/9/195, 6
332020/2/45, 6432020/6/35, 6532020/10/15, 6
342020/2/165, 6442020/6/155, 6542020/10/135, 6
352020/2/285, 6452020/6/275, 6552020/10/255, 6
362020/3/115, 6462020/7/95, 6562020/11/65, 6
372020/3/235, 6472020/7/215, 6572020/11/185, 6
382020/4/45, 6482020/8/25, 6582020/11/305, 6
392020/4/165, 6492020/8/145, 6592020/12/125, 6
402020/4/285, 6502020/8/265, 6602020/12/245, 6
Year: 2020, Relative Orbit: 84
No.Acquisition DateSliceNo.Acquisition DateSliceNo.Acquisition DateSlice
612020/1/11712020/4/301812020/8/281
622020/1/131722020/5/121822020/9/211
632020/1/251732020/5/241832020/10/31
642020/2/61742020/6/51842020/10/151
652020/2/181752020/6/171852020/10/271
662020/3/11762020/6/291862020/11/81
672020/3/131772020/7/111872020/11/201
682020/3/251782020/7/231882020/12/21
692020/4/61792020/8/41892020/12/141
702020/4/181802020/8/161902020/12/261
Year: 2023, Relative Orbit: 157
No.Acquisition DateSliceNo.Acquisition DateSliceNo.Acquisition DateSlice
12023/1/21, 2112023/5/21, 2212023/8/301, 2
22023/1/141, 2122023/5/141, 2222023/9/111, 2
32023/1/261, 2132023/5/261, 2232023/9/231, 2
42023/2/71, 2142023/6/71, 2242023/10/51, 2
52023/2/191, 2152023/6/191252023/10/171, 2
62023/3/31, 2162023/7/11262023/11/101, 2
72023/3/151, 2172023/7/131272023/11/221, 2
82023/3/271, 2182023/7/251282023/12/41, 2
92023/4/81, 2192023/8/61292023/12/161, 2
102023/4/201, 2202023/8/181, 2302023/12/281, 2
Year: 2023, Relative Orbit: 55
No.Acquisition DateSliceNo.Acquisition DateSliceNo.Acquisition DateSlice
312023/1/75, 6412023/5/75, 6462023/9/45, 6
322023/1/195, 6422023/5/195, 6472023/9/165, 6
332023/1/315, 6432023/5/315, 6482023/9/285, 6
342023/2/125, 6 5, 6492023/10/105, 6
352023/2/245, 6 5, 6502023/10/225, 6
362023/3/85, 6442023/7/65, 6512023/11/35, 6
372023/3/205, 6 5, 6522023/11/155, 6
382023/4/15, 6 5, 6532023/11/275, 6
392023/4/135, 6 5, 6542023/12/95, 6
402023/4/255, 6452023/8/235, 6552023/12/215, 6
Year: 2023, Relative Orbit: 84
No.Acquisition DateSliceNo.Acquisition DateSliceNo.Acquisition DateSlice
562023/1/91652023/5/91752023/9/61
572023/1/211662023/5/211762023/9/181
1672023/6/21772023/9/301
582023/2/141682023/6/141782023/10/121
592023/2/261692023/6/261792023/10/241
602023/3/101702023/7/81802023/11/51
612023/3/221712023/7/201812023/11/171
622023/4/31722023/8/11822023/11/291
632023/4/151732023/8/131832023/12/111
642023/4/271742023/8/251842023/12/231
Table A2. The parameters for RiceLSTM, RiceMU, and RiceTS models.
Table A2. The parameters for RiceLSTM, RiceMU, and RiceTS models.
ModelInput SizeReshaped SizeOuput SizeLoss FunctionTotal ParametersEstimated Total
Size (MB)
RiceLSTM256 × 256 × 15 × 265,536 × 15 × 2256 × 256 × 2CrossEntropyLoss134,4021016.08
RiceLSTMWith
Attention
256 × 256 × 15 × 265,536 × 15 × 2256 × 256 × 2CrossEntropyLoss151,0422022.78
RiceMU256 × 256 × 15 × 2256 × 256 × 30256 × 256 × 2CrossEntropyLoss2,964,234840.36
RiceTS256 × 256 × 15 × 265,536 × 15 × 2256 × 256 × 2CrossEntropyLoss3,133,0664520.43

References

  1. Singha, M.; Dong, J.; Zhang, G.; Xiao, X. High resolution paddy rice maps in cloud-prone Bangladesh and Northeast India using Sentinel-1 data. Sci. Data 2019, 6, 26. [Google Scholar] [CrossRef] [PubMed]
  2. Elert, E. Rice by the numbers: A good grain. Nature 2014, 514, S50–S51. [Google Scholar] [CrossRef]
  3. Zhang, X.; Wu, B.; Ponce-Campos, G.; Zhang, M.; Chang, S.; Tian, F. Mapping up-to-Date Paddy Rice Extent at 10 M Resolution in China through the Integration of Optical and Synthetic Aperture Radar Images. Remote Sens. 2018, 10, 1200. [Google Scholar] [CrossRef]
  4. Xu, L.; Zhang, H.; Wang, C.; Wei, S.; Zhang, B.; Wu, F.; Tang, Y. Paddy Rice Mapping in Thailand Using Time-Series Sentinel-1 Data and Deep Learning Model. Remote Sens. 2021, 13, 3994. [Google Scholar] [CrossRef]
  5. Bazzi, H.; Baghdadi, N.; El Hajj, M.; Zribi, M.; Dinh, H.T.M.; Ndikumana, E.; Courault, D.; Belhouchette, H. Mapping Paddy Rice Using Sentinel-1 SAR Time Series in Camargue, France. Remote Sens. 2019, 11, 887. [Google Scholar] [CrossRef]
  6. Yang, H.; Pan, B.; Li, N.; Wang, W.; Zhang, J.; Zhang, X. A systematic method for spatio-temporal phenology estimation of paddy rice using time series Sentinel-1 images. Remote Sens. Environ. 2021, 259, 112394. [Google Scholar] [CrossRef]
  7. Song, X.; Huang, W.; Hansen, M.C.; Potapov, P. An evaluation of Landsat, Sentinel-2, Sentinel-1 and MODIS data for crop type mapping. Sci. Remote Sens. 2021, 3, 100018. [Google Scholar] [CrossRef]
  8. Fan, X.; Wang, Z.; Zhang, H.; Liu, H.; Jiang, Z.; Liu, X. Large-Scale Rice Mapping Based on Google Earth Engine and Multi-Source Remote Sensing Images. J. Indian Soc. Remote Sens. 2023, 51, 93–102. [Google Scholar] [CrossRef]
  9. Waleed, M.; Mubeen, M.; Ahmad, A.; Habib-Ur-Rahman, M.; Amin, A.; Farid, H.U.; Hussain, S.; Ali, M.; Qaisrani, S.A.; Nasim, W.; et al. Evaluating the efficiency of coarser to finer resolution multispectral satellites in mapping paddy rice fields using GEE implementation. Sci. Rep. 2022, 12, 13210. [Google Scholar] [CrossRef]
  10. Xia, L.; Zhao, F.; Chen, J.; Yu, L.; Lu, M.; Yu, Q.; Liang, S.; Fan, L.; Sun, X.; Wu, S.; et al. A full resolution deep learning network for paddy rice mapping using Landsat data. ISPRS J. Photogramm. Remote Sens. 2022, 194, 91–107. [Google Scholar] [CrossRef]
  11. Hu, J.; Chen, Y.; Cai, Z.; Wei, H.; Zhang, X.; Zhou, W.; Wang, C.; You, L.; Xu, B. Mapping Diverse Paddy Rice Cropping Patterns in South China Using Harmonized Landsat and Sentinel-2 Data. Remote Sens. 2023, 15, 1034. [Google Scholar] [CrossRef]
  12. Crisóstomo De Castro Filho, H.; Abílio De Carvalho Júnior, O.; Ferreira De Carvalho, O.L.; Pozzobon de Bem, P.; dos Santos de Moura, R.; Olino de Albuquerque, A.; Rosa Silva, C.; Guimaraes Ferreira, P.H.; Fontes Guimarães, R.; Trancoso Gomes, R.A. Rice Crop Detection Using LSTM, Bi-LSTM, and Machine Learning Models from Sentinel-1 Time Series. Remote Sens. 2020, 12, 2655. [Google Scholar] [CrossRef]
  13. Ribbes, F.; Le Toan, T. Rice field mapping and monitoring with RADARSAT data. Int. J. Remote Sens. 1999, 20, 745–765. [Google Scholar] [CrossRef]
  14. Shen, G.; Nie, C. Mapping Rice Area Using Sentinel-1 SAR Data and Deep Learning. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Pasadena, CA, USA, 16–21 July 2023. [Google Scholar] [CrossRef]
  15. Zhang, X.; Shen, R.; Zhu, X.; Pan, B.; Fu, Y.; Zheng, Y.; Chen, X.; Peng, Q.; Yuan, W. Sample-free automated mapping of double-season rice in China using Sentinel-1 SAR imagery. Front. Environ. Sci. 2023, 11, 1207882. [Google Scholar] [CrossRef]
  16. Dong, Y.; Sun, G.; Pang, Y. Monitoring of rice crop using ENVISAT ASAR data. Sci. China Ser. D Earth Sci. 2006, 49, 755–763. [Google Scholar] [CrossRef]
  17. Chen, J.; Lin, H.; Pei, Z. Application of ENVISAT ASAR Data in Mapping Rice Crop Growth in Southern China. IEEE Geosci. Remote Sens. Lett. 2007, 4, 431–435. [Google Scholar] [CrossRef]
  18. Yang, S.; Shen, S.; Li, B.; Le Toan, T.; He, W. Rice Mapping and Monitoring Using ENVISAT ASAR Data. IEEE Trans. Geosci. Remote Sens. 2008, 5, 108–112. [Google Scholar] [CrossRef]
  19. Bouvet, A.; Le Toan, T.; Nguyen, L.D. Monitoring of the Rice Cropping System in the Mekong Delta Using ENVISAT/ASAR Dual Polarization Data. IEEE Trans. Geosci. Remote Sens. 2009, 47, 517–526. [Google Scholar] [CrossRef]
  20. Bouvet, A.; Thuy, L.T. Use of ENVISAT/ASAR wide-swath data for timely rice fields mapping in the Mekong River Delta. Remote Sens. Environ. 2011, 115, 1090–1101. [Google Scholar] [CrossRef]
  21. Duy, B.N.; Clauss, K.; Cao, S.; Naeimi, V.; Kuenzer, C.; Wagner, W. Mapping Rice Seasonality in the Mekong Delta with Multi-Year Envisat ASAR WSM Data. Remote Sens. 2015, 7, 15868–15893. [Google Scholar] [CrossRef]
  22. Zhang, Y.; Wang, C.; Wu, J.; Qi, J.; Salas, W.A. Mapping paddy rice with multitemporal ALOS/PALSAR imagery in southeast China. Int. J. Remote Sens. 2009, 30, 6301–6315. [Google Scholar] [CrossRef]
  23. Li, K.; Brisco, B.; Yun, S.; Touzi, R. Polarimetric decomposition with RADARSAT-2 for rice mapping and monitoring. Can. J. Remote Sens. 2012, 38, 169–179. [Google Scholar] [CrossRef]
  24. Hoang, H.K.; Bernier, M.; Duchesne, S.; Tran, Y.M. Rice Mapping Using RADARSAT-2 Dual- and Quad-Pol Data in a Complex Land-Use Watershed: Cau River Basin (Vietnam). IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 3082–3096. [Google Scholar] [CrossRef]
  25. Koppe, W.; Gnyp, M.L.; Hütt, C.; Yao, Y.; Miao, Y.; Chen, X.; Bareth, G. Rice monitoring with multi-temporal and dual-polarimetric TerraSAR-X data. Int. J. Appl. Earth Obs. Geoinf. 2013, 21, 568–576. [Google Scholar] [CrossRef]
  26. Lopez-Sanchez, J.M.; David Ballester-Berman, J.; Hajnsek, I. First Results of Rice Monitoring Practices in Spain by Means of Time Series of TerraSAR-X Dual-Pol Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2011, 4, 412–422. [Google Scholar] [CrossRef]
  27. Pei, Z.; Zhang, S.; Guo, L.; Mcnairn, H.; Shang, J.; Jiao, X. Rice identification and change detection using TerraSAR-X data. Can. J. Remote Sens. 2011, 37, 151–156. [Google Scholar] [CrossRef]
  28. Shen, G.; Fu, W.; Guo, H.; Liao, J. Water body mapping using long time series Sentinel-1 SAR data in Poyang Lake. Water 2022, 14, 1902. [Google Scholar] [CrossRef]
  29. Clauss, K.; Ottinger, M.; Kuenzer, C. Mapping rice areas with Sentinel-1 time series and superpixel segmentation. Int. J. Remote Sens. 2018, 39, 1399–1420. [Google Scholar] [CrossRef]
  30. N, K.; Salma, S.; Dodamani, B.M. Identifying Rice Crop Flooding Patterns Using Sentinel-1 SAR Data. J. Indian Soc. Remote Sens. 2022, 50, 1569–1584. [Google Scholar] [CrossRef]
  31. Li, H.; Fu, D.; Huang, C.; Su, F.; Liu, Q.; Liu, G.; Wu, S. An Approach to High-Resolution Rice Paddy Mapping Using Time-Series Sentinel-1 SAR Data in the Mun River Basin, Thailand. Remote Sens. 2020, 12, 3959. [Google Scholar] [CrossRef]
  32. Chang, L.; Chen, Y.T.; Wang, J.H.; Chang, Y.L. Rice-Field Mapping with Sentinel-1A SAR Time-Series Data. Remote Sens. 2021, 13, 103. [Google Scholar] [CrossRef]
  33. Zhu, A.; Zhao, F.; Pan, H.; Liu, J. Mapping Rice Paddy Distribution Using Remote Sensing by Coupling Deep Learning with Phenological Characteristics. Remote Sens. 2021, 13, 1360. [Google Scholar] [CrossRef]
  34. He, Z.; Li, S. Research progress on radar remote sensing for rice growth monitoring. J. Remote Sens. 2023, 27, 2363–2382. [Google Scholar] [CrossRef]
  35. Gao, X.; Chi, H.; Huang, J.; Ling, F.; Han, Y.; Jia, X.; Li, Y.; Huang, D.; Dong, J. Review of paddy rice mapping with remote sensing technology. Natl. Remote Sens. Bull. 2024, 28, 2144–2169. [Google Scholar] [CrossRef]
  36. Wei, P.; Chai, D.; Lin, T.; Tang, C.; Du, M.; Huang, J. Large-scale rice mapping under different years based on time-series Sentinel-1 images using deep semantic segmentation model. ISPRS J. Photogramm. Remote Sens. 2021, 174, 198–214. [Google Scholar] [CrossRef]
  37. Zhao, R.; Li, Y.; Ma, M. Mapping Paddy Rice with Satellite Remote Sensing: A Review. Sustainability 2021, 13, 503. [Google Scholar] [CrossRef]
  38. Xu, S.; Zhu, X.; Chen, J.; Zhu, X.; Duan, M.; Qiu, B.; Wan, L.; Tan, X.; Xu, Y.N.; Cao, R. A robust index to extract paddy fields in cloudy regions from SAR time series. Remote Sens. Environ. 2023, 285, 113374. [Google Scholar] [CrossRef]
  39. Hinton, G.E.; Salakhutdinov, R.R. Reducing the Dimensionality of Data with Neural Networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef]
  40. Ma, X.; Huang, Z.; Zhu, S.; Fang, W.; Wu, Y. Rice Planting Area Identification Based on Multi-Temporal Sentinel-1 SAR Images and an Attention U-Net Model. Remote Sens. 2022, 14, 4573. [Google Scholar] [CrossRef]
  41. Li, Q.; Tian, J.; Tian, Q. Deep Learning Application for Crop Classification via Multi-Temporal Remote Sensing Images. Agriculture 2023, 13, 906. [Google Scholar] [CrossRef]
  42. Ndikumana, E.; Ho Tong Minh, D.; Baghdadi, N.; Courault, D.; Hossard, L. Deep Recurrent Neural Network for Agricultural Classification using multitemporal SAR Sentinel-1 for Camargue, France. Remote Sens. 2018, 10, 1217. [Google Scholar] [CrossRef]
  43. Wu, M.; Alkhaleefah, M.; Chang, L.; Chang, Y.; Shie, M.; Liu, S.; Chang, W. Recurrent Deep Learning for Rice Fields Detection from SAR Images. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA, 26 September–2 October 2020; pp. 1548–1551. [Google Scholar] [CrossRef]
  44. Jo, H.W.; Lee, S.; Park, E.; Lim, C.H.; Song, C.; Lee, H.; Ko, Y.; Cha, S.; Yoon, H.; Lee, W.K. Deep Learning Applications on Multitemporal SAR (Sentinel-1) Image Classification Using Confined Labeled Data: The Case of Detecting Rice Paddy in South Korea. IEEE Trans. Geosci. Remote Sens. 2020, 58, 7589–7601. [Google Scholar] [CrossRef]
  45. Zhou, Y.; Luo, J.; Feng, L.; Yang, Y.; Chen, Y.; Wu, W. Long-short-term-memory-based crop classification using high-resolution optical images and multi-temporal SAR data. GISci. Remote Sens. 2019, 56, 1170–1191. [Google Scholar] [CrossRef]
  46. Thorp, K.R.; Drajat, D. Deep machine learning with Sentinel satellite data to map paddy rice production stages across West Java, Indonesia. Remote Sens. Environ. 2021, 265, 112679. [Google Scholar] [CrossRef]
  47. Lin, Z.; Zhong, R.; Xiong, X.; Guo, C.; Xu, J.; Zhu, Y.; Xu, J.; Ying, Y.; Ting, K.C.; Huang, J.; et al. Large-Scale Rice Mapping Using Multi-Task Spatiotemporal Deep Learning and Sentinel-1 SAR Time Series. Remote Sens. 2022, 14, 699. [Google Scholar] [CrossRef]
  48. Sun, C.; Zhang, H.; Xu, L.; Wang, C.; Li, L. Rice Mapping Using a BiLSTM-Attention Model from Multitemporal Sentinel-1 Data. Agriculture 2021, 11, 977. [Google Scholar] [CrossRef]
  49. Sun, C.; Zhang, H.; Ge, J.; Wang, C.; Li, L.; Xu, L. Rice Mapping in a Subtropical Hilly Region Based on Sentinel-1 Time Series Feature Analysis and the Dual Branch BiLSTM Model. Remote Sens. 2022, 14, 3213. [Google Scholar] [CrossRef]
  50. He, X.; Zhou, Y.; Zhao, J.; Zhang, D.; Yao, R.; Xue, Y. Swin Transformer Embedding UNet for Remote Sensing Image Semantic Segmentation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4408715. [Google Scholar] [CrossRef]
  51. Xie, Y.; Xu, L.; Zhang, H.; Song, M.; Ge, J.; Wu, F. Tropical Rice Mapping Using Time-Series SAR Images and ESF-Seg Model in Hainan, China, from 2019 to 2023. Remote Sens. 2025, 17, 209. [Google Scholar] [CrossRef]
  52. Teimouri, N.; Dyrmann, M.; Jørgensen, R.N. A Novel Spatio-Temporal FCN-LSTM Network for Recognizing Various Crop Types Using Multi-Temporal Radar Images. Remote Sens. 2019, 11, 990. [Google Scholar] [CrossRef]
  53. Chang, Y.L.; Tatini, N.B.; Chen, T.H.; Wu, M.C.; Chuah, J.H.; Chen, Y.T.; Chang, L. Convlstm Neural Network for Rice Field Classification from Sentinel-1A Sar Images. In Proceedings of the IGARSS 2022-2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 5047–5050. [Google Scholar] [CrossRef]
  54. Yang, L.; Huang, R.; Huang, J.; Lin, T.; Wang, L.; Mijiti, R.; Wei, P.; Tang, C.; Shao, J.; Li, Q.; et al. Semantic Segmentation Based on Temporal Features: Learning of Temporal–Spatial Information From Time-Series SAR Images for Paddy Rice Mapping. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4403216. [Google Scholar] [CrossRef]
  55. Wei, S.; Zhang, H.; Wang, C.; Xu, L.; Wu, F.; Zhang, B. Large-Scale Rice Mapping of Thailand using Sentinel-1 Multi-Temporal SAR Data. In Proceedings of the 2019 SAR in Big Data Era (BIGSARDATA), Beijing, China, 5–6 August 2019; pp. 1–4. [Google Scholar] [CrossRef]
  56. Ehiemere, C.I.; Okeke, F.I.; Ehiemere, N.D. Time-series Sentinel-1A SAR remote sensing of rice planting methods in Ebonyi State, Nigeria. Sci. Afr. 2023, 22, e01929. [Google Scholar] [CrossRef]
  57. Pan, B.; Zheng, Y.; Shen, R.; Ye, T.; Zhao, W.; Dong, J.; Ma, H.; Yuan, W. High Resolution Distribution Dataset of Double-Season Paddy Rice in China. Remote Sens. 2021, 13, 4609. [Google Scholar] [CrossRef]
  58. Shen, R.; Pan, B.; Peng, Q.; Dong, J.; Chen, X.; Zhang, X.; Ye, T.; Huang, J.; Yuan, W. High-resolution distribution maps of single-season rice in China from 2017 to 2022. Earth Syst. Sci. Data 2023, 15, 3203–3222. [Google Scholar] [CrossRef]
  59. You, N.; Dong, J.; Huang, J.; Du, G.; Zhang, G.; He, Y.; Yang, T.; Di, Y.; Xiao, X. The 10-m crop type maps in Northeast China during 2017–2019. Sci. Data 2021, 8, 41. [Google Scholar] [CrossRef]
  60. Sun, L.; Yang, T.; Lou, Y.; Shi, Q.; Zhang, L. Paddy Rice Mapping Based on Phenology Matching and Cultivation Pattern Analysis Combining Multi-Source Data in Guangdong, China. J. Remote Sens. 2024, 4, 0152. [Google Scholar] [CrossRef]
  61. Han, J.; Zhang, Z.; Luo, Y.; Cao, J.; Zhang, L.; Zhuang, H.; Cheng, F.; Zhang, J.; Tao, F. Annual paddy rice planting area and cropping intensity datasets and their dynamics in the Asian monsoon region from 2000 to 2020. Agric. Syst. 2022, 200, 103437. [Google Scholar] [CrossRef]
  62. Tan, S.; Wu, B.; Zhang, X. Mapping Paddy Rice in the Hainan Province Using both Google Earth Engine and Remote Sensing Images. J. Geo-Inf. Sci. 2019, 21, 937–947. [Google Scholar]
  63. Wang, C.; Xing, Z.; Lu, J.; Cao, F.; Sun, J.; Chai, X.; Liu, X.; Xiong, X. Research on Remote Sensing Intelligent Extraction Method of Tropical Rice Planting Area based on Deep Learning: A Case Study of Haikou City, Hainan Province. Remote Sens. Technol. Appl. 2024, 39, 1106–1114. [Google Scholar]
  64. Wang, B.; Chen, X.M.; Zhong, M.Q.; Zou, H.P.; Qian, K.; Liu, S.J. Spatiotemporal Change of Rice Growth Period in Hainan and Its Response to Climate Warming. Chin. J. Trop. Crop. 2017, 38, 415–420. [Google Scholar] [CrossRef]
  65. Graves, A.; Schmidhuber, J. Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Netw. 2005, 18, 602–610. [Google Scholar] [CrossRef] [PubMed]
  66. Marjani, M.; Mahdianpari, M.; Mohammadimanesh, F. CNN-BiLSTM: A Novel Deep Learning Model for Near-Real-Time Daily Wildfire Spread Prediction. Remote Sens. 2024, 16, 1467. [Google Scholar] [CrossRef]
  67. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.U.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  68. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Munich, Germany, 5–9 October 2015; Volume 9351. [Google Scholar] [CrossRef]
  69. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar] [CrossRef]
  70. Statistical Bureau of Hainan Province; Survey Office of National Bureau of Statistics in Hainan. Hainan Statistical Yearbook (2020); China Statistics Press: Beijing, China, 2021. [Google Scholar]
  71. Huang, D.; Xu, L.; Zou, S.; Liu, B.; Li, H.; Pu, L.; Chi, H. Mapping Paddy Rice in Rice–Wetland Coexistence Zone by Integrating Sentinel-1 and Sentinel-2 Data. Agriculture 2024, 14, 345. [Google Scholar] [CrossRef]
  72. Mullissa, A.; Vollrath, A.; Odongo-Braun, C.; Slagter, B.; Balling, J.; Gou, Y.; Gorelick, N.; Reiche, J. Sentinel-1 SAR Backscatter Analysis Ready Data Preparation in Google Earth Engine. Remote Sens. 2021, 13, 1954. [Google Scholar] [CrossRef]
  73. Jamali, A.; Mahdianpari, M. Swin Transformer and Deep Convolutional Neural Networks for Coastal Wetland Classification Using Sentinel-1, Sentinel-2, and LiDAR Data. Remote Sens. 2022, 14, 359. [Google Scholar] [CrossRef]
Figure 1. Spatial location and extent of the study area of Hainan Island. (a) Indicates the administrative divisions of China, highlighting the study area of Hainan Island with a blue polygon. (b) Indicates the Sentinel-1 SAR images (R: σ 0 ( V H ) , G: σ 0 ( V V ) , B: σ 0 ( V H ) ) from relative orbits of 55 (left bottom part), 157 (middle part), and 84 (top right part) acquired on 11, 6, and 13 January 2020. (c) Indicates the Copernicus Digital Elevation Model (DEM) data. (d) Indicates the Google Earth high-resolution images (true color composite) used to generate training samples. (e) Indicates the 10m resolution paddy rice planting distribution product in 2020 from the Geographic Remote Sensing Ecological Network platform (http://www.gisrs.cn/shopdata?id=fa14d54e-19a6-4f35-b045-9dc166918060 (accessed on 19 July 2024)), which is also used to generate training samples.
Figure 1. Spatial location and extent of the study area of Hainan Island. (a) Indicates the administrative divisions of China, highlighting the study area of Hainan Island with a blue polygon. (b) Indicates the Sentinel-1 SAR images (R: σ 0 ( V H ) , G: σ 0 ( V V ) , B: σ 0 ( V H ) ) from relative orbits of 55 (left bottom part), 157 (middle part), and 84 (top right part) acquired on 11, 6, and 13 January 2020. (c) Indicates the Copernicus Digital Elevation Model (DEM) data. (d) Indicates the Google Earth high-resolution images (true color composite) used to generate training samples. (e) Indicates the 10m resolution paddy rice planting distribution product in 2020 from the Geographic Remote Sensing Ecological Network platform (http://www.gisrs.cn/shopdata?id=fa14d54e-19a6-4f35-b045-9dc166918060 (accessed on 19 July 2024)), which is also used to generate training samples.
Remotesensing 17 01033 g001
Figure 2. The temporal distributions (day of year) of Sentinel-1 SAR observations in 2020 (a) and 2023 (b).
Figure 2. The temporal distributions (day of year) of Sentinel-1 SAR observations in 2020 (a) and 2023 (b).
Remotesensing 17 01033 g002
Figure 3. The flowchart showing the overall methods used in this study.
Figure 3. The flowchart showing the overall methods used in this study.
Remotesensing 17 01033 g003
Figure 4. The flowchart of the RiceLSTM model used in this study. RiceLSTM is based on the BiLSTM model, with the number of features in the hidden state of 64 and the number of LSTM layers of 2. When the attention layer is applied, the entire BiLSTM output ( 15 × 128 ) is passed to the attention layer, which reduces it to a 1 × 128 tensor. The attention-enhanced output is then fed into the fully connected layer to estimate the paddy rice probability for each pixel. In the absence of the attention layer, only the last BiLSTM output sequence ( 1 × 128 ) is directly passed to the fully connected layer for classification, which produces a final output of 256 × 256 , corresponding to the paddy rice classification map.
Figure 4. The flowchart of the RiceLSTM model used in this study. RiceLSTM is based on the BiLSTM model, with the number of features in the hidden state of 64 and the number of LSTM layers of 2. When the attention layer is applied, the entire BiLSTM output ( 15 × 128 ) is passed to the attention layer, which reduces it to a 1 × 128 tensor. The attention-enhanced output is then fed into the fully connected layer to estimate the paddy rice probability for each pixel. In the absence of the attention layer, only the last BiLSTM output sequence ( 1 × 128 ) is directly passed to the fully connected layer for classification, which produces a final output of 256 × 256 , corresponding to the paddy rice classification map.
Remotesensing 17 01033 g004
Figure 5. The flowchart of the RiceTS model used in this study. The RiceLSTM layer is based on RiceLSTM model (Figure 4) and generates an output size of 65,536 ×   64 . This output is reshaped into a 256 × 256 × 64 feature map and subsequently fed into the RiceMU layer (green dashed box), which processes the spatial features and produces a final output of 256 × 256 , corresponding to the paddy rice classification map.
Figure 5. The flowchart of the RiceTS model used in this study. The RiceLSTM layer is based on RiceLSTM model (Figure 4) and generates an output size of 65,536 ×   64 . This output is reshaped into a 256 × 256 × 64 feature map and subsequently fed into the RiceMU layer (green dashed box), which processes the spatial features and produces a final output of 256 × 256 , corresponding to the paddy rice classification map.
Remotesensing 17 01033 g005
Figure 6. The time-series VV and VH polarization backscatter distribution map for different land cover types.
Figure 6. The time-series VV and VH polarization backscatter distribution map for different land cover types.
Remotesensing 17 01033 g006
Figure 7. The variations in backscatter from Sentinel-1 SAR data for different land cover types and VV and VH polarization.
Figure 7. The variations in backscatter from Sentinel-1 SAR data for different land cover types and VV and VH polarization.
Remotesensing 17 01033 g007
Figure 8. The evaluation indices for the classification results of the proposed models for early and late paddy rice in 2020.
Figure 8. The evaluation indices for the classification results of the proposed models for early and late paddy rice in 2020.
Remotesensing 17 01033 g008
Figure 9. The comparison between the paddy rice product, the results derived from the RiceLSTM-Attention, RiceMU, and RiceTS models, (a,f) for GEE high-resolution image (true color composite), (b,g) for the paddy rice product (Figure 1e), (c,h) for the result derived from RiceLSTM-Attention model, (d,i) for the result derived from RiceMU model, and (e,j) for the result derived from RiceTS model.
Figure 9. The comparison between the paddy rice product, the results derived from the RiceLSTM-Attention, RiceMU, and RiceTS models, (a,f) for GEE high-resolution image (true color composite), (b,g) for the paddy rice product (Figure 1e), (c,h) for the result derived from RiceLSTM-Attention model, (d,i) for the result derived from RiceMU model, and (e,j) for the result derived from RiceTS model.
Remotesensing 17 01033 g009
Figure 10. The paddy rice distribution map in 2020 derived from the RiceTS model: 0 for non-rice area, 1 for only late paddy rice, 2 for only early paddy rice, and 3 for both.
Figure 10. The paddy rice distribution map in 2020 derived from the RiceTS model: 0 for non-rice area, 1 for only late paddy rice, 2 for only early paddy rice, and 3 for both.
Remotesensing 17 01033 g010
Figure 11. The paddy rice distribution map in 2023 derived from the RiceTS model: 0 for non-rice area, 1 for only late paddy rice, 2 for only early paddy rice, and 3 for both.
Figure 11. The paddy rice distribution map in 2023 derived from the RiceTS model: 0 for non-rice area, 1 for only late paddy rice, 2 for only early paddy rice, and 3 for both.
Remotesensing 17 01033 g011
Table 1. Accuracy assessment for classification results generated from different models for early and late paddy rice in 2020.
Table 1. Accuracy assessment for classification results generated from different models for early and late paddy rice in 2020.
ModelAccuracyPrecisionRecallF1 ScoreKappa
RiceLSTM (Early)0.89710.86360.81420.83820.7628
RiceLSTM (Late)0.90670.86500.84720.85600.7870
RiceLSTM With
Attention (Early)
0.91820.88390.86350.87350.8131
RiceLSTM With
Attention (Late)
0.92450.88090.88980.88530.8291
RiceMU (Early)0.96420.77870.95170.85660.8364
RiceMU (Late)0.96040.74740.98120.84840.8264
RiceTS (Early)0.96560.75120.92850.83050.8116
RiceTS (Late)0.98080.84650.98120.90880.8982
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shen, G.; Liao, J. Paddy Rice Mapping in Hainan Island Using Time-Series Sentinel-1 SAR Data and Deep Learning. Remote Sens. 2025, 17, 1033. https://doi.org/10.3390/rs17061033

AMA Style

Shen G, Liao J. Paddy Rice Mapping in Hainan Island Using Time-Series Sentinel-1 SAR Data and Deep Learning. Remote Sensing. 2025; 17(6):1033. https://doi.org/10.3390/rs17061033

Chicago/Turabian Style

Shen, Guozhuang, and Jingjuan Liao. 2025. "Paddy Rice Mapping in Hainan Island Using Time-Series Sentinel-1 SAR Data and Deep Learning" Remote Sensing 17, no. 6: 1033. https://doi.org/10.3390/rs17061033

APA Style

Shen, G., & Liao, J. (2025). Paddy Rice Mapping in Hainan Island Using Time-Series Sentinel-1 SAR Data and Deep Learning. Remote Sensing, 17(6), 1033. https://doi.org/10.3390/rs17061033

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop