1. Introduction
Surface water is a crucial component of Earth’s water resources, playing an indispensable role in various hydrological cycles [
1,
2,
3,
4,
5]. Accurate detection of surface water is vital for sustainable water resource management, environmental protection, and socio-economic development [
6,
7,
8,
9,
10]. Against this backdrop, the automatic and precise extraction of the water body using Synthetic Aperture Radar (SAR) data has begun to attract widespread scholarly attention [
11,
12,
13,
14,
15].
As an active microwave sensor, SAR is capable of sensing targets during both day and night, irrespective of sunlight conditions. Its signal can penetrate cloud cover, rain, and fog, overcoming the limitations faced by optical remote sensing sensors under adverse weather conditions. Therefore, SAR can provide stable and reliable time-series data for surface water-body detection. However, the challenge of automatically and accurately extracting water-body information from vast amounts of SAR data lies in the need for a substantial amount of annotated water-body data. The acquisition of this data typically relies on manual annotation, which undoubtedly amplifies the workload involved in this process [
16,
17,
18,
19,
20,
21].
Deep learning, a powerful branch of artificial intelligence, excels in automatic learning and feature extraction. In recent years, water-body detection methods based on Convolutional Neural Network (CNN) models have shown promising results [
22,
23,
24,
25]. In [
26], Li et al. proposed a high-precision water boundary extraction method for SAR images, combining super-resolution recovery based on lightweight residual CNNs with traditional extraction methods. In [
27], Zhang et al. proposed a segmentation network architecture based on deep learning that extracts high-dimensional features from SAR images using dense depth-separable convolutions and dilated convolutions. It achieved exceptional accuracy, robustness, and speed, surpassing traditional methods. In [
28], Hertel et al. compared the performance of two probabilistic CNNs for water segmentation based on SAR using Sentinel-1 data. However, the implementation of deep learning typically necessitates a substantial amount of labeled water-body data, the acquisition of which usually relies on manual annotation, undoubtedly amplifying the workload involved in this process [
29,
30,
31,
32,
33,
34,
35]. Therefore, unsupervised learning methods, such as clustering algorithms, can be considered to reduce the huge workload and time required for manual labeling.
Clustering algorithms, as a type of unsupervised learning method, offer numerous advantages, leading to their wide application in early water-body detection research [
35,
36]. This confers an advantage in creating water-body labels, presumably reducing the workload of manual annotation and enhancing the efficiency of label generation. However, the segmentation results of clustering algorithms are strongly influenced by the choice of initial cluster centers. To address this issue, the application of swarm intelligence optimization algorithms becomes particularly important.
Swarm intelligence optimization algorithms can explore complex solution spaces, allowing them to converge to the optimal or near-optimal solution. Their global search capability also enables them to adapt to changing data distributions. This provides a new idea for the selection of initial cluster centers in clustering algorithms.
In this paper, we utilize the swarm intelligence optimization algorithm, Dung Beetle Optimizer (DBO), to optimize the initial cluster centers of the K-means clustering algorithm, naming it the DBO-K-means (DK) method. The DK method categorizes the training dataset into four different classes, generating water-body labels needed for deep learning without any manual intervention. This unsupervised method significantly reduces the workload associated with manual labeling in traditional supervised deep-learning methods. Subsequently, Classifier is used to perform pixel-level dense classification tasks, yielding an initial result image with vague water-body boundaries. Finally, the optimizer is used to perform fine-grained optimization on the initial result image, yielding a final result image with clear water-body boundaries. Experiments conducted in the study area demonstrate that the proposed method not only effectively improves the accuracy of water body detection but also significantly reduces the workload of manual labeling by shifting the detection process from supervised to unsupervised [
36,
37,
38,
39,
40,
41,
42].
2. Materials and Methods
2.1. Study Area and SAR Data
The study area is East Dongting Lake, located in Yueyang City, Hunan Province, China, with coordinates ranging from 29°00′ N to 29°38′ N latitude and 112°43′ E to 113°14′ E longitude, as shown in
Figure 1. The study area has a subtropical humid monsoon climate, characterized by abundant sunshine and significant annual variations in precipitation. The region is home to the Hunan East Dongting Lake National Nature Reserve, which joined the Ramsar Convention on Wetlands of International Importance in 1992 and was upgraded to a national nature reserve in 1994. It represents a typical wetland ecosystem in China.
In recent years, climate change has been severe, and drought and floods have occurred frequently in Hunan. East Dongting Lake serves as an essential flood storage lake in the Yangtze River system, and its seasonal characteristics are particularly striking. There is a great urgent need to have an insight into the situation of the surface water body for the ecological protection of this area.
Sentinel-1 SAR sensor works in the C-band and exhibits better performance for discriminating water bodies with the appropriate spatiotemporal resolution. Two SAR images from it are used in this study, and their information is shown in
Table 1. One is for training, and the other is for testing. The training image consists of 33,097 × 21,287 pixels. To avoid the computational burden, this image is cropped into 509 patches, with a size of 512 × 512 pixels.
The preprocessing of SAR images involves orbit correction, radiometric calibration, speckle filtering (Refined Lee), range-Doppler terrain correction, and converting from linear to dB.
2.2. Method
The proposed method consists of four steps, as shown in
Figure 2. First, the DK method is employed to divide the training dataset into four categories. Then, the water body is extracted from these images to serve as the necessary labels for the Classifier-Optimizer (CO) training. Second, the classifier is used for pixel-level dense prediction, generating an initial result image with fuzzy water-body boundaries. Third, the optimizer takes the initial result image and the original SAR image as inputs for fine-grained optimization, producing a result image with clearer water-body boundaries.
2.2.1. DK Processing
To address the limitations of the K-means algorithm, such as initial centroid selection and convergence speed, the DBO is selected to optimize the results of K-means by mimicking the behavior of dung beetles. There are four categories of agents in DBO: ball-rolling dung beetles, brood balls, small dung beetles, and thieves [
34]. During the rolling process, the position of the ball-rolling dung beetle is updated and can be expressed as:
where
represents the current iteration number,
denotes the position information of the
th dung beetle,
denotes a constant value which indicates the deflection coefficient,
is a natural coefficient that is assigned
or
, and
indicates the global worst position.
When an obstacle is encountered, the dung beetle dances to reorient itself, and its position is updated as follows:
where
is the deflection angle.
To simulate the area where female dung beetles lay eggs a boundary selection strategy was used, which was defined as
where
denotes the best position,
represents the lower bound, and
represents the upper bound.
The position of the ovoid changes dynamically during the iterative process, as defined below.
where
is the position information of brood ball, and
and
are random vectors by size
.
Additionally, it is necessary to identify the ideal foraging area to effectively guide the small dung beetle, thereby simulating its foraging behavior more simply. The optimal foraging area is defined as follows:
where
denotes the global best position. As a result, the position of the small dung beetle is updated as follows.
where
represents a random number that follows normally distributed, and
denotes a random vector belonging to
.
Assuming that
is the best place to compete for food, the location of dung beetles with stealing behavior is thus updated as described below:
where
is a random vector thar follows normally distributed;
indicates a constant value.
Though the iteration of DBO, the initial clustering centers for K-means are determined. Subsequently, each pixel within the dataset is assigned to a specific clustering category. The center of each cluster is updated and calculated as the mean of all points within that cluster. This procedure is repeated until every data point is closest to the center of its corresponding cluster, ensuring optimal clustering.
2.2.2. CO Processing
In our approach, the classifier employs the DeepLabV3+ network architecture specifically for surface water-body detection in SAR images using pixel-wise dense prediction tasks [
35]. DeepLabV3+ features a complete encoder-decoder structure, where the encoder captures the semantic information of water in the SAR image, and the decoder recovers this information to pixel-level resolution, facilitating accurate detection of water boundaries.
To address the issue of reduced spatial resolution in traditional CNNs, DeepLabV3+ employs the atrous convolution, which expands its receptive field without increasing computational complexity or parameter count. This enhancement allows the network to efficiently capture multi-scale contextual information, essential for detecting water bodies of varying sizes and shapes in SAR images.
Furthermore, we leverage DeepLabV3+’s ability to handle objects of different scales by using multiple parallel atrous convolution layers with different sampling rates. The features extracted at each sampling rate are further processed and fused in different branches to generate the final result, utilizing the Atrous Spatial Pyramid Pooling (ASPP) module, as shown in
Figure 3.
By applying the DeepLabV3+ network in our classifier, we can perform pixel-level classification tasks on SAR images to accurately identify water bodies. After training the classifier with the DK-generated labels, it can effectively classify pixels and distinguish water bodies with varying scales and shapes, crucial for water resource management and environmental protection. Finally, the output generated by the classifier is further refined by the optimizer to obtain a more accurate and clear representation of water-body boundaries, enabling more precise monitoring and analysis of the surface water body.
In the classifier, there has always been a trade-off between classification accuracy and boundary segmentation accuracy. Models with deeper layers and multiple max-pooling layers perform well in classification tasks. However, as the number of max-pooling layers increases, spatial information gradually disappears, resulting in coarse boundary segmentation outcomes. Therefore, the classifier can predict the presence and approximate location of objects, but it is not capable of accurately delineating their boundaries. Therefore, following the classifier segmentation, an initial result image with fuzzy water-body boundary is generated. FC-CRF is employed for fine-grained refinement of the initial segmentation image, yielding sharper water-body boundaries. The energy function used is presented in Equation (8) [
14].
where
represents the pixel’s label assignment,
is one-dimensional potential, and
denotes the label assignment probability of pixel
, as calculated by the DeeplabV3+.
The pairwise potential adopts a structure that enables efficient inference when utilizing a fully-connected graph, that is, connecting every pair of image pixels
and
. Specifically, we employ the expression as described in Equation (9):
where
equals
if
and
have the same value, signifying that only nodes with distinct labels face penalties. The remaining expression incorporates two Gaussian kernels in distinct feature spaces. The appearance kernel relies on both pixel positions and grayscale value, while the smoothness kernel depends solely on pixel positions. The hyperparameters
,
, and
dictate the scale of the Gaussian kernels. The first kernel encourages pixels with similar grayscale value and position to have corresponding labels, whereas the second kernel takes into account only spatial closeness when promoting smoothness.
3. Experimental and Results
3.1. Qualitative Evaluation
In the comparative experiment, we select three unsupervised and two deep-learning methods of water-body detection in SAR images to evaluate the performance and effectiveness, namely SegNet [
36], U-Net [
37], Markov Random Field (MRF) [
38], Fuzzy C-means (FCM) [
39], and Active Contour Model (ACM) [
40]. Their effectiveness and accuracy are demonstrated through both qualitative and quantitative assessments.
To provide a more accurate and intuitive comparison of the six methods’ results, we select two subregions within the study area for a detailed presentation. As shown in
Figure 4, the green box represents Region I, while the red box represents Region II. The upper-right corner of Region I contains numerous vessels, which can cause significant interference in water-body detection; selecting this area allows for the assessment of the model’s ability to resist such interference. Region II primarily consists of small ponds and narrow water bodies, and selecting this area allows for the evaluation of the model’s ability to identify these smaller water bodies. As seen in the image, the scattering energy of the water body in Region I is higher than that in Region II; by choosing these two regions, it is possible to simultaneously determine whether the model training is sufficient and if it can accurately detect different types of water-body. The specific results are shown in
Figure 5 and
Figure 6.
Figure 5 shows the comparison results of various methods in Region I. As can be seen from the blue circles in
Figure 5b,c, the clarity of the water-body edges within the blue circles is significantly improved after the optimization by the optimizer. At the same time, the water-body extraction appears more complete in
Figure 5b,c compared to
Figure 5a. In
Figure 5d,e, although most of the water bodies are detected, compared with
Figure 5a, it can be seen that the SegNet and U-Net have limited ability to extract small water body without an optimizer. In
Figure 5f–h, the most noticeable areas are the red circles on the right. Unsupervised segmentation methods may struggle to accurately identify water bodies in the original SAR images, possibly due to the presence of many ships in the river. In
Figure 5c, the water-body edge within the left red circle is quite clear. However, in
Figure 5f,h, the water body is not completely extracted. In
Figure 5g, the water-body boundary within the left red circle is very blurry.
Figure 6 shows the comparison results of various methods in Region II. In
Figure 6b,c, after optimization by the optimizer, the small water body within the bellow blue circle is also detected. The water-body edges within the upper blue circle have become very clear and smooth. At the same time, the water-body extraction appears more complete in
Figure 6b,c compared to
Figure 6a. In
Figure 6d,e, the results are similar to those in
Figure 5; although most of the water-body are detected, compared to
Figure 6a,c, it can be seen that the deep-learning method also has limited ability to extract small water body without an optimizer. In
Figure 6f–h, the small pond within the right red circle is not detected well. In fact, the water body is not detected at all within the right red circles in
Figure 6g,h. In
Figure 6b,c, the water-body edges within the left red circle are very clear, especially in
Figure 6c where the detection errors are corrected after optimization. However, in
Figure 6f–h, the water-body boundaries within the left red circle are extremely blurred. Moreover, in
Figure 6g,h, the water body within the left red circle is not completely detected.
3.2. Quantitative Evaluation
In quantitative evaluation, we compare the resulting images obtained by various methods with the ground truth labels to generate confusion matrices, and the results for Region I and Region II are shown in
Figure 7 and
Figure 8. As shown in
Table 2, the definitions of the four indicators were given in the confusion matrix. ACC, Precision, F1-score, Recall, and Kaapa are chosen as evaluation indicators. The calculations of them are illustrated in Equations (10)–(14), respectively. Based on the confusion matrix, we perform a quantitative evaluation of the different methods. The averaged results are listed in
Table 3.
Among them, the proposed DK-CO achieves the best accuracy. The ACC score is 97.22%, the precision score is 93.45%, the F1-Score is 95.21%, the recall score is 97.05%, and the Kappa coefficient score is 93.25%. Unsupervised classification methods cannot effectively establish mapping relationships for classifying each pixel, resulting in lower metric values. Although the deep-learning method without an optimizer can detect major water-body, the ability to detect small water bodies needs to be improved. Compared to machine learning methods and deep learning without an optimizer method, the metrics are improved for the proposed method in this paper.
In order to objectively evaluate the performance of each method, a visual comparison of the accuracy assessment metrics, including accuracy (ACC), precision, F1-Score, recall, and Kappa, was presented in
Figure 9. The methods can be ranked in terms of ACC as follows: DK-CO > SubPixelACM > MRF >FCM. This indicates that the DK-CO method surpasses the other approaches in the task of water body detection, showing its superior performance.
4. Discussion
In the community of water-body detection using SAR images, the automatic generation of sample labels significantly optimizes the training process based on deep-learning network models. Traditional deep-learning methods usually rely on a large amount of manually labeled data, and the process is time-consuming and cumbersome. However, the DBO-K-means-based method proposed in this paper achieves the automatic extraction of water-body labels by optimizing the initial cluster centers and segmenting the training images into four categories. This method without human intervention greatly reduces the workload of manual labeling and improves the efficiency of label generation. On the other hand, this method also improves the accuracy and consistency of label generation, since machine learning algorithms are able to maintain consistent data better than humans. Therefore, this advancement has far-reaching implications for large-scale or real-time water-body detecting projects, especially when large amounts of SAR image data need to be processed.
The method proposed in this paper combines the advantages of optimization algorithms and deep-learning techniques to make it perform well in the task of water-body detection. First, through a refined pixel-level classification task, the classifier generates initial result images with blurred water-body boundaries. The optimizer then refines this initial result to produce a final result image with sharp water-body boundaries. This combined method takes full advantage of the capabilities of optimization algorithms in exploring complex solution spaces, adapting to different data distributions, and finding global optimal or near-optimal solutions, as well as the advantages of deep learning in automatic learning and feature extraction. In addition, the accuracy of water-body detection is evaluated by several performance metrics, including ACC, precision, F1-Score, recall, and Kappa coefficient. The results show that all indicators exceeded 93%, which further indicated the accuracy and reliability of this method in water-body detection. Therefore, in future work, we plan to expand the dataset and test the model in more complex scenarios to further improve the performance and applicability of the model. At the same time, we will also look for more effective optimization algorithms and deep-learning network models to improve the accuracy and speed of water-body detection.
Although the method proposed in this paper performs well in experiments, there are still some limitations and open issues. First, our method is mainly tested in simple scenarios and may require further validation in more complex scenarios. Second, our dataset is relatively small, which may affect the generalization ability of the model. In addition, although the method in this paper reduces the workload of manual labeling, it still requires certain computing resources to perform the DBO-K-means and training of deep-learning models.
Additionally, this paper explored the time consumption of DK-CO, SegNet, U-Net, MRF, ACM, and FCM. Through ten experiments, it was concluded that the time consumption of traditional machine learning methods was relatively less, and the time consumption of deep learning is similar. The result is shown in
Figure 10. For research purposes, it is worthwhile to use an optimizer to improve accuracy compared with traditional machine learning and deep learning without an optimizer.
5. Conclusions
A water-body detection method based on SAR images is proposed. The labels required for CO training are obtained by using the DK method. Then, the labels and training dataset are brought into CO. Finally, CO is used to perform a dense pixel-by-pixel classification task and fine-grained optimization of the water-body boundary to obtain a result map with a clear water-body boundary. The ACC, precision, F1-score, recall, and Kappa coefficients reach 97.22%, 93.45%, 95.21%, 97.05%, and 93.25%, respectively. The results show the effectiveness of the proposed method. Moreover, the method is applicable to other areas, where fine water-body and complex terrain can be detected. However, our study also has some limitations. First, our method is mainly tested in simple scenarios and may require further validation in more complex scenarios. In addition, our dataset is relatively small, which may have affected the generalization ability of the model. In future work, we plan to expand the dataset and test the model in more complex scenarios to further improve the performance and applicability of the model. This study can provide a new idea and method for remote sensing detection of water bodies.
Author Contributions
Analysis, Q.Y. and Y.X.; methodology, Y.X., H.Z. and K.Y.; validation, Y.X., H.Z. and Q.Y.; resources, Y.X., H.Z. and Q.Y.; writing—original draft preparation, Q.Y., Y.X. and H.Z.; writing—review and editing, Q.Y., C.Y. All authors have read and agreed to the published version of the manuscript.
Funding
This work was supported by the National Natural Science Foundation of China (42101386), the Key Laboratory of Natural Resources Monitoring and Regulation in Southern Hilly Region, Ministry of Natural Resources of the People’s Republic of China (NRMSSHR2022Z01) and the Key Laboratory of Land Satellite Remote Sensing Application, Ministry of Natural Resources of the People’s Republic of China (KLSMNR-202302).
Data Availability Statement
Not applicable.
Acknowledgments
Thanks to all anonymous reviewers and editors for their comments and suggestions, making this paper’s content more rigorous and meaningful.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Wu, L.; Niu, S.; Guo, Z.; Li, N. A Novel Waterline Mapping Method of SAR Image with Sub-Pixel Accuracy. Radar Sci. Technol. 2020, 18, 494–500. [Google Scholar]
- Guo, Z.; Wu, L.; Huang, Y.; Guo, Z.; Zhao, J.; Li, N. Water-Body Segmentation for SAR Images: Past, Current, and Future. Remote Sens. 2022, 14, 1752. [Google Scholar] [CrossRef]
- Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
- Sun, Z.; Dai, M.; Leng, X.; Lei, Y.; Xiong, B.; Ji, K.; Kuang, G. An anchor-free detection method for ship targets in high-resolution SAR images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 7799–7816. [Google Scholar] [CrossRef]
- Rizaev, I.G.; Karakuş, O.; Hogan, S.J.; Achim, A. Modeling and SAR imaging of the sea surface: A review of the state-of-the-art with simulations. ISPRS J. Photogramm. Remote Sens. 2022, 187, 120–140. [Google Scholar] [CrossRef]
- Guo, H.; Yang, X.; Wang, N.; Gao, X. A CenterNet++ model for ship detection in SAR images. Pattern Recognit. 2021, 112, 107787. [Google Scholar] [CrossRef]
- Jiang, S.; Tarabalka, Y.; Yao, W.; Hong, Z.; Feng, G. Space-to-speed architecture supporting acceleration on VHR image processing. ISPRS J. Photogramm. Remote Sens. 2023, 198, 30–44. [Google Scholar] [CrossRef]
- Zhao, Y.; Zhao, L.; Xiong, B.; Kuang, G. Attention receptive pyramid network for ship detection in SAR images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 2738–2756. [Google Scholar] [CrossRef]
- Jiang, S.; Yao, W.; Wong, M.S.; Li, G.; Hong, Z.; Kuc, T.Y.; Tong, X. An optimized deep neural network detecting small and narrow rectangular objects in Google Earth images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 1068–1081. [Google Scholar] [CrossRef]
- Geng, J.; Jiang, W.; Deng, X. Multi-scale deep feature learning network with bilateral filtering for SAR image classification. ISPRS J. Photogramm. Remote Sens. 2020, 167, 201–213. [Google Scholar] [CrossRef]
- Chen, J.; Lv, J.; Li, N.; Wang, Q.; Wang, J. External Groundwater Alleviates the Degradation of Closed Lakes in Semi-Arid Regions of China. Remote Sens. 2020, 12, 45. [Google Scholar] [CrossRef] [Green Version]
- Pradhan, B.; Sameen, M.I.; Kalantar, B. Optimized Rule-Based Flood Mapping Technique Using Multitemporal RADARSAT-2 Images in the Tropical Region. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 3190–3199. [Google Scholar] [CrossRef]
- Ai, J.; Tian, R.; Luo, Q.; Jin, J.; Tang, B. Multi-Scale Rotation-Invariant Haar-Like Feature Integrated CNN-Based Ship Detection Algorithm of Multiple-Target Environment in SAR Imagery. IEEE Trans. Geosci. Remote Sens. 2019, 57, 10070–10087. [Google Scholar] [CrossRef]
- Gong, M.; Zhao, J.; Liu, J.; Miao, Q.; Jiao, L. Change Detection in Synthetic Aperture Radar Images Based on Deep Neural Networks. IEEE Trans. Neural Netw. Learn. Syst. 2016, 27, 125–138. [Google Scholar] [CrossRef] [PubMed]
- Wu, L.; Sun, M.; Min, L.; Zhao, J.; Li, N.; Guo, Z. An improved method of algal-bloom discrimination in Taihu Lake using Sentinel-1A data. In Proceedings of the 2019 6th Asia-Pacific Conference on Synthetic Aperture Radar (APSAR), Xiamen, China, 26–29 November 2019; pp. 1–5. [Google Scholar]
- Boulze, H.; Korosov, A.; Brajard, J. Classification of Sea Ice Types in Sentinel-1 SAR Data Using Convolutional Neural Networks. Remote Sens. 2020, 12, 2165. [Google Scholar] [CrossRef]
- Ronneberger, O.; Philipp, F.; Thomas, B. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015. Part III 18. [Google Scholar]
- Zhang, Z.; Liu, Q.; Wang, Y. Road extraction by deep residual u-net. IEEE Geosci. Remote Sens. Lett. 2018, 15, 749–753. [Google Scholar] [CrossRef] [Green Version]
- Pai, M.; Mehrotra, V.; Aiyar, S.; Verma, U.; Pai, R.M. Automatic segmentation of river and land in sar images: A deep learning approach. In Proceedings of the IEEE Second International Conference on Artificial Intelligence and Knowledge Engineering (AIKE), Sardinia, Italy, 3–5 June 2019. [Google Scholar]
- Pai, M.; Mehrotra, V.; Verma, U.; Pai, R.M. Improved semantic segmentation of water bodies and land in SAR images using generative adversarial networks. Int. J. Semant. Comput. 2020, 14, 55–69. [Google Scholar] [CrossRef]
- Zhang, L.; Xia, J. Flood detection using multiple Chinese satellite datasets during 2020 China summer floods. Remote Sens. 2021, 14, 51. [Google Scholar] [CrossRef]
- Katiyar, V.; Nopphawan, T.; Masahiko, N. Near-real-time flood mapping using off-the-shelf models with SAR imagery and deep learning. Remote Sens. 2021, 13, 2334. [Google Scholar] [CrossRef]
- Dong, Z.; Wang, G.; Amankwah, S.O.Y.; Wei, X.; Hu, Y.; Feng, A. Monitoring the summer flooding in the Poyang Lake area of China in 2020 based on Sentinel-1 data and multiple convolutional neural networks. Int. J. Appl. Earth Obs. Geoinf. 2021, 102, 102400. [Google Scholar] [CrossRef]
- Li, J.; Wang, C.; Wang, S.; Zhang, H.; Fu, Q.; Wang, Y. Gaofen-3 sea ice detection based on deep learning. In Proceedings of the 2017 Progress in Electromagnetics Research Symposium–Fall (PIERS–FALL), Singapore, 19–22 November 2017; pp. 933–939. [Google Scholar]
- Wang, C.; Zhang, H.; Wang, Y.; Zhang, B. Sea Ice Classification with Convolutional Neural Networks Using Sentinel-L Scansar Images. In Proceedings of the IGARSS 2018–2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 7125–7128. [Google Scholar]
- Li, N.; Niu, S. High-precision SAR image water segmentation method based on local super-resolution reconstruction. J. Radars 2020, 9, 174–184. [Google Scholar]
- Zhang, J.; Xing, M.; Sun, G. A SAR Image Water Segmentation Algorithm Based on Dense Depth Separation Convolution. J. Radars 2019, 8, 400–412. [Google Scholar]
- Hertel, V.; Chow, C.; Wani, O.; Wieland, M.; Martinis, S. Probabilistic SAR-based water segmentation with adapted Bayesian convolutional neural network. Remote Sens. Environ. 2023, 285, 113388. [Google Scholar] [CrossRef]
- Wieland, M.; Martinis, S.; Kiefl, R.; Gstaiger, V. Semantic segmentation of water bodies in very high-resolution satellite and aerial images. Remote Sens. Environ. 2023, 287, 113452. [Google Scholar] [CrossRef]
- Qin, X.; Zhang, Z.; Huang, C.; Gao, C.; Dehghan, M.; Jagersand, M. Basnet: Boundary-aware salient object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7479–7489. [Google Scholar]
- Zhang, J.; Xing, M.; Sun, G.C.; Chen, J.; Li, M.; Hu, Y.; Bao, Z. Water Body Detection in High-Resolution SAR Images With Cascaded Fully-Convolutional Network and Variable Focal Loss. IEEE Trans. Geosci. Remote Sens. 2021, 59, 316–332. [Google Scholar] [CrossRef]
- Jin, S. Analysis of the Distribution and Seasonal Variability of the South China Sea Water Masses based on the K-Means Cluster Method; First Institute of Oceanography MNR: Qingdao, China, 2017. [Google Scholar]
- Gao, C. Water Mass Division Algorithm Based on K-means Clustering and Visualization; Ocean University of China: Qingdao, China, 2008. [Google Scholar]
- Xue, J.; Shen, B. Dung beetle optimizer: A new meta-heuristic algorithm for global optimization. J. Supercomput. 2022, 79, 1–32. [Google Scholar] [CrossRef]
- Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [Green Version]
- Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
- Lopez-Martinez, C.; Fabregas, X. Model-based Polarimetric SAR Speckle Filter. IEEE Trans. Geosci. Remote Sens. 2008, 46, 3894–3907. [Google Scholar] [CrossRef]
- Deng, H.; Clausi, D. Unsupervised segmentation of synthetic aperture Radar sea ice imagery using a novel Markov random field model. IEEE Trans. Geosci. Remote Sens. 2005, 43, 528–538. [Google Scholar] [CrossRef]
- Hahmann, T.; Wessel, B. Surface Water Body Detection in High-Resolution TerraSAR-X Data using Active Contour Models. In Proceedings of the 8th European Conference on Synthetic Aperture Radar, Aachen, Germany, 7–10 June 2010; pp. 1–4. [Google Scholar]
- Hu, H.; Dai, G.; Wang, M.; Yi, P. Detection of river in SAR images. In Proceedings of the 2012 8th International Conference on Information Science and Digital Content Technology (ICIDT2012), Jeju Island, Republic of Korea, 26–28 June 2012; pp. 632–635. [Google Scholar]
- Kumar, A.; Mishra, V.; Panigrahi, R.K.; Martorella, M. Application of Hybrid-pol SAR in Oil-spill Detection. IEEE Geosci. Remote Sens. Lett. 2023, 20, 1–5. [Google Scholar] [CrossRef]
- Liu, J.; Wang, Y. Water Body Extraction in Remote Sensing Imagery Using Domain Adaptation-Based Network Embedding Selective Self-Attention and Multi-Scale Feature Fusion. Remote Sens. 2022, 14, 3538. [Google Scholar] [CrossRef]
Figure 1.
SAR and optical images of the study area. (a) SAR image on 1 August 2022. (b). The optical image on 29 September 2022.
Figure 1.
SAR and optical images of the study area. (a) SAR image on 1 August 2022. (b). The optical image on 29 September 2022.
Figure 2.
Flowchart of the proposed method.
Figure 2.
Flowchart of the proposed method.
Figure 3.
Flowchart of the DeepLabV3+.
Figure 3.
Flowchart of the DeepLabV3+.
Figure 4.
Schematic diagram of subregions selection.
Figure 4.
Schematic diagram of subregions selection.
Figure 5.
Result images of Region I. (a) SAR image. (b,c) represent the results of CO’s classifier and optimizer, respectively. (d–h) are the results of SegNet, U-Net MRF, ACM, and FCM, respectively.
Figure 5.
Result images of Region I. (a) SAR image. (b,c) represent the results of CO’s classifier and optimizer, respectively. (d–h) are the results of SegNet, U-Net MRF, ACM, and FCM, respectively.
Figure 6.
Result images of Region II. (a) SAR image. (b,c) represent the results of CO’s classifier and optimizer, respectively. (d–h) are the results of SegNet, U-Net, MRF, ACM, and FCM, respectively.
Figure 6.
Result images of Region II. (a) SAR image. (b,c) represent the results of CO’s classifier and optimizer, respectively. (d–h) are the results of SegNet, U-Net, MRF, ACM, and FCM, respectively.
Figure 7.
Confusion matrix result for Region I. (a) results of DK-CO, (b) results of SegNet, (c) results of U-Net, (d) results of MRF, (e) results of ACM, and (f) results of FCM.
Figure 7.
Confusion matrix result for Region I. (a) results of DK-CO, (b) results of SegNet, (c) results of U-Net, (d) results of MRF, (e) results of ACM, and (f) results of FCM.
Figure 8.
Confusion matrix result for Region II. (a) results of DK-CO, (b) results of SegNet, (c) results of U-Net, (d) results of MRF, (e) results of ACM, and (f) results of FCM.
Figure 8.
Confusion matrix result for Region II. (a) results of DK-CO, (b) results of SegNet, (c) results of U-Net, (d) results of MRF, (e) results of ACM, and (f) results of FCM.
Figure 9.
Visual comparison of evaluation indicators for different methods.
Figure 9.
Visual comparison of evaluation indicators for different methods.
Figure 10.
Time consumption of various methods.
Figure 10.
Time consumption of various methods.
Table 1.
Detailed information on SAR data.
Table 1.
Detailed information on SAR data.
ID | Time (M/D/Y) | Range Spacing (m) | Azimuth Spacing (m) | Orbit Direction | Processing Level |
---|
1 | 24 July 2022 | 10 | 10 | Ascending | L1-GRD (IW) |
2 | 1 August 2022 | 10 | 10 | Ascending | L1-GRD (IW) |
Table 2.
Confusion matrix.
Table 2.
Confusion matrix.
| | Generated Label |
---|
| | Water | Non-Water |
---|
Ground truth | Water | True Position (TP) | False Negative (FN) |
Non-water | False Positive (FP) | True Negative (TN) |
Table 3.
The accuracy of water-body detection.
Table 3.
The accuracy of water-body detection.
Method | ACC | Precision | F1-Score | Recall | Kappa (%) |
---|
(%) | (%) | (%) | (%) |
---|
DK-CO | 97.22 | 93.45 | 95.21 | 97.05 | 93.25 |
SegNet | 95.68 | 93.15 | 94.35 | 95.58 | 92.65 |
U-Net | 95.81 | 94.32 | 94.53 | 94.32 | 92.93 |
MRF | 95.92 | 88.99 | 93.14 | 97.86 | 90.26 |
SubPixelACM | 96.29 | 93.62 | 93.35 | 93.12 | 90.76 |
FCM | 95.79 | 97.57 | 92.04 | 92.66 | 89.18 |
| Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).