Next Article in Journal
The Survey of Lava Tube Distribution in Jeju Island by Multi-Source Data Fusion
Next Article in Special Issue
Detection of Flash Flood Inundated Areas Using Relative Difference in NDVI from Sentinel-2 Images: A Case Study of the August 2020 Event in Charikar, Afghanistan
Previous Article in Journal
Coseismic Slip Model of the 2021 Maduo Earthquake, China from Sentinel-1 InSAR Observation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Flash Flood Water Depth Estimation Using SAR Images, Digital Elevation Models, and Machine Learning Algorithms

Civil Engineering Department, College of Engineering, Najran University, King Abdulaziz Road, Najran P.O. Box 1988, Saudi Arabia
Remote Sens. 2022, 14(3), 440; https://doi.org/10.3390/rs14030440
Submission received: 22 December 2021 / Revised: 7 January 2022 / Accepted: 10 January 2022 / Published: 18 January 2022
(This article belongs to the Special Issue Remote Sensing in Urban Flooding Monitoring)

Abstract

:
In this article, the local spatial correlation of multiple remote sensing datasets, such as those from Sentinel-1, Sentinel-2, and digital surface models (DSMs), are linked to machine learning (ML) regression algorithms for flash floodwater depth retrieval. Edge detection filters are applied to remote sensing images to extract features that are used as independent features by ML algorithms to estimate flood depths. Data of dependent variables were obtained from the Hydrologic Engineering Center’s River Analysis System (HEC-RAS 2D) simulation model, as applied to the New Cairo, Egypt, post-flash flood event from 24–26 April 2018. Gradient boosting regression (GBR), random forest regression (RFR), linear regression (LR), extreme gradient boosting regression (XGBR), multilayer perceptron neural network regression (MLPR), k-nearest neighbors regression (KNR), and support vector regression (SVR) were used to estimate floodwater depths; their outputs were compared and evaluated for accuracy using the root-mean-square error (RMSE). The RMSE accuracy for all ML algorithms was 0.18–0.22 m for depths less than 1 m (96% of all test data), indicating that ML models are relatively portable and capable of computing floodwater depths using remote sensing data as an input.

1. Introduction

Floodwater depth identification during or after flash flood events is critical in determining hazard degrees and risk zone maps for the economy and human life [1,2]. Compared with direct surveying methods, measurement techniques, such as side-scan and multi-beam sonar, hydrologic modeling, and flow water depth, based on remote sensing are fast, large-scale, and quasi-synchronous with high spatial resolutions. Furthermore, direct surveying methods to determine floodwater depth can be extremely precise, but they are greatly influenced by weather conditions and costly, and surveying field crews are not authorized to reach sensitive flooded areas.
In addition, optical and synthetic aperture radar (SAR) images, and the digital elevation model (DEM) based on airborne light detection and ranging (LiDAR), have been integrated and classified for floodwater surface identification [3,4]. Although SAR data are superior to optical satellite data, as SARs can penetrate cloud cover, they suffer from a long revisit time [5,6,7].
Studies have employed a variety of hydrodynamic 1D, 2D, and 3D software to simulate water levels and floodwater depths, including HEC-RAS, Delft-3D, and LISFLOOD-FP [8,9]. These models require rainfall, soil moisture, flood maps, gauge discharge, cross-sections, and other hydrological inputs to simulate water depth. The disadvantage of employing hydrodynamic models to simulate floodwater depth is the requirement of a large input dataset and extensive computation and calibration.
Recently, the growth of urban regions and infrastructures such as highways and trains has extended the flood-prone regions in New Cairo, Egypt, which has increased the severity of floods [10]. Several studies have mapped flood inundation in New Cairo, but none have addressed floodwater depth. Numerous hydrological characteristics and flood models, such as the Hydraulic Engineering Center’s River Analysis System (HEC-RAS), have been applied to map and forecast flood inundation [11,12,13], and machine learning (ML) algorithms have been used to map flood inundation [14,15,16]. No study has attempted to quantify floodwater depth and duration during flood disasters in Egypt so far.
Supervised ML regression algorithms are used to learn a function that combines a set of feature data (independent variables) to predict a dependent variable [17,18]. Random forest regression (RFR) is a kind of ensemble learning [19], i.e., a decision tree supervised ML algorithm based on a set of rules [17].
Furthermore, ML has been used for data extraction, pattern recognition, regression, and classification problems since the start of the 21st century. Research has shown that ML algorithms, such as support vector machine (SVM), RFR, and extreme gradient boosting (XGBR), can efficiently produce spatial predictions [19,20]. RF and XGBR use both decision trees and the bagging technique [21]. Some researchers have trained ML algorithms to estimate water depths using the pixel reflectance values of satellite data, and validated the results through field observation [22].
The current work aims to estimate flash flood water depths by concatenating the spectral information of remote sensing data, such as Sentinel-1 SAR data, digital surface models (DSMs), and land-use maps, at feature levels, and apply a regression algorithm using ML techniques, such as gradient boosting regression (GBR), RFR, linear regression (LR), XGBR, multilayer perceptron neural network regression (MLPR), k-nearest neighbors regression (KNR), and support vector regression (SVR). ML has many properties that make it appropriate for obtaining water depths from remote sensing images. For example, ML algorithms are ideal for processing locally joined data, such as raster data with a spatial grid structure. Moreover, the floodwater levels of unknown locations can be considered the weighted averages of nearby known water depths, as obtained by geographic interpolation. Therefore, I investigate the impact of adjacent pixels of remote sensing data on floodwater depth prediction through ML algorithms. An informative and appropriate number of features should be derived for the subsequent regression ML algorithms. The performance of extracted features and their importance in improving accuracy and accelerating the algorithm are investigated.

2. Study Area and Dataset

2.1. Study Area

New Cairo is in the southeast of Egypt’s Cairo Province. It was established in 2000, with a land area of approximately 70,000 acres. A 90-square-kilometer area with altitudes ranging from 200 m to 420 m above the mean sea level was chosen to model flash floodwater depths. Figure 1 shows a true-color Sentinel-2 image of the study area, from which a land-use map was created for use in calculating roughness values.

2.2. Rainfall Intensity Data

Cairo has a four-season hot desert climate. The study area receives the majority of its rainfall from November through April. Daily rainfall data for the area were obtained from the NASA Prediction of Worldwide Energy Resources (POWER) project, which delivers global weather data with a spatial resolution of 0.5°. Based on historical meteorological information collected from the POWER website for 1981–2018, the average annual precipitation in the watershed is 51.2 mm/y (https://power.larc.nasa.gov/data-access-viewer/ (accessed on 20 August 2021)). Figure 2 shows the hourly rain intensity during the stormy period, which was used to simulate the water depth of the study area using the HEC-RAS 2D program.

2.3. DSM Data Preparation

A DSM is a raster map representing the above-ground elements. A DEM, without objects such as trees and buildings, shows the shape of the bare earth. Many DSM and DEM data sources, varying from low- to high-resolution, have been produced in the past 30 years [23,24]. In particular, the Shuttle Radar Topography Mission (SRTM)-30 m and Advanced Spaceborne Thermal Emission and Reflection Radiometer (ASTER)-30 m are freely available and can be used for engineering applications, such as flood mapping. Data from the Advanced Land Observing Satellite (ALOS)-30 m DSM for the current study area can be downloaded from the ALOS website (https://scihub.copernicus.eu/ (accessed on 10 August 2021)). The region of interest (ROI) was clipped for the study area. Figure 3 shows the DSM with Sentinel-2 true color as the background.

2.4. Sentinel-1 Data

SAR penetrates clouds and dense vegetation, which makes it suitable for detecting water areas in all weather conditions. Sentinel-1 data, as a SAR product of the European Copernicus program, are freely accessible through the Copernicus Open Access Hub (https://step.esa.int/main/toolboxes/snap/ (accessed on 15 August 2021)), and cover the globe with a six-day temporal resolution and 10 m spatial resolution. The level-1 Interferometric Wide Swath (IW) product, acquired on 27 April 2018, was used to extract wet areas and estimate water depths during a flash flood in the study area. The Sentinel Application Platform (SNAP) software (https://step.esa.int/main/toolboxes/snap/ (accessed on 1 July 2021)) deals with all Sentinel-1 data products. Fundamental SAR image-processing steps include radiometric calibration, speckle filtering, terrain correction, and sigma naught ( σ 0 ) value calculation for vertical horizontal (VH) polarization. To improve the visualization of water areas, the pixel values were converted to backscattering in decibels (dB) of σ V H 0 . The final SAR image was re-sampled to a 30 m ground sample distance (GSD) by nearest-neighbor interpolation and reprojected into the WGS84/UTM coordinate system.

2.5. Land Use

The land-use map was prepared using the ArcGIS 10.4 software by applying the maximum likelihood (ML) supervised classification method for the Sentinel-2 image, which entails four steps: (1) determine the number of layers; (2) select the training sample for each class; (3) estimate the mean vector and covariance matrix for each training sample layer; and (4) classify each pixel in the satellite image based on the covariance matrix and mean vector. Figure 4 shows the obtained classification land-use map. Table 1 shows that the overall classification accuracy was 87.1%, with a kappa coefficient of 0.815. The kappa coefficient indicates how well the classification results and truth values agree. A kappa value of 1 indicates complete agreement, whereas a value of 0 indicates no agreement.

2.6. Water Depth Extraction (Dependent Variable)

There were no direct measurements of water depth during flood times, so the DSM, land-use map, and rainfall data were used to analyze the unsteady flow using HEC-RAS 2D version 5.0.6. to estimate the water depth. The water depth from 24–27 April 2018 was generated using the modeling software. The digital map of the modeled maximum floodwater depth over 24–27 April is shown in Figure 5.

3. Methodology and Data Preparation

3.1. Research Methodology

An ML algorithm consists of a dependent variable (Y) and an independent variable (X). In our study, Y represents floodwater depth, and X represents DSM, land-use map, and SAR image information. Water depth values will be predicted from a set of predictors or independent variables. In addition, the water depth and independent data images should match spatially and have the same resolution. The algorithm randomly divides the data into training and test sets before training the model. The ML algorithm updates its input parameters and generates the water depth model throughout the training phase. The model is validated by predicting new water depths using test data and calculating the root mean square error (RMSE) between the predicted and test values. Figure 6 shows a full flowchart for this procedure. I used 80% of the collected elevation data to train the algorithm, and 20% to validate the solutions.

3.2. Machine Learning Data Preparation

Any ML predictive modeling project is unique, but there are common basic processes, such as identifying the problem, preparing the data, and assessing and finalizing models [25]. The current project, predicting the depth of floodwater, involves a continuous output quantity rather than a discrete class label. Hence, I used ML regression modeling. Data preparation focuses on converting the gathered raw data to a format that can be used in modeling [26]. I selected appropriate metrics to evaluate the model and optimum hyperparameter tuning as part of the model evaluation. Independent variables (features) are used as the input for ML algorithms. The next two subsections describe how to obtain and manipulate the ML data input.

3.2.1. Dependent Feature Extraction and Preparation (Y)

During the training and testing of the chosen ML algorithm, the water depth map displayed in Figure 5 (width = 458 and height = 274 pixels) was used as a dependent variable. Before data are used, they must be smoothed by a Gaussian filter to remove noise and outliers [27]. Figure 7 shows a histogram of water depth before and after Gaussian smoothing with σ = 1. Values were then reshaped from 2D (770 × 700) to a 1D Pandas DataFrame [28].

3.2.2. Independent Feature Extraction and Preparation (X)

The total number of independent features used to train the ML algorithms was 33, as extracted from SAR images, DSM models, and land-use maps. By applying some edge detection filters to the SAR image in addition to its original pixel values, 30 features were extracted. The earth surface slope and stream order calculated based on a DSM map were added as two more features. The land-use map (Figure 4) was used as another independent feature. Raster images were resized to a raster size of 458 × 274 pixels and a resolution of 30 m before any computations. The total number of sample datasets used after reshaping the 2D raster maps to a 1D vector data form was 125,492.

3.2.3. Independent Feature Extraction Algorithms and Methods

The digital filters and equations used to collect independent variables for input to ML algorithms will be discussed.
Raster SAR images represent pixels with a 30 m ground sample distance, and each pixel has a σ 0 value for σ V H 0   polarization. Gabor kernel values and those from additional edge detection filters, such as Canny, Sobel, Roberts, and Prewitt, were estimated as independent features based on SAR image pixel values. Gabor filters are directional filters used for edge detection and analysis if an image has a sudden sharp increase [29,30], and are determined as
G ( x , y ; λ , θ , ψ , σ , γ ) = exp [ x 2 + γ 2 y 2 2 σ 2 ] exp [ i ( 2 π x λ + ψ ) ] ,
where
x = x   c o s θ + y   s i n θ   a n d   y = x   s i n θ + y   c o s θ
λ is the wavelength, σ is the standard deviation of the Gaussian function in the x and y directions, θ indicates the orientation of the filter, and ψ is the phase offset. The shape of a Gabor filter depends on the aspect ratio γ. If γ is equal to one, then the filter appears as a circle. The shape of the filter will gradually change from an ellipsoid to a straight line when γ is close to zero [31]. By varying the Gabor parameters λ, θ, ψ, σ, and γ, different orientations are used to analyze the texture or obtain features from images. Figure 8 shows the source python code used to extract the Gabor filter parameters using multi for loops. Accordingly, 32 Gabor labels were generated using a Python code, each having the values of its own parameters.
The OpenCV open-source code library was used to calculate the values of the Gabor features and other edge detection filters [32,33].
The surface slope indicates the steepness of the ground’s surface. The slope between two points on the Earth’s surface can be calculated by dividing their elevation difference by the horizontal distance between them. For the DSM surface, the surface slope in degrees can be calculated as
S l o p e = t a n 1 [ ( d z d x ) 2 + ( d z d y ) 2 ] ,
where ( d z d x ,   d z d x ) is the perpendicular rate of change for the center cell of a moving 3 × 3 pixel window for grid-based DSM [34,35,36]. Figure 9 shows the pixel values of a moving 3 × 3 window, where the neighbors of the center cell e are identified by the letters a to i. The perpendicular rate of change for the x and y directions is calculated as
d z d x = ( c + 2 f + i ) ( a + 2 d + g ) 8 x _ c e l l s i z e   ;   d z d y = ( g + 2 h + i ) ( a + 2 b + c ) 8 y _ c e l l s i z e ,
The ArcGIS 10.7 software was used to calculate the slope for the current study. In hydrology, the stream or waterbody order is a positive, whole number that indicates the degree of splitting in a stream channel. Using ArcGIS software version 10.7, a raster image representing the stream order for the current study site was delineated from the DSM [37]. Subsequently, Sentinel-2 remote sensing satellite photos were classified to obtain land-cover features, including water, roads, green areas, buildings, and bare soil, identified by integers 1–5, respectively. Figure 10 and Table 2 show the 32 features extracted to train the ML algorithms and predict water depths. Some features were removed because their output values were constant at a given value and did not vary at each pixel. As a result, the overall number of features dropped from 33 to 32.
The collected data were randomly divided into a training dataset (80%, n = 100,393) to generate water depths using various interpolation ML algorithms, and a validation dataset (20%, n = 25,099) to calculate the accuracy of each model.

3.3. Quality Assessment

Of the measures for assessing derived ML models, RMSE is commonly used when comparing predicted and actual independent values. The RMSE uses the squared error; hence, greater errors have a stronger influence. The RMSE is calculated as
R M S E = 1 n   j = 0 n ( y i y ^ ) 2
where y i and y ^ are the actual and predicted water depths, respectively. The RMSE was used to compare ML models.

4. Results

Several ML regression techniques, including support vector machine (SVR), random forest (RF), k-nearest neighbors (KNR), and extreme gradient boosting (Xgboost), have been suggested for the spatial interpolation of environmental variables, and several hybrid methods have been adopted.

4.1. Machine Learning Hyperparameter Tuning

The precision and accuracy of ML algorithms are determined by the input variables. Hyperparameter tuning is the process of finding the optimal hyperparameters to achieve high precision and accuracy. GridSearch is a function in the Scikit-learn package that is used to find the optimum parameters by building and evaluating multiple models with different hyperparameter combinations. Other parameter tuning methods, such as Random search and Bayesian optimization, can be implemented using Scikit-learn [38]. Figure 11A illustrates the influence of the number of estimators on the RF model’s accuracy, which was steady and did not improve after using more than 15 estimators. Figure 11B depicts the impact of using the number of nearby points on the tree model’s accuracy. The accuracy of the KNR model was stable and did not improve after employing more than eight neighboring points.
The optimal results can be summarized as follows: the best number of estimators for the RF algorithm was 12; the k-value for KNN was 7; for xgboost, the optimal number of estimators was 60; and for SVR, the kernel was ‘C’: 1, ‘epsilon’: 0.02, ‘gamma’: 0.01, ‘kernel’: ‘rbf’. The C value was used to adjust the error or margin, gamma was used with Gaussian rbf kernel, and epsilon was used to smooth the algorithm response.

4.2. Accuracy of Obtained ML Algorithms

When the number of input features is large, it is preferable to use the most significant ones during ML training to reduce processing time, enhance output accuracy, and make model interpretation and understanding easier. Some ML algorithms, such as RFR, assess the effectiveness of their input features [39,40]. Figure 12 shows the degree of importance of each data feature.

5. Discussion

As shown by the comparison results, the most effective feature was the surface slope, followed by the SAR backscattering values. Figure 13 shows 30 sample data outputs for the different used algorithms.
Although the RMSE is heavily influenced by the distribution of validation points, it is nevertheless a useful metric for assessing the accuracy of predicted water depths. The RMSEs of multiple water depth ranges are shown in Table 3. The GBR and XGBRFR approaches had the best accuracy for the overall test data, and MLPR and KNR had the worst accuracy. The SVR and XGBR algorithms had the most accurate models, with RMSE of 0.20 m at water depths of 0–1 m, contributing to 96% of the test data. Figure 14 shows the RMSE values for all conducted regression algorithms at different water depths.
Figure 14 shows the RMSE values for all conducted regression algorithms at different water depths. Previous methods depended on DEM and SAR data and could only be used for flat or gently sloping terrain [41]. The technique I propose is straightforward because it works on any terrain and simply requires SAR and DEM data.

6. Conclusions

In this study, water depth estimation after a flash flood event in New Cairo City, Egypt, was investigated using multiple ML algorithms. Several training datasets and ML techniques were combined. The backscattering spectral band of the SAR data was used to extract features to be used as independent inputs for ML algorithms. The water depth (dependent input) was extracted by hydrodynamic HEC-RAS 2D software to obtain 250,099 data samples and 33 features to train and validate the used algorithms; 80% of the input data were used for training and 20% for testing. I compared the accuracy of the obtained models. The water depths were classified into five groups, and 96.27% of total water depths fell between 0 and 1 m. All ML methods produced RMSE values of 0.18–0.22 m over this water depth range. The XGBR and SVR models had the best accuracy, and the KNR model had the worst. For deep water depths (more than 6 m), four sample datasets were found. Therefore, the accuracy of all approaches was reduced, as the RMSE values ranged from 7.20 to 6.77 m. Moreover, for all water depths, the prediction results were more consistent, and the GBR and XGBR models achieved the best accuracy. For effective forecasting of water depth from satellite data to build an emergency plan in case of floods, I recommend integrating various training datasets and machine learning algorithms.

Funding

The project was funded by the Deanship of Scientific Research at Najran University, project number (NU/SERC/10/550).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the author upon reasonable request.

Acknowledgments

The authors are thankful to the Deanship of Scientific Research at Najran University for funding this work under the General Research Funding program grant code (NU/SERC/10/550).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Townsend, P.A.; Walsh, S.J. Modeling floodplain inundation using an integrated GIS with radar and optical remote sensing. Geomorphology 1998, 21, 295–312. [Google Scholar] [CrossRef]
  2. Vishnu, C.L.; Sajinkumar, K.S.; Oommen, T.; Coffman, R.A.; Thrivikramji, K.P.; Rani, V.R.; Keerthy, S. Satellite-based assessment of the August 2018 flood in parts of Kerala, India. Geomat. Nat. Hazards Risk 2019, 10, 758–767. [Google Scholar] [CrossRef] [Green Version]
  3. Irwin, K.; Beaulne, D.; Braun, A.; Fotopoulos, G. Fusion of SAR, optical imagery and airborne LiDAR for surface water detection. Remote Sens. 2017, 9, 890. [Google Scholar] [CrossRef] [Green Version]
  4. Musa, Z.N.; Popescu, I.; Mynett, A. A review of applications of satellite SAR, optical, altimetry and DEM data for surface water modelling, mapping and parameter estimation. Hydrol. Earth Syst. Sci. 2015, 19, 3755–3769. [Google Scholar] [CrossRef] [Green Version]
  5. Bovenga, F.; Bovenga, F.; Belmonte, A.; Refice, A.; Pasquariello, G.; Nutricato, R.; Nitti, D.O.; Chiaradia, M.T. Performance analysis of satellite missions for multi-temporal SAR interferometry. Sensors 2018, 18, 1359. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Bioresita, F.; Puissant, A.; Stumpf, A.; Malet, J.P. A method for automatic and rapid mapping of water surfaces from Sentinel-1 imagery. Remote Sens. 2018, 10, 217. [Google Scholar] [CrossRef] [Green Version]
  7. Alsdorf, D.E.; Rodríguez, E.; Lettenmaier, D.P. Measuring surface water from space. Rev. Geophys. 2007, 45. [Google Scholar] [CrossRef]
  8. Yalcin, E. Two-dimensional hydrodynamic modelling for urban flood risk assessment using unmanned aerial vehicle imagery: A case study of Kirsehir, Turkey. J. Flood Risk Manag. 2019, 12, e12499. [Google Scholar] [CrossRef]
  9. Costabile, P.; Costanzo, C.; Ferraro, D.; Barca, P. Is HEC-RAS 2D accurate enough for storm-event hazard assessment? Lessons learnt from a benchmarking study based on rain-on-grid modelling. J. Hydrol. 2021, 603, 126962. [Google Scholar] [CrossRef]
  10. El Afandi, G.; Morsy, M. Developing an Early Warning System for Flash Flood in Egypt: Case Study Sinai Peninsula. In Advances in Science, Technology and Innovation; Springer: Berlin/Heidelberg, Germany, 2020; pp. 45–60. [Google Scholar]
  11. Abdeldayem, O.M.; Eldaghar, O.; KMostafa, M.; MHabashy, M.; Hassan, A.A.; Mahmoud, H.; Morsy, K.M.; Abdelrady, A.; Peters, R.W. Mitigation plan and water harvesting of flashflood in arid rural communities using modelling approach: A case study in Afouna village, Egypt. Water 2020, 12, 2565. [Google Scholar] [CrossRef]
  12. Sadek, M.; Li, X.; Mostafa, E.; Dossou, J.F. Monitoring flash flood hazard using modeling-based techniques and multi-source remotely sensed data: The case study of Ras Ghareb City, Egypt. Arab. J. Geosci. 2021, 14, 2030. [Google Scholar] [CrossRef]
  13. Elkhrachy, I.; Pham, Q.B.; Costache, R.; Mohajane, M.; Rahman, K.U.; Shahabi, H.; Linh, N.T.T.; Anh, N.T. Sentinel-1 remote sensing data and Hydrologic Engineering Centres River Analysis System two-dimensional integration for flash flood detection and modelling in New Cairo City, Egypt. J. Flood Risk Manag. 2021, 14, e12692. [Google Scholar] [CrossRef]
  14. El-Haddad, B.A.; Youssef, A.M.; Pourghasemi, H.R.; Pradhan, B.; El-Shater, A.H.; El-Khashab, M.H. Flood susceptibility prediction using four machine learning techniques and comparison of their performance at Wadi Qena Basin, Egypt. Nat. Hazards 2021, 105, 83–114. [Google Scholar] [CrossRef]
  15. El-Magd, S.A.A.; Pradhan, B.; Alamri, A. Machine learning algorithm for flash flood prediction mapping in Wadi El-Laqeita and surroundings, Central Eastern Desert, Egypt. Arab. J. Geosci. 2021, 14, 323. [Google Scholar] [CrossRef]
  16. Mudashiru, R.B.; Sabtu, N.; Abustan, I. Quantitative and semi-quantitative methods in flood hazard/susceptibility mapping: A review. Arab. J. Geosci. 2021, 14, 941. [Google Scholar] [CrossRef]
  17. Friedman, J.; Hastie, T.; Tibshirani, R. The Elements of Statistical Learning; Springer series in statistics; Springer: New York, NY, USA, 2001; Volume 1, No. 10. [Google Scholar]
  18. Ghorpade, P.; Gadge, A.; Lende, A.; Chordiya, H.; Gosavi, G.; Mishra, A.; Hooli, B.; Ingle, Y.S.; Shaikh, N. Flood Forecasting Using Machine Learning: A Review. In Proceedings of the 2021 8th International Conference on Smart Computing and Communications: Artificial Intelligence, AI Driven Applications for a Smart World, ICSCC, Kochi, Kerala, India, 1–3 July 2021; Volume 2021, pp. 32–36. [Google Scholar] [CrossRef]
  19. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  20. Sekulić, A.; Kilibarda, M.; Heuvelink, G.; Nikolić, M.; Bajat, B. Random forest spatial interpolation. Remote Sens. 2020, 12, 1687. [Google Scholar] [CrossRef]
  21. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  22. Wu, Z.; Mao, Z.; Shen, W. Integrating Multiple Datasets and Machine Learning Algorithms for Satellite-Based Bathymetry in Seaports. Remote Sens. 2021, 13, 4328. [Google Scholar] [CrossRef]
  23. Elkhrachy, I. Vertical accuracy assessment for SRTM and ASTER Digital Elevation Models: A case study of Najran city, Saudi Arabia. Ain Shams Eng. J. 2018, 9, 1807–1817. [Google Scholar] [CrossRef]
  24. Mesa-Mingorance, J.L.; Ariza-López, F.J. Accuracy assessment of digital elevation models (DEMs): A critical review of practices of the past three decades. Remote Sens. 2020, 12, 2630. [Google Scholar] [CrossRef]
  25. Kuhn, M.; Johnson, K. Applied Predictive Modeling; Springer: Berlin/Heidelberg, Germany, 2013; Volume 26. [Google Scholar]
  26. Liu, H. Feature Engineering for Machine Learning and Data Analytics; O’Reilly Media, Inc.: Sevastopol, CA, USA, 2018. [Google Scholar]
  27. Davies, E.R. Machine Vision: Theory, Algorithms, Practicalities; Elsevier: Amsterdam, The Netherlands, 2004. [Google Scholar]
  28. McKinney, W.; Team, P.D. Pandas-Powerful python data analysis toolkit. Pandas—Powerful Python Data Anal. Toolkit 2015, 1625. Available online: https://pandas.pydata.org/docs/pandas.pdf (accessed on 15 September 2021).
  29. Fogel, I.; Sagi, D. Gabor filters as texture discriminator. Biol. Cybern. 1989, 61, 103–113. [Google Scholar] [CrossRef]
  30. Grigorescu, S.E.; Petkov, N.; Kruizinga, P. Comparison of texture features based on Gabor filters. IEEE Trans. Image Process. 2002, 11, 1160–1167. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Palm, C.; Lehman, T. Classification of color textures by gabor filtering. Mach. Graph. Vis. 2002, 11, 195–220. [Google Scholar]
  32. OpenCV, L. Computer Vision with the OpenCV Library; O’Reilly Media, Inc.: Sevastopol, CA, USA, 2008. [Google Scholar]
  33. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. Available online: http://jmlr.csail.mit.edu/papers/v12/pedregosa11a.html%5Cnhttp://arxiv.org/abs/1201.0490 (accessed on 15 September 2021).
  34. Skidmore, A.K. A comparison of techniques for calculating gradient and aspect from a gridded digital elevation model. Int. J. Geogr. Inf. Syst. 1989, 3, 323–334. [Google Scholar] [CrossRef]
  35. Zhou, Q.; Liu, X. Analysis of errors of derived slope and aspect related to DEM data properties. Comput. Geosci. 2004, 30, 369–378. [Google Scholar] [CrossRef]
  36. Cone, J. Principles of Geographical Information Systems by Peter A; Oxford University Press: Oxford, UK, 1998; Volume 54. [Google Scholar]
  37. Jenson, S.K. Applications of hydrologic information automatically extracted from digital elevation models. Hydrol. Process. 1991, 5, 31–44. [Google Scholar] [CrossRef]
  38. Buitinck, L.; Louppe, G.; Blondel, M.; Pedregosa, F.; Mueller, A.; Grisel, O.; Niculae, V.; Prettenhofer, P.; Gramfort, A.; Grobler, J.; et al. API design for machine learning software: Experiences from the scikit-learn project. arXiv 2013, arXiv:1309.0238. [Google Scholar]
  39. Hall, M.A. Correlation-based Feature Selection for Discrete and Numeric Class Machine Learning. Eff. Br. Mindfulness Interv. Acute Pain ExAn. Exam. Individ. Differ. 2015, 1, 1689–1699. Available online: https://researchcommons.waikato.ac.nz/handle/10289/1024 (accessed on 21 August 2021).
  40. Genuer, R.; Poggi, J.M.; Tuleau-Malot, C. Variable selection using random forests. Pattern Recognit. Lett. 2010, 31, 2225–2236. [Google Scholar] [CrossRef] [Green Version]
  41. Cian, F.; Marconcini, M.; Ceccato, P.; Giupponi, C. Flood depth estimation by means of high-resolution SAR images and lidar data. Nat. Hazards Earth Syst. Sci. 2018, 18, 3063–3084. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Ten-meter spatial resolution true-color image based on Sentinel-2 satellite data for the study area.
Figure 1. Ten-meter spatial resolution true-color image based on Sentinel-2 satellite data for the study area.
Remotesensing 14 00440 g001
Figure 2. Rainfall intensity during a flash flood on 24 April 2018.
Figure 2. Rainfall intensity during a flash flood on 24 April 2018.
Remotesensing 14 00440 g002
Figure 3. DSM of study area projected by WGS84/UTM coordinate system.
Figure 3. DSM of study area projected by WGS84/UTM coordinate system.
Remotesensing 14 00440 g003
Figure 4. Land-use map of the study area.
Figure 4. Land-use map of the study area.
Remotesensing 14 00440 g004
Figure 5. Simulated water depth using HEC-RAD 2D.
Figure 5. Simulated water depth using HEC-RAD 2D.
Remotesensing 14 00440 g005
Figure 6. Flowchart of water depth estimation.
Figure 6. Flowchart of water depth estimation.
Remotesensing 14 00440 g006
Figure 7. Histograms of the dependent variable (water depth): (A) before smoothing; (B) after smoothing.
Figure 7. Histograms of the dependent variable (water depth): (A) before smoothing; (B) after smoothing.
Remotesensing 14 00440 g007
Figure 8. Parameters of Gabor filters.
Figure 8. Parameters of Gabor filters.
Remotesensing 14 00440 g008
Figure 9. Moving 3 × 3-pixel window to calculate the surface slope.
Figure 9. Moving 3 × 3-pixel window to calculate the surface slope.
Remotesensing 14 00440 g009
Figure 10. Features used to model water depth.
Figure 10. Features used to model water depth.
Remotesensing 14 00440 g010
Figure 11. ML hyperparameter tuning: (A) number of estimators in RF; (B) k-values for KNR.
Figure 11. ML hyperparameter tuning: (A) number of estimators in RF; (B) k-values for KNR.
Remotesensing 14 00440 g011
Figure 12. Input feature importance.
Figure 12. Input feature importance.
Remotesensing 14 00440 g012
Figure 13. Input feature importance.
Figure 13. Input feature importance.
Remotesensing 14 00440 g013
Figure 14. Accuracy of different regression algorithms.
Figure 14. Accuracy of different regression algorithms.
Remotesensing 14 00440 g014
Table 1. Classification accuracy of land-use map.
Table 1. Classification accuracy of land-use map.
Bare SoilBuildingsGreenRoadsWaterSumUser’s Accuracy
Bare soil107013211394.7%
Buildings2421214887.5%
Green116111060.0%
Roads2132823677.8%
Water101261060.0%
Sum11344123612217
Producer’s accuracy94.7%95.5%50.0%77.8%50.0%
Total Accuracy87.1%
a26,714
b32,790
Kappa81.5%
Table 2. Sample of the extracted feature values.
Table 2. Sample of the extracted feature values.
ML DataBased On:FeaturesPixel ID (Sample Number)
IDFilter12125491125492
Independent data (X)SAR image1Original Image19.2318.4516.5915.29
2Gabor391.0090.0084.0081.00
3Gabor451.0051.0046.0044.00
4Gabor529.0029.0027.0026.00
5Gabor617.0016.0015.0014.00
6Gabor712.0012.0011.0011.00
7Gabor87.007.006.006.00
8Gabor1162.0062.0058.0055.00
9Gabor1257.0057.0053.0051.00
10Gabor19255.00255.00255.00255.00
11Gabor20156.00155.00140.00132.00
12Gabor2110.0010.0010.0010.00
13Gabor2334.0034.0032.0031.00
14Gabor247.007.006.006.00
15Gabor27206.00205.00188.00180.00
16Gabor28159.00158.00144.00136.00
17Gabor298.007.008.008.00
18Gabor304.004.004.004.00
19Gabor3114.0014.0014.0013.00
20Gabor328.008.008.008.00
21GMM1.000.000.000.00
22Canny Edge0.000.000.000.00
23Roberts0.010.000.000.00
24Sobel0.010.000.000.00
25Scharr0.010.000.000.00
26Prewitt0.010.000.000.00
27Gaussian s315.0015.0015.0014.00
28Gaussian s715.0015.0015.0015.00
29Median s318.0018.0016.0015.00
30Otsu255.00255.000.000.00
31Slope_Per6.256.094.482.98
32Str_Ord0.000.000.000.00
33Land_Use3.003.003.003.00
1Water_depth0.170.160.100.12
31Slope_Per6.256.094.482.98
DSM image32Str_Ord0.000.000.000.00
Sentinel-2 image33Land_Use3.003.003.003.00
Dependent data (Y)HEC-RAS results1Water_depth0.170.160.100.12
Table 3. Comparison of RMSE and regression techniques at various water depths.
Table 3. Comparison of RMSE and regression techniques at various water depths.
Water DepthNumber
of Points
Percentages (%)RMSE (m)
ML Algorithm
GBRRFRLRXGBRMLPRKNRSVR
>6 m40.027.206.957.436.947.476.776.98
3–6 m560.223.423.463.623.433.813.443.57
2–3 m1570.632.142.192.172.192.272.182.22
1–2 m7182.861.121.131.181.121.261.171.21
0–1 m24,16496.270.190.190.190.180.190.220.18
Overall25,0991000.360.370.380.360.390.390.37
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Elkhrachy, I. Flash Flood Water Depth Estimation Using SAR Images, Digital Elevation Models, and Machine Learning Algorithms. Remote Sens. 2022, 14, 440. https://doi.org/10.3390/rs14030440

AMA Style

Elkhrachy I. Flash Flood Water Depth Estimation Using SAR Images, Digital Elevation Models, and Machine Learning Algorithms. Remote Sensing. 2022; 14(3):440. https://doi.org/10.3390/rs14030440

Chicago/Turabian Style

Elkhrachy, Ismail. 2022. "Flash Flood Water Depth Estimation Using SAR Images, Digital Elevation Models, and Machine Learning Algorithms" Remote Sensing 14, no. 3: 440. https://doi.org/10.3390/rs14030440

APA Style

Elkhrachy, I. (2022). Flash Flood Water Depth Estimation Using SAR Images, Digital Elevation Models, and Machine Learning Algorithms. Remote Sensing, 14(3), 440. https://doi.org/10.3390/rs14030440

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop