Next Article in Journal
Automatic Generation of Seamless Mosaics Using Invariant Features
Next Article in Special Issue
Digital Elevation Models: Terminology and Definitions
Previous Article in Journal
Semantic Segmentation of Urban Buildings Using a High-Resolution Network (HRNet) with Channel and Spatial Attention Gates
Previous Article in Special Issue
Comparison of Gridded DEMs by Buffering
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Enhanced Double-Filter Deep Residual Neural Network for Generating Super Resolution DEMs

1
School of Resource and Environment Science, Wuhan University, Wuhan 430079, China
2
Spatial Sciences Institute, University of Southern California, Los Angeles, CA 90089, USA
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(16), 3089; https://doi.org/10.3390/rs13163089
Submission received: 25 June 2021 / Revised: 30 July 2021 / Accepted: 2 August 2021 / Published: 5 August 2021
(This article belongs to the Special Issue Perspectives on Digital Elevation Model Applications)

Abstract

:
High-resolution DEMs are important spatial data, and are used in a wide range of analyses and applications. However, the high cost to obtain high-resolution DEM data over a large area through sensors with higher precision poses a challenge for many geographic analysis applications. Inspired by the convolution neural network (CNN) excellent performance in super-resolution (SR) image analysis, this paper investigates the use of deep residual neural networks and low-resolution DEMs to generate high-resolution DEMs. An enhanced double-filter deep residual neural network (EDEM-SR) method is proposed, which uses filters with different receptive field sizes to fuse and extract features and reconstruct a more realistic high-resolution DEM. The results were compared with those generated with the bicubic, bilinear, and EDSR methods. The numerical accuracy and terrain feature preserving effects of the EDEM-SR method can generate reconstructed DEMs that better match the original DEMs, show lower MAE and RMSE, and improve the accuracy of the terrain parameters. MAE is reduced by about 30 to 50% compared with traditional interpolation methods. The results show how the EDEM-SR method can generate high-resolution DEMs using low-resolution DEMs.

Graphical Abstract

1. Introduction

With the increasing use of DEMs in many fields, such as 3D terrain visualization, hydrological, ecological, and geomorphological analysis [1,2,3,4], it is now necessary to obtain high resolution DEMs for large areas. A high-resolution DEM contains more information and can better reflect the actual surface, which plays a crucial role in the correct derivation of terrain factors such as slope, aspect, and the topographic wetness index [5,6]. However, it is difficult to obtain large-scale high-resolution DEMs using sensors with high precision. There are some open-access low-resolution DEMs with global coverage, including SRTM and ASTER GDEM. Therefore, this paper explores methods for generating high-resolution DEMs from low-resolution DEMs to provide an alternative way to obtain high-resolution DEMs.
There are usually two ways to obtain high-resolution DEMs, one is to generate DEMs using high-precision equipment, and the other uses one or more algorithms to reconstruct high-resolution DEMs from low-resolution DEMs. The main sources of DEM generation are ground survey, GPS, and remote sensing [7]. In particular, the emergence of LiDAR technology has made an important contribution to the acquisition of high-resolution DEMs [7,8]. Many ground filtering algorithms have been proposed to improve the accuracy of DEM generation from LiDAR data under various conditions [9,10,11,12]. However, using LiDAR data to generate DEM is expensive and labor-intensive, such that it cannot meet the need for large-scale, high-precision, and high-resolution DEMs. On the other hand, super-resolution (SR) DEMs offer an alternative way to generate large area, high-resolution DEMs from low-resolution DEMs.
SR DEMs can be generated using interpolation, reconstruction, and learning-based methods. Interpolation is one of the most commonly used methods, which use continuously curved surfaces to fit the terrain surface, including inverse distance, Kriging, bilinear, and bicubic interpolation [13,14]. The performance of these methods varies across different terrain conditions, and the accuracy is unstable [15,16]. Moreover, the terrain features of the interpolated DEMs will be over-smoothed. The reconstruction methods rely on data fusion and use the complementary information of multi-source DEMs to generate SR DEMs [17,18,19]. The learning-based methods can improve the super-resolution effect theoretically and practically by learning some repeated and similar patterns of the original DEMs and introducing high-frequency information into the super-resolution versions of low-resolution DEMs [20].
Although the relief of the Earth’s surface described by DEMs is different in every place, the local topographic features will likely share some similarities. Therefore, we can use learning-based methods to build the mapping model from the high and low-resolution DEMs of a certain region, and then reconstruct DEMs in regions lacking high-resolution DEMs. CNN is widely used in the field of computer vision because it offers good performance in image recognition [21], image classification [22], and image super-resolution [23]. SRCNN [23] was the first model to apply CNN to generate super-resolution DEMs. In this model, the low-resolution is enlarged to a specified size by interpolation, and then the image quality is improved by three convolution modules. In order to improve the computational efficiency and avoid introducing additional errors in advance, some follow-up studies have carried out up-sampling via a deconvolutional layer or sub-pixel convolution layer at the end, instead of the interpolation inverse before the input [24,25]. Many modules, such as the residual [26], residual dense block [27], and generation countermeasure modules [28], have been applied to image SR. The EDSR [29], WDSR [30], RDN [27], ESRGAN [31], MSRN [32], RCAN [33], and CARN [34] approaches have achieved good performance in image SR.
Gridded DEMs are similar to images. Using CNN for SR DEMs can be regarded as an extension of image SR. Chen et al. [35], used SRCNN to reconstruct SR DEMs in the first such application. Several others have used CNNs to reconstruct SR DEMs [36,37,38]. Some of the classical models in image SR such as EDSR, SRGAN, and ESRGAN, which have achieved good results, have been used in SR DEM applications [39,40,41].
However, the SR DEM methods mentioned above show some shortcomings that may limit their applicability. Interpolation and reconstruction methods do not introduce high-frequency information into the SR process, and the reconstructed high-resolution DEMs will ignore many terrain details and generate overly smooth DEMs. The method based on CNN is still in the exploratory stage, and the models deployed are often relatively simple or migrated from image SR applications.
Therefore, to overcome these problems, this study proposed an EDEM-SR method that uses high- and low-resolution DEMs as the training data to solve the problem of insufficient high-resolution DEM data. In this paper, a double-filter deep residual CNN is proposed to extract and fuse features in super-resolution (SR) DEM reconstruction. The model employs a double-filter residual block, which has two parallel filters with different receptive fields that can better use neighborhood information. Two interpolation methods, bicubic and bilinear, and EDSR were chosen as reference models to evaluate the performance of the proposed EDEM-SR model.

2. Methodology

The EDEM-SR method includes: (1) data pre-processing; (2) setting up the structure of the EDEM-SR network; (3) model training; and (4) super-resolution DEM generation and evaluation. The workflow of the paper is summarized in Figure 1.

2.1. Data Pre-Processing

The corresponding data from the high- and low-resolution DEMs are used as training data, test data, and validation data. Because DEM data from different sources often have different projections, it is not easy to form the corresponding high- and low-resolution data with scales of 2, 3, and 4. In order to facilitate the experiment, the 12.5 m DEMs obtained from the PALSAR sensor of the ALOS satellite were used as the original high-resolution DEMs. We used ArcGIS to pre-process the DEMS. First, the 12.5 m DEMs was down-sampled to scales of 2, 3, and 4 by bicubic interpolation to form the corresponding low-resolution datasets. The ratio of training data, test data, and validation data was 8:1:1. Second, the original DEM data were cut out, the areas with poor data quality and urban areas were removed, and the DEM data of high quality were retained. Finally, the cropped high- and low-resolution DEM data were segmented and named systematically to form a labeled data set. In more detail, the preprocessing of the DEM data of high- and low-resolution proceeded as follows:
(1)
Bicubic down-sampling was used with the original DEM data with a resolution of 12.5 m to construct DEMs with resolutions of 25 m, 37.5 m, and 50 m.
(2)
A natural area with obvious terrain features was selected as the research area, and the high- and low-resolution DEM data were clipped to match the area in (1).
(3)
The cropped high- and low-resolution DEM data were divided into corresponding regions. For example, if the high-resolution DEM data is divided into images with 192 × 192 pixels, then the low-resolution data of scales of 2, 3, and 4 need to be divided into images with 96, 64, and 48 pixel sizes.

2.2. The Structure of the EDEM-SR Network

The architecture of the proposed network is shown in Figure 2 and Figure 3, and is composed of three main sub-networks, namely the feature extraction, residual, and up-sampling modules. The specific structure and functions of the three modules are as follows:
(1)
The feature extraction module consists of a convolution layer with 256 channels. Its function is to extract the features of the low-resolution DEM input data through the convolution layer, and output 256 feature maps.
(2)
The residual module consists of 32 small residual modules. The function of these residual modules is to increase the depth of the network, and further extract and fuse the feature maps generated by filters with different receptive fields. Each small residual module includes two different filter branches (i.e., 3 × 3, 5 × 5), each branch includes two convolution layers, and the activation function links the two convolution layers, as shown in Figure 3. We concatenate the feature map extracted from the two branches at the end of each residual module.
(3)
The up-sampling module consists of a sub-pixel convolution layer and a convolution layer. The function of the sub-pixel convolution module is to enlarge the size of the output image. The sub-pixel layer is used to extract s 2 feature maps and splice them together to enlarge the resolution of the image by s times. Then, the final super-resolution image is obtained by fusing the expected size feature images through the subsequent convolution layer. Compared with the traditional use of the deconvolution layer for up-sampling, the sub-pixel convolution layer can avoid the introduction of irrelevant information and reduce the amount of calculation.
In order to achieve better reconstruction results, the residual module of the proposed EDEM-SR has two filters with different size convolution kernels, which have different receptive fields. The size of the receptive field determines the scale of observation and affects the super-resolution. The traditional deep learning model often uses a fixed filter size, so the observation scale is limited. The multi-scale technology integration method can fuse the features extracted by different receptive field filters and better fuse the neighborhood features, which is advantageous in practice. Considering the complexity of the network, residual learning is introduced to train the network. Generally speaking, residual learning is carried out through a series of residual modules with the same structure. Different from the general convolution neural network which directly establishes the input-output mapping, the residual network calculates the difference between the desired output and input. For example, if the input is x and the target output is f(x), the goal of residual learning is to fit g(x) = f(x) − x and add the input X and G(x) to get the desired output f(x). In addition, because each layer of the convolution neural network extracts 256 feature maps, too many feature maps will lead to numerical instability in the training process. Therefore, this paper sets the residual scale to 0.1, and then multiplies the output of the convolution layer by 0.1 and adds the input x, which helps to stabilize the training process.
In summary, the EDEM-SR network can be divided into three modules. The first module extracts the feature mapping of the input DEM. The second module consists of a series of residual modules with filters of different kernel size, and further describes the extracted features as complex feature maps. These feature maps are up-sampled and high-resolution images are generated in the third and final module.

2.3. Model Training

The convolution operation can be formulated as:
a ( x ) = σ ( ω x + b )
where σ ( · ) denotes the activation function, ω denotes the convolution kernel for feature extraction, and b denotes the bias.
The proposed network takes the rectified linear unit (ReLU) [42] as the activation function, formulated as σ ( x ) = max ( 0 , x ) , owing to its effectiveness in nonlinear mapping. For the output X of CNN, after the activation function ReLU, the positive value remains the same, and the negative value becomes 0. ReLU has the advantages of simple calculation, fast speed and avoiding over-fitting.
In order to optimize these parameters, CNN is trained by minimizing the error between the output and the real high-resolution DEM. The error is calculated using L1 Loss as the loss function, which is defined as:
L 1 = 1 n i = 1 n ( F ( X i Y i ) )
where { X i } denotes the input low-resolution DEM, F ( · ) denotes the mapping function from low resolution to high resolution based on the CNN, { Y i } denotes the input high-resolution DEM, and n denotes the number of training samples. A certain number of DEMs is taken from the training data as a batch, and a fixed size patch is randomly cropped from each picture in the batch. These patches are then concatenated together as batch data for error backpropagation. After calculating the error between the output and the actual data, the error is reduced by adaptive moment estimation using Adam [43], which is realized by updating the parameters of each layer of the neural network through the error backpropagation and iterating repeatedly. The iterative formula can be expressed as:
v W l = β 1 v W l + ( 1 β 1 ) δ F δ W l ,   v W l c o r r e c t e d = v w l 1 β 1 t ,
s W l = β 2 s W l + ( 1 β 2 ) ( δ F δ W l ) 2 ,   s W l c o r r e c t e d = s w l 1 β 2 t ,
W l = W l α v W l c o r r e c t e d s W l c o r r e c t e d + ε
where t denotes the number of iterations, v W l denotes the momentum, s W l denotes the weighted average of differential squares, and W l denotes the filters of L-th layer.
Theoretically, the use of this transfer learning strategy based on the use of natural images to pre-train the network can make the network converge faster and reduce the amount of training sample data needed. Hence, the preliminary experiments show that the training samples used in this paper are sufficient to obtain good performance.

2.4. DEM Super-Resolution and Evaluation

According to the EDEM-SR model trained with the training dataset, the test dataset is input into the model, and the reconstructed high-resolution DEM data is output after a series of transformations. After that, the accuracy of these reconstructed high-resolution DEMs is evaluated and compared with that of other methods.
The results of the method are compared with those of the bicubic, bilinear, and EDSR methods to evaluate the super-resolution DEM performance. The bicubic and bilinear models rely on interpolation to approximate the original DEM. The EDSR model is based on convolutional neural networks, using a similar network structure to our method.
The mean absolute error (MAE), root mean squared error (RMSE), maximum elevation error ( E m a x ), and mean error of terrain parameters ( E t p ) are used to evaluate the performance of each model. In addition, all methods are tested using the test dataset.
The evaluation metrics are given by:
MAE = 1 m i = 1 m | ( y i y i ¯ ) |
RMSE = 1 m i = 1 m ( y i y i ¯ ) 2
E m a x = 1 m ( i = 1 m max ( abs ( y i y i ¯ ) ) )
E t p = 1 m i = 1 m | ( t i t i ¯ ) |
where m denotes the number of pixels in the test samples, y i denotes the values in the original high-resolution DEM, y i ¯ denotes the values in the reconstructed high resolution DEM, E t p denotes the error of the terrain parameters, t i denotes the values of the terrain parameters generated with the original high-resolution DEM, and t i ¯ denotes the values of the terrain parameters generated with the reconstructed high resolution DEM.

3. Experiments and Results

To prove the performance of the proposed method, the results were evaluated using 800 test DEMs. The test data were not used in the neural network training, so they can well reflect the generalization ability of the network. In the experiment, the networks were implemented based on the Pytorch framework. The operating system is Ubuntu 16.04, the CPU is an Intel(R) Core(TM) i7-8700, and the GPU is a GeForce GTX 2080 equipped with 16 GB of memory. The relevant experimental configurations are listed in Table 1.
In the experiment, DEM with resolutions of 25 m, 37.5 m, and 50 m were used to reconstruct a 12.5 m high-resolution DEM to evaluate the performance of the proposed model. It should be pointed out that due to the lack of real data sets of 2, 3, and 4 scales, we generated these three data sets by down-sampling the 12.5 m data in the same period, so as to ensure consistency of the data evaluation benchmarks. Four sets of results were generated with the proposed method and the bicubic, bilinear, and EDSR methods.

3.1. Study Area and Data

The Loess Plateau is more than 1000 km long from east to west and 750 km wide from south to north. It is located on the second step of China, with an altitude of 800–3000 m. The Loess Plateau is the largest loess area in the world. It has long been the focus of research because of its unique geomorphological features. The DEM is an important geographical data theme, however, only a small part of the region has a high-resolution DEM, which is not conducive to the analysis of a large-scale study of the Loess Plateau. A part of the Loess Plateau was selected as the study area. First, we chose the region with high-resolution DEM data on the edge of the Loess Plateau as the training data. Then, we took another area of the Loess Plateau as the test dataset and input it into the model for super-resolution DEM reconstruction. The distribution of training and test data is shown in Figure 4. The test DEM is shown in Figure 5.
The original high-resolution DEM data is divided into 192 × 192 cells. By bicubic interpolation, these high-resolution DEM are down-sampled to scales of 2, 3, and 4 times to obtain low-resolution DEM. Each sample includes a high-resolution DEM and a low-resolution DEM. We generated 8155 such samples. In this experiment, part of the Loess Plateau is used as the training set to train the model, and the trained model can be used to generate super-resolution DEMs in other regions of the Loess Plateau. For DEM reconstruction in other locations, the network can be initialized with the parameters of the existing model and then optimized with a small number of DEMs from the locations to be reconstructed to obtain good reconstruction results.

3.2. Visual Assessment

Figure 6 shows the 3D surfaces for the original DEM (Figure 6a) and the SR DEM generated with the EDEM-SR method (Figure 6b). These two images show that the DEM generated with the proposed method looks close to the original DEM and retains most of the terrain details.
Figure 7 shows the DEM samples generalized by the different methods at a scale of 4. Naturally, with a decrease in resolution, the DEMs become rough, reflecting less terrain detail, and the difficulty of reconstruction increases. We chose the SR DEM at this scale because the differences between the four methods would be more obvious, and it is very difficult to reconstruct the 50 m low-resolution DEMs in this case. Figure 7b shows the LR DEM used to reconstruct the HR DEM. It can be seen that the LR DEM at the down-sampling scale of 4 is very rough in detail. Figure 7c shows the DEM generated with the bilinear method and the artificial texture shows the poor performance of this method for generating SR DEMs, which confirmed the previously noted disadvantage of interpolation methods losing terrain features. Figure 7d shows the DEM generated with the bicubic method that was the best performing of the two interpolation methods. It can be seen that although the DEM reconstructed by this method is close to the original DEM, it has some artificial texture. Figure 7e,f shows the DEMs generated by the two methods using CNN. The DEMs generated with this pair of methods show few visual differences and both can restore the terrain features well.
In general, bilinear has the worst reconstruction effect, with obvious artificial textures, and while the bicubic performs better and has no obvious artificial textures, the DEM reconstructed with this method has smooth terrain and poor terrain detail recovery. The methods based on deep learning provide better performance in terms of detail restoration. As expected, the high-resolution DEM reconstructed by the EDEM-SR method well reflects the valley and ridge line details.

3.3. Overall Accuracy

Table 2 shows how well the DEMs reconstructed by the different methods perform with reference to the original 12.5 m ALOS high-resolution DEM data using MAE, RMSE, and E m a x . These results show that the performance of different methods varies at different scales. With the increase of scales, the performance of all methods decreases, because the input low-resolution DEM contains less information. The performance of the bilinear method decreases the most, and the performance of the two deep learning methods decreases less than that of the two interpolation methods. Table 2 also shows that the proposed EDEM-SR method achieves the best results on scales of 2, 3, 4, given the lowest MAE, RMSE, and E m a x values. At the same time, the bilinear interpolation performs the worst and has the highest MAE, RMSE, and E m a x at all scales. The MAE of the proposed EDEM-SR method is about 50% lower than that of the worst bilinear method on the scale of 4. The proposed EDEM-SR model also has better results compared with the deep learning model EDSR, which has the same number of feature maps and the same number of residual modules.

3.4. Terrain Parameters Maintenance

Figure 8 show the terrain parameters for DEMs reconstructed by different methods at the scale of 4. The color difference between Figure 8a, Figure 9a and Figure 10a and other images series shows the ability of different methods to maintain the terrain features. Figure 8c, Figure 9c and Figure 10c show the slope, aspect, and curvature for a DEM reconstructed by bilinear interpolation. From its color change, it can be seen that the bilinear method eliminates the slope decline, connects several adjacent areas with similar aspects, and the curvature is quite different from the original curvature. Specifically, the color transition of Figure 8c is not obvious, the regions separated in Figure 9a are connected in Figure 9c, and Figure 10c has an obvious artificial texture. Figure 8d, Figure 9d and Figure 10d show the terrain parameters for a sample DEM reconstructed by bicubic interpolation. The method’s performance is better than bilinear, but it has similar problems maintaining the terrain parameters. Interpolation methods all have shortcomings associated with the preservation of fuzzy and smooth terrain features. Figure 8e, Figure 9e and Figure 10e and Figure 8f, Figure 9f and Figure 10f show the results of EDSR and EDEM-SR, respectively. It can be seen that the color changes for these two methods are close to the original DEM, which confirmed that the terrain features reconstructed by the deep learning methods are close to those in the original DEM. Compared with the DEMs reconstructed by interpolation methods, the DEMs reconstructed by our method show many of the same features as the original DEM.
Table 3 shows the numerical accuracy of slope, aspect, and curvature and indicates that the terrain parameters calculated by the DEM generated by the proposed EDEM-SR method have the lowest error. The errors accompanying the EDSR method are higher than for the EDEM-SR method. The numerical results in Table 3 confirm the trend evident in the maps reproduced in Figure 8. The results in Table 3 mirror those in Table 2, indicating that the performance of the interpolation methods decreases more obviously than that of the deep learning methods with an increase in scale

4. Discussion

The proposed EDEM-SR model reconstructs super-resolution DEMs at three scales. The model parameters were adjusted to improve the super-resolution performance. Through many experiments, the EDEM-SR method shows three advantages. First, compared with other interpolation methods, such as bicubic interpolation, bilinear interpolation, the DEM reconstructed by the proposed EDEM-SR method has higher accuracy, especially at large SR scales. Second, the residual module of the EDEM-SR model has two branches of filters with different convolution kernel sizes, which fully integrate the information of neighborhoods with different sizes. As a result, compared with other CNN methods, such as EDSR, the proposed EDEM-SR model shows better accuracy. Third, the EDEM-SR model shows excellent performance in preserving the details of the reconstructed DEM terrain parameters.

4.1. The Evaluation of the Precision of Using the EDEM-SR Method

Compared with the bicubic, bilinear, and EDSR methods, the proposed EDEM-SR method offered higher overall accuracy when generating SR DEMs. As shown in Table 2, the DEM reconstructed by the EDEM-SR method at the scales of 2, 3, and 4 are closer to the original high-resolution DEM with smaller MAE, RMSE, and E m a x , especially in the edge region, where the interpolation method cannot obtain sufficient information and the accuracy drops significantly. In addition, with an increase of the reconstruction scale, the accuracy of the interpolation method decreases more significantly than that of the deep learning method. This is because with the increase of scale, the pixels in the low-resolution data become less, and the information obtained by the interpolation method is more limited, while deep learning methods such as EDEM-SR learn the mapping relationship between low- and high-resolution data through a large number of training samples. In super-resolution DEM reconstruction, the high-frequency information learned from other high-resolution DEM data is introduced, so these models have significant advantages in large-scale super-resolution work tasks. Table 4 shows the reconstruction effect after training the model with different sizes of training samples. It can be seen that with an increase in the number of training samples, the reconstruction effect gradually improves, and finally converges to a better result.

4.2. The Advantages of Using Double-filters

The proposed EDEM-SR method performs better than EDSR, which also is based on residual CNNs because the complete terrain structure in the real world often has different sizes. These terrain structures are represented in DEMs as sets of pixels with different extents. In order to collect complete information, a filter with an appropriate receptive field size is needed. A CNN with a single filter usually extracts information from a fixed receptive field, which cannot take every situation present in the DEM into account. Therefore, this paper proposes a residual structure with double filters, as shown in Figure 3. The information of different sized neighborhoods is collected through different size receptive fields, and then this information is fused to obtain better super-resolution performance. As shown in Table 2, the ensemble learning of the multi-filter CNN achieves the highest overall accuracy. Because multiple filters can help to obtain more important abstract features at different scales, larger filter sizes can consider more information about neighbors than smaller filter sizes and obtain the overall trend of terrain, whereas smaller filter sizes can capture the details of the object and improve the numerical accuracy. Thus, the proposed EDEM-SR model achieves better numerical accuracy compared with the EDSR method.

4.3. Retention of Terrain Parameters

From the perspective of the terrain analysis, in addition to acceptable MAE, RMSE, and E m a x , maintaining terrain features is also a key measure. Figure 8, Figure 9 and Figure 10 show the terrain parameters generated by the reconstructed DEM using different methods at a scale of 4. Table 3 shows the error for mean slope ( E s l o p e ), aspect ( E a s p e c t ) and curvature ( E c u r v a t u r e ) of different algorithms at different super-resolution scales.
From Figure 8, Figure 9 and Figure 10, one can see from the local maps that among the four methods, bilinear does the worst in maintaining terrain parameters. Bicubic is the best of the two interpolation methods and can restore details of the terrain to some extent, but it cannot restore the overall trend of the terrain very well. From the visual results, the two methods based on deep learning are better than the interpolation methods in the restoration of terrain details and the overall trend of terrain.
Table 3 quantitatively compares the performance of each method in the terrain feature preservation based on numerical accuracy, and shows how the EDEM-SR method did best in terms of maintaining the terrain parameters. All four algorithms showed good performance in the retention of terrain parameters at the scale of 2, confirmed by the small numerical errors of slope, aspect, and curvature calculated by each algorithm. But the errors in the terrain parameters obtained by the interpolation methods increased significantly at scales of 3 and 4, and the differences between the interpolation methods and the proposed EDEM-SR model became larger. The proposed EDEM-SR method offers the best performance, especially at the larger super-resolution scales.

4.4. Limitations and Future Enhancements

The proposed EDEM-SR method for super-resolution DEM reconstruction shows excellent performance. However, there are still some limitations. Although EDEM-SR achieves the lowest MAE, RMSE and E m a x , at scales of 2, 3 and 4, it is not capable of training a mapping model for super-resolution at any scale. Therefore, it is necessary to consider a network architecture for full-scale super-resolution reconstruction. Although the use of double-filters can make full use of the neighborhood information, the amount of calculation increases and the computational efficiency is reduced.
In addition, the proposed model requires the training of a separate model for each type of terrain. For example, in this paper, the model trained by the proposed method has achieved good performance using mountainous terrain on the Loess Plateau. When reconstructing DEMs for other terrain types, such as flat or urbanized areas, we need to take the current model as the initial model, and use a small number of DEMs of these types to adjust model parameters to obtain the best results. Besides, for urban DEMs, LiDAR data with a resolution of around 1m is generally used, and for this type of high-resolution data, more data needs to be used to train a better model compared with reconstructing a natural area DEM with a resolution of 10 m.
Finally, the EDEM-SR method still focuses on the overall characteristics of the terrain, without considering the high-resolution terrain feature points in local areas. The method may offer better results if the reconstruction results of the super-resolution model were constrained by local terrain features.
Future work will focus on training terrain features into the network to constrain the super-resolution results, so that the SR DEM can retain more terrain features. We also plan to use higher resolution DEMs and open-access low-resolution DEMs to generate training data, instead of just down-sampling high-resolution DEMs to obtain training data.

5. Conclusions

In this paper, an EDEM-SR method with double filters was proposed for super-resolution DEM reconstruction. Based on the data fusion between different neighborhoods of low-resolution DEMs, this method can reconstruct a high-resolution DEM with better numerical accuracy and preservation of the original DEM terrain features. Comparing the accuracy of the high-resolution DEMs reconstructed by different methods, the EDEM-SR model achieves the best performance at scales of 2, 3, and 4. The experimental results show that, with the same residual module and the same number of feature maps, the reconstruction accuracy is improved by fusing the feature maps of filters with different convolution kernel sizes.
The results show that using the EDEM-SR method to build a mapping model from high- and low-resolution data of a certain region offers a promising solution for generating SR DEMs. The high-resolution DEMs of other regions can be reconstructed using the mapping model with low-resolution DEMs. In future work, local terrain features should be considered to constrain the results of super-resolution DEM reconstruction.

Author Contributions

A.Z. collected and processed the data, performed analysis, and wrote the paper; J.P.W. helped to write and edit the article; A.Z., Y.C. and H.S. analyzed the results; Z.X. and Q.C. contributed to the validation. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Nature Science Foundation of China: [Grant No. 41671380].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hayat, K.; Puech, W.; Gesquiere, G.; Chaumont, M. Wavelet-based data hiding of DEM in the context of real-time 3D visualization. In Proceedings of the Visualization and Data Analysis 2007, San Jose, CA, USA, 29–30 January 2007; Volume 6495, p. 64950N. [Google Scholar]
  2. Andreani, L.; Stanek, K.P.; Gloaguen, R.; Krentz, O.; Domínguez-González, L. DEM-based analysis of interactions between tectonics and landscapes in the ore mountains and eger rift (East Germany and NW Czech Republic). Remote Sens. 2014, 6, 7971–8001. [Google Scholar] [CrossRef] [Green Version]
  3. Wilson, J.P. Environmental Applications of Digital Terrain Modeling; Wiley-Blackwell: Oxford, UK, 2018. [Google Scholar]
  4. Simpson, A.L.; Balog, S.; Moller, D.K.; Strauss, B.H.; Saito, K. An urgent case for higher resolution digital elevation models in the world’s poorest and most vulnerable countries. Front. Earth Sci. 2015, 3, 1–2. [Google Scholar] [CrossRef] [Green Version]
  5. Grohmann, C.H. Effects of spatial resolution on slope and aspect derivation for regional-scale analysis. Comput. Geosci. 2015, 77, 111–117. [Google Scholar] [CrossRef] [Green Version]
  6. Schumann, G.J.-P.; Bates, P.D. The need for a high-accuracy, open-access global DEM. Front. Earth Sci. 2018, 6, 1–5. [Google Scholar] [CrossRef]
  7. Liu, X. Airborne LiDAR for DEM generation: Some critical issues. Prog. Phys. Geogr. 2008, 32, 31–49. [Google Scholar] [CrossRef]
  8. Shan, J.; Sampath, A. Urban DEM generation from raw lidar data: A labeling algorithm and its performance. Photogramm. Eng. Remote Sens. 2005, 71, 217–226. [Google Scholar] [CrossRef] [Green Version]
  9. Meng, X.; Currit, N.; Zhao, K. Ground filtering algorithms for airborne LiDAR data: A review of critical issues. Remote Sens. 2010, 2, 833–860. [Google Scholar] [CrossRef] [Green Version]
  10. Mongus, D.; Žalik, B. Parameter-free ground filtering of LiDAR data for automatic DTM generation. ISPRS J. Photogramm. Remote Sens. 2012, 67, 1–12. [Google Scholar] [CrossRef]
  11. Zhao, X.; Guo, Q.; Su, Y.; Xue, B. Improved progressive TIN densification filtering algorithm for airborne LiDAR data in forested areas. ISPRS J. Photogramm. Remote Sens. 2016, 117, 79–91. [Google Scholar] [CrossRef] [Green Version]
  12. Zhang, W.; Qi, J.; Wan, P.; Wang, H.; Xie, D.; Wang, X.; Yan, G. An easy-to-use airborne LiDAR data filtering method based on cloth simulation. Remote Sens. 2016, 8, 501. [Google Scholar] [CrossRef]
  13. Reuter, H.I.; Nelson, A.; Jarvis, A. An evaluation of void-filling interpolation methods for SRTM data. Int. J. Geogr. Inf. Sci. 2007, 21, 983–1008. [Google Scholar] [CrossRef]
  14. Rees, W.G. The accuracy of digital elevation models interpolated to higher resolutions. Int. J. Remote Sens. 2000, 21, 7–20. [Google Scholar] [CrossRef]
  15. Shi, W.Z.; Li, Q.Q.; Zhu, C.Q. Estimating the propagation error of DEM from higher-order interpolation algorithms. Int. J. Remote Sens. 2005, 26, 3069–3084. [Google Scholar] [CrossRef]
  16. Chaplot, V.; Darboux, F.; Bourennane, H.; Leguédois, S.; Silvera, N.; Phachomphon, K. Accuracy of interpolation techniques for the derivation of digital elevation models in relation to landform types and data density. Geomorphology 2006, 77, 126–141. [Google Scholar] [CrossRef]
  17. Li, X.; Shen, H.; Feng, R.; Li, J.; Zhang, L. DEM generation from contours and a low-resolution DEM. ISPRS J. Photogramm. Remote Sens. 2017, 134, 135–147. [Google Scholar] [CrossRef]
  18. Yue, L.; Shen, H.; Zhang, L.; Zheng, X.; Zhang, F.; Yuan, Q. High-quality seamless DEM generation blending SRTM-1, ASTER GDEM v2 and ICESat/GLAS observations. ISPRS J. Photogramm. Remote Sens. 2017, 123, 20–34. [Google Scholar] [CrossRef] [Green Version]
  19. Yue, L.; Shen, H.; Yuan, Q.; Zhang, L. Fusion of multi-scale DEMs using a regularized super-resolution method. Int. J. Geogr. Inf. Sci. 2015, 29, 2095–2120. [Google Scholar] [CrossRef]
  20. Xu, Z.; Wang, X.; Chen, Z.; Xiong, D.; Ding, M.; Hou, W. Nonlocal similarity based DEM super resolution. ISPRS J. Photogramm. Remote Sens. 2015, 110, 48–54. [Google Scholar] [CrossRef]
  21. Chu, T.; Chen, Y.; Huang, L.; Xu, Z.; Tan, H. A grid feature-point selection method for large-scale street view image retrieval based on deep local features. Remote Sens. 2020, 12, 3978. [Google Scholar] [CrossRef]
  22. Sun, Y.; Zhang, X.; Xin, Q.; Huang, J. Developing a multi-filter convolutional neural network for semantic segmentation using high-resolution aerial imagery and LiDAR data. ISPRS J. Photogramm. Remote Sens. 2018, 143, 3–14. [Google Scholar] [CrossRef]
  23. Yoon, Y.; Jeon, H.G.; Yoo, D.; Lee, J.Y.; Kweon, I.S. Learning a deep convolutional network for light-field image super-resolution. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 57–65. [Google Scholar] [CrossRef]
  24. Dong, C.; Loy, C.C.; Tang, X. Accelerating the super-resolution convolutional neural network. Lect. Notes Comput. Sci. 2016, 9906, 391–407. [Google Scholar] [CrossRef] [Green Version]
  25. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.; Bishop, R.; Rueckert, D.; Wang, Z. Real-time single image and ideo super-esolution using an efficient sub-pixel convolutional neural network. Comput. Vis. Patter Recognit. 2016, 1874–1883. [Google Scholar]
  26. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  27. Xu, J.; Chae, Y.; Stenger, B.; Datta, A. Residual dense network for image super-resolution. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 71–75. [Google Scholar] [CrossRef]
  28. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.P.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; Volume 2, p. 4. [Google Scholar]
  29. Lim, B.; Son, S.; Kim, H.; Nah, S.; Lee, K.M. Enhanced Deep Residual Networks for Single Image Super-Resolution. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 1132–1140. [Google Scholar] [CrossRef] [Green Version]
  30. Yu, J.; Fan, Y.; Yang, J.; Xu, N.; Wang, Z.; Huang, T.; Wang, X. Wide activation for efficient and accurate image super-resolution. arXiv 2018, arXiv:1808.08718. [Google Scholar]
  31. Wang, X.; Yu, K.; Wu, S.; Gu, J.; Liu, Y.; Dong, C.; Loy, C.C.; Qiao, Y.; Tang, X. ESRGAN: Enhanced super-resolution generative adversarial networks. In Computer Vision—ECCV 2018 Workshops. ECCV 2018; Leal-Taixé, L., Roth, S., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2018; Volume 11133. [Google Scholar] [CrossRef] [Green Version]
  32. Li, J.; Fang, F.; Mei, K.; Zhang, G. Multi-scale Residual Network for Image Super-Resolution. In Computer Vision—ECCV 2018. ECCV 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2018; Volume 11212. [Google Scholar] [CrossRef]
  33. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. Lect. Notes Comput. Sci. 2018, 11211, 294–310. [Google Scholar] [CrossRef] [Green Version]
  34. Ahn, N.; Kang, B.; Sohn, K.A. Fast, accurate, and lightweight super-resolution with cascading residual network. arXiv 2018, arXiv:1803.08664. [Google Scholar]
  35. Chen, Z.; Wang, X.; Xu, Z.; Hou, W. Convolutional neural network based dem super resolution. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. ISPRS Arch. 2016, 41, 247–250. [Google Scholar] [CrossRef] [Green Version]
  36. Jiang, L.; Hu, Y.; Xia, X.; Liang, Q.; Soltoggio, A. A multi-scale mapping approach based on a deep learning CNN model for reconstructing high-resolution urban DEMs. arXiv 2020, arXiv:1907.12898. [Google Scholar]
  37. Zhu, D.; Cheng, X.; Zhang, F.; Yao, X.; Gao, Y.; Liu, Y. Spatial interpolation using conditional generative adversarial neural networks. Int. J. Geogr. Inf. Sci. 2020, 34, 735–758. [Google Scholar] [CrossRef]
  38. Shin, D.; Spittle, S. LoGSRN: Deep super resolution network for digital elevation model. In Proceedings of the 2019 IEEE International Conference on Systems, Man and Cybernetics (SMC), Bari, Italy, 6–9 October 2019; pp. 3060–3065. [Google Scholar] [CrossRef] [Green Version]
  39. Xu, Z.; Chen, Z.; Yi, W.; Gui, Q.; Hou, W.; Ding, M. Deep gradient prior network for DEM super-resolution: Transfer learning from image to DEM. ISPRS J. Photogramm. Remote Sens. 2019, 150, 80–90. [Google Scholar] [CrossRef]
  40. Demiray, B.Z.; Sit, M.; Demir, I. D-SRGAN: DEM super-resolution with generative adversarial networks. arXiv 2020, arXiv:2004.04788. [Google Scholar]
  41. Wu, Z.; Ma, P. ESRGAN-based DEM super-resolution for enhanced slope deformation monitoring in lantau island of Hong Kong. In Proceedings of the International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences-ISPRS Archives, Nice, France, 14–20 June 2020; Volume 43, pp. 351–356. [Google Scholar]
  42. Geoffrey, E.H.; Vinod, N. Rectified linear units improve restricted boltzmann machines vinod nair. Int. Conf. Mach. Learn. 2010, 807–814. [Google Scholar] [CrossRef]
  43. Kingma, D.P.; Ba, J.L. Adam: A method for stochastic optimization. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015-Conference Track Proceedings, San Diego, CA, USA, 7–9 May 2015; pp. 1–15. [Google Scholar]
Figure 1. The EDEM-SR workflow.
Figure 1. The EDEM-SR workflow.
Remotesensing 13 03089 g001
Figure 2. An enhanced double-filter residual neural network (EDEM-SR) to generate SR DEMs. The residual block has a double-filter of different kernel size as detailed in Figure 3; Conv denotes convolutional layers.
Figure 2. An enhanced double-filter residual neural network (EDEM-SR) to generate SR DEMs. The residual block has a double-filter of different kernel size as detailed in Figure 3; Conv denotes convolutional layers.
Remotesensing 13 03089 g002
Figure 3. The structure of the double-filter residual block.
Figure 3. The structure of the double-filter residual block.
Remotesensing 13 03089 g003
Figure 4. Location of the study area with locations of the DEM data used for the experiments.
Figure 4. Location of the study area with locations of the DEM data used for the experiments.
Remotesensing 13 03089 g004
Figure 5. DEM of the test area.
Figure 5. DEM of the test area.
Remotesensing 13 03089 g005
Figure 6. The surfaces modeled by the original DEM (a) and the proposed EDEM-SR method (b).
Figure 6. The surfaces modeled by the original DEM (a) and the proposed EDEM-SR method (b).
Remotesensing 13 03089 g006
Figure 7. DEM reconstruction results: (a) is the high-resolution 12.5 m DEM; (b) is the low-resolution DEM with 50 m resolution; (c,d) are the bilinear and bicubic results; (e) is the EDSR result, and (f) is the EDEM-SR result. The above reconstruction results were all generated using an up-scaling factor of 4.
Figure 7. DEM reconstruction results: (a) is the high-resolution 12.5 m DEM; (b) is the low-resolution DEM with 50 m resolution; (c,d) are the bilinear and bicubic results; (e) is the EDSR result, and (f) is the EDEM-SR result. The above reconstruction results were all generated using an up-scaling factor of 4.
Remotesensing 13 03089 g007
Figure 8. Example slope map visualizations with an up-scaling factor of 4.
Figure 8. Example slope map visualizations with an up-scaling factor of 4.
Remotesensing 13 03089 g008
Figure 9. Example aspect maps with an up-scaling factor of 4.
Figure 9. Example aspect maps with an up-scaling factor of 4.
Remotesensing 13 03089 g009
Figure 10. Example curvature maps with an up-scaling factor of 4.
Figure 10. Example curvature maps with an up-scaling factor of 4.
Remotesensing 13 03089 g010
Table 1. Experimental environmental configuration.
Table 1. Experimental environmental configuration.
EnvironmentVersion
Operating SystemUbuntu 16.04 (64-bit)
GPUNvidia GTX 2080 (8G)
CPUInter (R) Core i7-6800K
Batch size16
Patch size96
Number of ResBlock32
OptimizerAdam
Table 2. Quantitative evaluation of the reconstruction effect of the test dataset.
Table 2. Quantitative evaluation of the reconstruction effect of the test dataset.
ScaleMethodMAE (m)RMSE (m) E max ( m )
2Bicubic0.30480.61416.60
Bilinear0.54820.84477.02
EDSR0.25300.50834.07
EDEM-SR0.25200.50703.12
3Bicubic0.64531.106813.05
Bilinear0.96131.456313.37
EDSR0.42730.74147.69
EDEM-SR0.42590.73947.15
4Bicubic1.18241.774219.00
Bilinear1.71002.342219.28
EDSR0.82091.209912.99
EDEM-SR0.80121.193912.48
Table 3. Quantitative evaluation of terrain parameter retention.
Table 3. Quantitative evaluation of terrain parameter retention.
ScaleMethod E slope E aspect E curvature
2Bicubic0.84505.83350.8836
Bilinear1.09236.96560.9659
EDSR0.73735.64830.7617
EDEM-SR0.72425.61940.7640
3Bicubic1.548710.26091.1521
Bilinear1.953713.38711.3649
EDSR1.06137.67580.8272
EDEM-SR1.01457.33250.7999
4Bicubic2.215114.89861.1520
Bilinear2.748317.87781.2238
EDSR1.677211.46720.8848
EDEM-SR1.650411.24040.8655
Table 4. Performance of different models trained with different amounts of training data.
Table 4. Performance of different models trained with different amounts of training data.
ScaleMethod25% Training Data (MAE)50% Training Data (MAE)100% Training Data (MAE)
2EDSR0.37130.27450.2530
EDEM-SR0.37450.26580.2520
3EDSR0.54270.44870.4273
EDEM-SR0.57410.44510.4259
4EDSR0.90440.83330.8209
EDEM-SR0.93990.84520.8012
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhou, A.; Chen, Y.; Wilson, J.P.; Su, H.; Xiong, Z.; Cheng, Q. An Enhanced Double-Filter Deep Residual Neural Network for Generating Super Resolution DEMs. Remote Sens. 2021, 13, 3089. https://doi.org/10.3390/rs13163089

AMA Style

Zhou A, Chen Y, Wilson JP, Su H, Xiong Z, Cheng Q. An Enhanced Double-Filter Deep Residual Neural Network for Generating Super Resolution DEMs. Remote Sensing. 2021; 13(16):3089. https://doi.org/10.3390/rs13163089

Chicago/Turabian Style

Zhou, Annan, Yumin Chen, John P. Wilson, Heng Su, Zhexin Xiong, and Qishan Cheng. 2021. "An Enhanced Double-Filter Deep Residual Neural Network for Generating Super Resolution DEMs" Remote Sensing 13, no. 16: 3089. https://doi.org/10.3390/rs13163089

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop