Next Article in Journal
Bringing Life Back into Former Mining Sites: A Mini-Review on Soil Remediation Using Organic Amendments
Next Article in Special Issue
Prior Knowledge-Based Deep Convolutional Neural Networks for Fine Classification of Land Covers in Surface Mining Landscapes
Previous Article in Journal
Sulphate Removal in Industrial Effluents Using Electrocoagulation Sludge as an Adsorbent
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Three-Stream and Double Attention-Based DenseNet-BiLSTM for Fine Land Cover Classification of Complex Mining Landscapes

1
Faculty of Computer Science, China University of Geosciences, Wuhan 430074, China
2
Hubei Key Laboratory of Intelligent Geo-Information Processing, China University of Geosciences, Wuhan 430074, China
3
Wuhan Centre of China Geological Survey, Wuhan 430205, China
*
Author to whom correspondence should be addressed.
Sustainability 2022, 14(19), 12465; https://doi.org/10.3390/su141912465
Submission received: 5 August 2022 / Revised: 27 September 2022 / Accepted: 27 September 2022 / Published: 30 September 2022
(This article belongs to the Special Issue Innovation and Sustainable Development of Remote Sensing Technology)

Abstract

:
The fine classification of land cover around complex mining areas is important for environmental protection and sustainable development. Although some advances have been made in the utilization of high-resolution remote sensing imagery and classification algorithms, the following issues still remain: (1) how the multimodal spectral–spatial and topographic features can be learned for complex mining areas; (2) how the key features can be extracted; and (3) how the contextual information can be captured among different features. In this study, we proposed a novel model comprising the following three main strategies: (1) design comprising a three-stream multimodal feature learning and post-fusion method; (2) integration of deep separable asymmetric convolution blocks and parallel channel and spatial attention mechanisms into the DenseNet architecture; and (3) use of a bidirectional long short-term memory (BiLSTM) network to further learn cross-channel context features. The experiments were carried out in Wuhan City, China using ZiYuan-3 imagery. The proposed model was found to exhibit a better performance than other models, with an overall accuracy of 98.65% ± 0.05% and an improvement of 4.03% over the basic model. In addition, the proposed model yielded an obviously better visual prediction map for the entire study area. Overall, the proposed model is beneficial for multimodal feature learning and complex landscape applications.

1. Introduction

The illegal over-exploitation of open pit mines can easily lead to surface subsidence and desertification, which aggravates soil erosion and vegetation reduction. Soil erosion and reduced vegetation can lead to changes in the surface land cover types. According to recent studies, changes in surface land cover types have a significant impact on the surface temperature of the region. Furthermore, changes in the surface temperature increase with changes in the surface land cover [1]. The increase in surface temperature creates a series of environmental problems that affect the goal of sustainable development.
Real-time monitoring of open pit mining activities is imperative. An effective approach for monitoring the activities of open pit mines involves the detection of changes in land cover types in and around the mine. The fine classification of the different types of land cover in mining areas and their vicinity is of great significance for environmental protection, sustainable development, and scientific mining. The detection of land cover is an important method for reflecting the ecological environment [2], and it can effectively prevent the threat of dust, environmental pollution, and vegetation damage caused by mineral resource development. However, owing to the complex topography near the open pit mine and damage to the land by mining activities, performing manual observations of the complex open pit mining areas and detecting their changes in real-time are impractical [3].
Multispectral images provide rich spectral information and have broad application prospects in the field of remote sensing [4,5]. In recent years, with the development of multispectral imaging technology, the coverage of remote sensing images has increased [6], and the spatial resolution of remote sensing images has continuously improved [7,8]. However, such advancements have increased the information complexity of multispectral remote sensing images [9,10,11] and has led to a reduction in the spectral differences between different types of objects. The spectral difference within the class increases [12,13], and the spectral–spatial information cannot be fully utilized [14,15]. Because open pit mining areas have complex landscapes and significant three-dimensional topographic features [16,17], the gathering of multimodal spectral–spatial and topographic information to improve the land cover classification (LCC) accuracy of complex open pit mining areas is an issue.
Friedl and Brodley [18] used three decision tree algorithms (i.e., univariate, multivariate, and mixed decision trees) to conduct LCC studies. In a prior study [19], the researchers performed pixel- and object-based experiments on MS images acquired by the HJ-1B and ALOS satellites using the support vector machine (SVM) and random forest (RF) algorithms, respectively. Machine learning algorithms (MLAs) are effective for LCC and fine LCC (FLCC) applications. However, MLAs generally have shortcomings when classifying remote sensing images in complex landscapes. Due to the utilization of feature engineering with a low representation ability, MLAs cannot fully extract image information [15]. As a result, the corresponding classification accuracies are unsatisfactory.
In the past few years, deep learning (DL) algorithms have been widely used for FLCC applications due to their powerful feature learning abilities [20,21], and they have achieved good results in data mining, natural language processing, and other fields [22]. The restricted Boltzmann machine in the deep belief network (DBN) can be trained while unsupervised to learn the features that the original data contain to the greatest extent; while supervised, it can be fine-tuned using the back-propagation algorithm to further learn data features, which requires little labeled data for the FLCCs of complex mining areas [23]. Li et al. [24] proposed a multi-level output-based DBN model for FLCCs in complex mining areas. Helber et al. [25] used a deep convolutional neural network (DCNN) to perform an LCC of Sentinel-2 satellite images. Xu et al. [26] used a multiscale DCNN to achieve good classification results. However, the DCNN method has a deep network structure and numerous model parameters. Accordingly, gradient disappearance and network degradation can easily be induced. The ResNet architecture increases the network depth through fast identity mapping [27], thereby improving the feature extraction ability and avoiding the above-mentioned issues of the DCNN method. However, ResNet connects the input and residual feature maps by summing them in the skip connection stage, which may block the information flow. ResNet also generates numerous feature maps in each layer and trains with too many parameters, leading to excessive training time. These deficiencies can be solved by the DenseNet architecture [28], as each layer of DenseNet is connected to all subsequent layers in the forward feedback mode, and its input includes the output of the previous layer and the input of all layers before the current output layer. The shallow features that are directly extracted from the original data can be directly used by the deep layers; therefore, DenseNet enhances feature reuse, reduces the number of parameters, and achieves stronger performance than the ordinary CNN method. Therefore, many researchers have introduced it into the field of remote sensing. Tao et al. [29] added a classifier at a certain interval between the input and output layers in the DenseNet architecture to improve the feature extraction ability of the network. Li et al. [30] inputted the features of the last dense block of DenseNet121 into the global branch of the key filter bank to obtain global features; the authors also adopted the right branch of the key flow to extract features from key regions, improving the classification accuracy of hyperspectral imagery (HSI).
Multimodal data-based multistream CNNs are widely used in the field of remote sensing. For example, Wu et al. [31] designed a dual-stream CNN to extract the features of HSI and light detection and ranging (LiDAR) data. Hang et al. [32] designed a dual-stream CNN structure with mutually coupled convolutional layers to extract and fuse HSI and LiDAR features.
Multispectral remote sensing images have spectral features that are related to different spectral bands as well as spatial features that are related to different neighborhoods. Making full use of the spectral spatial features simultaneously is key to improving the classification accuracy. Xu et al. [26] used a three-dimensional CNN to mine spectral–spatial features. To further improve the classification accuracy of the HSI, Zhao and Du [33] designed a multiscale convolutional network to build an image pyramid for each spectral band by fusing the extracted cross-scale spatial features with spectral features. Similarly, Gao et al. [34] designed a deep cross-attention module to extract complementary information from spectral–spatial features. Researchers also introduced making use of generative adversarial networks (GANs) to mine spectral features. Li et al. [35] proposed a deep translation-based change detection network (DTCDN) to automatically extract the variation area between optical and synthetic aperture radar remote sensing images. To establish a more fluent and stable training, Zhang et al. [36] optimized the training process and loss function for GANs and proposed a hyperspectral image classification optimization training method based on GAN. The proposed method can obtain good training results with only a small amount of labeled training data.
Most existing spectral–spatial fusions are based on adjacent pixel feature extraction methods, which do not maximize the contextual information of the spatial–spectral features. A long short-term memory (LSTM) network can establish interdependence between the input sequences and is widely used in natural language processing, video processing, machine translation, and other fields [37]. In the past few years, this particular network has demonstrated excellent performance in remote sensing image processing. Xu et al. [38] designed a band grouping strategy that enabled LSTM networks to better learn the contextual information between adjacent channels of the HSI. To obtain combined spectral–spatial results, researchers have proposed a spectral–spatial LSTM model. Zhou et al. [39] designed a spectral–spatial LSTM to extract the spectral–spatial features of the HSI, which captured the dependencies of adjacent spectral–spatial neighborhoods. Liu et al. [12] proposed a spectral–spatial bidirectional LSTM (BiLSTM) network to simultaneously extract spectral–spatial features using recursive and convolution operators. Similarly, Yin et al. [40] used a 3D CNN to extract spatial features and a band grouping strategy-based BiLSTM to extract spectral features.
Although DCNNs have a strong feature extraction ability, their ability to extract key features in a complex landscape is not strong enough. The attention mechanism can focus on the different aspects of each output of the inputs, thereby improving its ability to extract the most relevant information. Chen et al. [41] proposed a framework based on global context spatial attention and the DenseNet architecture to extract multiscale global scene features. Some researchers are beginning to determine the effects of different attention mechanisms on key feature extraction. Chen et al. [42] proposed generative adversarial networks based on the joint mechanism of channel and spatial attention for land cover classification, which assigns higher weights to more important features and reduces the interference of invalid features. Tong et al. [43] proposed a channel attention-based DenseNet method for scene classification. The channel attention mechanism adaptively strengthens the weights of the important feature channels and suppresses the secondary feature channels. To represent the importance weights of different local regions of each image, Sumbul and Demİr [44] proposed a novel multi-attention strategy-based BiLSTM network to define a global descriptor for each image.
In summary, three issues exist in the FLCCs of complex mining areas: (1) how the multimodal spectral–spatial and topographic features can be learned, (2) how the key features can be extracted, and (3) how the contextual information can be captured among different features. To improve the classification accuracy of FLCCs in complex mining areas, this study proposed a three-stream, double attention network (3S-A2 DenseNet-BiLSTM). This network comprises three strategies: (1) A three-stream multimodal feature learning and post-fusion method. Briefly, the multispectral imagery and topographic data were first fed into subsequent DL networks. Thereafter, the low-level spectral–spatial features were fused with the extracted multimodal deep features; (2) An integration of deep separable asymmetric convolution blocks (ACB-DS), parallel channel, and spatial attention mechanisms into the DenseNet architecture for key feature extraction; and (3) the use of the BiLSTM network to further learn the cross-channel context features from the outputs of the A2 DenseNet. The proposed model was tested in Wuhan City, China, using Ziyuan-3 (ZY-3) stereo satellite imagery.
The remainder of this paper is organized as follows: Section 2 introduces the study area and data sources; Section 3 describes the proposed network structure, feature fusion strategy, parameter optimization, and evaluation indicators; Section 4 presents the analysis of the experimental results, including the experimental equipment, parameter optimization results, and accuracy evaluation; Section 5 discusses the effectiveness of the model and compares it with models used in previous research; and Section 6 presents the conclusions.

2. Study Area and Remote Sensing Datasets

In this study, the same research area was adopted as previous studies [7,23,24], which is located in the Wulongquan mining area, covering an area of 109.4 km2 in the Jiangxia District, Wuhan City, Hubei Province, China. The area is a large ore production base and the mining resources in this area are distributed in contiguous and concentrated areas. Different complex open pit mining areas and agricultural development landscapes exist at different phenological stages.
At present, commonly used satellite images include Sentinel-3, Landsat, Sentinel-2, and HJ-1A [45]; however, ZY-3 satellite images are widely used in FLCCs due to their higher spatial resolution. According to prior studies, the higher the spatial resolution is in multispectral satellites, the more accurate the classification of land cover types in the region [46]. Figure 1 shows a satellite image captured by the ZY-3 satellite on 20 June 2012. As the front- and back-facing cameras had a resolution of 3.6 m, the digital elevation model (DEM) data had a resolution of 10 m. The nadir-looking panchromatic camera had a resolution of 2.1 m, and the multispectral camera had a resolution of 5.8 m. We used multispectral imagery from an optical satellite in this study as the type of satellite images (that is, the ZY-3 satellite). The four bands of blue visible light, green visible light, red visible light, and near-infrared light, which were fused to form the multispectral imagery by Gram–Schmidt spectral sharpening; a fused multispectral imagery with a resolution of 2.1 m was obtained. Subsequently, the DEM was resampled to a 2.1 m resolution to match the multispectral imagery.
In this study, a two-level land cover classification system was designed [47] using twenty-eight field survey samples, including seven first-level and twenty second-level land covers, as shown in Table 1. The second-level land cover type is a detailed classification of the first-level land cover type. The FLCC was based on the second-level land cover type. The design of the two-level land cover classification system is conducive to the classification of land cover on a fine scale.
Based on our previous studies [7,23,24,47,48], the training, validation, and test sets were constructed using data polygons. Each class of the training set contained 2000 samples, and there were 40,000 in total. Each class of the validation set had 500 samples, 10,000 in total; the test set had the same number of samples. The sample proportion of the training, validation, and test sets was at a ratio of 4:1:1, respectively. The training set was used to fit and train multiple classification models. To identify the model with the best performance, each trained model was used to predict the validation set, and we selected the model with the best classification effect. The test set was used to evaluate the model’s accuracy.

3. Methods

3.1. Description of the Proposed 3S-A2 DenseNet-BiLSTM

As shown in Figure 2, the neighborhoods of the N × N pixels were first extracted from the multispectral image and topographic data, respectively, and the neighborhood size was subjectively determined according to the actual situation. Thereafter, these pixel neighborhoods were inputted into the A2 DenseNet-BiLSTM network using an asymmetric separable convolution block and double attention mechanism to extract the cross-channel and context-dependent information. The two branches adopted a parameter sharing method to learn network weights and extract similar features at the same location of different input data, that is, the deep spectral–spatial and topographic features. Subsequently, the low-level spectral–spatial features extracted from the multispectral images were fused with the depth features and inputted into Softmax for classification. By comparing the prediction results with the real labels, the batch size loss was calculated according to the loss function. Finally, a back-propagation algorithm was used to optimize the model.

3.2. A2 DenseNet for Key Feature Extraction

The A2 DenseNet (Figure 3a) network includes the ACB-DS block (Figure 3c), dense block (Figure 3d), convolution block, and transition layer. The dense connection mechanism of DenseNet connects each layer in the dense block, and the input of each layer is the output of all previous layers. If the dense block has L layers, the number of connections can reach L ( L + 1 ) 2 . The nonlinear transformation function is H L ( · ) , and the output of the L layer is X L . For a standard CNN, the output of the L layer is given by:
X L = H L ( X L 1 )
For the dense block, the output of its L layer is given by:
X L = H L ( [ X 0 , X 1 , · · · , X L 1 ] )
The deeper the layers of input data and gradient information transfer, the easier the occurrence of the network in the gradient disappearance phenomenon. Such a dense connection mechanism allows for a more effective transfer of features and gradients and alleviates the problem of gradient disappearance caused by the network being too deep. This mechanism has a regularizing effect that reduces the computational parameters in the network and alleviates the overfitting phenomenon [43].
As shown in Figure 3a, each dense block in DenseNet is followed by a transition layer that consists of a 1 × 1 convolution and average pooling operation. The 1 × 1 convolution can reduce the number of output feature maps and fix the number of channels in output feature maps to reduce the image dimensions, reduce the number of parameters in the network, and reduce the model complexity.
The H L ( · ) structure of the classic DenseNet comprises BN, ReLU, and convolution, with a kernel of 3 × 3. In this study, ACB-DS (Figure 3c) was used to replace the original convolution, which included ACB and deep separable (DS) convolution. For ACB (Figure 3b), the output feature map is the sum of the input images after 3 × 3, 1 × 3, and 3 × 1 convolutions, which are equivalently fused into a square convolution kernel to replace a single convolution block. ACB strengthens the feature representation and feature extraction capabilities of the convolution layer, making the extracted features more discriminative and feature extraction more effective [49].
The channel attention (Figure 4a) and spatial attention (Figure 4b) mechanisms were introduced into the dense block (Figure 3c). These mechanisms were embedded in a parallel form, where the channel attention was implemented by a squeeze-and-excitation net (SENet). An SENet focuses on the inter-channel relationship and explicitly models the inter-channel interdependence to adaptively recalibrate the channel feature response [50]. This network includes two main operations: squeezing and excitation. For the squeeze operation, an adaptive average pooling operation was performed on the spatial dimension. During the compression of the spatial dimension, the average value of the corresponding dimension can be obtained, which can suppress unnecessary features. According to the structure in Figure 4a, the size of the original feature map is H × W × C , denoted as u c , where H and W denote the spatial dimension of the feature map and C is the number of channels. The squeeze operation compresses the feature map H × W × C into a one-dimensional feature vector of 1 × 1 × C , denoted as z c , which is obtained as follows:
z c = F s q u e e z e ( u c ) = i = 1 H j = 1 W u c ( i , j ) H × W
For the excitation operation, a multilayer perceptron was first used to learn the channel weight, which had two layers: a fully connected (FC) layer with the activation function of ReLU (represented by δ) and a FC layer with a sigmoid function (represented by σ). Additionally, s denotes the output weight matrix, g denotes the first FC layer, and Fexcitation denotes the second FC layer. The relations among them are expressed as follows:
s = F e x c i t a i o n ( z c ) = σ ( W 2 δ ( W 1 z c ) )
The obtained weight was multiplied by the original feature map to obtain the final weighted feature map; this was performed to learn the features that were useful for the current task.
The spatial attention mechanism (Figure 4b) can cause the network to focus more on the spatial location of important features. First, the maximum pooling and average pooling operations were performed on the input feature map (represented by F) of H × W × C . Thereafter, the two feature maps were cascaded according to the channel. Finally, the convolution operation f with activation function σ was conducted as follows:
M ( F ) = σ ( f ( A v g P o o l ( F ) , M a x P o o l ( F ) ) )
The spatial attention mechanism can enhance the spatial structure features and spatial neighborhood-related information [51]. The channel and spatial attention branches were merged and fused. The fused feature map has more discriminative spectral–spatial features.

3.3. BiLSTM for Extraction of Cross-Channel Context Features

LSTM is a special recurrent neural network (RNN). Back-propagation and long-term dependencies easily lead to RNN gradient disappearance and explosion. Use of an LSTM introduces a gate mechanism and memory unit to improve RNN [52].
C t denotes the memory unit of the LSTM and memorizes all sequence information up to time t, C t ˜ denotes the update value of the memory cell, x t denotes the input information at time t, b denotes the offset value, W denotes the weights between the LSTM nodes, and h t denotes the hidden layer output. The forget gate f t controls the forgetting of historical information, the input gate i t controls the stored information, and the output gate o t controls the output information. The mathematical illustrations of the LSTM information flow are given by:
f t = σ ( W f · [ h t 1 , x t ] + b f )
i t = σ ( W i · [ h t 1 , x t ] + b i )
C t ˜ = t a n h ( W C · [ h t 1 , x t ] + b C )
C t = f t * C t 1 + i t * C t ˜
o t = σ ( W o · [ h t 1 , x t ] + b o )
h t = o t * t a n h ( C t )
In this study, the feature map X R H × W × C extracted by the A2 DenseNet ( H × W is the feature map size and C is the number of channels) was flattened and fully connected to obtain a one-dimensional feature vector. The feature sequence corresponding to each spectral band was inputted into the BiLSTM (Figure 5). Accordingly, the [ 1 , C ] group sequence { X ( 1 ) , X ( 2 ) , · · · , X ( C ) } was employed as the input of the BiLSTM module to extract the context information across channels. The use of BiLSTM can not only establish the forward correlation between the input sequences, but also learn the dependency between them in the backward sequence to fully extract the context information between different channels.

3.4. Three-Stream Multimodal Feature Learning and Post-Fusion Strategy

The multispectral data included four channels with a neighborhood size of 16 × 16 × 4, and the topographic data included three channels with an input size of 16 × 16 × 3. Based on the multispectral imagery and DEM data, six types of spectral–spatial and topographic features were extracted (Table 2), and the feature dimension was 106. The topographic data included the DEM, slope, and aspect. There were 103 low-level spectral–spatial features. Based on the A2 DenseNet-BiLSTM results for the deep features from the multispectral imagery and topographic data branches, a post-fusion was conducted with low-level features.

3.5. Model Construction and Parameter Optimization

This study sought to design three groups of comparative experiments and one group of the overall experiment.
Comparative Experiment 1: Feature fusion method based on multispectral images, topographic data, and two DenseNet branches.
Comparative Experiment 2: Feature fusion method based on multispectral images, topographic data, and two A2 DenseNet branches.
Comparative Experiment 3: Feature fusion method based on multispectral images, topographic data, and two A2 DenseNet-BiLSTM branches.
Overall Experiment: The proposed 3S-A2 DenseNet-BiLSTM. The three-stream feature fusion methods are based on multispectral images, topographic data, and low-level spectral–spatial features. In addition, there are two A2 DenseNet-BiLSTM branches.
For the four sets of comparative experiments, a total of 4 types of parameters were optimized in this study, as shown in Table 3. Some parameters were optimized by the trial-and-error method and included a batch size of 64, input size of 16 × 16, dense block number of three, BiLSTM layer of two, and A2 DenseNet growth rate of 32.

3.6. Accuracy Assessment Metrics

In this study, four accuracy evaluation metrics, namely the overall accuracy (OA), kappa, F1-score, and F1-measure, were selected to evaluate the performance of the proposed models. The F1-measure was used to assess the accuracy of each land cover type. The F1-score was the average value of all F1-measures.

4. Results

4.1. Results of Parameter Optimization

The four sets of models ran on a machine with the Centos 7 operating system and two NVIDIA 1080Ti GPUs with 11 GB of memory.
Five sets of experiments were repeated five times on the training and validation sets, each with 200 iterations. The results were averaged, and 12 sets of parameters (Table 3) were optimized for each set of experiments. The optimization results of the four experimental groups are shown in Figure 6. The Overall Experiment achieved the best effect when the numbers of the ACB-DS blocks and FC layers in the dense block were set to 1, and when DenseNet121 was used. The corresponding average validation accuracy reached 98.85% ± 0.08%. Comparative Experiment 1 achieved the best effect when DenseNet201 was used, and the numbers of the ACB-DS blocks and FC layers were 4 and 1, respectively. The average validation accuracy of Comparative Experiment 1 reached 93.70% ± 1.23%. Comparative Experiment 2 achieved the best effect when using the same parameter combination as that of the Overall Experiment. The average validation accuracy reached 97.71% ± 0.09%. Comparative Experiment 3 also achieved the best effect when using the same parameter combination as that of the Overall Experiment. The corresponding accuracy was 97.69% ± 0.27%.
As shown in Figure 7, as the epoch increased, the training and validation accuracies of the four sets of experiments displayed an upward trend, and the losses gradually decreased in a small range.
For the Overall Experiment (Figure 7a), the model converged at an epoch of 175 and the OA reached 98.65%. Comparative Experiment 1 (Figure 7b) converged at an epoch of 178, and the OA was 89.10%. Comparative Experiment 2 (Figure 7c) converged at 160, and the OA was 94.83%. Comparative experiment 3 (Figure 7d) converged at 174, with an OA of 97.70%.
After the introduction of the ACB-DS block into DenseNet, the convergence speed of the network was markedly accelerated. Although the training speed of the network decreased after the addition of the BiLSTM module and low-level spectral–spatial features, it remained smaller than the convergence rounds presented in Figure 7b. The Overall Experiment conducted on the proposed network also converged at a relatively fast speed. Other comparative experiments have one or two step-by-step increases in both the training and validation accuracy curves, and the network may jump out of the local optima. The training and validation accuracy curve trends of the Overall Experiment were relatively flat, and local optimization was not easily achieved. The ACB-DS block significantly reduced the computational parameters. Furthermore, the proposed network does not introduce too many redundant modules. The addition of ACB-DS blocks and low-level spectral–spatial features can effectively reduce the starting point of the training and validation loss curves and result in a smoother curve trend.
Overall, although the 3S-A2 DenseNet-BiLSTM model proposed in this study has a markedly more complex structure and a deeper network than DenseNet, it still rapidly converges. The training and validation losses decreased more smoothly than the remaining three sets of comparative experiments, and the training and validation accuracies of the remaining experiments displayed a very steep upward trend. Therefore, the 3S-A2 DenseNet-BiLSTM model had a better performance and more stable properties.

4.2. Results of Accuracy Assessment

The different experiments had different performances in terms of the F1-score, kappa, and OA; the results are presented in Table 4. For the four groups of experiments, the OA values appeared in the order of 98.65% ± 0.05% (Overall Experiment), 97.70% ± 0.08% (Comparative Experiment 3), 97.14% ± 0.28% (Comparative Experiment 2), and 94.83% ± 0.04% (Comparative Experiment 1). The same trend was observed for the other metrics. The experimental results show that the 3S-A2 DenseNet-BiLSTM method is effective for FLCCs in complex mining areas.
Compared with Comparative Experiment 1, after the introduction of the ACB-DS block and parallel attention mechanism in Comparative Experiment 2, the F1-score, kappa, and OA increased by 2.43%, 2.43%, and 2.31%. Compared with Comparative Experiment 2, after the introduction of the BiLSTM module in Comparative Experiment 3, the F1-score, kappa, and OA were increased by 0.56%, 0.58%, and 0.56%, respectively. Compared with Comparative Experiment 3, after introducing the low-level spectral–spatial feature and three-stream feature fusion strategy into the Overall Experiment, the F1-score, kappa, and OA increased by 1.08%, 1.08%, and 0.95%, respectively. The superior performance of the proposed model has been demonstrated. Our previous study [24] concluded that statistical testing depends on the size of the test set. Owing to the small sample size of the test set in this study, a statistical test was not performed. Test sets with different sizes and corresponding statistical tests will be investigated in future studies.
As shown in Table 5, the classification accuracies of each land cover type reached a high level. In fact, even the average F1-measures of woodland, white roof, fallow land, and shrubbery, with the worst classification effects, were above 87%.
In Comparative Experiment 1, the F1-measures of most classes exceeded 94%, with seven exceptions: woodland, white roof, fallow land, green dry land, shrubbery, nursery, and ore processing sites. Among them, fallow land and shrubbery had the lowest F1-measures (lower than 90%). However, the accuracies for dark road and greenhouse in Comparative Experiment 1 reached 100%.
In Comparative Experiment 2, the F1-measures of most classes exceeded 97%, with only six exceptions: woodland, white roof, fallow land, green dry land, nursery, and shrubbery. Generally, the F1-measures of most classes in Comparative Experiment 2 were higher than those of Comparative Experiment 1. Only dark road, greenhouse, and mine pit pond lagged behind Comparative Experiment 1, with accuracy decreases of 0.30%, 0.45%, and 0.05%, respectively. Furthermore, the F1-measures of the other 17 classes were 0.05–6.60% higher than that of Comparative Experiment 1.
In Comparative Experiment 3, the F1-measures of most classes exceeded 97%, with only five exceptions. The accuracy of woodland, white roof, fallow land, and green dry land was still greater than 93%, and only shrubbery had an accuracy of 89.70%. Dark roads and greenhouses were correctly classified. Generally, most types of F1-measures in Comparative Experiment 3 were higher than those in Comparative Experiment 2. Blue roof, white roof, red roof, and mine pit pond lagged by accuracy decreases of only 0.05%, 1.10%, 0.05%, and 0.05%, respectively.
The 3S-A2 DenseNet-BiLSTM method had the best classification effect for almost all 20 types, except for Comparative Experiment 3, which had the best classification effect for nurseries (with a narrow margin of 0.20%). In the Overall Experiment, the F1-measures of most classes exceeded 98%, and the classifier could correctly identify all dark roads, greenhouses, and mine pit ponds, leading to the same conclusion as in Section 4.1: the 3S-A2 DenseNet-BiLSTM model has a significant effect on the FLCCs of complex mining areas.

4.3. Results of Visual Prediction

The four sub-images in Figure 8 clearly show the performances of the four experimental groups. These images comprise 20 different colors that represent 20 secondary land cover classes. Compared with the real image (Figure 1), the four prediction maps were found to be visually accurate. In particular, the visual prediction maps of Comparative Experiment 1 and the Overall Experiment show that the results of the visual prediction are quite different, ultimately confirming the results presented in Table 5. The best algorithm (Figure 8b) is not only better than the worst algorithm (Figure 8a), but is also better by a large gap. Ultimately, the 3S-A2 DenseNet-BiLSTM model can be proven to have a superior performance.
According to Figure 8d, the 3S-A2 DenseNet-BiLSTM model obtained better classification results than the other three models. In particular, the open pit, light gray road, green dryland, ore processing site, all-dark road, greenhouse, and mine pit ponds features were correctly classified. The four groups of experiments had some classification errors for woodland, fallow land, nursery, and shrubbery. Notably, the features of these four types of ground objects might be similar; that is, the spectral–spatial and topographic differences are small. The reasons for this analysis are as follows.
(1)
Compared with Comparative Experiment 1, Comparative Experiment 2 introduces a parallel channel, spatial attention mechanism, and ACB-DS block, which proves the accuracy of 20 classes of land cover on a large scale for the first iteration.
(2)
Compared with Comparative Experiment 2, Comparative Experiment 3 adds BiLSTM based on A2 DenseNet, allowing the model to extract correlation information among the spectral channels. Therefore, the classification accuracy of fallow land, woodland, and shrubbery with similar spectral features was markedly improved for the second iteration.
(3)
Compared with Comparative Experiment 3, Comparative Experiment 4 introduces 106 dimensional low-level features and a multistream post-fusion strategy based on A2 DenseNet-BiLSTM. The features that were manually extracted cover the shortage of depth features extracted by the depth model in detail and have a higher resolution. Therefore, the accuracy of the 20 land cover classes was improved to a certain extent.

5. Discussion

5.1. Effectiveness of the Proposed 3S-A2 DenseNet-BiLSTM Model

According to the results presented in Section 4, the 3S-A2 DenseNet-BiLSTM model has the best FLCC performance for complex mining areas relative to the remaining three classification models proposed in this study.
Many studies have confirmed the effectiveness of the various network modules and fusion strategies introduced in this study. For the ACB-DS module, Lo et al. [53] proposed an efficient dense module with an ACB and dilated convolution structure, which was 2.7 times faster than the comparison network. According to Wang et al. [54], the spatial symmetry of existing convolutional blocks limits the ability of the network to learn the spatial location of features. Thus, a multiscale, spatially asymmetric recalibration network was designed to extract feature maps with spatial asymmetry. Zhu et al. [55] used ACB to reduce the number of floating-point operations and parameters in the hyperspectral image classification process, and simultaneously solved the problem where traditional convolution blocks could not capture features with arbitrary sizes and shapes. Wang et al. [56] designed a spectral and spatial SENet for channel and spatial attention mechanisms, which was used to model the interdependencies between channels and spaces to recalibrate feature responses, with good performance in HSI classification. Roy et al. [57] and Roy et al. [58] used parallel spatial and channel SENet modules to extract richer spatial and channel features. Zhang et al. [59] introduced SENet after each convolutional layer to construct a channel-feature reweighted DenseNet. According to Zhang et al. [60], the convolution operation would ignore the influence of feature points far away from the current region and thus used SENet for feature point channel recalibration. For the multistream feature fusion strategy, Liu et al. [61] used a two-stream CNN to extract the spectral–spatial information of HSI and a cascade network to extract the spatial information of multispectral imagery, forming an effective three-stream network. Ge et al. [62] utilized a multistream 2D CNN to extract features from multi-source datasets. Based on the above studies, the modules and strategies used in the 3S-A2 DenseNet-BiLSTM method are important for improving the accuracy of the FLCCs. The detailed analyses were as follows:
(1)
3S-A2 DenseNet-BiLSTM can learn the joint representation of low-level spectral–spatial, deep spectral–spatial, and deep topographic features. It can also capture the complete information of different emphases of the same object landscape under different imaging methods that cannot be perceived by a single form of data.
(2)
Compared with the classic DenseNet method, the A2 DenseNet module used the ACB-DS block and a double attention mechanism to extract more discriminative features. Simultaneously, the number of parameter calculations was reduced, and the speed of network convergence was accelerated. As the spatial regions of various ground objects are not always symmetrical, traditional symmetric convolution was not suitable for extracting the features of irregularly shaped land covers. However, the ACB-DS method solved the problem well, and was able to extract richer and more detailed spatial information.
(3)
BiLSTM was used to model the contextual features of correlations across different channels. It supplemented the global features extracted by A2 DenseNet and outputted more abundant spectral–spatial and topographic features, respectively.
(4)
A multistream post-fusion strategy was used to further fuse the low-level spectral–spatial features and multimodal deep features extracted by the A2 DenseNet-BiLSTM model. This strategy took full advantage of the spectral, spatial, and topographic information to obtain joint representations.
This study had some limitations in the broadest context:
(1)
The model training time was too long. During the performance of the FLCC experiments, each model training process took approximately 30 h, which is unfavorable for the real-time monitoring of mining activities. A large amount of time hinders the process of industrializing the research. To assess the complexity of the model, both the performance of the model and the difficulty of training must be considered. On the one hand, the model must be as accurate as possible, which requires it to have a higher expression ability; therefore, it is easier to achieve this goal using a model with a higher complexity. On the other hand, if the complexity of the deep model is too high, it will increase the difficulty of training, thus wasting computing resources. Therefore, reducing the training time as much as possible without reducing the expression ability of the deep model requires further research.
(2)
An insufficient sample size was employed. For the data used for model training, only 2000 sample points were used for each class of land cover. For more complex deep models, the nonglobal features of the training data can be learned by the model. However, in the field of remote sensing, it is usually difficult and costly to obtain labeled data; thus, reducing the complexity of the model without reducing its classification accuracy requires further research.
(3)
The generalization performance is unknown. Normally, many illegal mining activities may be carried out at night; however, the current study only monitored mining activities during the day. The performance of the model on remote sensing satellite images at night still requires further transfer learning research.

5.2. Effects of Different Models

To validate the performance of the proposed model, comparisons were made based on the results of previous studies [7,23,24] (Table 6), including traditional MLAs (i.e., RF, SVM, and feature subset-based SVM (FS-SVM)) and some DL methods, such as multilevel output-based DBN (DBN-ML), DBN-RF, and CNN.
For the models based on MLAs, the OA values decreased for RF (88.90% ± 0.20%) [23], SVM (77.88% ± 0.53%) [23], and FS-SVM (91.77% ± 0.57%) [23]. Kwan et al. [63] demonstrated that SVM can achieve better LCC results, and Goldblatt et al. [64] found that the RF model using Landsat 8 satellite data performed better than when using Landsat 7 data. However, when the normalized vegetation and building indices were added to the Landsat 7 data, the performance improvement of the SVM was the most significant.
The DL-based models have been demonstrated to be significantly better than traditional MLAs. The order of OA values observed was: multimodal remote sensing data and multiscale kernel-based multistream CNN with the selected value of parameter k (3M-CNN-Magnify) (96.60%) [7]; multimodal remote sensing data and multiscale kernel-based multistream CNN (3M-CNN) (95.11%) [7]; DBN-ML (95.10%) [24]; DCNN (95.02%) [24]; DBN-SVM (94.74% ± 0.35%) [24]; DBN-S (94.23% ± 0.67%) [23]; DBN-RF (94.07% ± 0.34%) [23]; single-scale CNN(93.76% ± 0.76%) [7]; and CNN (90.20% ± 1.64%) [24]. Compared with the best MLA of FS-SVM, all deep models had better classification performances. Among them, the best model, 3M-CNN-Magnify, had an improved OA of 4.83%. Other studies involving DL-based models have reached similar conclusions. For example, Zhao et al. [65] proved that CNN can effectively handle rice mapping in complex landscape areas. Several researchers have also proposed DBN-based models. For example, Li et al. [66] proposed a multilabel electrical signal classification method based on DBN-RF, which had better classification accuracies and recognition efficiencies than the naive Bayes, k-nearest neighbor (KNN), SVM, and RF models. By comparing DBN-S with the classical KNN and SVM algorithms, Jiang et al. [67] found that the proposed hybrid algorithm can achieve more satisfactory results. Wang et al. [68] demonstrated that the SVM algorithm outperformed single classification methods that used DBN, SVM, and maximum likelihood estimation. Meanwhile, Li et al. [24] revealed that DBN-ML can produce better results in multiscale feature extraction for complex surface conditions while effectively solving the overfitting problem of the DBN models.
The proposed 3S-A2 DenseNet-BiLSTM not only achieved the highest classification accuracy among the comparison models as demonstrated in Figure 9, but also achieved the visual effect closest to the real satellite image (Figure 1), especially for the open pit, bare surface land, and other classes. As shown in Figure 9a–c, they have been misclassified to varying degrees. DBN-S and DBN-SVM also misclassified dark roads as bare surface land and open pits. Notably, none of the models were suitable for the classification of fallow land, woodland, and shrubbery. The reasons for this are summarized as follows:
(1)
Open pit, bare surface land, fallow land, woodland, and shrubbery have relatively similar spectral features, but neither the DBN-based model nor the 3M-CNN-Magnify considers the contextual correlation among spectral bands. Accordingly, it is impossible to distinguish between land covers with similar spectral features and visually similar colors.
(2)
The DBN-based model classifies dark roads into land covers that are similar to this class in terms of geographical space. Although they do not have similar spectral features, the dark road passes through bare surface land and open pits. 3M-CNN-Magnify uses a multiscale kernel-based convolution block, which can select different convolution kernel sizes according to the multi-size data input, making full use of spatial neighborhood information. The proposed 3S-A2 DenseNet-BiLSTM model introduces an ACB-DS block, which can extract the features of irregularly shaped land covers, making the extracted spatial features more abundant. However, DBN-S and DBN-SVM do not have a structure that is more conducive to the extraction of spatial features; thus, they cannot distinguish land covers with strong spatial correlation.
(3)
All models confuse the three classes of land cover: fallow land, woodland, and shrubbery; this may be due to the small spectral differences between them. Moreover, it is difficult to perform visual distinction; however, their height difference is large. Therefore, the model can only extract the discriminative features according to the topographic height difference of the DEM data. However, the 3S-A2 DenseNet-BiLSTM model only extracts the neighborhood size of topographic data instead of a complete image, which may weaken the advantage of topographic information.
In summary, the proposed 3S-A2 DenseNet-BiLSTM model yielded better performance for the FLCCs of complex mining areas.

6. Conclusions

In this study, a three-stream double attention network, 3S-A2 DenseNet-BiLSTM, was proposed for the FLCC of complex mining areas.
Four sets of experiments were performed, and the proposed model was proven to have the best classification performance. Compared with the basic model, the proposed model had an overall accuracy of 98.65% ± 0.05% and obtained better results on the visual prediction map, with an increase in accuracy of 4.03%. Accordingly, the proposed model facilitates the study of complex landscapes. The following conclusions were drawn: (1) the three-stream multimodal feature learning and post-fusion method was effective for learning and fusing low-level spectral–spatial, deep spectral–spatial, and topographic features; (2) integrating ACB-DS blocks into DenseNet can extract richer spatial information; (3) the double attention mechanism can adaptively select more important features; and (4) BiLSTM can extract cross-channel context features.
In conclusion, the proposed 3S-A2 DenseNet-BiLSTM model was determined to be effective for the FLCC of complex mining areas. In the future, we will apply the main study area algorithm to other auxiliary study areas and carry out temporal and spatial transfer learning of the proposed model with the newest remote sensing data and different remote sensing satellite imagery such as Sentinel-3, Landsat, Sentinel-2, HJ-1A, and Gaofen-7 to test the generalization capability of the model. Meanwhile, we will attempt to use the proposed model for the fine classification of other complex geographic environments, as well as build related large remote sensing datasets.

Author Contributions

Conceptualization, D.Z., X.L. and W.H.; Data curation, J.L. and W.C.; Investigation, D.Z., J.L., X.L., W.H. and W.C.; Methodology, D.Z., J.L., X.L., W.H. and W.C.; Software, D.Z. and J.L.; Validation, D.Z. and J.L.; Visualization, D.Z., J.L., X.L., W.H. and W.C.; Writing—original draft, D.Z. and J.L.; Writing—review & editing, D.Z., J.L., X.L., W.H. and W.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China: 42071430; Department of Natural Resources of Hubei Province: ZRZY2021KJ04; National Natural Science Foundation of China: U21A2013; College Students’ Innovation and Entrepreneurship Training Program: S202110491130. And The APC was funded by National Natural Science Foundation of China: 42071430.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hussain, S.; Mubeen, M.; Ahmad, A.; Majeed, H.; Qaisrani, S.A.; Hammad, H.M.; Amjad, M.; Ahmad, I.; Fahad, S.; Ahmad, N.; et al. Assessment of land use/land cover changes and its effect on land surface temperature using remote sensing techniques in Southern Punjab, Pakistan. Environ. Sci. Pollut. Res. 2022, 1–17. [Google Scholar] [CrossRef] [PubMed]
  2. Chen, W.; Wang, Y.; Li, X.; Zou, Y.; Liao, Y.; Yang, J. Land use/land cover change and driving effects of water environment system in Dunhuang Basin, northwestern China. Environ. Earth Sci. 2016, 75, 1027. [Google Scholar] [CrossRef]
  3. Wang, S.; Lu, X.; Chen, Z.; Zhang, G.; Ma, T.; Jia, P.; Li, B. Evaluating the feasibility of illegal open-pit mining identification using insar coherence. Remote Sens. 2020, 12, 367. [Google Scholar] [CrossRef]
  4. Pan, Z.-W.; Shen, H.-L. Multispectral image super-resolution via RGB image fusion and radiometric calibration. IEEE Trans. Image Process. 2018, 28, 1783–1797. [Google Scholar] [CrossRef]
  5. Wang, C.; Guo, Z.; Wang, S.; Wang, L.; Ma, C. Improving hyperspectral image classification method for fine land use assessment application using semisupervised machine learning. J. Spectrosc. 2015, 2015, 969185. [Google Scholar] [CrossRef]
  6. Li, X.; Chen, G.; Liu, J.; Chen, W.; Cheng, X.; Liao, Y. Effects of RapidEye imagery’s red-edge band and vegetation indices on land cover classification in an arid region. Chin. Geogr. Sci. 2017, 27, 827–835. [Google Scholar] [CrossRef]
  7. Qian, M.; Sun, S.; Li, X. Multimodal Data and Multiscale Kernel-Based Multistream CNN for Fine Classification of a Complex Surface-Mined Area. Remote Sens. 2021, 13, 5052. [Google Scholar] [CrossRef]
  8. Wu, C.; Li, X.; Chen, W.; Li, X. A review of geological applications of high-spatial-resolution remote sensing data. J. Circuits Syst. Comput. 2020, 29, 2030006. [Google Scholar] [CrossRef]
  9. Tong, X.-Y.; Xia, G.-S.; Lu, Q.; Shen, H.; Li, S.; You, S.; Zhang, L. Land-cover classification with high-resolution remote sensing images using transferable deep models. Remote Sens. Environ. 2020, 237, 111322. [Google Scholar] [CrossRef]
  10. Shao, Z.; Tang, P.; Wang, Z.; Saleem, N.; Yam, S.; Sommai, C. BRRNet: A Fully Convolutional Neural Network for Automatic Building Extraction From High-Resolution Remote Sensing Images. Remote Sens. 2020, 12, 1050. [Google Scholar] [CrossRef] [Green Version]
  11. Li, X.; Chen, W.; Cheng, X.; Liao, Y.; Chen, G. Comparison and integration of feature reduction methods for land cover classifi-cation with RapidEye imagery. Multimed. Tools Appl. 2017, 76, 23041–23057. [Google Scholar] [CrossRef]
  12. Liu, Q.; Zhou, F.; Hang, R.; Yuan, X. Bidirectional-Convolutional LSTM Based Spectral-Spatial Feature Learning for Hyperspectral Image Classification. Remote Sens. 2017, 9, 1330. [Google Scholar] [CrossRef]
  13. Zhou, G.; Chen, W.; Gui, Q.; Li, X.; Wang, L. Split Depth-Wise Separable Graph-Convolution Network for Road Extraction in Complex Environments from High-Resolution Remote-Sensing Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5614115. [Google Scholar] [CrossRef]
  14. Zheng, H.; Shen, L.; Jia, S. Joint spatial and spectral analysis for remote sensing image classification. MIPPR 2011 Multispectral Image Acquis. Process. Anal. 2011, 8002, 352–356. [Google Scholar] [CrossRef]
  15. Wang, L.; Zhang, J.; Liu, P.; Choo, K.-K.R.; Huang, F. Spectral–spatial multi-feature-based deep learning for hyperspectral remote sensing image classification. Soft Comput. 2017, 21, 213–221. [Google Scholar] [CrossRef]
  16. Chen, W.; Li, X.; He, H.; Wang, L. A Review of Fine-Scale Land Use and Land Cover Classification in Open-Pit Mining Areas by Remote Sensing Techniques. Remote Sens. 2018, 10, 15. [Google Scholar] [CrossRef]
  17. Chen, W.; Li, X.; He, H.; Wang, L. Assessing Different Feature Sets’ Effects on Land Cover Classification in Complex Sur-face-Mined Landscapes by ZiYuan-3 Satellite Imagery. Remote Sens. 2018, 10, 23. [Google Scholar] [CrossRef]
  18. Friedl, M.A.; Brodley, C.E. Decision tree classification of land cover from remotely sensed data. Remote Sens. Environ. 1997, 61, 399–409. [Google Scholar] [CrossRef]
  19. Wang, X.; Guo, Y.; He, J.; Du, L. Fusion of HJ1B and ALOS PALSAR data for land cover classification using machine learning methods. Int. J. Appl. Earth Obs. Geoinf. 2016, 52, 192–203. [Google Scholar] [CrossRef]
  20. Kussul, N.; Lavreniuk, M.; Skakun, S.; Shelestov, A. deep learning classification of land cover and crop types using remote sensing data. IEEE Geosci. Remote Sens. Lett. 2017, 14, 778–782. [Google Scholar] [CrossRef]
  21. Liu, J.; Xiang, J.; Jin, Y.; Liu, R.; Yan, J.; Wang, L. Boost Precision Agriculture with Unmanned Aerial Vehicle Remote Sensing and Edge Intelligence: A Survey. Remote Sens. 2021, 13, 4387. [Google Scholar] [CrossRef]
  22. Zhang, C.; Sargent, I.; Pan, X.; Li, H.; Gardiner, A.; Hare, J.; Atkinson, P. Joint Deep Learning for land cover and land use classi-fication. Remote Sens. Environ. 2019, 221, 173–187. [Google Scholar] [CrossRef]
  23. Li, X.; Tang, Z.; Chen, W.; Wang, L. Multimodal and Multi-Model Deep Fusion for Fine Classification of Regional Complex Landscape Areas Using ZiYuan-3 Imagery. Remote Sens. 2019, 11, 2716. [Google Scholar] [CrossRef]
  24. Li, M.; Tang, Z.; Tong, W.; Li, X.; Chen, W.; Wang, L. A multi-level output-based dbn model for fine classification of complex geo-environments area using ziyuan-3 TMS Imagery. Sensors 2021, 21, 2089. [Google Scholar] [CrossRef] [PubMed]
  25. Helber, P.; Bischke, B.; Dengel, A.; Borth, D. EuroSAT: A novel dataset and deep learning benchmark for land use and land cover classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 2217–2226. [Google Scholar] [CrossRef]
  26. Xu, S.; Mu, X.; Zhao, P.; MA, J. Scene classification of remote sensing image based on multi-scale feature and deep neural network. Acta Geod. Cartogr. Sin. 2016, 45, 834. [Google Scholar]
  27. Chen, W.; Zhou, G.; Liu, Z.; Li, X.; Zheng, X.; Wang, L. NIGAN: A Framework for Mountain Road Extraction Integrating Remote Sensing Road-Scene Neighborhood Probability Enhancements and Improved Conditional Generative Adversarial Network. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5626115. [Google Scholar] [CrossRef]
  28. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  29. Tao, Y.; Xu, M.; Lu, Z.; Zhong, Y. DenseNet-based depth-width double reinforced deep learning neural network for high-resolution remote sensing image per-pixel classification. Remote Sens. 2018, 10, 779. [Google Scholar] [CrossRef]
  30. Li, F.; Feng, R.; Han, W.; Wang, L. High-resolution remote sensing image scene classification via key filter bank based on con-volutional neural network. IEEE Trans. Geosci. Remote Sens. 2020, 58, 8077–8092. [Google Scholar] [CrossRef]
  31. Wu, X.; Hong, D.; Chanussot, J. Convolutional neural networks for multimodal remote sensing data classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5517010. [Google Scholar] [CrossRef]
  32. Hang, R.; Li, Z.; Ghamisi, P.; Hong, D.; Xia, G.; Liu, Q. Classification of hyperspectral and LIDAR data using coupled CNNs. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4939–4950. [Google Scholar] [CrossRef] [Green Version]
  33. Zhao, W.; Du, S. Learning multiscale and deep representations for classifying remotely sensed imagery. ISPRS J. Photogramm. Remote Sens. 2016, 113, 155–165. [Google Scholar] [CrossRef]
  34. Gao, Y.; Li, W.; Zhang, M.; Wang, J.; Sun, W.; Tao, R.; Du, Q. Hyperspectral and multispectral classification for coastal wetland using depthwise feature interaction network. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5512615. [Google Scholar] [CrossRef]
  35. Li, X.; Du, Z.; Huang, Y.; Tan, Z. A deep translation (GAN) based change detection network for optical and SAR remote sensing images. ISPRS J. Photogramm. Remote Sens. 2021, 179, 14–34. [Google Scholar] [CrossRef]
  36. Zhang, F.; Bai, J.; Zhang, J.; Xiao, Z.; Pei, C. An optimized training method for GAN-based hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2020, 18, 1791–1795. [Google Scholar] [CrossRef]
  37. Sak, H.; Senior, A.; Beaufays, F. Long short-term memory based recurrent neural network architectures for large vocabulary speech recognition. arXiv 2014, arXiv:1402.1128. [Google Scholar]
  38. Xu, Y.; Du, B.; Zhang, L.; Zhang, F. A Band grouping based LSTM algorithm for hyperspectral image classification. In Proceedings of the CCF Chinese Conference on Computer Vision, Tianjin, China, 11–14 October 2017; pp. 421–432. [Google Scholar] [CrossRef]
  39. Zhou, F.; Hang, R.; Liu, Q.; Yuan, X. Hyperspectral image classification using spectral-spatial LSTMs. Neurocomputing 2019, 328, 39–47. [Google Scholar] [CrossRef]
  40. Yin, J.; Qi, C.; Chen, Q.; Qu, J. Spatial-Spectral Network for Hyperspectral Image Classification: A 3-D CNN and Bi-LSTM Framework. Remote Sens. 2021, 13, 2353. [Google Scholar] [CrossRef]
  41. Chen, W.; Ouyang, S.; Tong, W.; Li, X.; Zheng, X.; Wang, L. GCSANet: A global context spatial attention deep learning network for remote sensing scene classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 1150–1162. [Google Scholar] [CrossRef]
  42. Chen, W.; Ouyang, S.; Yang, J.; Li, X.; Zhou, G.; Wang, L. JAGAN: A Framework for Complex Land Cover Classification Using Gaofen-5 AHSI Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 1591–1603. [Google Scholar] [CrossRef]
  43. Tong, W.; Chen, W.; Han, W.; Li, X.; Wang, L. Channel-attention-based DenseNet network for remote sensing image scene classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 4121–4132. [Google Scholar] [CrossRef]
  44. Sumbul, G.; Demİr, B. A deep multi-attention driven approach for multi-label remote sensing image classification. IEEE Access 2020, 8, 95934–95946. [Google Scholar] [CrossRef]
  45. Yang, X.; Li, Y.; Wei, Y.; Chen, Z.; Xie, P. Water Body Extraction from Sentinel-3 Image with Multiscale Spatiotemporal Super-Resolution Mapping. Water 2020, 12, 2605. [Google Scholar] [CrossRef]
  46. Waleed, M.; Mubeen, M.; Ahmad, A.; Habib-ur-Rahman, M.; Amin, A.; Farid, H.U.; Hussain, S.; Ali, M.; Qaisrani, S.A.; Nasim, W.; et al. Evaluating the efficiency of coarser to finer resolution multi-spectral satellites in mapping paddy rice fields using GEE implementation. Sci. Rep. 2020, 12, 13210. [Google Scholar] [CrossRef] [PubMed]
  47. Li, X.; Chen, W.; Cheng, X.; Wang, L. A comparison of machine learning algorithms for mapping of complex surface-mined and agricultural landscapes using ZiYuan-3 stereo satellite imagery. Remote Sens. 2016, 8, 514. [Google Scholar] [CrossRef]
  48. Guan, R.; Li, Z.; Li, T.; Li, X.; Yang, J.; Chen, W. Classification of Heterogeneous Mining Areas Based on ResCapsNet and Gaofen-5 Imagery. Remote Sens. 2022, 14, 3216. [Google Scholar] [CrossRef]
  49. Ding, X.; Guo, Y.; Ding, G.; Han, J. ACNet: Strengthening the Kernel Skeletons for Powerful CNN via Asymmetric Convolution Blocks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019. [Google Scholar] [CrossRef]
  50. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  51. Mei, X.; Pan, E.; Ma, Y.; Dai, X.; Huang, J.; Fan, F.; Du, Q.; Zheng, H.; Ma, J. Spectral-spatial attention networks for hyperspectral image classification. Remote Sens. 2019, 11, 963. [Google Scholar] [CrossRef]
  52. Shi, X.; Chen, Z.; Wang, H.; Yeung, D.; Wong, W.; Wang, C. Convolutional LSTM network: A machine learning approach for precipitation nowcasting. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; p. 28. [Google Scholar]
  53. Lo, S.-Y.; Hang, H.-M.; Chan, S.-W.; Lin, J.-J. Efficient dense modules of asymmetric convolution for real-time semantic segmentation. In Proceedings of the 2019 ACM Multimedia Asia, Beijing, China, 15–18 December 2019; pp. 1–6. [Google Scholar] [CrossRef]
  54. Wang, Y.; Xie, L.; Qiao, S.; Zhang, Y.; Zhang, W.; Yuille, A.L. Multi-scale spatially-asymmetric recalibration for image classification. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 523–539. [Google Scholar] [CrossRef]
  55. Zhu, M.; Fan, J.; Yang, Q.; Chen, T. SC-EADNet: A Self-Supervised Contrastive Efficient Asymmetric Dilated Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5519517. [Google Scholar] [CrossRef]
  56. Wang, L.; Peng, J.; Sun, W. Spatial–spectral squeeze-and-excitation residual network for hyperspectral image classification. Remote Sens. 2019, 11, 884. [Google Scholar] [CrossRef]
  57. Roy, S.K.; Dubey, S.R.; Chatterjee, S.; Chaudhuri, B.B. FuSENet: Fused squeeze-and-excitation network for spectral-spatial hy-perspectral image classification. IET Image Process. 2020, 14, 1653–1661. [Google Scholar] [CrossRef]
  58. Roy, A.G.; Navab, N.; Wachinger, C. Concurrent spatial and channel ‘squeeze & excitation’ in fully convolutional networks. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Granada, Spain, 16–20 September 2018; pp. 421–429. [Google Scholar]
  59. Zhang, K.; Guo, Y.; Wang, X.; Yuan, J.; Ding, Q. Multiple feature reweight densenet for image classification. IEEE Access 2019, 7, 9872–9880. [Google Scholar] [CrossRef]
  60. Zhang, K.; Guo, Y.; Wang, X.; Yuan, J.; Ma, Z.; Zhao, Z. Channel-wise and feature-points reweights densenet for image classifi-cation. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 410–414. [Google Scholar]
  61. Liu, C.; Tao, R.; Li, W.; Zhang, M.; Sun, W.; Du, Q. Joint classification of hyperspectral and multispectral images for mapping coastal wetlands. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 14, 982–996. [Google Scholar] [CrossRef]
  62. Ge, C.; Gu, I.Y.; Jakola, A.S.; Yang, J. Deep learning and multi-sensor fusion for glioma classification using multistream 2D con-volutional networks. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; pp. 5894–5897. [Google Scholar]
  63. Kwan, C.; Gribben, D.; Ayhan, B.; Li, J.; Bernabe, S.; Plaza, A. An accurate vegetation and non-vegetation differentiation approach based on land cover classification. Remote Sens. 2020, 12, 3880. [Google Scholar] [CrossRef]
  64. Goldblatt, R.; You, W.; Hanson, G.; Khandelwal, A.K. Detecting the boundaries of urban areas in India: A dataset for pixel-based image classification in google earth engine. Remote Sens. 2016, 8, 634. [Google Scholar] [CrossRef]
  65. Zhao, S.; Liu, X.; Ding, C.; Liu, S.; Wu, C.; Wu, L. Mapping rice paddies in complex landscapes with convolutional neural net-works and phenological metrics. Gisci. Remote Sens. 2019, 57, 37–48. [Google Scholar] [CrossRef]
  66. Li, K.; Yu, N.; Li, P.; Song, S.; Wu, Y.; Li, Y.; Liu, M. Multi-label spacecraft electrical signal classification method based on DBN and random forest. PLoS ONE 2017, 12, e0176614. [Google Scholar] [CrossRef]
  67. Jiang, M.; Liang, Y.; Feng, X.; Fan, X.; Pei, Z.; Xue, Y.; Guan, R. Text classification based on deep belief network and softmax regression. Neural Comput. Appl. 2016, 29, 61–70. [Google Scholar] [CrossRef]
  68. Wang, H.; Wu, Z.; Ma, S.; Lu, S.; Zhang, H.; Ding, G.; Li, S. Deep learning for signal demodulation in physical layer wireless communications: Prototype platform, open dataset, and analytics. IEEE Access 2019, 7, 30792–30801. [Google Scholar] [CrossRef]
Figure 1. ZY-3 satellite remote sensing image of the study area (revised from [23]).
Figure 1. ZY-3 satellite remote sensing image of the study area (revised from [23]).
Sustainability 14 12465 g001
Figure 2. The algorithm flow chart of the 3S-A2 DenseNet-BiLSTM.
Figure 2. The algorithm flow chart of the 3S-A2 DenseNet-BiLSTM.
Sustainability 14 12465 g002
Figure 3. Structure of the A2 DenseNet network (a), asymmetric convolution block (ACB) (b), ACB-DS block (c), and dense block (d).
Figure 3. Structure of the A2 DenseNet network (a), asymmetric convolution block (ACB) (b), ACB-DS block (c), and dense block (d).
Sustainability 14 12465 g003
Figure 4. Structure of the channel attention module (a) and spatial attention module (b).
Figure 4. Structure of the channel attention module (a) and spatial attention module (b).
Sustainability 14 12465 g004
Figure 5. Structure of BiLSTM.
Figure 5. Structure of BiLSTM.
Sustainability 14 12465 g005
Figure 6. Average and standard deviation values of the validation accuracy for the four groups of comparative experiments.
Figure 6. Average and standard deviation values of the validation accuracy for the four groups of comparative experiments.
Sustainability 14 12465 g006
Figure 7. The average and standard deviation values of the training and validation accuracies and losses at each epoch for the Overall Experiment (a), Comparative Experiment 1 (b), Comparative Experiment 2 (c), and Comparative Experiment 3 (d).
Figure 7. The average and standard deviation values of the training and validation accuracies and losses at each epoch for the Overall Experiment (a), Comparative Experiment 1 (b), Comparative Experiment 2 (c), and Comparative Experiment 3 (d).
Sustainability 14 12465 g007
Figure 8. Visual prediction maps of the study area for Comparative Experiment 1 (a), Comparative Experiment 2 (b), Comparative Experiment 3 (c), and the Overall Experiment (d).
Figure 8. Visual prediction maps of the study area for Comparative Experiment 1 (a), Comparative Experiment 2 (b), Comparative Experiment 3 (c), and the Overall Experiment (d).
Sustainability 14 12465 g008
Figure 9. Visual prediction maps of the study area for the DBN-S (a), DBN-SVM (b), 3M-CNN-Magnify (c), and 3S-A2 DenseNet-BiLSTM (d) models.
Figure 9. Visual prediction maps of the study area for the DBN-S (a), DBN-SVM (b), 3M-CNN-Magnify (c), and 3S-A2 DenseNet-BiLSTM (d) models.
Sustainability 14 12465 g009
Table 1. The two-level land cover classification system [47].
Table 1. The two-level land cover classification system [47].
First-Level Land Cover TypeSecond-Level Land Cover TypeDescription
Mining landOpen pit
Ore processing site
Dumping ground
An ore deposit formed by stripping the soil and rock covering the upper part of the ore body
A factory that processes ore
A place where mining wastes are discharged in a centralized manner
FarmlandPaddy
Greenhouse
Green dry land
Fallow land
Cultivated land for planting rice
Facilities that can transmit light and keep warm, which are used to cultivate plants
Cultivated land not planted with rice
Uncultivated land
WoodlandWoodland
Shrubbery
Coerced forest
Nursery
Tall macrophanerophytes
Low vegetation
Vegetation with restricted growth
Economic trees artificially cultivated in nurseries and orchards
WatersPond and stream
Mine pit pond
Natural waters
A place where groundwater is discharged during mining
RoadDark road
Light gray road
Bright road
Asphalt driveway
Dirt road
Cement road
Residential landBlue roof
White roof
Red roof
Urban land
Land for rural residents
Other construction land
Unused landBare surface land
Table 2. Low-level multimodal features.
Table 2. Low-level multimodal features.
Feature Parameter TypeFeature Parameter NameNumber
Spectral featuresSpectral bands4
Principal component featuresFirst and second principal components of spectral bands2
Vegetation indexNormalized vegetation index eliminating, with the difference between the two channel reflectors1
Filter featuresGaussian low-pass, mean, and standard deviation filtering in the spectral band with a core size of 3 × 3, 5 × 5, and 7 × 7 pixels36
Texture featuresGray level co-occurrence matrix texture in the spectral band, including the contrast, correlation, angular second moment, homogeneity and entropy. Core size 3 × 3, 5 × 5, and 7 × 7 pixels60
Topographic featuresDEM, slope, aspect3
Table 3. Parameter optimization scheme.
Table 3. Parameter optimization scheme.
Comparative Experiment 1Other Experiments
DenseNet depthDenseNet121ACB-DS1, 2, 3, 4
DenseNet161Convergent epoch1–200
DenseNet169Fully connected layer1, 2, 3
DenseNet201
Convergent epoch1–200
Fully connected layer1, 2, 3
Table 4. The average and standard deviation values of the four groups of experiments on the metrics of F1-score, kappa, and OA (%).
Table 4. The average and standard deviation values of the four groups of experiments on the metrics of F1-score, kappa, and OA (%).
ExperimentF1-ScoreKappaOA
Comparative Experiment 194.82 ± 0.0494.56 ± 0.0494.83 ± 0.04
Comparative Experiment 297.13 ± 0.2996.99 ± 0.3097.14 ± 0.28
Comparative Experiment 397.69 ± 0.0897.57 ± 0.0897.70 ± 0.08
Overall Experiment98.65 ± 0.0598.65 ± 0.0598.65 ± 0.05
Table 5. Average and standard deviation values of the F1-measures for the four groups of experiments (%).
Table 5. Average and standard deviation values of the F1-measures for the four groups of experiments (%).
ClassOverall
Experiment
Comparative Experiment 1Comparative Experiment 2 Comparative Experiment 3
Open pit99.75 ± 0.2295.80 ± 0.2097.90 ± 1.0297.90 ± 0.50
Ore processing site99.45 ± 0.0992.92 ± 0.2897.35 ± 1.1798.20 ± 0.20
Dumping ground99.70 ± 0.1796.90 ± 0.1099.20 ± 0.1499.20 ± 0.00
Paddy99.75 ± 0.0996.20 ± 0.2098.75 ± 0.7099.00 ± 0.20
Greenhouse100.00 ± 0.00100.00 ± 0.0099.55 ± 0.09100.00 ± 0.00
Green dry land 99.35 ± 0.2293.20 ± 2.0096.85 ± 1.1996.90 ± 0.10
Fallow land96.90 ± 0.5987.70 ± 0.1091.80 ± 2.7894.00 ± 0.60
Woodland95.30 ± 0.3090.80 ± 0.2092.20 ± 1.0793.00 ± 0.20
Shrubbery92.35 ± 0.5280.30 ± 0.7086.90 ± 2.3489.70 ± 0.10
Coerced forest99.55 ± 0.0996.20 ± 0.0099.00 ± 0.4599.00 ± 0.00
Nursery98.20 ± 0.6290.70 ± 2.1096.85 ± 1.3698.40 ± 0.20
Pond and stream 99.15 ± 0.4195.40 ± 0.2098.35 ± 0.5799.10 ± 0.10
Mine pit pond100.00 ± 0.0099.90 ± 0.1099.85 ± 0.0999.80 ± 0.00
Dark road100.00 ± 0.00100.00 ± 0.0099.70 ± 0.17100.00 ± 0.00
Light gray road99.70 ± 0.1797.10 ± 0.5098.50 ± 0.7198.80 ± 0.20
Bright road99.95 ± 0.0998.70 ± 0.5098.75 ± 0.8099.60 ± 0.20
Blue roof99.80 ± 0.0099.10 ± 0.1099.75 ± 0.0999.70 ± 0.10
White roof94.85 ± 0.4391.80 ± 0.4094.50 ± 1.6893.40 ± 0.20
Red roof99.75 ± 0.0997.40 ± 0.4099.55 ± 0.2699.50 ± 0.10
Bare surface land99.55 ± 0.1796.50 ± 0.3097.15 ± 2.1798.70 ± 0.10
Table 6. Accuracy evaluation results of the 13 classification models on the ZY-3 dataset (%).
Table 6. Accuracy evaluation results of the 13 classification models on the ZY-3 dataset (%).
ModelF1-ScoreKappaOADescription
3S-A2 DenseNet-BiLSTM98.65 ± 0.0598.65 ± 0.0598.65 ± 0.05The proposed network
Single-scale CNN [7] 93.76 ± 0.76Multimodal and single-scale kernel-based multistream CNN
3M-CNN [7] 95.11 ± 0.48Multimodal and multiscale kernel-based multistream CNN
3M-CNN-Magnify [7] 96.60 ± 0.22Multimodal and multiscale kernel-based multistream CNN with the selected parameter value
DBN-ML [24]95.0794.8495.10Multi-level output-based deep belief network
RF [23]88.85 ± 0.2288.31 ± 0.2288.90 ± 0.20Random forest
SVM [23]77.79 ± 0.5476.72 ± 0.5577.88 ± 0.53Support vector machine
FS-SVM [23]91.75 ± 0.5791.34 ± 0.6091.77 ± 0.57SVM with feature fusion method
DBN-S [23]94.22 ± 0.6793.93 ± 0.7094.23 ± 0.67DBN with Softmax classifier
DBN-RF [23]94.05 ± 0.3493.76 ± 0.3694.07 ± 0.34DBN with feature fusion method
DBN-SVM [23]94.72 ± 0.3594.46 ± 0.3794.74 ± 0.35DBN with SVM classifier
CNN [24]90.15 ± 1.6689.68 ± 1.7590.20 ± 1.64VGG network
DCNN [24]95.0094.7695.02VGG with deformable convolutions
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, D.; Leng, J.; Li, X.; He, W.; Chen, W. Three-Stream and Double Attention-Based DenseNet-BiLSTM for Fine Land Cover Classification of Complex Mining Landscapes. Sustainability 2022, 14, 12465. https://doi.org/10.3390/su141912465

AMA Style

Zhang D, Leng J, Li X, He W, Chen W. Three-Stream and Double Attention-Based DenseNet-BiLSTM for Fine Land Cover Classification of Complex Mining Landscapes. Sustainability. 2022; 14(19):12465. https://doi.org/10.3390/su141912465

Chicago/Turabian Style

Zhang, Diya, Jiake Leng, Xianju Li, Wenxi He, and Weitao Chen. 2022. "Three-Stream and Double Attention-Based DenseNet-BiLSTM for Fine Land Cover Classification of Complex Mining Landscapes" Sustainability 14, no. 19: 12465. https://doi.org/10.3390/su141912465

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop