Next Article in Journal
Spatiotemporal Downscaling of GRACE Total Water Storage Using Land Surface Model Outputs
Next Article in Special Issue
Ghost Elimination via Multi-Component Collaboration for Unmanned Aerial Vehicle Remote Sensing Image Stitching
Previous Article in Journal
Improved Boreal Forest Wildfire Fuel Type Mapping in Interior Alaska Using AVIRIS-NG Hyperspectral Data
Previous Article in Special Issue
CF2PN: A Cross-Scale Feature Fusion Pyramid Network Based Remote Sensing Target Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SSCNN-S: A Spectral-Spatial Convolution Neural Network with Siamese Architecture for Change Detection

1
Jiangsu Key Construction Laboratory of Audit Information Engineering, Nanjing Audit University, Nanjing 211815, China
2
School of Information Engineering, Nanjing Audit University, Nanjing 211815, China
3
School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing 210094, China
4
School of Electronic Information, Qingdao University, Qingdao 266071, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(5), 895; https://doi.org/10.3390/rs13050895
Submission received: 9 January 2021 / Revised: 23 February 2021 / Accepted: 23 February 2021 / Published: 27 February 2021
(This article belongs to the Special Issue Semantic Segmentation of High-Resolution Images with Deep Learning)

Abstract

:
In this paper, a spectral-spatial convolution neural network with Siamese architecture (SSCNN-S) for hyperspectral image (HSI) change detection (CD) is proposed. First, tensors are extracted in two HSIs recorded at different time points separately and tensor pairs are constructed. The tensor pairs are then incorporated into the spectral-spatial network to obtain two spectral-spatial vectors. Thereafter, the Euclidean distances of the two spectral-spatial vectors are calculated to represent the similarity of the tensor pairs. We use a Siamese network based on contrastive loss to train and optimize the network so that the Euclidean distance output by the network describes the similarity of tensor pairs as accurately as possible. Finally, the values obtained by inputting all tensor pairs into the trained model are used to judge whether a pixel belongs to the change area. SSCNN-S aims to transform the problem of HSI CD into a problem of similarity measurement for tensor pairs by introducing the Siamese network. The network used to extract tensor features in SSCNN-S combines spectral and spatial information to reduce the impact of noise on CD. Additionally, a useful four-test scoring method is proposed to improve the experimental efficiency instead of taking the mean value from multiple measurements. Experiments on real data sets have demonstrated the validity of the SSCNN-S method.

1. Introduction

Due to the development of remote sensing technology it is possible to obtain hyperspectral images (HSIs) of the same area at different time points. Change detection (CD) using multitemporal remote sensing data has an important application value in disaster assessment [1], terrain change analysis [2], urban change analysis [3] and resource auditing. The rich spectral and spatial information of HSIs, which contain hundreds of bands, provides a more powerful data source for object observation. In [4], the author divides CD into the following categories: anomaly detection [5,6,7], binary and multiclass CD [8,9,10,11] and CD based on time series data [12,13].
Many researchers have studied the multispectral CD task with a low number of bands and proposed a few CD algorithms. Change vector analysis (CVA) [14] is often combined with other methods. By calculating the spectral change vector corresponding to a pixel, the magnitude and angle of the spectral change of the pixel are analyzed. Multivariate alteration detection (MAD) [15] and iteratively reweighted multivariate alteration detection (IR-MAD) [16] are based on canonical correlation analysis (CCA) [17]. The change area is determined by calculating the values and their weights of MAD variables. In addition, it is also a feasible strategy to classify HSIs at different times and to compare the classification results to determine the change area. This strategy can introduce excellent algorithms in the field of HSI classification into CD [18,19]. The aforementioned methods use algebraic and statistical theories to extract the features of the spectral vector or spectral change vector and have achieved good results with low-dimensional space.
However, the CD algorithm—which is suitable for low-dimensional space—does not work well in the high-dimensional space of HSIs [20,21]. One important reason for this is that the limited calculation accuracy of computers will cause certain calculation errors. While these calculation errors have a limited impact on the final result in low-dimensional space, a few vector and matrix operations performed in high-dimensional space (e.g., solving the inverse matrix and eigenvectors of high-dimensional matrices) may be greatly affected by calculation errors.
Additionally, due to the strong correlation between adjacent bands in HSIs, a large amount of redundant information also increases the difficulty of feature extraction [22,23,24]. Therefore, reducing the dimension of the HSI is an important topic. The aforementioned CCA algorithm achieves the goal of dimension reduction by finding two typical variables of a lower dimension to represent the original vector (of a higher dimension). Principal component analysis (PCA) [25] solves the eigenvalues and eigenvectors of the covariance matrix and selects the eigenvectors corresponding to the largest k eigenvalues to form a linear transformation matrix to map the original data to the specified dimension k. Considering the high correlation between adjacent hyperspectral bands, it is also prudent to select a certain number of bands from the original hyperspectral image for feature extraction. Ma et al. [26] improved the effect of CD by selecting bands with more change information and processing them to suppress noise. In addition to linear dimension reduction, manifold learning as a non-linear dimension reduction method is also applied in hyperspectral image processing. Yu et al. [27] improved the neighborhood rough set, proposed the local neighborhood selection method combined with local linear embedding (INRSLLE) and effectively reduced the dimension by using local manifold learning. With the development of deep learning [28,29,30], it has been applied to reduce the dimension and extract the features [31,32,33]. Chen et al. [34] constructed a 1D convolution network to extract the features of the spectral vectors corresponding to a single pixel and achieved the effect of hyperspectral dimension reduction. Lv et al. [35] suppressed noise in synthetic aperture radar (SAR) images by stacking contractive autoencoder (sCAE) and extracting features to improve the accuracy of CD.
Noise is unavoidable in hyperspectral images and originates from the internal noise of the hyperspectral imager itself and external factors such as atmospheric scattering. Li et al. [36] mentioned that CD algorithms that only extract the spectral information of pixels will be affected by noise, thereby resulting in poor CD results. Therefore, it is important to extract HSI features other than spectral features. Wang et al. [37] introduced the endmember abundance information obtained by unmixing into an affinity matrix and used a convolution neural network (CNN) for CD to achieve good results. Spatial information is an important source of information for HSIs. Notably, a few advanced algorithms have introduced the extraction of spatial features. Wang et al. [38] used 1D and 2D convolution networks to extract the spectral and spatial features of HSIs, respectively. Huang et al. [39] proposed the tensor-based hyperspectral remote sensing images underlying features change information model (TFS-Cube) for feature extraction, which also included spatial information. Furthermore, Ran et al. [40] extracted neighborhood spatial information using three different combinations of filters. Roy et al. [41] proposed a hybrid spectral CNN (HybridSN) to blend 2D convolution networks with 3D convolution networks for more abstract spatial features.
Tensor-based methods are widely used in HSI processing to extract spectral and spatial information simultaneously. Similar to CVA, we can obtain the change tensor by subtracting two tensors. However, this results in losing the original spatial information of the HSI. While we can also classify the two tensors separately and determine whether the pixel belongs to the change area based on the classification results, this requires prior knowledge of which categories the tensor can be classified into.
Based on the face and action recognition task [42,43,44,45,46], we can segment the change area using a Siamese network. A Siamese network measures the similarity of tensors and can accept two inputs. The more similar the two inputs, the smaller the corresponding output value and vice versa. A Siamese network can solve the aforementioned problems: it can not only retain complete spatial information but also does not require any prior information about categories.
Therefore, a spectral-spatial network with Siamese architecture (SSCNN-S) is proposed to solve the problem of hyperspectral binary CD. First, for each pixel to be detected, the spectral information of the pixel and its neighborhoods are extracted from the two HSIs recorded at different time points to form a tensor pair as the input of the network. The vectors in the tensor pair are then incorporated into a Siamese network composed of the spectral module and the spatial module to obtain the corresponding spectral-spatial vectors. The Euclidean distance of the two spectral-spatial vectors is used as the similarity of the two tensors in the tensor pair. After obtaining the similarity for each pixel, the similarity is binarized by the threshold method to generate the final CD result.
The method proposed in this paper uses the theory of deep learning to measure the similarity of high-dimensional tensors and detect changes, which has the following main advantages. Using the Siamese network, the problem of CD is transformed into a problem of measuring the similarity of two tensors, which retains the complete spectral and spatial information of HSIs compared with the differential method. Extracting the spectral characteristics of the tensor through a 1D convolution network and reducing the dimensions in the Siamese network can effectively reduce the parameters of the network and increase the CD speed. Moreover, a 2D convolution network extracts spatial characteristics from the reduced dimension tensor, which can reduce the impact of noise and improve the CD accuracy. Experiments using three real data sets show that the SSCNN-S method proposed in this paper shows good performance in solving the problem of CD in HSIs. The main contributions of this work can be summarized as follows:
(1)
1D and 2D convolutional neural networks are used to extract spectral features and spatial features while local tensors are converted into spectral-spatial vectors. In this manner, the spectral and the spatial features are combined to increase detection speed.
(2)
The introduction of the Siamese network helps to retain the original spatial features and can introduce advanced hyperspectral classification methods [47] into CD without the prior information of the number of categories and other processing methods [48].
(3)
The four-test scoring method is proposed. This method is mainly used in parameter selection experiments with uncertain results. For each set of parameters, the method can give the final results based on the results of two to four independent experiments as the basis for parameter selection.
The other parts of this paper are arranged as follows. The second part details the proposed hyperspectral CD method (SSCNN-S) based on the Siamese network and the spectral-spatial combination method. The third part describes the experiments performed in this study by outlining the utilized data set and evaluation index as well as comparing and analyzing existing algorithms. The final part summarizes this paper.

2. Materials and Methods

2.1. Establishing the Sample Set

Let X ( t ) h × w × c represent the HSI at the t t h time where h and w correspond to the height and width of the hyperspectral spatial dimension, respectively, and c represents the number of bands in the HSI t { 1 , 2 } . Considering the i t h pixel in X ( 1 ) , to extract both the spectral and spatial features of this pixel, we extract the spectral information of this pixel and its neighborhood to form a hyperspectral tensor X i ( 1 ) b × b × c where b represents the spatial dimension of the hyperspectral tensor. The corresponding tensors from each pixel in X ( 1 ) and X ( 2 ) are extracted to form a tensor set X ( 1 ) = { X 1 ( 1 ) , X 2 ( 1 ) , , X h w ( 1 ) } and X ( 2 ) = { X 1 ( 2 ) , X 2 ( 2 ) , , X h w ( 2 ) } . To easily use the Siamese network to measure the similarity between tensors, we need to pair the tensors corresponding to the same pixels in X ( 1 ) and X ( 2 ) to form a tensor pair set X = { ( X 1 ( 1 ) , X 1 ( 2 ) ) , ( X 2 ( 1 ) , X 2 ( 2 ) ) , , ( X h w ( 1 ) , X h w ( 2 ) ) } . Label Y = { Y 1 , Y 2 , , Y h w } corresponds to X where:
Y i = { 0 the   i t h   pixel   is   unchanged 1 the   i t h   pixel   is   changed
Thus, ( X , Y ) can be considered as the sample set. The sample set is randomly divided into the training set ( X t r a i n , Y t r a i n ) , validation set ( X v a l , Y v a l ) and test set ( X t e s t , Y t e s t ) .

2.2. Extract Spectral-Spatial Features of the Hyperspectral Tensor

Similar to a previous work [38], we extracted the spectral and spatial features of the hyperspectral tensor X i ( t ) using a 1D and 2D convolution neural network (Figure 1). Unlike the classification task, the network does not output a specific category but a spectral-spatial vector that combines spectral and spatial features. The network is divided into a spectral module and a spatial module where the spectral module is used to extract the spectral features of hyperspectral tensors while reducing the dimension of the tensors and the spatial module is used to extract the spatial features of the reduced tensors.

2.2.1. Spectral Module

HSIs have abundant spectral information. Most methods consider the extraction of spectral features as an important research topic. A hyperspectral tensor X i ( t ) b × b × c contains b × b spectral vectors and each has a dimension of c . Therefore, a 1D convolutional network can be used to extract the features of each spectral vector [34] and combine them to obtain the spectral features of the tensor. Another important role of 1D convolution is to reduce the spectral dimension of the tensor to k. In each convolution layer, different convolution kernels correspond to different methods of dimension reduction and these dimension reduction parameters can be automatically learned during network training. This can reduce the number of model parameters, which save computing resources and storage space while effectively improving the model training speed.
In network training, the overall distribution of the activation function value of the hidden layer shifts greatly, which causes the gradient to disappear and the training speed to decrease. To avoid this problem, we used batch normalization (BN) [49] in the spectral module to force the overall distribution back to the standard normal distribution.
The activation function used in the spectral module is the parametric rectified linear unit (PReLU) [50] whose function expression is as follows:
PReLU ( x p ) = max { 0 , x p } + a p min { 0 , x p }
where x p represents the input of the activation function for the p th channel and a p is a very small positive number. While a p is not set artificially, it is updated during model training as follows:
Δ a p m Δ a p + l r l o s s a p
where m represents momentum and l r represents the learning rate. HSIs contain values less than 0. Compared with the rectified linear unit (ReLU) function [51], the PReLU function does not set the activation values less than 0 directly to 0. Instead, it compresses them to a close negative value of 0. Although a few parameters that require training are added, faster model training is more advantageous.

2.2.2. Spatial Module

Although the spatial resolution of the HSI is not high, extracting the spatial features of the HSI as an adjunct basis for CD can help reduce the interference of noise in the spectral information. After dimension reduction, the hyperspectral tensor X s i ( t ) b × b × k contains k tensors of size b × b , each with a large amount of spatial information [52]. We extract spatial information [34] using a 2D convolution network with a kernel size of n × n . In the spatial module, we still use PReLU and BN after each convolution layer.

2.2.3. Achieve Spectral-Spatial Vector

After the spatial module, a hyperspectral tensor X s s i ( t ) b × b × k is obtained that combines the spectral and spatial features. X s s i ( t ) is then transformed into a vector input into the fully connected layer through pooling, flattening and dropout. The output vector s s v i ( t ) k is used as the spectral-spatial vector of the input tensor.

2.3. Contrastive Loss in the Siamese Network

Consider the i t h tensor group in tensor set ( X i ( 1 ) , X i ( 2 ) ) whose corresponding spectral-spatial vectors are s s v i ( 1 ) and s s v i ( 2 ) (labeled Y i ). To describe the similarity between s s v i ( 1 ) and s s v i ( 2 ) , we calculate the Euclidean distance between them:
d i s t ( s s v i ( 1 ) , s s v i ( 2 ) ) = s s v i ( 1 ) s s v i ( 2 ) 2
We want the d i s t ( s s v i ( 1 ) , s s v i ( 2 ) ) to be as large as possible when Y i = 1 and the d i s t ( s s v i ( 1 ) , s s v i ( 2 ) ) to be as small as possible when Y i = 0 . Therefore, we use contrastive loss [51] as the loss function for model optimization:
L o s s = 1 N i = 1 N [ ( 1 Y i ) d i s t ( s s v i ( 1 ) , s s v i ( 2 ) ) + Y i m a x { m d i s t ( s s v i ( 1 ) , s s v i ( 2 ) ) , 0 } ]
where N is the number of training samples and m is a boundary value that is set artificially. When Y i = 0 , the loss function becomes d i s t ( s s v i ( 1 ) , s s v i ( 2 ) ) . Only s s v i ( 1 ) and s s v i ( 2 ) are close enough to decrease the loss. When Y i = 1 , the loss function becomes m a x { m d i s t ( s s v i ( 1 ) , s s v i ( 2 ) ) , 0 } . Only s s v i ( 1 ) and s s v i ( 2 ) are far enough to decrease the loss. When the d i s t ( s s v i ( 1 ) , s s v i ( 2 ) ) > m , loss is regarded as 0, which limits the influence of the tensor pair with too large a distance on overall loss.

2.4. Proposed SSCNN-S Method

The CD method of SSCNN-S is presented in Figure 2. SSCNN-S trains a Siamese network using training sets ( X t r a i n , Y t r a i n ) . During training, the features of each tensor pair in X t r a i n are extracted through the feature extraction network to obtain the corresponding spectral-spatial vector and the Euclidean distances of the two spectral-spatial vectors are calculated as the output of the network.
To segment the change region, we also need to find a binarization method that converts the output distance to 0 or 1. In SSCNN-S, we use a threshold-based binarization method; that is, to find a threshold θ for each d i s t we use the binarization function:
B i n ( d i s t , θ ) = { 0 d i s t θ 1 d i s t > θ
Notably, the threshold θ can be determined in several ways. In SSCNN-S, we select the best threshold by traversing all possible thresholds in the validation set. This method is chosen because neither the validation set nor the test set participate in model training while the randomness of the validation set and test set generation allows them to be regarded as having approximately the same distribution. After binarizing the distances for each pixel, a CD result map can be obtained.
The detailed algorithm of SSCNN-S is shown in Algorithm 1.
Algorithm 1: Algorithm of SSCNN-S for hyperspectral image (HSI) change detection (CD).
Input:
1. Two HSIs of the same region at different times with ground truthing.
2. The number of training pairs N t and the number of validation pairs N v .
Step 1: Construct the corresponding tensor sets X ( 1 ) and X ( 2 ) for two HSIs and pair them to form a tensor pair X and generate the sample set ( X , Y ) according to the change situation reflected by the ground truthing.
Step 2: Randomly select N t pairs in X as the training set G t r a i n and randomly select N v pairs in X as the validation set G v a l i d a t i o n .
Step 3: Input G t r a i n and G v a l i d a t i o n to the network.
Step 4: Train the model and obtain the optimal parameters.
Step 5: Traverse all possible thresholds in the validation set to select the optimal threshold θ .
Step 6: Calculate the distance for each pixel. If the distance is greater than θ , it is considered as a changed pixel; otherwise, it is considered an unchanged pixel.
Output:
1. Change map.

3. Results

To verify the effect of SSCNN-S on CD, we first introduce three real hyperspectral data sets used for experiments and then provide indexes for evaluating the effects of different algorithms. Finally, we provide the experimental results and corresponding analyses on each data set.

3.1. Data Sets

We selected three real hyperspectral data sets (Figure 3): Farmland [53], River [37] and USA [54]. All three data sets were collected using Earth Observing-1 (EO-1) Hyperion data. The EO-1 hyperspectral imager covers electromagnetic waves with wavelengths ranging from 0.4 to 2.5 micrometers. EO-1 has a spectral resolution of 10 nm and a spatial resolution of about 30 m, with a total of 242 different bands.
The first data set, Farmland, was selected from a farmland area in Yancheng, Jiangsu Province, China. The data set primarily depicted changes in cultivated land. The time points of the two hyperspectral images were 3 May 2006 and 23 April 2007. Their spatial size was 450 × 140 pixels with 155 bands after removing low signal-noise-ratio (SNR) bands. Referring to the number of samples set in [37], we randomly selected 13,200 pixels as the training set of which 4400 were changed pixels and 8800 were unchanged pixels. Additionally, 6600 pixels were randomly selected as the verification set.
The second data set, River, was selected from a river region in Jiangsu Province, China. This data set mainly reflected material changes in the river. The selected time points were 3 May 2013 and 31 December 2013. The spatial size was 463 × 241 pixels with 198 bands. Referring also to the settings in [37], 3750 pixels were randomly selected as the training set of which 1250 were changed pixels and 2500 were unchanged pixels. Additionally, 1875 pixels were randomly selected as the verification set.
The third data set, USA, was from irrigated farmland in Hermiston, Umatilla County, Oregon, USA. It covered soil, irrigation areas, rivers and other terrain. The spatial size of the images was 307 × 241 pixels with 154 bands. Referring to the settings in [54], we randomly selected 7232 pixels as the training set of which 3313 were changed pixels and 3919 were unchanged pixels. A total of 3616 pixels were used for the validation sets.

3.2. Evaluation Index

We evaluated the difference between the CD result of the algorithm and the ground truth value to evaluate the results. For a binary CD problem, we supposed the total number of pixels to be tested was T. There were only one of four possible scenarios for each pixel: correctly classifying the changing pixels, whose number was denoted as TP; incorrectly classifying the changing pixels, whose number was denoted as FP; correctly classifying the changing pixels, whose number was denoted as TN; incorrectly classifying the changing pixels, whose number was denoted as FN. Then we obtained:
T = T P + F P + T N + F N
The first index introduced was overall accuracy (OA), which was calculated by:
O A = T P + T N T
The OA was used to measure the proportion of pixels correctly classified by the algorithm. In this index, TP and TN had the same impact on the OA.
However, in the case of extremely unbalanced data sets, only using the OA as an evaluation index was problematic. Taking the River data set as an example, it contained 111,583 pixels. According to the ground truth, the actual percentage of changed pixels was not greater than 10%. This implied that if a model classified all of the pixels as unchanged, then the OA would exceed 90%; however, this was not the model we required. Therefore, we introduced a second evaluation index, a Kappa coefficient, which was calculated as follows:
K a p p a = O A p e 1 p e
where p e could be calculated by:
p e = ( T P + F P ) ( T P + F N ) T 2 + ( T N + F P ) ( T N + F N ) T 2
Effectively introducing the Kappa coefficient solved the problem of the consistency of model predictions. The relatively small number of changed pixels in the three data sets implied that the Kappa coefficient imposed greater penalties on the FP.

3.3. Experimental Results

All of the experiences in this paper were carried out on a personal computer that was equipped with an Intel Core i7-9700K Central Processing Unit (CPU) and an independent Graphics Processing Unit (GPU) of NVIDIA GeForce RTX 2080. We set the batch size to 128. We chose the Adam optimizer and its parameters were the default parameters. The boundary value m in contrastive loss was set to 1.5. Due to the randomness of SSCNN-S, we ran it four times on each data set to calculate the mean and standard deviation of each index as the final result. To analyze the effect of the spatial module in CD, the method of extracting spectral features X s i ( t ) obtained after dimension reduction for CD was also included in the experiment, which was the spectral convolution neural network with Siamese architecture (SCNN-S).
To evaluate the effectiveness of SSCNN-S, we selected several existing algorithms for comparison: CVA [14], PCACVA [25], a support vector machine (SVM) [19,55], patch-based CNN (PBCNN) [56], Hybrid Spectral CNN (HybridSN) [41] and GETNET [37], which used different remote sensing image CD strategies and techniques. Figure 4, Figure 5 and Figure 6 intuitively show the segmentation results of the different algorithms on the three sets of data sets. Table 1 quantitatively compares the effects of various algorithms using evaluation indexes. In order to evaluate the time complexity of each algorithm, in Table 2, we provide the running time of different algorithms on the three data sets for reference.

4. Discussion

4.1. Selecting Parameters

4.1.1. Parameters in the Spectral Module

In the spectral module of SSCNN-S, the spatial dimension b of X i ( t ) and the dimension b of s s v i ( t ) are two important parameters. Moreover, b and k should also be selected differently for different data sets. We used SCNN-S to test different parameter combinations ( b , k ) on each data sets and observed the Kappa coefficient of CD results to determine which parameter combinations to use. In the parameter experiment, the value range of b was { 3 , 5 , 7 , 9 } and the value range of k was { k | k = 15 + 5 p , p N } . Due to SCNN-S also being random, in order to find a Kappa value to represent the CD effect of parameter combination more accurately and quickly, we proposed a four-test scoring method that was analogous to the subjective scoring system of the Chinese College Entrance Examination. For each parameter combination, the four-test scoring method provided a comprehensive Kappa value for that parameter combination after two to four experiments.
Algorithm 2: Four-test scoring method.
Input:
1. Error threshold ϑ 0 .
Step 1: Perform the first experiment and obtain the result R 1 .
Step 2: Perform the second experiment and obtain the result R 2 .
Step 3: If | R 1 R 2 | ϑ , let the final result be F R = R 1 + R 2 2 . The algorithm is aborted.
Step 4: Otherwise, perform the third experiment and obtain the result R 3 .
Step 5: If | R 1 R 3 | ϑ , let the final result be F R = R 1 + R 3 2 . The algorithm is aborted.
Step 6: Otherwise, if | R 2 R 3 | ϑ , let the final result be F R = R 2 + R 3 2 . The algorithm is aborted.
Step 7: Otherwise, perform the final experiment and obtain the result R 4 .
Step 8: Let the final result be F R = R 1 + R 2 + R 3 + R 4 4 .
Output:
1. Final result F R .
In the parameter experiment of this paper, we set ϑ = 0.001 .
The results of the Farmland parameter selection experiment using SCNN-S are shown in Figure 7. Although the experimental results of different combinations were somewhat different, they were generally close. Based on the experimental results, we chose a combination of b = 7 , k = 75 for our experiments on the Farmland data set.
The results of the River parameter selection experiment using SCNN-S are shown in Figure 8. Based on the experimental results, we chose a combination of b = 3 , k = 85 for our experiments on the River data set.
The results of the USA parameter selection experiment using SCNN-S are shown in Figure 9. Based on the experimental results, we chose a combination of b = 5 , k = 80 for our experiments on the USA data set.

4.1.2. Parameters in the Spatial Module

In the spatial module, the kernel size of 2D convolution was a parameter that had a greater impact on CD. We conducted the parameter selection experiment of the spatial module on the basis of the results of the parameter selection of the spectral module. The range of the kernel size value n was { 3 , 5 , 7 , 9 , 11 } . As per the method in Section 4.1.1, we used the four-test scoring method to measure the results of the different parameters.
The experiment results of the parameter selection for selecting the kernel size n of the 2D convolution kernel on the Farmland data set are shown in Figure 10. Based on the experimental results, we chose n = 9 for our experiments on the Farmland data set.
The experiment results of the parameter selection for selecting the kernel size n of the 2D convolution kernel on the River data set are shown in Figure 11. Based on the experimental results, we chose n = 7 for our experiments on the River data set.
The experiment results of the parameter selection for selecting the kernel size n of the 2D convolution kernel on the USA data set are shown in Figure 12. Based on the experimental results, we chose n = 9 for our experiments on the USA data set.

4.2. Discussion on Farmland Experiment

The Farmland data set was a more standardized data set. The ratio of changed pixels to unchanged pixels was balanced and the changed areas were relatively concentrated and regular in shape. Moreover, there was not a large number of scatter areas. The spectral information was rich and the influence of noise was small. Based on the experiment result, all methods involved in the comparison achieved good results and the main change areas were detected. The CVA, PCACVA, SVM and CNN were all more susceptible to noise, which was evident from the CD results. The HybridSN, through the extracting of spatial features, coped well with situations where different types of changes had similar spectral features. The GETNET incorporated more sub-pixel information than a CNN. SCNN-S and SSCNN-S extracted spectral features through a deep network that greatly reduced the impact of noise and achieved excellent CD results.

4.3. Discussion on River Experiment

The River data set was very challenging. Based on the ground truth map, it was evident that the River data set had a wide variety of variations, irregular areas of variation and a large number of scatters. Additionally, the ratio of changing pixels to total pixels in the River data set was less than 10%. This extremely unbalanced data set made classifying the changed into the unchanged greatly affect the Kappa coefficients. Based on the experiment results, SSCNN-S achieved the best accuracy and PCACVA achieved the best Kappa coefficient, which was similar to the results of [26]. This result demonstrated that SSCNN-S was more stringent in judging changed pixels while spatial information supplementation effectively improved the accuracy and Kappa value compared with SCNN-S. Overall, the Kappa coefficients for all methods were low, which demonstrated the high complexity of this data set.

4.4. Discussion on USA Experiment

The USA data set was a more comprehensive data set that contained a lot of circular change areas and a few curve and scatter areas. Based on the experimental results, the area of curve change representing the river boundary was a watershed that affected the detection results of the different methods. Of all of the methods involved in the comparison, only SCNN-S and SSCNN-S depicted changes in the river boundary very well, indicating that the spectral module could extract the spectral features of the tensor more effectively. Compared with SCNN-S, SSCNN-S significantly improved the OA and Kappa coefficients, indicating that the auxiliary features of the spatial module made up for the noise interference of the spectral features.

5. Conclusions

In this paper, we proposed a spectral-spatial convolution neural network with Siamese architecture (SSCNN-S) for CD. SSCNN-S extracted the spectral-spatial features of hyperspectral tensors through spectral and spatial modules and converted the hyperspectral tensors into spectral-spatial vectors. SSCNN-S introduced contrastive loss into CD and transformed the CD problem into a similarity measurement problem of spectral-spatial vectors. The distance function was used to calculate the distance between two spectral-spatial vectors to describe the similarity of two tensors and this was used as the basis of CD. Considering the high spectral resolution and low spatial resolution of the HSIs, SSCNN-S used a spectral-spatial combination method to extract features of the hyperspectral tensor, which mainly relied on the spectral features and supplemented the spatial features. Based on the full extraction of spectral features, the auxiliary spatial features weakened the influence of noise on the spectral features. The introduction of the Siamese network architecture into CD bridged a gap between hyperspectral classification and CD. A few advanced hyperspectral classification techniques could be applied to CD based on this architecture. From the experience, SCNN-S and SSCNN-S achieved good results on three real data sets.
In our future work, we will continue to explore the following two aspects. The first aspect is the HSI. In addition to the real HSI, the use of synthetic data may bring further discoveries. The second aspect is the further research on the spatial module. We tested the effect of different sizes of 2D convolution kernels on the results in our experiments. In fact, the spatial module can have more architectures to extract the spatial features of HSIs as fully as possible.

Author Contributions

Conceptualization, T.Z., B.S., Y.X. and Z.W.; methodology, M.W. and X.W.; software, B.S. and G.Y.; validation, T.Z., B.S. and Y.X.; formal analysis, M.W.; investigation, M.W.; resources, G.Y.; data curation, B.S.; writing—original draft preparation, T.Z. and B.S.; writing—review and editing, B.S.; visualization, B.S.; supervision, T.Z.; project administration, T.Z. and Z.W.; funding acquisition, T.Z. and Z.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under Grant 61976117, 62071233, 61876213, 61772274 and 61772277 in part by the Natural Science Foundation of Jiangsu Province under Grant BK20191409, BK20201397, BK20180018 and BK20171494, in part by the Key Projects of University Natural Science Fund of Jiangsu Province under Grant 19KJA360001 and 18KJA520005, in part of the National Key R&D Program under Grant 2017YFC0804002, in part by the Fundamental Research Funds for the Central Universities under Grant 30917015104, 30919011103 and 30919011402, in part by the Collaborative Innovation Center of Audit Information Engineering and Technology under Grant 18CICA09, in part by the Young Teacher Research and Cultivation Project of Nanjing Audit University under Grant 18QNPY015 and in part by the Postgraduate Research and Practice Innovation Program of Jiangsu Province under Grant KYCX20_1680.

Acknowledgments

We are grateful to Wang Qi and Hasanlou Mahdi who provided the data for this research.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses or interpretation of data; in the writing of the manuscript or in the decision to publish the results.

References

  1. Li, Y.; Hu, C.; Ao, D. Rapid surface large-change monitoring by repeat-pass GEO SAR multibaseline interferometry. IEEE Geosci. Remote Sens. Lett. 2020, 1–5. [Google Scholar] [CrossRef]
  2. Yasir, M.; Sheng, H.; Fan, H.; Nazir, S.; Niang, A.J.; Salauddin, M.; Khan, S. Automatic coastline extraction and changes analysis using remote sensing and GIS technology. IEEE Access 2020, 8, 180156–180170. [Google Scholar] [CrossRef]
  3. Ansari, R.A.; Buddhiraju, K.M.; Malhotra, R. Urban change detection analysis utilizing multiresolution texture features from polarimetric SAR images. Remote Sens. Appl. Soc. Environ. 2020, 20, 100418. [Google Scholar] [CrossRef]
  4. Liu, S.; Marinelli, D.; Bruzzone, L.; Bovolo, F. A review of change detection in multitemporal hyperspectral images: Current techniques, applications, and challenges. IEEE Geosci. Remote Sens. Mag. 2019, 7, 140–158. [Google Scholar] [CrossRef]
  5. Padŕon-Hidalgo, J.A.; Pérez-Suay, A.; Nar, F.; Laparra, V.; Camps-Valls, G. Efficient kernel Cook’s Distance for remote sensing anomalous change detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 5480–5488. [Google Scholar] [CrossRef]
  6. Wu, C.; Zhang, L.; Du, B. Hyperspectral anomaly change detection with slow feature analysis. Neurocomputing 2015, 151, 175–187. [Google Scholar] [CrossRef]
  7. Hou, Y.; Zhu, W.; Wang, E. Hyperspectral Mineral Target Detection Based on Density Peak. Intell. Autom. Soft Comput. 2019, 25, 805–814. [Google Scholar] [CrossRef]
  8. Dalmiya, C.P.; Santhi, N.; Sathyabama, B. A novel feature descriptor for automatic change detection in remote sensing images. Egypt. J. Remote Sens. Space Sci. 2019, 22, 183–192. [Google Scholar] [CrossRef]
  9. Chen, Z.; Wang, B. Spectrally-spatially regularized low-rank and sparse decomposition: A novel method for change detection in multitemporal hyperspectral images. Remote Sens. 2017, 9, 1044. [Google Scholar] [CrossRef] [Green Version]
  10. Zhao, J.; Yang, J.; Lu, Z.; Li, P.; Liu, W. Change detection based on similarity measure and joint classification for polarimetric SAR images. In Proceedings of the 2017 IEEE International Geoscience Remote Sensing Symposium (IGARSS), Fort Worth, TX, USA, 23–28 July 2017; pp. 1896–1899. [Google Scholar]
  11. Wu, C.; Zhang, L.; Zhang, L. A scene change detection framework for multi-temporal very high resolution remote sensing images. Signal. Process. 2016, 124, 184–197. [Google Scholar] [CrossRef]
  12. Lu, L.; Weng, Q.; Guo, H.; Feng, S.; Li, Q. Assessment of urban environmental change using multi-source remote sensing time series (2000–2016): A comparative analysis in selected megacities in Eurasia. Sci. Total Environ. 2019, 684, 567–577. [Google Scholar] [CrossRef]
  13. Henrot, S.; Chanussot, J.; Jutten, C. Dynamical spectral unmixing of multitemporal hyperspectral images. IEEE Trans. Image Process. 2016, 25, 3219–3232. [Google Scholar] [CrossRef] [Green Version]
  14. MALILA, W. Change vector analysis: An approach for detecting forest changes with Landsat. In Proceedings of the Machine Processing of Remotely Sensed Data Symposium, West Lafayette, IN, USA, 29 June–1 July 1980; pp. 326–335. [Google Scholar]
  15. Nielsen, A.A.; Conradsen, K.; Simpson, J.J. Multivariate alteration detection (MAD) and MAF postprocessing in multispectral, bitemporal image data: New approaches to change detection studies. Remote Sens. Environ. 1998, 64, 1–19. [Google Scholar] [CrossRef] [Green Version]
  16. Nielsen, A.A. The regularized iteratively reweighted MAD method for change detection in multi-and hyperspectral data. IEEE Trans. Image Process. 2007, 16, 463–478. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Hardoon, D.R.; Szedmak, S.; Shawe-Taylor, J. Canonical correlation analysis: An overview with application to learning methods. Neural Comput. 2004, 16, 2639–2664. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Bovolo, F.; Marchesi, S.; Bruzzone, L. A framework for automatic and unsupervised detection of multiple changes in multitemporal images. IEEE Trans. Geosci. Remote Sens. 2011, 50, 2196–2212. [Google Scholar] [CrossRef]
  19. Nemmour, H.; Chibani, Y. Multiple support vector machines for land cover change detection: An application for mapping urban extensions. ISPRS J. Photogramm. Remote Sens. 2006, 61, 125–133. [Google Scholar] [CrossRef]
  20. Bruzzone, L.; Liu, S.; Bovolo, F.; Du, P. Change detection in multitemporal hyperspectral images. In Multitemporal Remote Sensing; Springer: Cham, Swizerland, 2016; pp. 63–88. [Google Scholar]
  21. Liu, S.; Du, Q.; Tong, X.; Samat, A.; Pan, H.; Ma, X. Band selection-based dimensionality reduction for change detection in multi-temporal hyperspectral images. Remote Sens. 2017, 9, 1008. [Google Scholar] [CrossRef] [Green Version]
  22. Ye, Q.; Yang, J.; Liu, F.; Zhao, C.; Ye, N.; Yin, T. L1-Norm Distance Linear Discriminant Analysis Based on an Effective Iterative Algorithm. IEEE Trans. Circuits Syst. Video Technol. 2018, 28, 114–129. [Google Scholar] [CrossRef]
  23. Fu, L.; Li, Z.; Ye, Q.; Yin, H.; Liu, Q.; Chen, X.; Fan, X.; Yang, W.; Yang, G. Learning Robust Discriminant Subspace Based on Joint L2,p- and L2,s-Norm Distance Metrics. IEEE Trans. Neural Netw. Learn. Syst. 2020, 1–15. [Google Scholar] [CrossRef] [PubMed]
  24. Ye, Q.; Li, Z.; Fu, L.; Yang, W.; Yang, G. Nonpeaked Discriminant Analysis for Data Representation. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3818–3832. [Google Scholar] [CrossRef]
  25. Zhang, H.; Gong, M.; Zhang, P.; Su, L.; Shi, J. Feature-level change detection using deep representation and feature change analysis for multispectral imagery. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1666–1670. [Google Scholar] [CrossRef]
  26. Ma, L.; Jia, Z.; Yu, Y.; Yang, J.; Kasabov, N.K. Multi-spectral image change detection based on band selection and single-band iterative weighting. IEEE Access 2019, 7, 27948–27956. [Google Scholar] [CrossRef]
  27. Yu, W.; Zhang, M.; Shen, Y. Learning a local manifold representation based on improved neighborhood rough set and LLE for hyperspectral dimensionality reduction. Signal. Process. 2019, 164, 20–29. [Google Scholar] [CrossRef]
  28. Xu, F.; Zhang, X.; Xin, Z.; Yang, A. Investigation on the Chinese text sentiment analysis based on ConVolutional neural networks in deep learning. Comput. Mater. Con. 2019, 58, 697–709. [Google Scholar] [CrossRef] [Green Version]
  29. Guo, Y.; Li, C.; Liu, Q. R2N: A novel deep learning architecture for rain removal from single image. Comput. Mater. Con. 2019, 58, 829–843. [Google Scholar] [CrossRef] [Green Version]
  30. Wu, H.; Liu, Q.; Liu, X. A review on deep learning approaches to image classification and object segmentation. Comput. Mater. Con. 2019, 60, 575–597. [Google Scholar] [CrossRef] [Green Version]
  31. Zhang, X.; Lu, W.; Li, F.; Peng, X.; Zhang, R. Deep feature fusion model for sentence semantic matching. Comput. Mater. Con. 2019, 61, 601–616. [Google Scholar] [CrossRef]
  32. Mohanapriya, N.; Kalaavathi, B. Adaptive image enhancement using hybrid particle swarm optimization and watershed segmentation. Intell. Autom. Soft Comput. 2019, 25, 663–672. [Google Scholar] [CrossRef]
  33. Hung, C.; Mao, W.; Huang, H. Modified PSO algorithm on recurrent fuzzy neural network for system identification. Intell. Autom. Soft Comput. 2019, 25, 329–341. [Google Scholar]
  34. Chen, L.; Wei, Z.; Xu, Y. A Lightweight Spectral–Spatial Feature Extraction and Fusion Network for Hyperspectral Image Classification. Remote Sens. 2020, 12, 1395. [Google Scholar] [CrossRef]
  35. Lv, N.; Chen, C.; Qiu, T.; Sangaiah, A.K. Deep learning and superpixel feature extraction based on contractive autoencoder for change detection in SAR images. IEEE Trans. Ind. Inform. 2018, 14, 5530–5538. [Google Scholar] [CrossRef]
  36. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Exploiting spatial information in semi-supervised hyperspectral image segmentation. In Proceedings of the 2010 2nd Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing, Reykjavik, Iceland, 14–16 June 2010; pp. 1–4. [Google Scholar]
  37. Wang, Q.; Yuan, Z.; Du, Q.; Li, X. GETNET: A general end-to-end 2-D CNN framework for hyperspectral image change detection. IEEE Trans. Geosci. Remote Sens. 2018, 57, 3–13. [Google Scholar] [CrossRef] [Green Version]
  38. Wang, W.; Dou, S.; Jiang, Z.; Sun, L. A fast dense spectral–spatial convolution network framework for hyperspectral images classification. Remote Sens. 2018, 10, 1068. [Google Scholar] [CrossRef] [Green Version]
  39. Huang, F.; Yu, Y.; Feng, T. Hyperspectral remote sensing image change detection based on tensor and deep learning. J. Vis. Commun. Image Represent. 2019, 58, 233–244. [Google Scholar] [CrossRef]
  40. Ran, Q.; Zhao, S.; Li, W. Change Detection Combining Spatial-spectral Features and Sparse Representation Classifier. In Proceedings of the 2018 Fifth International Workshop on Earth Observation Remote Sensing Applications (EORSA), Xi’an, China, 18–20 June 2018; pp. 1–4. [Google Scholar]
  41. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN feature hierarchy for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2020, 17, 277–281. [Google Scholar] [CrossRef] [Green Version]
  42. Heidari, M.; Fouladi-Ghaleh, K. Using Siamese Networks with Transfer Learning for Face Recognition on Small-Samples Datasets. In Proceedings of the 2020 International Conference on Machine Vision and Image Processing (MVIP), Tehran, Iran, 18–20 February 2020; pp. 1–4. [Google Scholar]
  43. Cui, W.; Zhan, W.; Yu, J.; Sun, C.; Zhang, Y. Face Recognition via Convolutional Neural Networks and Siamese Neural Networks. In Proceedings of the 2019 International Conference on Intelligent Computing, Automation and Systems (ICICAS), Chongqing, China, 6–8 December 2019; pp. 746–750. [Google Scholar]
  44. Zhou, S.; Ke, M.; Luo, P. Multi-camera transfer GAN for person re-identification. J. Vis. Commun. Image Represent. 2019, 59, 393–400. [Google Scholar] [CrossRef]
  45. Zhou, S.; Chen, L.; Sugumaran, V. Hidden Two-Stream Collaborative Learning Network for Action Recognition. CMC-Comput. Mater. Contin. 2020, 63, 1545–1561. [Google Scholar] [CrossRef]
  46. Li, W.; Li, X.; Bourahla, O.E.; Huang, F.; Wu, F.; Liu, W.; Liu, H. Progressive Multi-Stage Learning for Discriminative Tracking. IEEE Trans. Cybern. 2020. Available online: https://arxiv.org/abs/2004.00255 (accessed on 1 April 2020).
  47. Sun, L.; Ma, C.; Chen, Y.; Zheng, Y.; Shim, H.J.; Wu, Z.; Jeon, B. Low rank component induced spatial-spectral kernel method for hyperspectral image classification. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 3829–3842. [Google Scholar] [CrossRef]
  48. Wei, W.; Yongbin, J.; Yanhong, L.; Ji, L.; Xin, W.; Tong, Z. An advanced deep residual dense network (DRDN) approach for image super-resolution. Int. J. Comput. Intell. Syst. 2019, 12, 1592–1601. [Google Scholar] [CrossRef] [Green Version]
  49. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  50. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 11–18 December 2015; pp. 1026–1034. [Google Scholar]
  51. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  52. Hadsell, R.; Chopra, S.; LeCun, Y. Dimensionality reduction by learning an invariant mapping. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17–22 June 2006; pp. 1735–1742. [Google Scholar]
  53. Yuan, Y.; Lv, H.; Lu, X. Semi-supervised change detection method for multi-temporal hyperspectral images. Neurocomputing 2015, 148, 363–375. [Google Scholar] [CrossRef]
  54. Hasanlou, M.; Seydi, S.T. Hyperspectral change detection: An experimental comparative study. Int. J. Remote Sens. 2018, 39, 7029–7083. [Google Scholar] [CrossRef]
  55. Ye, Q.; Zhao, H.; Li, Z.; Yang, X.; Gao, S.; Yin, T.; Ye, N. L1-norm Distance Minimization Based Fast Robust Twin Support Vector k-plane Clustering. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 4494–4503. [Google Scholar] [CrossRef] [PubMed]
  56. Sharma, A.; Liu, X.; Yang, X.; Shi, D. A patch-based convolutional neural network for remote sensing image classification. Neural Netw. 2017, 95, 19–28. [Google Scholar] [CrossRef]
Figure 1. Network for extracting spectral-spatial features of hyperspectral tensors.
Figure 1. Network for extracting spectral-spatial features of hyperspectral tensors.
Remotesensing 13 00895 g001
Figure 2. Overview of the spectral-spatial convolution neural network with Siamese architecture (SSCNN-S).
Figure 2. Overview of the spectral-spatial convolution neural network with Siamese architecture (SSCNN-S).
Remotesensing 13 00895 g002
Figure 3. Experimental data sets: (a) Farmland data set on 3 May 2006. (b) Farmland data set on 21 April 2006. (c) River data set on 3 May 2013. (d) River data set on 31 December 2013. (e) USA data set on 1 May 2004. (f) USA data set on 8 May 2007.
Figure 3. Experimental data sets: (a) Farmland data set on 3 May 2006. (b) Farmland data set on 21 April 2006. (c) River data set on 3 May 2013. (d) River data set on 31 December 2013. (e) USA data set on 1 May 2004. (f) USA data set on 8 May 2007.
Remotesensing 13 00895 g003
Figure 4. Change detection (CD) result on Farmland. (a) Change vector analysis (CVA) [14] (b) Principal component analysis CVA (PCACVA) [25] (c) Support vector machine (SVM) [19] (d) Patch-based convolution neural network (PBCNN) [56] (e) GETNET [37] (f) Hybrid spectral CNN (HybridSN) [41] (g) SCNN-S (h) SSCNN-S (i) Ground truth.
Figure 4. Change detection (CD) result on Farmland. (a) Change vector analysis (CVA) [14] (b) Principal component analysis CVA (PCACVA) [25] (c) Support vector machine (SVM) [19] (d) Patch-based convolution neural network (PBCNN) [56] (e) GETNET [37] (f) Hybrid spectral CNN (HybridSN) [41] (g) SCNN-S (h) SSCNN-S (i) Ground truth.
Remotesensing 13 00895 g004aRemotesensing 13 00895 g004b
Figure 5. CD result on River. (a) CVA [14] (b) PCACVA [25] (c) SVM [19] (d) PBCNN [56] (e) GETNET [37] (f) HybridSN [41] (g) SCNN-S (h) SSCNN-S (i) Ground truth.
Figure 5. CD result on River. (a) CVA [14] (b) PCACVA [25] (c) SVM [19] (d) PBCNN [56] (e) GETNET [37] (f) HybridSN [41] (g) SCNN-S (h) SSCNN-S (i) Ground truth.
Remotesensing 13 00895 g005aRemotesensing 13 00895 g005b
Figure 6. CD result on USA. (a) CVA [14] (b) PCACVA [25] (c) SVM [19] (d) PBCNN [56] (e) GETNET [37] (f) HybridSN [41] (g) SCNN-S (h) SSCNN-S (i) Ground truth.
Figure 6. CD result on USA. (a) CVA [14] (b) PCACVA [25] (c) SVM [19] (d) PBCNN [56] (e) GETNET [37] (f) HybridSN [41] (g) SCNN-S (h) SSCNN-S (i) Ground truth.
Remotesensing 13 00895 g006
Figure 7. Parameter selection experiment for selecting b and k on the Farmland data set.
Figure 7. Parameter selection experiment for selecting b and k on the Farmland data set.
Remotesensing 13 00895 g007
Figure 8. Parameter selection experiment for selecting b and k on the River data set.
Figure 8. Parameter selection experiment for selecting b and k on the River data set.
Remotesensing 13 00895 g008
Figure 9. Parameter selection experiment for selecting b and k on the USA data set.
Figure 9. Parameter selection experiment for selecting b and k on the USA data set.
Remotesensing 13 00895 g009
Figure 10. Parameter selection experiment for selecting n on the Farmland data set.
Figure 10. Parameter selection experiment for selecting n on the Farmland data set.
Remotesensing 13 00895 g010
Figure 11. Parameter selection experiment for selecting n on the River data set.
Figure 11. Parameter selection experiment for selecting n on the River data set.
Remotesensing 13 00895 g011
Figure 12. Parameter selection experiment for selecting n on the USA data set.
Figure 12. Parameter selection experiment for selecting n on the USA data set.
Remotesensing 13 00895 g012
Table 1. Experimental results of different algorithms on three data sets (the optimal results are highlighted in bold).
Table 1. Experimental results of different algorithms on three data sets (the optimal results are highlighted in bold).
MethodIndexExperiment Data Sets
FarmlandRiverUSA
CVA [14]OA0.95230.92670.9200
Kappa0.88550.65750.7410
PCACVA [25]OA0.96680.95160.9153
Kappa0.92020.74770.7225
SVM [19]OA0.93760.94240.8810
Kappa0.84830.70660.6848
PBCNN [56]OA0.91850.91390.8902
Kappa0.79490.55850.6699
GETNET [37]OA0.9753 ± 0.00030.9499 ± 0.00540.9430 ± 0.0010
Kappa0.9394 ± 0.00080.7472 ± 0.02150.8249 ± 0.0030
HybridSN [41]OA0.9749 ± 0.00020.9614 ± 0.00190.9553 ± 0.0004
Kappa0.9392 ± 0.00040.7371 ± 0.01000.8701 ± 0.0015
SCNN-SOA0.9777 ± 0.00040.9610 ± 0.00190.9631 ± 0.0020
Kappa0.9445 ± 0.00120.7300 ± 0.00690.8848 ± 0.0059
SSCNN-SOA0.9774 ± 0.00030.9640 ± 0.00140.9651 ± 0.0010
Kappa0.9440 ± 0.00040.7431 ± 0.00340.8918 ± 0.0022
Table 2. Reference running time (in seconds) of different algorithms on three data sets.
Table 2. Reference running time (in seconds) of different algorithms on three data sets.
MethodExperiment Data Sets
FarmlandRiverUSA
CVA [14]296234
PCACVA [25]356738
SVM [19]35986207
PBCNN [56]9319551061
GETNET [37]515704568
HybridSN [41]403311297
SCNN-S615288432
SSCNN-S925328579
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhan, T.; Song, B.; Xu, Y.; Wan, M.; Wang, X.; Yang, G.; Wu, Z. SSCNN-S: A Spectral-Spatial Convolution Neural Network with Siamese Architecture for Change Detection. Remote Sens. 2021, 13, 895. https://doi.org/10.3390/rs13050895

AMA Style

Zhan T, Song B, Xu Y, Wan M, Wang X, Yang G, Wu Z. SSCNN-S: A Spectral-Spatial Convolution Neural Network with Siamese Architecture for Change Detection. Remote Sensing. 2021; 13(5):895. https://doi.org/10.3390/rs13050895

Chicago/Turabian Style

Zhan, Tianming, Bo Song, Yang Xu, Minghua Wan, Xin Wang, Guowei Yang, and Zebin Wu. 2021. "SSCNN-S: A Spectral-Spatial Convolution Neural Network with Siamese Architecture for Change Detection" Remote Sensing 13, no. 5: 895. https://doi.org/10.3390/rs13050895

APA Style

Zhan, T., Song, B., Xu, Y., Wan, M., Wang, X., Yang, G., & Wu, Z. (2021). SSCNN-S: A Spectral-Spatial Convolution Neural Network with Siamese Architecture for Change Detection. Remote Sensing, 13(5), 895. https://doi.org/10.3390/rs13050895

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop