Next Article in Journal
BEST—Blockchain-Enabled Secure and Trusted Public Emergency Services for Smart Cities Environment
Previous Article in Journal
A New Design of a CMOS Vertical Hall Sensor with a Low Offset
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dual-Coupled CNN-GCN-Based Classification for Hyperspectral and LiDAR Data

1
School of Computer Science, Shaanxi Normal University, Xi’an 710062, China
2
School of Mathematics and Computer Application, Shangluo University, Shangluo 726000, China
3
Engineering Research Center of Qinling Health Welfare Big Data, Universities of Shaanxi Province, Shangluo 726000, China
*
Authors to whom correspondence should be addressed.
Sensors 2022, 22(15), 5735; https://doi.org/10.3390/s22155735
Submission received: 21 June 2022 / Revised: 23 July 2022 / Accepted: 28 July 2022 / Published: 31 July 2022
(This article belongs to the Section Remote Sensors)

Abstract

:
Deep learning techniques have brought substantial performance gains to remote sensing image classification. Among them, convolutional neural networks (CNN) can extract rich spatial and spectral features from hyperspectral images in a short-range region, whereas graph convolutional networks (GCN) can model middle- and long-range spatial relations (or structural features) between samples on their graph structure. These different features make it possible to classify remote sensing images finely. In addition, hyperspectral images and light detection and ranging (LiDAR) images can provide spatial-spectral information and elevation information of targets on the Earth’s surface, respectively. These multi-source remote sensing data can further improve classification accuracy in complex scenes. This paper proposes a classification method for HS and LiDAR data based on a dual-coupled CNN-GCN structure. The model can be divided into a coupled CNN and a coupled GCN. The former employs a weight-sharing mechanism to structurally fuse and simplify the dual CNN models and extracting the spatial features from HS and LiDAR data. The latter first concatenates the HS and LiDAR data to construct a uniform graph structure. Then, the dual GCN models perform structural fusion by sharing the graph structures and weight matrices of some layers to extract their structural information, respectively. Finally, the final hybrid features are fed into a standard classifier for the pixel-level classification task under a unified feature fusion module. Extensive experiments on two real-world hyperspectral and LiDAR data demonstrate the effectiveness and superiority of the proposed method compared to other state-of-the-art baseline methods, such as two-branch CNN and context CNN. In particular, the overall accuracy (99.11%) on Trento achieves the best classification performance reported so far.

1. Introduction

In recent years, the rapid development of multi-sensor technologies has opened up the possibility of observing interest regions on the Earth’s surface from different perspectives. These technologies include mainly synthetic aperture radar (SAR) [1], light detection and ranging (LiDAR) [2], hyperspectral (HS) imagery [3] and multispectral (MS) imagery [4], which all contain several different types of information about the target area. Among them, HS, which is a passive remote sensing technique, samples the reflective part of the electromagnetic spectrum through hyperspectral sensors, resulting in rich and continuous spectral information. This spectral information can range from the visible region (0.4–0.7 μm) to the shortwave infrared region (nearly 2.4 μm) over hundreds of narrow, contiguous spectral channels (typically 10 nm wide). This detailed spectral information makes HS image a valuable data source for remote sensing image classification in complex scenes [5]. However, it is often difficult to obtain satisfactory and fine classification results from single-modal data [6]. For example, there are often many misclassifications in classification tasks of objects consisting of similar materials (e.g., grass, shrubs and trees). Therefore, there is a need to further refine or improve the classification results by combining information from other different sensors. LiDAR, as active remote sensing technology, uses lasers to detect and measure the target area, and obtains exact height and shape information about the scene. Therefore, the comprehensive utilization of HS and LiDAR data has become an active topic in the remote sensing community. Many studies have shown that LiDAR data can be used as an excellent complement to HS data to enhance or improve scene classification performance effectively [7,8]. Due to its significant advantages, multimodal data have also been applied in many other fields, such as crop cover [9] and forest fire management [10].
The fusion methods for multimodal data can currently be divided into two main categories: feature-level fusion and decision-level fusion [11,12]. The former first pre-processes and extracts features from different sources of remote sensing data, then fuses them through feature superposition or reconstruction, and finally sends them to a classifier to complete scene classification. The latter first employs multiple classifiers to classify remote sensing data from multiple sources independently, and finally fuses or integrates these multiple classification results (or decisions) to produce the final classification result. For example, Pedergnana et al. [13] proposed a technique for classifying features extracted from extended attribute profiles (EAPs) on optical and LiDAR images, thus enabling the fusion of spectral, spatial and elevation data. Rasti et al. [14] presented the fusion of the datasets using orthogonal total variance component analysis (OTVCA). It first automatically extracts spatial and elevation information from HS and rasterized LiDAR features, then uses the OTVCA feature fusion method to fuse the extracted spatial and elevation information with the spectral information to generate the final classification map. Rasti et al. [8] put forward a new sparse low-rank based multi-source feature fusion method. It first extracts spatial and elevation information from hyperspectral and LiDAR data using extinction profiles, respectively. Then, low-rank fusion features are estimated from the extracted features using sparse low-rank techniques, which are finally used to generate the final classification map. Sturari et al. [15] introduced a decision-level fusion method using elevation data and multispectral high-resolution imagery, thus enabling the joint use of multispectral, spatial and elevation data. However, most of these studies belong to shallow learning methods, which make it difficult to obtain more refined classification results in complex scenes.
In the last decades, deep learning techniques, represented by convolutional neural networks (CNN), have been able to model high-level feature representations through a hierarchical learning framework and have gradually replaced hand design-based approaches. It has been impressively successful in various fields of computer vision, such as image recognition [16], object detection [17] and natural language processing [18]. In particular, CNN can stack a series of specially designed layers (e.g., the convolutional layers, the batch normalization layer) to form deep models and automatically mine robust and high-level spatial features from a large number of samples, which are typically invariant to most local variations in the data. These deep learning techniques also significantly affect the field of scene classification for multimodal remote sensing data. For example, Xu et al. [19] adopted a two-branch convolutional neural network for extracting features from HS and LiDAR data, respectively, and fed the cascaded features into a standard classifier for pixel-level classification. Li et al. [20] proposed to use deep CNN to learn the features of a pixel pair composed of a central pixel and its surrounding pixels, and then determine the final label based on a voting strategy. Lee et al. [21] developed local contextual interactions by jointly exploiting the local spectral–spatial relationships of adjacent single-pixel vectors, and achieved significant results. In addition, Hong et al. [22] proposed a deep model (EndNet) based on an encoder-decoder structure to extract features from HS and LiDAR data in an unsupervised training manner, and achieved better results. However, deep learning-based methods usually require a large number of labelled samples, which is challenging to meet for remote sensing data.
In recent years, graph convolutional networks (GCN) have developed rapidly and have excelled in the field of unstructured data processing, e.g., social networks [23]. Based on the graph structure, GCN can aggregate and propagate information between any two nodes in a semi-supervised manner, and thus extracting structural features of the data in the middle- and long-range region. This is entirely different from CNN, which extracts neighborhood spatial features in the short-range region [24]. There have been several studies using GCN for the scene classification of HS data. For example, Qin et al. [25] developed a new method of constructing graphs based on combining spatial and spectral features, and improved the ability to classify remote sensing images. Wang et al. [26] designed an architecture to encode the multiple spectral contextual information in a spectral pyramid of multiple embedding spaces with a graph attention mechanism. Wan et al. [27,28] proposed a multi-scale graph convolution operation in the super-pixel technology, which allows for the adjacency matrix to be updated dynamically with network iteration. Besides, miniGCN, proposed by Hong et al. [24], adopted mini-batch learning to train GCN with a fixed scale, combining with local convolutional features for pixel-level classification, and obtaining satisfactory classification results. However, there is currently very little research on the direct use of GCN for multimodal data classification in the remote sensing community.
In summary, the nature of the data captured by different sensors is entirely different, which raises a severe challenge to the effective processing of multi-source remote sensing data [29]. However, it is still a problematic issue to design an effective deep classification model for multimodal data. In this paper, a hyperspectral and LiDAR data classification method based on a dual-coupled CNN-GCN structure (DCCG) is proposed. The model employs two sets of networks with similar structures for structure-level fusion, which can be divided into two parts: a coupled CNN and a coupled GCN. The former utilizes a weight-sharing mechanism to structurally fuse and simplify the dual CNN model and to extract spatial features from hyperspectral and LiDAR images, respectively. The latter first concatenates the HS and LiDAR data for constructing a uniform graph structure. Then, the dual GCN models perform structural fusion by sharing the graph structures and weight matrices of some layers to extract their structural information, respectively. Finally, the final hybrid features are fed into a standard classifier for the pixel-level classification task under a unified feature fusion module. The main contributions can be summarized as follows:
  • A dual-coupled network is proposed for the classification of hyperspectral and LiDAR images. The network adopts a weight-sharing strategy to significantly reduce the number of trainable parameters and alleviate the overfitting of the network. In addition, since CNN and GCN are used, both spatial and structural features can be extracted from hyperspectral and LiDAR data. This brings a richer set of input features to the subsequent classifier and helps to improve the classification performance of the whole network.
  • Several simple and effective multiple feature fusion strategies were developed to effectively utilize the four features extracted from the previous step. These strategies include traditional single fusion strategy and hybrid fusion strategy. Comparative analysis shows that a good feature fusion strategy can effectively improve classification performance.
  • Extensive experiments on two real-world HS and LiDAR data show that the proposed method exhibits significant superiority compared to state-of-the-art baseline methods, such as two-branch CNN and context CNN. In particular, the performance obtained on Trento data is comparable to that of the state-of-the-art methods reported so far.
The remainder of the paper is organized as follows. Section 2 describes the proposed methodology. The dataset and experimental results are presented in Section 3. The proposed method is analyzed and discussed in Section 4. Conclusions are summarized in Section 5.

2. Methodology

In this section, the details of our proposed model are described, and how it works is explained.

2.1. General Framework of the Proposed Method

In order to effectively extract features from multi-source remote sensing data (HS and LiDAR), we used convolutional neural networks and graph convolutional networks to extract neighborhood spatial features in a short-range region and structural features in a middle- and long-range region, respectively. It should be noted that these two types of features extracted from CNN and GCN are entirely different here, and this has been analyzed and explained in detail in several papers [24].
Additionally, we developed two sets of coupled models with similar structures for feature representation or learning by using a weight-sharing mechanism. This may bring two main benefits. On the one hand, it reduces the network parameters, alleviates overfitting and facilitates fast convergence of the model. On the other hand, implicit multi-source data fusion is also partially achieved in the feature extraction phase, as the same set of weights is used to learn the representation of HS and LiDAR data simultaneously.
So, we propose an HS and LiDAR data classification method based on a dual-coupled CNN-GCN structure (DCCG). Figure 1 illustrates the general framework of our proposed method. The model consists of an input module, a dual-coupled module, a feature fusion module and a classifier module. Note that the proposed dual-coupled module can be divided into two network structures: a coupled convolutional neural network (CCNet) and a coupled graph convolutional network (CGNet), which are described in detail in the subsequent Section 2.2 and Section 2.3. As shown in Figure 1, the inputs to the model are the spectral feature matrix X H w × h × D for HS data and the elevation feature matrix X L w × h for LiDAR data, respectively. w , h and D denote the original image’s width, height and the number of spectral bands, respectively. The overall output of the network is an end-to-end classification map. In the figure, ( C H ,   C L ,   G H ,   G L ) are the four features extracted by the dual-coupled module. F represents the fused features processed by the feature fusion module (see Section 2.4 for details).

2.2. Feature Extraction on Coupled CNN (CCNet)

It is well known that CNN have demonstrated impressive capabilities in image processing. Similarly, many CNN and their derivatives have been developed to capture the neighborhood spatial features from images in remote sensing classification tasks. These discriminative features will provide the possibility to distinguish target regions in complex scenes.
Figure 2 illustrates the main structure of the coupled CNN model. As can be seen from the figure, the input remote sensing data (i.e., X H or X L ) are first sent to a neighborhood patches sampler, which samples and generates a series of neighborhood patches (i.e., three-dimensional spectral cubes P H p × p × D or two-dimensional elevation pathes P L p × p ) from the original data according to a given neighborhood size ( p × p ). They are then fed into the corresponding CNN model for feature extraction or representation learning. Specifically, the model mainly consists of two almost identical CNN structures for HS and LiDAR data, respectively. The input to the CNN at the top of the figure is P H extracted from the hyperspectral image, while the input to the CNN at the bottom is P L extracted from the LiDAR image. Note that both CNN networks used here contain four convolutional blocks. Each convolutional block contains a similar structure, which consists of a convolutional layer, a batch normalization (BN) layer [30], a MaxPooling layer and a non-linear activation layer. For each CNN network, the number of convolutional kernels in these four convolutional layers is 32, 64, 128, and 128, respectively. The first three convolutional layers use a convolutional kernel size of 3 × 3 , and the last layer is 1 × 1 . C H N × 128 and C L N × 128 denote the short-range (or neighborhood) spatial features extracted from the CNN on the HS and LiDAR data, respectively.
Due to the weight-sharing mechanism, the two CNN models share the parameters and settings of the other three convolutional layers, except for the first convolutional layer. This will have the obvious benefit of a large reduction in the number of main parameters of the model. We discuss this in more detail in Section 4.3. Next, the convolution operation rules of the two CNN networks are given as follows:
H j l + 1 = f ( m H m l W j , m l + 1 + b j , m l + 1 )
where H m l is the m th feature map at the l th layer. W j , m l + 1 denotes the j th convolution kernel connected with the m th feature map at the ( l + 1 ) th layer, and b j , m l + 1 is the corresponding bias. Here, H 0 = P H   o r   P L . f ( · ) denotes the nonlinear activation function, namely ReLU [31]. Obviously, the size of the sampled neighborhood can significantly affect the final classification result. So, this point will be carefully discussed in Section 4.1.

2.3. Feature Extraction on Coupled GCN (CGNet)

A large amount of research shows that the neighborhood spatial features extracted by CNN are well suited for processing grid data like images, and have achieved impressive success [32]. However, unlike CNN, GCN can extract middle- and long-range structural features from images based on graph structure. This feature effectively complements CNN and has been used to improve the overall classification accuracy of images in complex scenes [24]. Therefore, naturally, we propose a multi-source remote sensing data classification method based on coupled graph convolutional networks (CGNet). The network attempts to use GCN for multi-source remote sensing data classification tasks and introduces a weight-sharing mechanism for reducing training parameters and preventing overfitting as described previously.
Since the classical GCN suffers from problems, such as difficulty in training on large size graphs, miniGCN [24] is chosen as the backbone to construct the coupled GCN in this paper. It can replace one full-batch training on the full graph with a series of mini-batch training on the sub-graphs.
Similar to CCNet, the proposed CGNet contains two GCN networks with similar structures for extracting (middle- and long-range) structural features from hyperspectral data and LiDAR data, respectively. Figure 3 illustrates the main structure of CGNet. As shown in the figure, the model first integrates the hyperspectral and LiDAR data to form a cascaded raw feature matrix [ X H , X L ] , which is used to construct a unified graph structure G N × N . N denotes the number of samples. In this paper, the following radial basis function (RBF) is employed as a similarity measure function for constructing graph structures:
G i j = e x p ( | | x i x j | | σ 2 )
where x i and x j denote the cascaded HS-LiDAR feature vectors of pixel i and j , respectively. σ is a parameter of RBF that controls the radial range (default: σ = 1). Only the similarities between each pixel and its k -nearest neighbours (default: k = 10) are computed to generate a sparse adjacency matrix.
Next, a joint subgraph sampler is created to allow for mini-batch training on the graph. This sampler can randomly sample from the full graph G at a fixed scale M ( M N ) to generate a series of subgraph sets, A = { A 1 , A 2 , , A s } ,   s = 1 , 2 , , N / M . A s M × M . As shown in the upper part of Figure 3, the generated subgraph set A and all spectral features V H N × D for each pixel are then fed into the GCN to complete the structural feature extraction for the HS data. Correspondingly, the same subgraph set A and the elevation features V L N for each pixel are also be fed into the GCN to perform the structural feature extraction for the LiDAR data, as shown in the lower part of Figure 3.
Note that the two GCN networks have nearly identical network structures. Each GCN consists of just two hidden layers with 32 and 128 units, respectively (too many hidden layers in a GCN would lead to overly smooth results [23]). Due to the difference in input layers, two GCN networks can only share their trainable weight matrix of the second hidden layer according to the weight-sharing strategy. Here, the rules for propagation between the layers of these two GCN networks are presented:
H ˜ s l + 1 = f ( D ˜ 1 2 A ˜ s D ˜ 1 2 H ˜ s l W l + 1 + b l + 1 )
H l + 1 = [ H ˜ 1 l + 1 , , H ˜ s l + 1 , , H ˜ N / M l + 1 ]
where s is not only the s th subgraph but also the s th batch in the network training process. A ˜ s = A s + I M , D ˜ i , i = j A ˜ i , j . A s is the adjacency matrix on the s th subgraph. I M is the identity matrix with M nodes. W l + 1 and b l + 1 represent the learnable weight parameter and bias of the ( l + 1 ) th hidden layer, respectively. H ˜ s l + 1 denotes the outputs of the s th subgraph in the l th hidden layer. H l + 1 denotes the total output of the l th layer on the whole graph, where H 0 = V H or V L . f ( · ) represents the ReLU activation function. G H N × 128 and G L N × 128 denote the middle- and long-range structural features extracted by the GCN on the HS and LiDAR data, respectively.

2.4. Training and Testing of the Model

After CCNet and CGNet processing, we can extract four features ( C H , C L , G H and G L ) from the hyperspectral and LiDAR data. An effective data fusion (i.e., fusion module) of these features is required for better classification before feeding into the classifier. Here, for simplicity, the fusion of the multi-source data is accomplished in a cascade fashion as follows:
F = [ C H , C L , G H , G L ]
where F N × 512 denotes the fused features. Note that more comparisons of fusion strategies will be discussed in Section 4.2.
Next, the fused features are fed into a standard softmax classifier, consisting of two fully connected layers containing 128 and C units, respectively. The final output of the whole network can be computed as:
y ^ = s o f t m a x ( W 2 f ( W 1 F + b 1 ) + b 2 )
where W 1 and W 2 denote the weight matrix of the two fully connected layers, respectively. b 1 and b 2 are their respective bias.
The loss of the whole model can be defined by the cross-entropy loss function as follows:
L = 1 N i = 1 N ( y i l o g y ^ i + ( 1 y i ) l o g ( 1 y ^ i ) )
where y i and y ^ i indicate the actual class label and the predicted label of pixel i , respectively. Using Adam [33], back propagation and parameter updates are performed on the whole model.

3. Experiments

To evaluate the effectiveness of the proposed method, we compare DCCG with several other methods, such as CNN [31], miniGCN [24], CCNet (see Section 2.2) and CGNet (see Section 2.3). Four metrics, such as per-class accuracy, overall accuracy (OA), average accuracy (AA) and kappa coefficient (κ), are used to measure classification accuracy quantitatively. Classification maps are adopted to evaluate the classification results qualitatively. Furthermore, to demonstrate the superiority of the proposed method, several state-of-the-art multimodal remote sensing classification models are compared in Section 4.4.

3.1. Data Description

All experiments in this paper were conducted on two real hyperspectral and LiDAR data, i.e., Trento and Houston.
  • Trento Data: The data are composed of the HSI and LiDAR images that were acquired over the area from the city of Trento, Italy. This scene consists of 166 × 600 pixels with 1 m spatial resolution. The HSI scene consists of 63 spectral bands with wavelength ranging from 0.42 to 0.99 μm including six classes [22]. Figure 4 visualizes the HS and LiDAR as well as the locations of training and testing samples on Trento data. Table 1 lists the information about the number of samples in different classes on Trento data.
  • Houston Data: The data are composed of the HSI and LiDAR images that were captured in 2012 by an airborne sensor over the area of University of Houston campus and the neighboring urban area. This scene consists of 349 × 1905 pixels with 2.5 m spatial resolution. The HSI scene consists of 144 spectral bands with wavelength ranging from 0.38 to 1.05 μm including 15 classes [34]. Figure 5 visualizes the HS and LiDAR as well as the locations of training and testing samples on Houston data. Table 2 lists the information about the number of samples in different classes on Houston data.

3.2. Experimental Setup

According to the properties of the methods, these classification methods involved in this experiment can be divided into two main categories: single model-based methods and hybrid model-based methods. The former includes CNN-H, CNN-L, miniGCN-H and miniGCN-L, while the latter includes CCNet-HL, CGNet-HL, DCCG-H, DCCG-L and DCCG-HL. After the symbol ‘-’, the letters ‘H’ and ‘L’ indicate HS and LiDAR data, respectively. Note that the network structure of the CNN and miniGCN models in the comparison experiments is the same as described in Section 2.2 and Section 2.3 of this paper. The maximum number of epochs, learning rate and batch size were set to 200, 0.001 and 32, respectively. Here, for comparison purposes, the neighborhood size p is set to 11 on both Trento data and Houston data. More neighborhood size settings will be discussed in Section 4.1. In particular, to facilitate computation and reduce data redundancy, as with most multimodal classification methods, we first performed PCA dimensionality reduction on the hyperspectral data in both experimental datasets before model training. Here, the first 20 dimensional features were empirically taken as model input before subsequent classification [35,36]. All deep models were implemented in the TensorFlow 1.5 framework. All experiments were carried out in the same hardware environment, i.e., Intel i5-12400, 16 GB memory and no GPU devices.

3.3. Results and Analysis of Trento Data

Table 3 lists the quantitative classification results obtained by the different methods on Trento data. From the table, some remarkable facts can be observed. Firstly, for all models using only single source (HS or LiDAR) data, the performance using only hyperspectral data is significantly higher than that of only LiDAR data. For example, miniGCN-H shows a 23.91% improvement on OA over miniGCN-L. This is mainly because hyperspectral images contain rich spectral information and fine neighborhood spatial information compared to LiDAR data, which provide many distinguishable features for effective classification. Furthermore, we observed that CNN significantly outperformed miniGCN in all single model-based methods. For instance, the overall accuracy was improved from 81.79% for miniGCN-H to 93.81% for CNN-H on hyperspectral data by 12.02%. One of the possible reasons for this phenomenon is that the CNN-extracted short-range (or neighborhood) features are more helpful for pixel classification than the GCN-extracted middle- and long-range structural features. Thirdly, the introduction of multiple source data and hybrid structures significantly improved the classification performance of the models compared to the single-source data and single-model approaches. For instance, CCNet-HL and CGNet-HL improved by 3.15% and 6.08% on OA, respectively, over their single-model approaches. This is partly due to the inclusion of multi-source data, making the classifier’s input more multi-source and abundant, while enhancing the dimensionality of the distinguishable features. More importantly, on the other hand, the adoption of a weight-sharing-based coupling structure allows for a significant reduction in the model parameters, which alleviates overfitting and improves the model’s generalization ability. Finally, our proposed method was found to achieve the best performance of all the methods in the experiment. It not only benefits from multi-source data and the coupling structure based on weight sharing, but also because the neighborhood spatial information and the middle- and long-range structural information are considered simultaneously in the model, resulting in the complementarity of information at the feature level. Specifically, DCCG-HL consists of CCNet-HL and CGNet-HL, but its classification performance is remarkably higher than the latter two by 4.65% and 6.85%, respectively. This also demonstrates the effectiveness of the proposed method from the side.
Figure 6 shows the full classification maps of the different methods on the Trento data. As can be seen, the inclusion of multiple-sources data significantly improves the classification results. Our proposed method retains the boundaries between lands while having smoother regions. Figure 7 illustrates the loss curves obtained by different methods for the Trento dataset. Similarly, our DCCG-HL successfully suppresses the overfitting and achieves the best test results.

3.4. Results and Analysis of the Houston Data

Table 4 reports the quantitative classification results obtained by the different methods on Houston data. From the table, a similar fact to Trento can be observed. Our proposed method still achieves the best classification results and outperforms the second place (CCNet-HL) by 1.99% on OA. It also approaches 100% accuracy in classes 1, 3, 7 and 15. In addition, an interesting phenomenon was observed. Specifically, in class 13, the classification accuracy of CCNet-HL and CGNet-HL was 74.46% and 69.92%, respectively. While our proposed DCCG-HL combined these two models, its OA improved by 15.32% to 89.78%. The main reason for this phenomenon is that the proposed method integrates well the neighborhood spatial features and the middle- and long-range structural features, achieving complementarity between the different features.
Figure 8 exhibits the full classification maps obtained by the different classification methods on the Houston data. On one hand, CNN-based methods (e.g., Figure 8a,e) can utilize the extracted neighborhood spatial features to make the classified region very smooth, but the edges and details are often ignored (e.g., roads). GCN-based methods (e.g., Figure 8c,f) can employ middle- and long-range structural features to retain more detailed information, but with more noise in the classified regions. This reflects the classification tendencies of the different classification methods. On the other hand, hyperspectral-based methods (e.g., Figure 8a,c,g) clearly provide more detailed information about the target using their rich spectral bands, which is very beneficial for the classification task. However, it has difficulty in solving the problem of classifying similar objects (e.g., grass and shrubs). LiDAR-based methods (e.g., Figure 8d) use elevation information to easily distinguish between targets with different heights, but targets with the same height are difficult to classify. This is a limitation of single-sensor classification. In contrast, our proposed DCCG-HL integrates both multi-sensor and multi-model techniques to effectively perform the classification task in complex scenes. Specifically, it can smooth the classified region while retain more classification detail information (e.g., roads) between regions. Figure 9 displays the loss curves obtained by different methods for the Houston dataset. Similar to Trento data, our proposed DCCG-HL method is smoother throughout the training and testing process and successfully suppresses the overfitting.

4. Discussion

4.1. Effect of Different Neighborhood Sizes on Classification Results

In the proposed model, one parameter (i.e., neighborhood size p × p ) needs to be determined in advance before training. We fixed the basic structure and main parameters of the network, picked the neighborhood size p from the candidate set {5,7,9,11,13,15,17,19,21} in turn, and tested the overall accuracy variation of the network. Figure 10 illustrates the effect of different neighborhood sizes on the classification accuracy of the Trento and Houston data, respectively. A similar phenomenon can be observed in the figure, where the OA first rises significantly and then slowly decreases as the neighborhood size increases. Specifically, the model achieves the highest classification accuracy of 99.06% on Trento data when p = 11 . Correspondingly, the model also achieves the highest classification accuracy of 93.79% on Houston data when p = 17 .

4.2. Effect of Different Fusion Strategies on Classification Results

The effective fusion of features for multi-source data classification is a key step. Here we discuss the impact of different fusion strategies on classification results. These strategies can be divided into traditional single-based fusion strategies and hybrid-based fusion strategies. Specifically, the former contains additive fusion (DCCG-A), pixel-wise multiplicative fusion (DCCG-M) and cascade fusion (DCCG-C). As for the latter, similar to Equation (5), we give definitions of three hybrid-based fusion strategies: DCCG-1 ( [ C h + C l , G h + G l ] ), DCCG-2 ( [ C h + G h , C l + G l ] ) and DCCG-3 ( [ C h + G l , C l + G h ] ).
Table 5 reports the effect of different fusion strategies on the classification results. As can be seen from the table, for the Houston data, the cascade fusion method (DCCG-C) slightly outperformed the DCCG-2 method and achieved the best OA. In contrast, for the Trento data, the DCCG-2 strategy slightly outperformed DCCG-1.

4.3. Analysis of the Number of Parameters

To demonstrate the effectiveness of the coupled structure, we counted the total number of parameters and the overall accuracy of DCCG before and after using the coupled structure (see Table 6). As can be seen from the table, the total number of parameters in the model was reduced from 301,936 to 188,848 on the Trento data and from 598,777 to 338,105 on the Houston data after adopting the weight-sharing module (i.e., coupled structure). In other words, the total parameters were decreased by 37% and 44% on the Trento and Houston data, respectively. Fortunately, the OA on these two datasets was slightly increased by 0.55% and 0.62%, respectively. This demonstrates that the weight-sharing module can both reduce the model parameters and improve the classification results in complex scenes.

4.4. Comparison with Other State-of-the-Art Methods

To verify the superiority of the proposed method, Table 7 compares several state-of-the-art multimodal classification methods: EndNet [22], MML(FC) [37], CNN-PPF(H + L) [20], Two-Branch CNN(H + L) [19], Context CNN(H + L) [21] and PToP CNN [38]. It should be noted that, except EndNet and MML(FC), the authors did not provide the original code for the implementation. Therefore, the classification results for CNN-PPF (H + L), Two-Branch CNN (H + L), Context CNN (H + L) and PToP CNN are taken directly from the best results in the relevant literature. For MML(FC), the original one-dimensional LiDAR is used here for testing as the authors do not provide a specific pre-processing procedure for Trento’s LiDAR data. The superiority of the proposed method can be seen in the table, which outperforms current state-of-the-art classification methods such as Two-Branch CNN and Context CNN on both Trento data and Houston data. Furthermore, we observe that the classification accuracy (99.11%) of the Trento data achieves the best classification performance reported so far.

4.5. Analysis on the Computation Cost

In order to evaluate the computation cost of the different methods, Table 8 and Table 9 report the computational time on the Trento and Houston data, respectively. Unfortunately, for the same reason as mentioned in Section 4.4, the authors’ source code was not available and thus some advanced methods, i.e., CNN-PPF (H + L), Two-Branch CNN (H + L), Context CNN (H + L) and PToP CNN, could not be evaluated in the same experimental setting in terms of computational time. Furthermore, it should be noted that all experiments were performed on the Tensorflow framework and without using any GPU acceleration techniques.
From the table, it can be observed that all single-source models take less time to train than their corresponding two-source models. This is mainly due to the fact that the single-source models deal with less data. In contrast, the two-source models need to process more remote sensing data from more sources and thus require more training time. However, once all models have been trained, they are tested very efficiently. For DCCG-HL, the introduction of weight sharing did not significantly increase the training and testing time.

5. Conclusions

This paper proposed a dual-coupled CNN-GCN-based classification method for hyperspectral and LiDAR data. The method can be divided into two sets of structurally similar networks: a coupled CNN and a coupled GCN. Such two networks can be employed to extract spatial-spectral features in a short-range region and structural features in a middle- and long-range region from hyperspectral and LiDAR data, respectively. Furthermore, these two classes of features are proven to achieve feature complementarity and performance enhancement in multi-source remote sensing data classification tasks. The coupling module using weight sharing makes the network with a substantially smaller number of parameters, alleviating overfitting and improving the model’s generalization performance. The effects of different neighborhood sizes and feature fusion strategies on the final classification results of the model are discussed in detail. Extensive experiments on two real-world hyperspectral and LiDAR images demonstrate the clear superiority of the proposed method compared to other state-of-the-art methods for multimodal data classification. In future research, we will continue to explore more efficient feature extraction methods and multimodal fusion methods in remote sensing data classification.

Author Contributions

Conceptualization, X.W. and L.W.; formal analysis, X.W.; Data curation, L.W.; Methodology, L.W.; Writing—original draft, L.W.; Writing—review and editing, L.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used to support the findings of this study are available from the co-first author upon request.

Acknowledgments

The authors are grateful to the editor and reviewers for their constructive comments that have helped improve this work significantly.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kang, J.; Hong, D.; Liu, J.; Baier, G.; Yokoya, N.; Demir, B. Learning Convolutional Sparse Coding on Complex Domain for Interferometric Phase Restoration. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 826–840. [Google Scholar] [CrossRef] [Green Version]
  2. Huang, R.; Hong, D.; Xu, Y.; Yao, W.; Stilla, U. Multi-Scale Local Context Embedding for LiDAR Point Cloud Classification. IEEE Geosci. Remote Sens. Lett. 2020, 17, 721–725. [Google Scholar] [CrossRef]
  3. Hang, R.; Liu, Q.; Hong, D.; Ghamisi, P. Cascaded Recurrent Neural Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5384–5394. [Google Scholar] [CrossRef] [Green Version]
  4. Yokoya, N.; Grohnfeldt, C.; Chanussot, J. Hyperspectral and Multispectral Data Fusion: A Comparative Review of the Recent Literature. IEEE Geosci. Remote Sens. Mag. 2017, 5, 29–56. [Google Scholar] [CrossRef]
  5. Ghamisi, P.; Gloaguen, R.; Atkinson, P.M.; Benediktsson, J.A.; Rasti, B.; Yokoya, N.; Wang, Q.; Hofle, B.; Bruzzone, L.; Bovolo, F.; et al. Multisource and Multitemporal Data Fusion in Remote Sensing: A Comprehensive Review of the State of the Art. IEEE Geosci. Remote Sens. Mag. 2019, 7, 6–39. [Google Scholar] [CrossRef] [Green Version]
  6. Khodadadzadeh, M.; Li, J.; Prasad, S.; Plaza, A. Fusion of Hyperspectral and LiDAR Remote Sensing Data Using Multiple Feature Learning. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2971–2983. [Google Scholar] [CrossRef]
  7. Xia, J.; Yokoya, N.; Iwasaki, A. Fusion of Hyperspectral and LiDAR Data With a Novel Ensemble Classifier. IEEE Geosci. Remote Sens. Lett. 2018, 15, 957–961. [Google Scholar] [CrossRef]
  8. Rasti, B.; Ghamisi, P.; Plaza, J.; Plaza, A. Fusion of Hyperspectral and LiDAR Data Using Sparse and Low-Rank Component Analysis. IEEE Trans. Geosci. Remote Sens. 2017, 55, 6354–6365. [Google Scholar] [CrossRef] [Green Version]
  9. Liu, X.; Bo, Y. Object-Based Crop Species Classification Based on the Combination of Airborne Hyperspectral Images and LiDAR Data. Remote Sens. 2015, 7, 922–950. [Google Scholar] [CrossRef] [Green Version]
  10. Ghamisi, P.; Hofle, B.; Zhu, X.X. Hyperspectral and LiDAR Data Fusion Using Extinction Profiles and Deep Convolutional Neural Network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 3011–3024. [Google Scholar] [CrossRef]
  11. Licciardi, G.; Pacifici, F.; Tuia, D.; Prasad, S.; West, T.; Giacco, F.; Thiel, C.; Inglada, J.; Christophe, E.; Chanussot, J.; et al. Decision Fusion for the Classification of Hyperspectral Data: Outcome of the 2008 GRS-S Data Fusion Contest. IEEE Trans. Geosci. Remote Sens. 2009, 47, 3857–3865. [Google Scholar] [CrossRef] [Green Version]
  12. Liao, W.; Bellens, R.; Pizurica, A.; Gautama, S.; Philips, W. Combining Feature Fusion and Decision Fusion for Classification of Hyperspectral and LiDAR Data. In Proceedings of the 2014 IEEE Geoscience and Remote Sensing Symposium, Quebec City, QC, Canada, 13–18 July 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 1241–1244. [Google Scholar]
  13. Pedergnana, M.; Marpu, P.R.; Dalla Mura, M.; Benediktsson, J.A.; Bruzzone, L. Classification of Remote Sensing Optical and LiDAR Data Using Extended Attribute Profiles. IEEE J. Sel. Top. Signal Process. 2012, 6, 856–865. [Google Scholar] [CrossRef]
  14. Rasti, B.; Ghamisi, P.; Gloaguen, R. Hyperspectral and LiDAR Fusion Using Extinction Profiles and Total Variation Component Analysis. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3997–4007. [Google Scholar] [CrossRef]
  15. Sturari, M.; Frontoni, E.; Pierdicca, R.; Mancini, A.; Malinverni, E.S.; Tassetti, A.N.; Zingaretti, P. Integrating Elevation Data and Multispectral High-Resolution Images for an Improved Hybrid Land Use/Land Cover Mapping. Eur. J. Remote Sens. 2017, 50, 1–17. [Google Scholar] [CrossRef]
  16. Liu, Q.; Zhang, N.; Yang, W.; Wang, S.; Cui, Z.; Chen, X.; Chen, L. A Review of Image Recognition with Deep Convolutional Neural Network. In Intelligent Computing Theories and Application; Huang, D.-S., Bevilacqua, V., Premaratne, P., Gupta, P., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2017; Volume 10361, pp. 69–80. ISBN 978-3-319-63308-4. [Google Scholar]
  17. Sultana, F.; Sufian, A.; Dutta, P. A Review of Object Detection Models Based on Convolutional Neural Network. In Intelligent Computing: Image Processing Based Applications; Mandal, J.K., Banerjee, S., Eds.; Advances in Intelligent Systems and Computing; Springer: Singapore, 2020; Volume 1157, pp. 1–16. ISBN 9789811542879. [Google Scholar]
  18. Chai, J.; Li, A. Deep Learning in Natural Language Processing: A State-of-the-Art Survey. In Proceedings of the 2019 International Conference on Machine Learning and Cybernetics (ICMLC), Kobe, Japan, 7–10 July 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–6. [Google Scholar]
  19. Xu, X.; Li, W.; Ran, Q.; Du, Q.; Gao, L.; Zhang, B. Multisource Remote Sensing Data Classification Based on Convolutional Neural Network. IEEE Trans. Geosci. Remote Sens. 2018, 56, 937–949. [Google Scholar] [CrossRef]
  20. Li, W.; Wu, G.; Zhang, F.; Du, Q. Hyperspectral Image Classification Using Deep Pixel-Pair Features. IEEE Trans. Geosci. Remote Sens. 2017, 55, 844–853. [Google Scholar] [CrossRef]
  21. Lee, H.; Kwon, H. Going Deeper With Contextual CNN for Hyperspectral Image Classification. IEEE Trans. on Image Process. 2017, 26, 4843–4855. [Google Scholar] [CrossRef] [Green Version]
  22. Hong, D.; Gao, L.; Hang, R.; Zhang, B.; Chanussot, J. Deep Encoder-Decoder Networks for Classification of Hyperspectral and LiDAR Data. IEEE Geosci. Remote Sens. Lett. 2020, 19, 5500205. [Google Scholar] [CrossRef]
  23. Kipf, T.N.; Welling, M. Semi-Supervised Classification with Graph Convolutional Networks. arXiv 2017, arXiv:1609.02907. [Google Scholar]
  24. Hong, D.; Gao, L.; Yao, J.; Zhang, B.; Plaza, A.; Chanussot, J. Graph Convolutional Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 5966–5978. [Google Scholar] [CrossRef]
  25. Qin, A.; Liu, C.; Shang, Z.; Tian, J. Spectral-Spatial Graph Convolutional Networks for Semel-Supervised Hyperspectral Image Classification. In Proceedings of the 2018 International Conference on Wavelet Analysis and Pattern Recognition (ICWAPR), Chengdu, China, 15–18 July 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 89–94. [Google Scholar]
  26. Wang, T.; Wang, G.; Tan, K.E.; Tan, D. Spectral Pyramid Graph Attention Network for Hyperspectral Image Classification. arXiv 2020, arXiv:2001.07108. [Google Scholar]
  27. Wan, S.; Gong, C.; Zhong, P.; Pan, S.; Li, G.; Yang, J. Hyperspectral Image Classification With Context-Aware Dynamic Graph Convolutional Network. IEEE Trans. Geosci. Remote Sens. 2021, 59, 597–612. [Google Scholar] [CrossRef]
  28. Wan, S.; Gong, C.; Zhong, P.; Du, B.; Zhang, L.; Yang, J. Multiscale Dynamic Graph Convolutional Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 3162–3177. [Google Scholar] [CrossRef] [Green Version]
  29. Li, J.; Hong, D.; Gao, L.; Yao, J.; Zheng, K.; Zhang, B.; Chanussot, J. Deep Learning in Multimodal Remote Sensing Data Fusion: A Comprehensive Review 2022. arXiv 2022, arXiv:2205.01380. [Google Scholar]
  30. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In Proceedings of the International conference on machine learning, Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  31. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  32. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Yu, P.S. A Comprehensive Survey on Graph Neural Networks. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 4–24. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2017, arXiv:1412.6980. [Google Scholar]
  34. Debes, C.; Merentitis, A.; Heremans, R.; Hahn, J.; Frangiadakis, N.; van Kasteren, T.; Liao, W.; Bellens, R.; Pizurica, A.; Gautama, S.; et al. Hyperspectral and LiDAR Data Fusion: Outcome of the 2013 GRSS Data Fusion Contest. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2405–2418. [Google Scholar] [CrossRef]
  35. Ling, J.; Li, L.; Wang, H. Improved Fusion of Spatial Information into Hyperspectral Classification through the Aggregation of Constrained Segment Trees: Segment Forest. Remote Sens. 2021, 13, 4816. [Google Scholar] [CrossRef]
  36. Du, Q.; Fowler, J.E. Hyperspectral Image Compression Using JPEG2000 and Principal Component Analysis. IEEE Geosci. Remote Sens. Lett. 2007, 4, 201–205. [Google Scholar] [CrossRef]
  37. Hong, D.; Gao, L.; Yokoya, N.; Yao, J.; Chanussot, J.; Du, Q.; Zhang, B. More Diverse Means Better: Multimodal Deep Learning Meets Remote-Sensing Imagery Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 4340–4354. [Google Scholar] [CrossRef]
  38. Zhang, M.; Li, W.; Du, Q.; Gao, L.; Zhang, B. Feature Extraction for Classification of Hyperspectral and LiDAR Data Using Patch-to-Patch CNN. IEEE Trans. Cybern. 2020, 50, 100–111. [Google Scholar] [CrossRef] [PubMed]
Figure 1. A general framework of proposed DCCG.
Figure 1. A general framework of proposed DCCG.
Sensors 22 05735 g001
Figure 2. Illustration of the structure of CCNet. Conv and Pool denote the convolution and maximum pooling operations, respectively. p is the given neighborhood size.
Figure 2. Illustration of the structure of CCNet. Conv and Pool denote the convolution and maximum pooling operations, respectively. p is the given neighborhood size.
Sensors 22 05735 g002
Figure 3. Illustration of the structure of CGNet. [ , ] indicates cascade operations.
Figure 3. Illustration of the structure of CGNet. [ , ] indicates cascade operations.
Sensors 22 05735 g003
Figure 4. Visualization of Trento dataset: (a) false-color image (using 50, 45 and 5 bands as R, G and B, respectively); (b) LiDAR image; (c) training samples; (d) test samples; (e) color code.
Figure 4. Visualization of Trento dataset: (a) false-color image (using 50, 45 and 5 bands as R, G and B, respectively); (b) LiDAR image; (c) training samples; (d) test samples; (e) color code.
Sensors 22 05735 g004
Figure 5. Visualization of Houston dataset: (a) false-color image (using 64, 43 and 22 bands as R, G and B, respectively); (b) LiDAR image; (c) training samples; (d) test samples; (e) color code.
Figure 5. Visualization of Houston dataset: (a) false-color image (using 64, 43 and 22 bands as R, G and B, respectively); (b) LiDAR image; (c) training samples; (d) test samples; (e) color code.
Sensors 22 05735 g005
Figure 6. Classification maps obtained by different methods for the Trento dataset. (a) CNN-H; (b) CNN-L; (c) miniGCN-H; (d) miniGCN-L; (e) CCNet-HL; (f) CGNet-HL; (g) DCCGNet-H; (h) DCCGNet-L; (i) DCCG-HL.
Figure 6. Classification maps obtained by different methods for the Trento dataset. (a) CNN-H; (b) CNN-L; (c) miniGCN-H; (d) miniGCN-L; (e) CCNet-HL; (f) CGNet-HL; (g) DCCGNet-H; (h) DCCGNet-L; (i) DCCG-HL.
Sensors 22 05735 g006
Figure 7. Loss curves obtained by different methods for the Trento dataset. (a) CNN-H; (b) CNN-L; (c) miniGCN-H; (d) miniGCN-L; (e) CCNet-HL; (f) CGNet-HL; (g) DCCGNet-H; (h) DCCGNet-L; (i) DCCG-HL.
Figure 7. Loss curves obtained by different methods for the Trento dataset. (a) CNN-H; (b) CNN-L; (c) miniGCN-H; (d) miniGCN-L; (e) CCNet-HL; (f) CGNet-HL; (g) DCCGNet-H; (h) DCCGNet-L; (i) DCCG-HL.
Sensors 22 05735 g007
Figure 8. Classification maps obtained by different methods for the Houston dataset. (a) CNN-H; (b) CNN-L; (c) miniGCN-H; (d) miniGCN-L; (e) CCNet-HL; (f) CGNet-HL; (g) DCCGNet-H; (h) DCCGNet-L; (i) DCCG-HL.
Figure 8. Classification maps obtained by different methods for the Houston dataset. (a) CNN-H; (b) CNN-L; (c) miniGCN-H; (d) miniGCN-L; (e) CCNet-HL; (f) CGNet-HL; (g) DCCGNet-H; (h) DCCGNet-L; (i) DCCG-HL.
Sensors 22 05735 g008
Figure 9. Loss curves obtained by different methods for the Houston dataset. (a) CNN-H; (b) CNN-L; (c) miniGCN-H; (d) miniGCN-L; (e) CCNet-HL; (f) CGNet-HL; (g) DCCGNet-H; (h) DCCGNet-L; (i) DCCG-HL.
Figure 9. Loss curves obtained by different methods for the Houston dataset. (a) CNN-H; (b) CNN-L; (c) miniGCN-H; (d) miniGCN-L; (e) CCNet-HL; (f) CGNet-HL; (g) DCCGNet-H; (h) DCCGNet-L; (i) DCCG-HL.
Sensors 22 05735 g009
Figure 10. Effect of different neighborhood sizes on OA.
Figure 10. Effect of different neighborhood sizes on OA.
Sensors 22 05735 g010
Table 1. List of the number of training and test samples for each class in Trento datasets.
Table 1. List of the number of training and test samples for each class in Trento datasets.
No.Class NameTrainingTest
1Apples1293905
2Buildings1252778
3Ground105374
4Woods1548969
5Wineyard18410,317
6Roads1223052
-Total81929,395
Table 2. List of the number of training and test samples for each class in Houston datasets.
Table 2. List of the number of training and test samples for each class in Houston datasets.
No.Class NameTrainingTest
1Healthy grass1981053
2Stressed grass1901064
3Synthetic grass192505
4Trees1881056
5Soil1861056
6Water182143
7Residential1961072
8Commercial1911053
9Road1931059
10Highway1911036
11Railway1811054
12Parking Lot 11921041
13Parking Lot 2184285
14Tennis Court181247
15Running Track187473
-Total283212,197
Table 3. Per-class accuracy, OA, AA and κ obtained by different methods on Trento data. The bold text in each row indicates the best classification result.
Table 3. Per-class accuracy, OA, AA and κ obtained by different methods on Trento data. The bold text in each row indicates the best classification result.
No.CNN-HCNN-LminiGCN-HminiGCN-LCCNet-HLCGNet-HLDCCG-HDCCG-LDCCG-HL
193.1788.4873.246.7598.2356.4598.0787.2499.61
274.2088.6169.8222.7885.8096.3173.7192.1396.20
390.0027.0550.424.7560.7177.2074.8927.6889.74
498.5795.2890.6276.26100.099.3599.3396.1899.99
598.5693.7880.9872.5499.0390.9798.0896.7099.48
683.8293.9187.036.8595.9897.5385.9086.4098.10
OA93.8191.2281.7957.8896.9687.8794.4192.2199.06
AA89.7281.1975.3531.6589.9586.3088.3381.0597.19
κ91.6988.1875.4043.1195.9284.0292.5089.5898.75
Table 4. Per-class accuracy, OA, AA and κ obtained by different methods on Houston data. The bold text in each row indicates the best classification result.
Table 4. Per-class accuracy, OA, AA and κ obtained by different methods on Houston data. The bold text in each row indicates the best classification result.
No.CNN-HCNN-LminiGCN-HminiGCN-LCCNet-HLCGNet-HLDCCG-HDCCG-LDCCG-HL
196.73 49.96 96.82 40.07 99.20 99.11 99.7746.71 99.32
280.51 24.85 92.91 0.00 86.00 92.03 92.9624.97 83.65
390.95 45.57 100.0033.29 93.01 98.24 98.83 47.34 99.80
496.64 83.04 93.64 13.92 97.6393.55 87.44 82.82 97.21
586.88 38.32 96.21 5.79 93.74 97.13 91.56 31.68 97.77
682.98 10.61 42.90 1.25 82.42 68.66 91.6112.14 86.79
789.15 66.67 98.92 39.08 93.99 98.71 95.85 66.74 99.17
882.90 67.13 76.57 26.98 82.86 68.36 84.30 67.85 91.43
982.78 41.68 62.32 25.00 86.1372.82 84.07 52.24 81.75
1077.71 80.08 80.00 0.00 91.8380.90 84.99 74.68 88.88
1170.01 88.75 85.96 45.93 92.81 94.22 87.70 89.29 97.13
1278.17 30.95 80.39 0.00 84.77 80.39 82.02 36.75 85.83
1375.54 44.93 73.68 3.63 74.46 69.92 69.48 52.44 89.78
1487.02 42.22 90.77 12.59 100.0096.08 89.17 31.68 96.86
1582.65 28.72 99.36 0.00 93.22 99.79 100.0022.50 99.79
OA83.79 52.50 85.61 23.24 90.43 87.42 89.23 52.39 92.42
AA84.04 49.57 84.70 16.50 90.14 87.33 89.32 49.32 93.01
κ82.39 48.60 84.39 17.90 89.61 86.35 88.31 48.52 91.77
Table 5. Effect of different fusion strategies on OA (%).
Table 5. Effect of different fusion strategies on OA (%).
DatasetsDCCG-ADCCG-MDCCG-CDCCG-1DCCG-2DCCG-3
Trento98.7498.8399.0699.0899.1198.91
Houston91.992.793.7993.1393.2992.86
Table 6. Changes in the number of model parameters by the weight-sharing module.
Table 6. Changes in the number of model parameters by the weight-sharing module.
DatasetsDCCG without Weight SharingDCCG with Weight Sharing
No. of ParametersOA (%)No. of ParametersOA (%)
Trento301,93698.56188,84899.11
Houston598,77793.17338,10593.79
Table 7. Comparison with state-of-the-art methods on OA (%).
Table 7. Comparison with state-of-the-art methods on OA (%).
DatasetsEndNetMML(FC)CNN-PPF (H + L)Two-Branch CNN (H + L)Context CNN (H + L)PToP CNNOur DCCG
Trento92.3491.6294.7697.9296.1198.3499.11
Houston88.4389.6983.3387.9886.9092.4893.79
Table 8. Computation time (seconds) of different methods on the Trento data.
Table 8. Computation time (seconds) of different methods on the Trento data.
Time (s)CNN-HCNN-LminiGCN-HminiGCN-LCCNet-HLCGNet-HL
Train192.72147.72388.81378.381062.55616.24
Test4.882.336.466.348.038.65
Time (s)DCCG-HDCCG-LDCCG-HL (without Weight Sharing)Our DCCG-HLEndNetMML(FC)
Train935.85729.921160.201198.1028.7187.87
Test12.699.0014.3414.700.190.34
Table 9. Computation time (seconds) of different methods on the Houston data.
Table 9. Computation time (seconds) of different methods on the Houston data.
Time (s)CNN-HCNN-LminiGCN-HminiGCN-LCCNet-HLCGNet-HL
Train321.21267.44132.76125.36944.73200.20
Test2.550.872.942.743.773.64
Time (s)DCCG-HDCCG-LDCCG-HL (without Weight Sharing)Our DCCG-HLEndNetMML(FC)
Train622.51589.55955.831003.7227.20100.49
Test5.283.766.406.590.070.15
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, L.; Wang, X. Dual-Coupled CNN-GCN-Based Classification for Hyperspectral and LiDAR Data. Sensors 2022, 22, 5735. https://doi.org/10.3390/s22155735

AMA Style

Wang L, Wang X. Dual-Coupled CNN-GCN-Based Classification for Hyperspectral and LiDAR Data. Sensors. 2022; 22(15):5735. https://doi.org/10.3390/s22155735

Chicago/Turabian Style

Wang, Lei, and Xili Wang. 2022. "Dual-Coupled CNN-GCN-Based Classification for Hyperspectral and LiDAR Data" Sensors 22, no. 15: 5735. https://doi.org/10.3390/s22155735

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop