Abstract
Many graph embedding methods are developed for dimensionality reduction (DR) of hyperspectral image (HSI), which only use spectral features to reflect a point-to-point intrinsic relation and ignore complex spatial-spectral structure in HSI. A new DR method termed spatial-spectral regularized sparse hypergraph embedding (SSRHE) is proposed for the HSI classification. SSRHE explores sparse coefficients to adaptively select neighbors for constructing the dual sparse hypergraph. Based on the spatial coherence property of HSI, a local spatial neighborhood scatter is computed to preserve local structure, and a total scatter is computed to represent the global structure of HSI. Then, an optimal discriminant projection is obtained by possessing better intraclass compactness and interclass separability, which is beneficial for classification. Experiments on Indian Pines and PaviaU hyperspectral datasets illustrated that SSRHE effectively develops a better classification performance compared with the traditional spectral DR algorithms.
1. Introduction
With spectral sampling from visible to short-wave infrared region, hyperspectral image (HSI) can provide a spatial scene in hundreds of narrow contiguous spectral channels [1,2]. HSI data with high spectral resolution can provide fine spectral details for different ground objects, and they have been widely applied in many fields such as geological survey, environmental monitoring, precision agriculture, and mineral exploration [3,4]. Classification of each pixel in HSI plays a crucial role in these real applications. However, the high dimensional characteristic of HSI poses a huge challenge to the traditional classification methods, and the Hughes effect may occur if only limited training samples are available [5,6].
In general, dimensionality reduction (DR) is an effective way to reduce the volume of high-dimensional data with minimum loss of useful information by feature extraction or band selection, and it brings benefits for classification by achieving discriminating embedding features [7,8,9]. Many DR methods based on feature extraction have been proposed to reduce the dimension of high-dimensional data. Principal component analysis (PCA) [10] and linear discriminant analysis (LDA) [11] are the most popular subspace methods, but the two linear methods cannot discover the underlying manifold structure embedding in the original high-dimensional space. Many manifold learning-based DR methods are introduced to reveal nonlinear structure in high-dimensional data, such as locally linear embedding (LLE) [12], Laplacian eigenmap (LE) [13], isometric mapping (ISOMAP) [14], neighborhood preserving embedding (NPE) [15], and locality preserving projection (LPP) [16]. The above methods can be unified under the graph embedding framework (GE), and the difference between them is how to define the similarity matrix of an intrinsic graph and the constraint matrix of a penalty graph [17,18,19,20]. On the basis of GE, some supervised DR methods are designed to exploit the prior knowledge of training samples for improving classification performance, such as marginal Fisher analysis (MFA) [21], local Fisher discriminant analysis (LFDA) [22], and regularized local discriminant embedding (RLDE) [23]. However, these direct graph-based DR methods only consider the pairwise relationship between data points, while HSI data usually possess complex relationships such as one sample versus multiple samples (different classes) or one class versus multiple samples. Therefore, the pairwise relation cannot discover complex relations in HSI, which limit the discriminability of embedding features for classification [24,25].
To explore the multiple adjacency relationships in high-dimensional data, hypergraph learning has been introduced to discover the complex geometric structure between HSI pixels [26,27]. In [28], discriminant hyper-Laplacian projection (DHLP) was proposed using the hypergraph Laplacian for exploring the high-order geometric relationship of samples. Semi-supervised hypergraph embedding (SHGE) learns the discriminant structure form both labeled and unlabeled data, and it reveals the complex relationships of HSI pixels by building a semi-supervised hypergraph [29]. For analyzing the intrinsic properties of HSI pixels, a hypergraph Laplacian sparse coding method was constituted to capture the similarity among data points within the same hyperedge [30]. In addition, the heterogeneous network is explored to measure the relatedness of heterogeneous objects. Pio et al. [31] introduced heterogeneous networks with an arbitrary structure to evaluate its performance for both clustering and classification tasks. Serafino et al. [32] proposed an ensemble learning approach to classify objects of different classes, which is based on the heterogeneous networks for extracting both correlation and autocorrelation that involve the observed objects.
The aforementioned methods are designed as spectral-based DR methods, in which the spatial relationship between a pixel and its spatial neighborhood is not taken into consideration for DR. Recent investigations show that incorporating spatial information into traditional spectral-based DR methods can further improve the performance of HSI classification [33,34,35,36,37]. Wu et al. presented a spatially adaptive model to extract the spectral and spatial-contextual information, which significantly enhances the land cover classification performance in both accuracy and computational efficiency [38]. Local pixel NPE (LPNPE) [23] and spatial consistency LLE [39] were proposed to reveal the local manifold structure in HSI data by using the distance between different spatial blocks instead Euclidean distance between pixels. As an extension of LPNPE, spatial and spectral information-based RLDE (SSRLDE) tries to maximize the ratio between local spatial-spectral data scatter and global spatial-spectral data scatter for enhancing the representation ability of embedding features [23]. The spatial-spectral coordination embedding (SSCE) method defines a spatial-spectral coordination distance for neighbor selection, and it can reduce the probability that heterogeneous objects are selected as nearest neighbors [40]. Discriminative spectral-spatial margin (DSSM) exploits spatial-spectral neighbors to obtain the low-dimensional embedding via preserving the local spatial-spectral relationship of HSI data [41]. These spatial-spectral DR methods have difficulty discovering the complex relationships in HSI data due to their pairwise nature.
Recently, the spatial information in HSI has been explored to construct a spatial-spectral hypergraph model [42,43,44]. Sun et al. proposed an adaptive hyperedge weight estimation scheme to preserve the prominent hyperedges, which is better for improving the classification accuracy [45]. Yuan et al. introduced a hypergraph embedding model for feature extraction, which can represent higher order relationships [46]. However, these spatial-spectral hypergraph methods are unsupervised and do not use prior information in HSI data, which is not conducive to extract discriminant features for enhancing the classification performance.
Motived by the above limitations, a new hypergraph embedding method termed spatial-spectral joint regularized sparse hypergraph embedding (SSRHE) is proposed for DR of HSI data. SSRHE explores sparse coefficients and label information of pixels to adaptively select neighbors for constructing a regularized sparse intraclass hypergraph and a regularized sparse interclass hypergraph, which can effectively represent the complex relationships in HSI data. Then, a local spatial neighborhood preserving scatter matrix and a total-scatter matrix are computed to preserve the neighborhood structure in spatial domain and the global structure in spectral domain, respectively. Finally, an optimal objective function is designed to extract spatial-spectral discriminant features, which not only preserves the local spatial structure of HSI data, but also compacts the samples belonging to the same class and separates the samples from different classes simultaneously. Therefore, embedding features achieve a good discriminative power for HSI classification.
The rest of this paper is organized as follows. In Section 2, some related works are briefly introduced. Section 3 gives a detailed description of the proposed SSRHE method. In Section 4, experimental results on two real HSI datasets are reported to demonstrate the effectiveness of SSRHE. Finally, Section 5 summarizes this paper and provides some suggestions for future work.
2. Related Works
In this section, we provide a brief review of GE framework and hypergraph model. For convenience, supposed a HSI dataset , where D is the number of spectral bands and N is the number of pixels. The label of denotes and c is the number of land cover classes. The goal of dimensionality reduction is to map , where . For the linear DR methods, Y can be obtained by with a projection matrix .
2.1. Graph Embedding
The graph embedding (GE) framework offers a unified view for understanding and explaining many popular DR algorithms such as PCA, LDA, ISOMAP, LLE, LE, NPE, and LPP. In GE, an intrinsic graph represents a certain desired statistical or geometrical properties of data, and a penalty graph describes some characteristics or relationships that should be avoided. and are the weight matrices of undirected graphs and . and describe the similarity and dissimilarity characteristics between vertices and in and , respectively.
The purpose of GE is to map each vertex of graph into a low-dimensional space that preserves similarities between the vertex pairs. The low-dimensional embedding can be obtained by solving the following objective function:
where is a diagonal matrix with , is the Laplacian matrix of , C is a constant, typically is a diagonal matrix for scale normalization, and it may be the Laplacian matrix of a penalty graph . That is, , where .
2.2. Hypergraph Model
The hypergraph can represent the complex relations between high-dimensional data, and every hyperedge connects multiple vertexes. A hypergraph can be constructed to formulate the relationships among data samples, where is a set of vertices, is a set of hyperedges, and is a diagonal matrix where each diagonal element denotes the weight of hyperedge. Each hyperedge is vested by a weight .
To represent the relationship of , an incidence matrix is denoted as follows:
Furthermore, the degree of hyper-edge and the degree of vertex can be defined as
There is an example to illustrate the hypergraph in Figure 1. A vertex is denoted as a circle (such as , , ). As shown in Figure 1a, the simple graph only holds two vertices per edge, which just describes a single one-to-one relationship (such as and , and and ). In Figure 1b, each hyper-edge is represented by a curve (such as a hypergraph is constructed by , , ), which represents complex multiple relations among pixels. That is, a hypergraph can describe the local neighborhood structure well and preserve the complex relationships within the neighborhood. This hypergraph consists of six vertices and four hyper-edges, and the corresponding incidence matrix is shown in Figure 1c. The incidence matrix intuitively represents affinity relationships between vertices and hyper-edges, the non-zero element in each row indicates that a hyper-edge is associated with the vertex, otherwise the vertex and the hyper-edge are not interrelated.
Figure 1.
An example for hypergraph: (a) simple graph; (b) hypergraph; and (c) incidence matrix H.
However, the hypergraph model only represents complex higher-order relations between pixels in spectral domain, and it fails to consider spatial information in hyperspectral image that limits the discriminant ability of embedding features for land cover classification.
3. SSRHE
To reveal the complex structure in HSI, we propose a new hypergraph learning method called spatial-spectral joint regularized sparse hypergraph embedding (SSRHE) for dimensionality reduction of HSI data. At first, SSRHE constructs a regularized sparse intraclass hypergraph and a regularized sparse interclass hypergraph by exploring sparse representation (SR) and prior knowledge. After that, it exploits the spatial consistency and global structure of HSI by computing a local spatial neighborhood preserving scatter and a total scatter, which brings benefits for combining spatial structure and spectral information for DR. Finally, an optimal objective function is designed to learn a spatial-spectral discriminant projection by minimizing the regularized sparse intraclass hypergraph scatter and the local spatial neighborhood preserving scatter, while maximizing the regularized sparse interclass hypergraph scatter and the total scatter of samples simultaneously. The flowchart of the proposed SSRHE method is shown in Figure 2.
Figure 2.
Flowchart of the proposed SSRHE method.
3.1. The Regularized Sparse Hypergraph Model
To discover the complex structure of HSI data, a hypergraph model is exploited to reveal the intrinsic relations between pixels. However, it remains difficult to choose a proper neighborhood size for constructing hypergraphs. Since sparse representation has natural discriminating power to adaptively reveal the inherent relationship of data [47,48,49], a sparse hypergraph model is designed based on SR theory.
Inspired by the observation that the most compact expression of a certain sample is generally given from similar samples, sparse coefficients are explored to find neighbors adaptively. Suppose is the sparse coefficient matrix of data samples. The sparse coefficients can be calculated as follows:
where denotes the sparse error, represents the sparse coefficients of pixel , which can be optimized by using Alternating Direction Method of Multipliers (ADMM) framework [50,51]. With the sparse coefficient matrix S, a hypergraph can be constructed with the criterion that nodes and are connected in a hyper-edge if sparse coefficient is not equal to 0. Since sparse coefficients can reflect the similarity between data, non-zero correspondence coefficient indicates the correlation between pixels, a large value indicates a high similarity. Compared with Euclidean metric, sparse coefficients can more effectively select neighbors of HSI data.
According to sparse coefficients and label information of samples, we construct a within-class sparse hypergraph and a between-class sparse hypergraph to characterize the intrinsic structure of HSI data. is denoted as the vertex set, and are the sets of intraclass hyper-edge and interclass hyper-edge. and are the weight matrices of hyper-edges of and , respectively.
In , the intraclass hyper-edge is formed by connecting pixel with its corresponding neighbors whose sparse coefficients are non-zero. The subedge weights of and can be defined as
in which is the sparse coefficient between pixel and and parameter () is used to enhance the contribution of samples from the same class for improving the discriminant power. Figure 3 illustrates the construction of sparse hypergraph on a simple classification model. As shown in Figure 3, a simple graph considers only pairwise relation between two observed samples, while the sparse hypergraph through sparse representation can select neighbors of samples adaptively and represent complex multiple relations among HSI pixels.
Figure 3.
A simple two classification model to explain the construction of sparse hypergraph.
The weight matrices corresponding to hyper-edge regions are defined as
The incidence matrix of intraclass hypergraph is calculated by
where is the heat-kernel parameter, that is the mean value of pixels in one hyper-edge, denotes the number of vertices in each hyper-edge region.
According to and , the degree of vertex and the degree of intraclass hyper-edge are computed by
For the between-class hypergraph , the between-class incidence matrix is defined as
Based on and , the degree of vertex and interclass hyper-edge can be defined as
In low-dimensional embedding space, the pixels from the same class should be as compact as possible for the intraclass sparse hypergraph, whereas pixels from different classes should be as separated as possible with the interclass sparse hypergraph. Therefore, the objective functions with intraclass and interclass sparse hypergraph constraint are constructed as
where is the intraclass sparse hypergraph Laplacian matrix and denotes the interclass sparse hypergraph Laplacian matrix. , , , , , with respect to , , , , , can be indicated as
Let represent the between-class scatter of . is the within-class scatter of . Then, the mapping matrix can be obtained through dealing with the following optimization problem:
In real applications, to avoid the singularity in the case of small samples, the above optimal objection can be further extended as
where is a tradeoff parameter. The regularization term is the maximal data variance, which is used to ensure that the diversity of HSI pixels. The diagonal regularization is introduced for overcoming the singularity problem when the number of training samples is small. With regularization, Equation (25) becomes more stable to effectively reserve the useful discriminant information.
3.2. Spatial-Spectral Hypergraph Embedding
Due to the spatial consistency of HSI, the pixels in HSI are usually spatially related, which means the pixels within a small neighborhood usually possess the spatial distribution consistency of ground objects. Therefore, neighborhood pixels can be utilized to learn spatial-spectral combined features.
Suppose that pixel is the center pixel, and T is the spatial neighborhood window. is the spatial coordinate of pixel in the image, and the spatial neighborhood set with the window size T (positive odd) can be recorded as
where , corresponds to the m-th pixel in the spatial neighborhood. has a total of pixels. Thus, the distance measure in the spatial neighborhood block can be defined as
where , which measures similarity between central pixel and its spectral-spatial neighbors. For all training samples in HSI data, the local spatial neighborhood preserving scatter matrix is calculated as follows:
To preserve the global structure of samples, a total scatter matrix is defined as
where is the mean of all samples.
To extract low-dimensional spatial-spectral joint features, a objective function should be designed to preserve the local neighborhood as well as compact the samples with interclass hypergraph and separate the samples with interclass hypergraph simultaneously. Therefore, Equations (25), (28), and (29) are transformed into the following optimization function:
The above optimal function can be further simplified as
According to Lagrange multiplier method, Equation (31) can be solved by
where is an eigenvalue set. With the d largest eigenvalues corresponding eigenvector, the optimal projection matrix can be denoted as . In the low-dimensional embedded space, the spatial-spectral embedding of test sample can be given as follows:
In summary, SSRHE compacts the samples from the same class and spatial neighborhood while separates interclass samples, and the embedding features possess stronger discriminative power that ensures good classification performance. The steps of the proposed SSRHE method are shown in Algorithm 1.
| Algorithm 1 SSRHE. |
Input: HSI dataset , corresponding class label set , tradeoff parameters , , weighted coefficient , spatial neighborhood size T, reduced dimensionality d.
|
In this paper, we adopt big O notation to analyze the computational complexity of SSRHE. Spatial neighborhood size and the number of sparse iterations are denoted as T and t, respectively. The sparse coefficient matrix is computed with the cost of . The weight of intraclass hyperedge and the weight of interclass hyperedge both take . The intraclass incidence matrix and the interclass incidence matrix are both calculated with . The cost of diagonal matrices , , , , , and are ). The intraclass hyper-Laplacian matrix and the interclass hyper-Laplacian matrix both cost . The local spatial neighborhood preserving scatter matrix takes . The extended total scatter matrix costs . The costs of matrix and are both . costs . It takes to deal with the generalized problem of Equation (32). The total time complexity of SSRHE is .
4. Experimental Results and Discussion
Some experiments were performed on two real HSI datasets to verity the effectiveness of the proposed method, and several state-of-the-art DR methods were compared with SSRHE in the experiments.
4.1. HSI Datasets
Indian Pines Dataset: This hyperspectral image is a scene of northwest Indiana collected by an airborne imaging spectrometer sensor in 1992. It consists of pixels and 220 spectral bands that cover the wavelength from 400 to 2450 nm. After removing the noise and water absorption bands, 200 spectral bands remained for use in the experiments. This image contains total 16 land cover types. The dataset in false color and its corresponding ground truth map are shown in Figure 4.
Figure 4.
Indian Pines hyperspectral image: (a) HSI in false color; (b) ground truth; and (c) spectral curve. (Note that the number of samples for each class is shown in brackets.)
University of Pavia (PaviaU) Dataset: This hyperspectral image covering the area of Pavia University in northern Italy was collected by ROSIS sensor in 2002. The spatial size of this image is pixels and the number of spectral band is 115. Due to the atmospheric affection, 12 bands were discarded and the remaining 103 spectral bands were used for experiment. The false color and ground truth map for this scene are shown in Figure 5.
Figure 5.
PaviaU hyperspectral image: (a) HSI in false color; (b) ground truth; and (c) spectral curve. (Note that the number of samples for each class is shown in brackets.)
4.2. Experimental Setup
In experiments, each HSI dataset was randomly divided into training and test sets. The training samples were utilized to learn a feature extracted model to obtain their low-dimensional embedding features. After that, the classifier was applied to obtain the class labels of test samples. The overall classification accuracy (OA), average classification accuracy of each class (AA), and kappa coefficient (KC) were employed to evaluate classification results.
To demonstrate the effectiveness of the proposed method, we compared SSRHE with four spectral-based approaches, LPP, LDA, MFA, and RLDE; two spatial-spectral combined approaches, LPNPE and SSCE; and a hypergraph method, DHLP. For all approaches, we tuned the parameters by cross-validation to achieve good results. In LPNPE, the size of spatial-spectral window was set to 15 for the PaviaU data and 11 for the Indian Pines data. In SSCE, the window scale and the local neighbor size were both set to 5 on both datasets. For RLDE and MFA, the intraclass neighbor size was set to 3 and 8, while the interclass neighbor size was set to 5 and 60, respectively. The number of nearest neighbor in DHLP was selected as 9.
The nearest neighbor (1-NN) classifier was applied for classification. Each experiment was randomly repeated ten times. All experiments were performed on a personal computer with i5-4590 central processing unit, 8-G memory, and 64-bit Windows 7 using MATLAB 2014a.
4.3. Parameters Selection
For the proposed SSRHE algorithm, there are four important parameters: spatial neighborhood size T, weighted coefficient , tradeoff parameters and . To select the optimal parameters, we performed some parameters selection experiments on the Indian Pines and PaviaU datasets. In each experiment, we randomly selected five samples from each class as training set and the remaining samples were used for testing.
To explore the effect of tradeoff parameters and , we turned them with a set of . Figure 6 and Figure 7 show the classification accuracies with respect to and on two datasets. As shown in Figure 6, an increasing produced subtle change of OAs under a fixed . The OAs first varied slightly and then declined significantly with the increase of on Indian Pines dataset, since a too large led to the algorithm losing the property of spatial-domain in HSI and the obtained features might fail to represent the intrinsic structure of hyperspectral image. For PaviaU dataset, the OAs first increased with increasing and , and then tended to decline. To balance the influence of spectral and spatial information for classification, we set parameters and for the Indian Pines dataset, and and for PaviaU dataset based on the results presented in Figure 6 and Figure 7.
Figure 6.
The OAs of SSRHE with different and on Indian Pines dataset.
Figure 7.
The OAs of SSRHE with different and on PaviaU dataset.
The parameters T and have a significant influence on the discrimination of SSRHE. The former is defined to determine the size of spatial neighborhood and the latter is exploited to adjust the compactness of the interclass neighbors. To analyze the influence of two parameters for SSRHE, we performed two experiments with respect to T and , and the corresponding experimental results on two HSI datasets are shown in Figure 8. In Figure 8a, it is clear that the OAs first quickly increased with the growth of T on two dataset because a larger spatial neighborhood was beneficial to preserving more useful spatial-domain information in HSI. However, if the value of T was too large, it might bring pixels from other classes into training model, as well as greatly increase computational complexity. According to the results in Figure 8a, the size of spatial neighborhood T was set to 7 for Indian Pines dataset and 15 for PaviaU dataset. As shown in Figure 8b, with the increasing of , the OAs improved until reaching a peak value due to further enhancing the compactness of intraclass samples in the embedding space. While kept increasing, the OAs declined since too large weakened the ability to extract between-class features. Therefore, we set for the Indian Pines dataset and for the PaviaU dataset.
Figure 8.
OAs with respect to parameters T and on different datasets: (a) the OAs of T; and (b) the OAs of .
4.4. Investigation of Embedding Dimension
To explore how the embedding dimension d affects the performance of the proposed SSRHE method, thirty samples were randomly selected from each class for training, and d was tuned from 5 to 50 with an interval of 5. Figure 9 shows the classification accuracy of SSRHE and other DR methods with different embedding dimensions on Indian Pines and PaviaU datasets. According to Figure 9, the OAs firstly rapidly rose and then tended to stabilize with an increasing d, because embedding features with high dimension contained more useful information for training, while too large dimension might lead to information saturation. The classification accuracies of the most DR methods were better than the RAW method since these methods can remove the redundant information in HSI. The proposed SSRHE achieved the best classification performance in comparison with other methods on two HSI datasets. The reason is that SSRHE discovered the complex relationships between samples by hypergraph model and fused the useful spatial information for extracting discriminant features for classification.
Figure 9.
OAs with different embedding d: (a) Indian Pines dataset; anf (b) Pavia U dataset.
4.5. Investigation of Classification Performance
To evaluate the performance of the proposed method under different training conditions, we selected samples per class for training, and 1-NN classifier was utilized to classify the remaining samples. Each experiment was repeated ten times, the averaged OA with standard deviations (STDs) and kappa coefficient (KC) of all DR methods on Indian Pines and PaviaU datasets are shown in Table 1 and Table 2.
Table 1.
Classification results using different methods with different classifiers for the Indian pines dataset. (OA ± STDs (%) (KC)).
Table 2.
Classification results using different methods with different classifiers for the PaviaU dataset. (OA ± STDs (%) (KC)).
As shown in Table 1 and Table 2, for all methods, the OA and KC greatly improved and the STDs decreased with the increasing of training samples, because a large number of training samples usually provide more useful information for training. The spatial-spectral combined methods, SSCE, LPNPE and SSRHE, produced better classification results than the spectral-based methods, because they exploit the spatial and spectral information in HSI to improve the representation ability of extracted features. Among these spectral-based methods, DHLP presented better classification performance under most training conditions, since it applies the hypergraph learning model to discover the intrinsic complex relationship between samples. Compared with all the other methods, SSRHE had better discriminant performance with different proportions of training samples on two HSI datasets, especially with small size of training set. SSRHE utilizes the hypergraph framework to discover the complex multivariate relationships between interclass samples and intraclass samples, and computes two spatial neighborhood scatters to reveal the spatial correlation between each pixels in HSI, which further enhances the discriminating power of low-dimensional features.
To explore the classification accuracy of the proposed method for each class, we randomly selected a percentage of samples from each class as the training set while the remaining samples were used for testing. In experiments, we set the training percent to 3% for the Indian Pines dataset and 5% for PaviaU dataset. Table 3 and Table 4 show the classification results of different DR methods on Indian Pines and PaviaU datasets, and the corresponding classification maps are displayed in Figure 10 and Figure 11.
Table 3.
Classification results of different classifiers on the Indian Pines dataset.
Table 4.
Classification results of different classifiers on the PaviaU dataset.
Figure 10.
Classification maps of different DR methods with 1-NN on the Indian Pines dataset. Note that the time cost of corresponding DR algorithms is marked in bracket.
Figure 11.
Classification maps of different DR methods with 1-NN on the PaviaU dataset. Note that the time cost of corresponding DR algorithms is marked in bracket.
According to Table 3 and Table 4, SSRHE obtained better classification results than other methods in most classes, and it achieved the best OA, AA and KC on two HSI datasets. This indicates that SSRHE possessed stronger ability to reveal the intrinsic complex geometric relations between samples and extract more useful the spatial-spectral combined features for classification. As shown in Figure 10 and Figure 11, the spatial-spectral methods generally produced smoother classification maps than traditional spectral-based methods, which demonstrates that the spatial information in HSI is effective for improving classification performance. Moreover, the classification maps of SSRHE possessed fewer misclassified pixels, especially in the areas Grass/Pasture, Wheat, and Stone-steel towers for Indian Pines dataset and Asphalt, Meadows, and Shadows for PaviaU dataset. In terms of running time, SSRHE cost more time than traditional spectral-based methods because it needs to build models in both spatial domain and spectral domain of HSI data. Compared with other spectral-spatial algorithms, the running time of SSRHE did not increase significantly and was far less than SSCE. Thus, SSRHE was more effective than other DR methods to weaken the influence of noisy points and develop the discrimination power of embedding features.
5. Conclusions
In this paper, a new dimensionality reduction method SSRHE is proposed based on sparse hypergraph learning and spatial-spectral information for HSI. SSRHE computes two spatial neighborhood scatters to reveal the spatial correlation between each pixels in HSI, and it also constructs two discriminant hypergraphs by sparse representation to discover the complex multivariate relationships between interclass samples and intraclass samples, respectively. Based on the spatial neighborhood scatters and the Laplacian scatters of hypergraphs, a spatial-spectral combined objective function is designed to obtain an optimal projection matrix mapping original HSI data to low-dimensional embedding space where the intrinsic spatial and spectral properties are well preserved. some experiments were conducted on Indian Pines and PaviaU datasets to demonstrate that the proposed method possesses better performance compared with some existing state-of-the-art DR methods. In the future, we will optimize the proposed algorithm to reduce the running time and extend it to other related fields such as multi-spectral images and very-high resolution images, and the sparse hypergraph model in this paper can be used in some application domains with high-dimensional data such as face images, gene expression and radiomics features. Furthermore, It would also be interesting to study integrating hypergraph learning and heterogeneous information networks when exploring high-dimensional data with multiple adjacency relationships.
Author Contributions
H.H. was primarily responsible for mathematical modeling and manuscript writing. M.C. contributed to the experimental design and experimental analysis. Y.D. provided important suggestions for improving the paper.
Funding
This work was supported in part by The Basic and Frontier Research Programmes of Chongqing under Grants cstc2018jcyjAX0093 and cstc2018jcyjAX0633, the Chongqing University Postgraduates Innovation Project under Grants CYB18048 and CYS18035, and the National Science Foundation of China under Grant 41371338.
Acknowledgments
The authors would like to thank the anonymous reviewers and associate editor for their valuable comments and suggestions to improve the quality of the paper.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Liao, D.; Qian, Y.; Tang, Y.Y. Constrained manifold learning for hyperspectral imagery visualization. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 1, 1213–1226. [Google Scholar] [CrossRef]
- Yuan, Q.Q.; Zhang, Q.; Li, J.; Li, J.; Zhang, L.P. Hyperspectral image denoising employing a spatial-spectral deep residual convolutional neural network. IEEE Trans. Geosci. Remote Sens. 2019, 57, 1205–1218. [Google Scholar] [CrossRef]
- Liu, Q.; Sun, Y.; Hang, R.; Song, H. Spatial-spectral locality-constrained low-rank representation with semi-supervised hypergraph learning for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 4171–4182. [Google Scholar] [CrossRef]
- Zhang, X.R.; Gao, Z.Y.; Jiao, L.C.; Zhou, H.Y. Multifeature hyperspectral image classification with local and nonlocal spatial information via markov random field in semantic space. IEEE Trans. Geosci. Remote Sens. 2016, 56, 1409–1424. [Google Scholar] [CrossRef]
- Xu, Y.H.; Zhang, L.P.; Du, B.; Zhang, F. Spectral-spatial unified networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5893–5909. [Google Scholar] [CrossRef]
- Kumar, B.; Dikshit, O. Spectral contextual classification of hyperspectral imagery with probabilistic relaxation labeling. IEEE Trans. Cybern. 2017, 47, 4380–4391. [Google Scholar] [CrossRef] [PubMed]
- Hang, R.; Liu, Q.; Sun, Y.; Yuan, X.; Pei, H.; Plaza, J.; Plaza, A. Robust matrix discriminative analysis for feature extraction from hyperspectral images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 2002–2011. [Google Scholar] [CrossRef]
- Luo, F.L.; Huang, H.; Ma, Z.Z.; Liu, J.M. Semi-supervised sparse manifold discriminative analysis for feature extraction of hyperspectral images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6197–6221. [Google Scholar] [CrossRef]
- Xu, J.; Yang, G.; Yin, Y.F.; Man, H.; He, H.B. Sparse representation based classification with structure preserving dimension reduction. Cogn. Comput. 2014, 6, 608–621. [Google Scholar] [CrossRef]
- Zabalza, J.; Ren, J.C.; Yang, M.Q.; Zhang, Y.; Wang, J.; Marshall, S.; Han, J.W. Novel folded-PCA for improved feature extraction and data reduction with hyperspectral imaging and SAR in remote sensing. ISPRS J. Photogramm. Remote Sens. 2014, 93, 112–122. [Google Scholar] [CrossRef]
- Li, W.; Prasad, S.; Fowler, J.E. Noise-adjusted subspace discriminant analysis for hyperspectral imagery classification. IEEE Geosci. Remote Sens. Lett. 2013, 10, 1374–1378. [Google Scholar] [CrossRef]
- Roweis, S.T.; Saul, L.K. Nonlinear dimensionality reduction by locally linear embedding. Science 2000, 290, 2323–2326. [Google Scholar] [CrossRef]
- Chen, C.; Zhang, L.; Bu, J.; Wang, C.; Chen, W. Constrained laplacian eigenmap for dimensionality reduction. Neurocomputing 2010, 73, 951–958. [Google Scholar] [CrossRef]
- Bachmann, C.M.; Ainsworth, T.L.; Fusina, R.A. Exploiting manifold geometry in hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2005, 43, 441–454. [Google Scholar] [CrossRef]
- He, X.F.; Cai, D.; Yan, S.C.; Zhang, H.J. Neighborhood preserving embedding. In Proceedings of the Tenth IEEE International Conference on Computer Vision (ICCV’05), Beijing, China, 17–21 October 2005; Volume 2, pp. 1208–1213. [Google Scholar]
- Zheng, Z.L.; Yang, F.; Tan, W.; Jia, J.; Yang, J. Gabor feature-based face recognition using supervised locality preserving projection. Signal Process. 2007, 87, 2473–2483. [Google Scholar] [CrossRef]
- Huang, H.; Yang, M. Dimensionality reduction of hyperspectral images with sparse discriminant embedding. IEEE Trans. Geosci. Remote Sens. 2015, 53, 5160–5169. [Google Scholar] [CrossRef]
- Jiang, X.; Song, X.; Zhang, Y.; Jiang, J.; Gao, J.; Cai, Z. Laplacian regularized spatial-aware collaborative graph for discriminant analysis of hyperspectral imagery. Remote Sens. 2019, 11, 29. [Google Scholar] [CrossRef]
- Hang, R.; Liu, Q. Dimensionality reduction of hyperspectral image using spatial regularized local graph discriminant embedding. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 3262–3271. [Google Scholar] [CrossRef]
- Feng, F.B.; Li, W.; Du, Q.; Zhang, B. Dimensionality reduction of hyperspectral image with graph-based discriminant analysis considering spectral similarity. Remote Sens. 2017, 9, 323. [Google Scholar] [CrossRef]
- Luo, F.L.; Huang, H.; Duan, Y.L.; Liu, J.M.; Liao, Y.H. Local geometric structure feature for dimensionality reduction of hyperspectral imagery. Remote Sens. 2017, 9, 790. [Google Scholar] [CrossRef]
- Sugiyama, M. Dimensionality reduction of multimodal labeled data by local fisher discriminant analysis. J. Mach. Learn. Res. 2007, 8, 1027–1061. [Google Scholar]
- Zhou, Y.C.; Peng, J.T.; Philip Chen, L.C. Dimension reduction using spatial and spectral regularized local discriminant embedding for hyperspectral image classification. IEEE Trans. Geosic. Remote Sens. 2015, 53, 1082–1095. [Google Scholar] [CrossRef]
- Huang, Y.; Liu, Q.; Zhang, S.; Metaxas, D. Image retrieval via probabilistic hypergraph ranking. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 3367–3383. [Google Scholar]
- Liu, Q.; Sun, Y.; Wang, C.; Liu, T.; Tao, D. Elastic net hypergraph learning for image clustering and semi-supervised classification. IEEE Trans. Image Process. 2017, 26, 452–463. [Google Scholar] [CrossRef] [PubMed]
- Yu, J.; Tao, D.; Wang, M. Adaptive hypergraph learning and its application in image classification. IEEE Trans. Image Process. 2012, 21, 3262–3272. [Google Scholar]
- Wang, W.H.; Qian, Y.T.; Tang, Y.Y. Hypergraph-regularized sparse NMF for hyperspectral unmixing. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 681–694. [Google Scholar] [CrossRef]
- Huang, S.; Yang, D.; Ge, Y.X.; Zhang, X.H. Discriminant hyper-Laplacian projection and its scalable extension for dimensionality reduction. Neurocomputing 2016, 173, 145–153. [Google Scholar] [CrossRef]
- Du, W.B.; Qiang, W.W.; Lv, M.; Hou, Q.L.; Zhen, L.; Jing, L. Semi-supervised dimension reduction based on hypergraph embedding for hyperspectral images. Int. J. Remote Sens. 2018, 39, 1696–1712. [Google Scholar] [CrossRef]
- Gao, S.H.; Tsang, I.W.H.; Chia, L.T. Laplacian sparse coding, hypergraph Laplacian sparse coding, and applications. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 92–104. [Google Scholar] [CrossRef] [PubMed]
- Pio, G.; Serafino, F.; Malerba, D.; Ceci, M. Multi-type clustering and classification from heterogeneous networks. Inf. Sci. 2018, 425, 107–126. [Google Scholar] [CrossRef]
- Serafino, F.; Pio, G.; Ceci, M. Ensemble learning for multi-type classification in heterogeneous networks. IEEE Trans. Knowl. Data Eng. 2018, 30, 2326–2339. [Google Scholar] [CrossRef]
- Liao, W.Z.; Mura, M.D.; Chanussot, J.; Pizurica, A. Fusion of spectral and spatial information for classification of hyperspectral remote sensed imagery by local graph. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 9, 583–594. [Google Scholar] [CrossRef]
- Zhao, J.; Zhong, Y.F.; Shu, H.; Zhang, L.P. High-resolution image classification integrating spectral-spatial-location cues by conditional random fields. IEEE Trans. Image Process. 2016, 25, 4033–4045. [Google Scholar] [CrossRef]
- Cao, J.; Wang, B. Embedding learning on spectral-spatial graph for semi-supervised hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1805–1809. [Google Scholar] [CrossRef]
- Zhao, W.; Du, S. Spectral-spatial feature extraction for hyperspectral image classification: A dimension reduction and deep learning approach. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4544–4554. [Google Scholar] [CrossRef]
- Li, Y.; Zhang, H.; Shen, Q. Spectral-spatial classification of hyperspectral imagery with 3D convolutional neural network. Remote Sens. 2017, 9, 67. [Google Scholar] [CrossRef]
- Wu, Z.B.; Shi, L.L.; Li, J.; Wang, Q.C.; Sun, L.; Wei, Z.H.; Plaza, J.; Plaza, A.J. GPU parallel implementation of spatially adaptive hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 1131–1143. [Google Scholar] [CrossRef]
- Mohan, A.; Sapion, G.; Bosch, E. Spatially coherent nonlinear dimensionality reduction and segmentation of hyperspectral images. IEEE Geosci. Remote Sens. Lett. 2007, 4, 206–210. [Google Scholar] [CrossRef]
- Huang, H.; Zheng, X.L. Hyperspectral image land cover classification algorithm based on spatial-spectral coordination embedding. Acta Geod. Cartogr. Sin. 2016, 4, 964–972. [Google Scholar]
- Feng, Z.X.; Yang, S.Y.; Wang, S.G.; Jiao, L.C. Discriminative spectral-spatial margin-based semisupervised dimensionality reduction of hyperspectral data. IEEE Geosci. Remote Sens. Lett. 2015, 12, 224–228. [Google Scholar] [CrossRef]
- Ji, R.R.; Gao, Y.; Hong, R.C.; Liu, Q.; Tao, D.C.; Li, X.L. Spectral-spatial constraint hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2014, 52, 1811–1824. [Google Scholar]
- Luo, F.L.; Du, B.; Zhang, L.P.; Zhang, L.F.; Tao, D.C. Feature learning using spatial-spectral hypergraph discriminant analysis for hyperspectral image. IEEE Trans. Cybern. 2019, 49, 2406–2419. [Google Scholar] [CrossRef]
- Tang, Y.Y.; Lu, Y.; Yuan, H. Hyperspectral image classification based on three-dimensional scattering wavelet transform. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2467–2480. [Google Scholar] [CrossRef]
- Sun, Y.B.; Wang, S.J.; Liu, Q.S.; Hang, R.L.; Liu, G.C. Hypergraph embedding for spatial-spectral joint feature extraction in hyperspectral images. Remote Sens. 2017, 9, 506. [Google Scholar]
- Yuan, H.L.; Tang, Y.Y. Learning with hypergraph for hyperspectral image feature extraction. IEEE Geosci. Remote Sens. Lett. 2015, 12, 1695–1699. [Google Scholar] [CrossRef]
- Sun, Y.; Hang, R.; Liu, Q.; Zhu, F.; Pei, H. Graph-rRegularized low rank representation for aerosol optical depth retrieval. Int. J. Remote Sens. 2016, 37, 5749–5762. [Google Scholar] [CrossRef]
- Liu, J.; Xiao, Z.; Chen, Y.; Yang, J. Spatial-Spectral graph regularized kernel sparse representation for hyperspectral image classification. ISPRS Int. J. Geo-Inf. 2017, 6, 258. [Google Scholar] [CrossRef]
- Zhang, S.Q.; Li, J.; Li, H.C.; Deng, C.Z.; Antonio, P. Spectral-spatial weighted sparse regression for hyperspectral image unmixing. IEEE Trans. Geosci. Remote Sens. 2018, 56, 3265–3276. [Google Scholar] [CrossRef]
- Sun, W.W.; Zhang, L.F.; Zhang, L.P.; Yenming, M.L. A dissimilarity-weighted sparse self-representation method for band selection in hyperspectral imagery classification. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2016, 9, 4374–4388. [Google Scholar] [CrossRef]
- Luo, M.N.; Zhang, L.L.; Liu, J.; Zheng, Q.H. Distributed extreme learning machine with alternating direction method of multiplier. Neurocomputing 2017, 261, 164–170. [Google Scholar] [CrossRef]
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).










