Next Article in Journal
Predicting Regional Outbreaks of Hepatitis A Using 3D LSTM and Open Data in Korea
Previous Article in Journal
A Multichannel Deep Learning Framework for Cyberbullying Detection on Social Media
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hyperspectral Remote Sensing Image Feature Representation Method Based on CAE-H with Nuclear Norm Constraint

1
School of Computer Science and Information Engineering, Harbin Normal University, Harbin 150025, China
2
College of Computer Science and Technology, Mudanjiang Normal University, Mudanjiang 157011, China
3
Department of Computer Science, Harbin Vocational and Technical College, Harbin 150081, China
*
Author to whom correspondence should be addressed.
Electronics 2021, 10(21), 2667; https://doi.org/10.3390/electronics10212667
Submission received: 1 October 2021 / Revised: 27 October 2021 / Accepted: 28 October 2021 / Published: 31 October 2021

Abstract

:
Due to the high dimensionality and high data redundancy of hyperspectral remote sensing images, it is difficult to maintain the nonlinear structural relationship in the dimensionality reduction representation of hyperspectral data. In this paper, a feature representation method based on high order contractive auto-encoder with nuclear norm constraint (CAE-HNC) is proposed. By introducing Jacobian matrix in the CAE of the nuclear norm constraint, the nuclear norm has better sparsity than the Frobenius norm and can better describe the local low dimension of the data manifold. At the same time, a second-order penalty term is added, which is the Frobenius norm of the Hessian matrix expressed in the hidden layer of the input, encouraging a smoother low-dimensional manifold geometry of the data. The experiment of hyperspectral remote sensing image shows that CAE-HNC proposed in this paper is a compact and robust feature representation method, which provides effective help for the ground object classification and target recognition of hyperspectral remote sensing image.

1. Introduction

Hyperspectral remote sensing images are rich in spatial, spectral and radiation information, including hundreds or even thousands of spectral bands, which can fully reflect the subtle features of the surface object spectrum and provide extremely rich information for the extraction of surface object information, which is beneficial to more detailed surface object classification [1,2,3,4]. In recent years, hyperspectral remote sensing images have attracted the attention of many scholars, and have been widely used in ecological monitoring [5,6], medical diagnosis [7,8], military reconnaissance [9] and other important fields. Due to the increase of spectral bands in hyperspectral remote sensing images, the problems of increased dimension and high data redundancy appear [10,11], resulting in the complexity of data processing [12]. In order to alleviate the above problems, it is usually necessary to reduce the dimension of hyperspectral data [13,14]. Hyperspectral data has a specific nonlinear structure in the high dimensional space, and this nonlinear structure is also the area where hyperspectral data are distributed and concentrated in high density [10,15]. The dimensional reduction representation of hyperspectral data can accurately describe the effective information in the data and keep the important information in the data only by maintaining the nonlinear structural relationship in the data [16,17]. Therefore, it is necessary to study the feature representation method that can keep the nonlinear structure relation in hyperspectral data.
Manifold learning algorithm can solve the problem of feature dimensionality reduction in hyperspectral remote sensing image. In many manifold learning algorithms, the local structure of manifold is represented by the local basis of changing direction, that is, the tangent plane of any point on the manifold. In order to construct the global manifold structure or global density, various methods of splicing these local tangent planes are proposed by different algorithms. In the past decade, domestic and foreign scholars have proposed many improved manifold algorithms. For example, Wang et al. proposed an improved ISOMAP algorithm for the analysis of hyperspectral image features. The improved ISOMAP algorithm selects neighborhood according to spectral Angle, avoiding neighborhood instability in high-dimensional spectral space [18]. Wang et al. proposed a method combining UVE and LLE, which used LLE to reduce the dimension of the image composed of effective wavelength, and used partial least squares discriminant analysis to establish the classification model [19]. In order to solve the local tangent space alignment in the adaptability of higher order information loss problems in the manifold, such as Yang et al. proposed a local neighborhood information extraction of the new algorithm optimization [20], through the optimization of the extraction of tangent vector, which can improve higher dimensional nonuniform distribution manifold dimensionality reduction effect, the proposed algorithm can effectively reconstruct density curve of low dimensional coordinates, the low dimensional high-dimensional image has good adaptability. Huang proposed a sparse discriminant embedded retained projection (SDE) [21], which takes advantage of the advantages of sparsity and manifold structure. It not only preserves the sparse reconstruction relation, but also promotes the manifold structure of discriminating data. Pan et al. put forward a kind of based on local keep rule of two-dimensional local projections (2 DLPP) directly from the image matrix to extract the features [22], the algorithm can better simulate the image characteristics inherent in the manifold structure, improves the characteristics of robustness, and reduces the computational complexity and the characteristics of the final dimension, on the recognition accuracy and recognition speed have achieved good results. However, a potentially serious limitation of these manifold learning algorithms is that they are based on local generalization, mainly using training points near points of interest to infer these local tangent planes. In order to overcome the disadvantage of local generalization, Reference [23] suggested neural network algorithm of Contractive Auto-encoder (CAE), the method by changing the main singular vectors of the Jacobian matrix, captured every input point around the local manifold structure, the corresponding singular value specified in related to the corresponding singular vectors direction how many local changes are credible, and keep in high density area of the input space. In order to maintain a smoother nonlinear manifold structure with data in a low-dimensional space, Rifai et al. proposed a method of Higher Order Contractive auto-encoders (CAE-H) based on CAE [24]. Yu et al. proposed a stacked contractive auto-encoder (SCAE) to improve the robustness of feature extraction through unsupervised training and learning [25]. Aamir et al. proposed an improved variant of CAE based on layered architecture following feed forward mechanism named as deep CAE. By encoding and decoding in each layer of CAE, reconstruction errors were reduced, so as to obtain more robust information features [26]. Ng et al. proposed a denoising-contractive auto-encoder (DCAE), which can learn robust feature representations from noisy and sparse feature vectors [27]. Zhang et al. proposed a new ensemble deep contractive auto-encoder (EDCAE), which automatically learns invariant feature representations by designing a variety of different DCAE models [28]. Due to the Jacobian penalty term in DCAE and different characteristics, these models can deal with various noisy data effectively. Finally, an effective EDCAE model is designed using the combination strategy. The above CAE models can effectively solve the invariance of characteristic jitter to varying degrees. However, due to the poor sparsity of the Frobenius norm of CAE model, the low-dimensional manifold structure cannot be effectively described, which reduces the expression ability of effective information in the data.
In view of the problem of high dimensionality and high data redundancy in hyperspectral remote sensing images, the dimensionality reduction representation of existing hyperspectral data cannot effectively maintain the nonlinear structural relationship in the data. Starting from the effective characterization of the nonlinear manifold structure of hyperspectral image data in low-dimensional space, this paper analyzes the Jacobian matrix Frobenius norm approximation of CAE, the geometric interpretation of CAE and the Hession matrix Frobenius norm approximation of CAE-H, and proposes a feature representation method of Higher Order Contractive Auto-Encoders With Nuclear Norm Constraint (CAE-HNC). By introducing Jacobian matrix in the CAE of the nuclear norm constraint, the nuclear norm has better sparsity than the Frobenius norm and can describe the local low dimension of the data manifold better. At the same time, a second-order penalty term is added, which is the Frobenius norm of the Hessian matrix expressed in the hidden layer of the input, encouraging a smoother low-dimensional manifold geometry of the data.

2. Contractive Auto-Encoders

From the point of view of manifold learning, the high-dimensional training data is located on a low-dimensional manifold [23]. Changes in the data correspond to local changes in the manifold (along the direction of the tangent plane), while changes in the data correspond to directions that are orthogonal to the manifold. Therefore, as long as we learn the changes and invariable directions in the data, the manifold structure of the high-dimensional data is also characterized. The goal of Contractive auto-encoders is to learn the manifold structure of high-dimensional data in low-dimensional space. The two driving forces for CAE learning are the contraction penalty term to keep the learned feature constant in all directions (shrinking in all directions), and the reconstruction error term to be able to reconstruct the learned feature back into the input. Therefore, during the learning process, the force of the contraction penalty term makes the change direction in the data (that is, the direction of the manifold tangent plane) able to resist the contraction, which is reflected in the large singular value in its corresponding Jacobian. The directions that do not resist the contractive ability correspond to the invariant directions in the data (orthogonal to the manifold’s direction), and the gradient in Jacobian becomes very small. It can be seen that the shrink autoencoder can effectively describe the low-dimensional manifold structure of the data so as to obtain a more compact data representation.

2.1. CAE Model

Contractive auto-encoders is a kind of regular auto-encoder, whose model structure is consistent with the traditional self-encoder. The main purpose of CAE is to suppress the disturbance of the training sample data in all directions and to achieve the effect of local space contraction by adding a penalty term on the target function of the traditional auto-encoder. The penalty term is the Frobenius norm of Jacobian matrix expressed in the hidden layer of input, whose purpose is to shrink the mapping of feature space near the training data, specifically expressed as follows:
J f ( x ) F 2 = i j ( h j ( x ) x i ) 2 ,  
where, J f ( x ) is Jacobian matrix of f ( x ) , represented as the sum of the partial derivatives of h j ( x ) with respect to x i . · F   is the Frobenius norm of the matrix. h j ( x ) contains j   function groups and x i contains i variables. CAE adds Equation (1) to the loss function as a penalty term to reduce the sensitivity of the model to small changes in input, so as to achieve the purpose of a good system. This formula is used as a penalty term because it has the following characteristics: when the penalty term has a relatively small first-order derivative, it indicates that the hidden layer expression corresponding to the input signal is relatively smooth. Then when the input changes to a certain extent, the hidden layer expression will not change much, which achieves the purpose of being insensitive to the input changes. Therefore, the loss function of CAE can be expressed as follows:
J C A E ( W , b ) = i = 1 N y i x i 2 2 + λ N i = 1 N J f ( x i ) F 2 ,
where W is the weight vector and b is the bias of the loss function, N is the number of training samples, λ is a super parameter to control the strength of the penalty term, and you can choose any value between 0 and 1. The former term of the loss function is to make the reconstruction error as small as possible, so that CAE can obtain all the information of the input signal as far as possible, and the latter term can be regarded as the information that CAE is discarding as much as possible. Therefore, CAE will finally obtain the disturbance information on the training data, making the model invariable to the disturbance. According to Equation (2), the optimization problem can be described as:
min W , b { J C A E ( W , b ) = i = 1 N y i x i 2 2 + λ N i = 1 N J f ( x i ) F 2 }   .
Learning parameters are:
W , b = a r g min W , b { J C A E ( W , b ) = i = 1 N y i x i 2 2 + λ N i = 1 N J f ( x i ) F 2 }   ,
J f ( x ) is the Jacobian matrix of the hidden layer relative to the input x, J f ( x ) = S W ,where S is the diagonalization of the first derivative function of the function, that is, S = d i a g   s ( y ) . When the stochastic gradient descent algorithm is used to solve the parameters, it is necessary to calculate the gradient of each parameter corresponding to the loss function.
CAE can describe the local complex manifold structure around each data point by the singular value decomposition (SVD) of input Jacobian matrix. The corresponding singular value specifies how much local variation is trusted in the direction associated with the corresponding singular vector, while remaining in the dense region of the input space. The penalty term of the CAE loss function excites the insensitivity of h ( x ) in all input space directions. This pressure is balanced by the need for accurate reconstruction, resulting in h ( x ) being essentially sensitive to only a few input directions and requiring training samples to distinguish closed input directions. When J ( x ) contains all the information to calculate the sensitivity of h = f ( x ) to the motion in any input direction, performing SVD produces a more direct orthonormal basis for the direction of information, ranking from the most sensitive to the least sensitive. The subset of the most sensitive directions in this orthogonal basis can be interpreted as a manifold generated from the tangent space at point x .

2.2. Jacobian’s Frobenius Norm Approximation

In order to solve the Frobenius norm of Jacobian expressed by the hidden layer of the input data in the penalty term in CAE, while calculating x f ( x ) 2 2 , this paper uses the method of ref [24] of the following:
x f ( x ) 2 2 = lim 1 σ 2 σ 0 E ϵ ~ N ( 0 , σ 2 I ) [ J f ( x ) J f ( x + ε ) F 2 ] ,
Where   ϵ ~ N ( 0 , σ 2 I ) is an isotropic gaussian distribution with variance of σ 2 . Theoretically, the smaller σ is the more accurate the random approximation. In practice, however, the larger σ is used because it actually allows regularization to explore relatively distant observation points from the data manifold. In fact, the above approximation use n   random sampling approximations, that is:
x J ( x ) F 2 1 n i = 1 n 1 σ 2 J f ( x ) J f ( x + ε i ) F 2 .
With parameter of Jacobian Frobenius norm of the gradient, about parameter Θ regular function is:
R ( Θ ) = 1 2 x J ( x ) F 2 = lim σ 0   E 1 2 σ 2 [ J f ( x ; Θ ) J f ( x + ε ; Θ ) F 2 ] .
It can be approximated as:
R ( Θ ) 1 2 n i = 1 n 1 σ 2 J f ( x ; Θ ) J f ( x + ε i ; Θ ) F 2 .
Purpose is to calculate gradient on Θ . Function R (Θ) about Θ differential for:
d R ( Θ ) 1 n σ 2 i = 1 n ( J f ( x ; Θ ) J f ( x + ε ; Θ ) ) T ( d J f ( x ; Θ ) d J f ( x + ε ; Θ ) ) .

2.3. Geometric Interpretation of CAE

The regular term in CAE encourages the hidden layer to encode h ( x ) in an input space that is insensitive in all directions (i.e., the activation function is saturated), which means that the points in the training set in the input space are not different (the so-called sensitivity is the difference). However, the reconstruction task in CAE needs to be able to identify different points in the training set, and this balance makes h ( x ) sensitive to only a few directions in the input space, so as to be able to distinguish different points near these phases in the training sample set. The geometric interpretation is that these sensitive directions span the locally tangent plane of the manifold. The tangent bundle of a smooth manifold consists of a set of tangent planes along all the sample points on the manifold. Each tangent plane corresponds to a Euclidean coordinate system or chart. In topology, atlas is a collection of such charts. Although the charts collection can form a non-Ou manifold, and each chart is Ou manifold.
Given data set D , h ( x ) , x D satisfies the necessary condition of local injectivity, and consider how to define the local chart around x by the property of h . Since h must be sensitive to the change of a sample x i from itself to one of its adjacent points x j , but not to other changes, we expect this sensitivity to be reflected in the Jacobian matrix J ( x ) = h ( x ) / x spectrum where training is concentrated at each x . Assuming that the rank of J ( x ) is k , h ( x + ϵ v ) and h ( x ) are, ideally, only different if v is the span of a singular vector corresponding to a nonzero singular value of J ( x ) . That is to say, the sensitive direction of the singular vector corresponding to the nonzero singular value of J ( x ) , is that is that when x becomes x , h ( x ) becomes h ( x ) . In fact,   J ( x ) has many smaller eigenvalues. Therefore, SVD decomposition of J ( x ) is used:
J T ( x ) = U ( x ) S ( x ) V T ( x ) .  
Define the partial chart of x . The tangent plane of sample x is defined as:
H x = { x + v v s p a n ( B x ) } .
where B x = { U k ( x ) S k k ( x ) > ϵ } , this is a larger k eigenvalues of the singular vector(columns vectors).The left singular value of Jacobian matrix transpose (gradient matrix) SVD decomposition spans the tangent plane. The coordinates of vector v s p a n ( B x ) on B x are the coordinates of vector v s p a n ( B x ) on the tangent plane H x . Based on the local linear approximation, the atlas described by the encoder function h is defined as:
A = { ( M x , x ) x D , x ( x ¯ ) = B x ( x ¯ x ) } ,
Given training set sample x x , sensitivity refers to h ( x ) h ( x ) , and insensitivity refers to h ( x ) = h ( x ) .

2.4. CAE-H Model and Its Norm Approximation

In order to improve the robustness of the input with small changes, CAE-H improves the objective function on the basis of CAE and adds a second-order penalty term, which is the Frobenius norm of the Hessian matrix expressed in the hidden layer of the input, as follows:
H f ( x ) F 2 = J f ( x ) x F 2 ,
Where, J f ( x ) is Jacobian matrix. The Frobenius norm constraint of Hessian matrix is used in CAE-H to punish curvature and encourage smoother manifold structure. From the above equation, the objective function of the CAE-H can be obtained as follows:
J C A E H ( W , b ) = i = 1 N y i x i 2 2 + λ N i = 1 N J f ( x i ) F 2 + γ N i = 1 N H f ( x i ) F 2 .
Because the second derivative is added, the complexity of the model is greatly increased. The second derivative is converted into the first derivative to reduce the computational complexity. The Hessian Frobenius norm H f ( x ) F 2 can be approximated as:
H f ( x ) F 2 = lim σ 0 1 σ 2 E [ J f ( x ) J f ( x + ε ) F 2 ] .
Therefore, the final objective function of CAE-H is as follows:
J C A E H ( W , b ) = i = 1 N y i x i 2 2 + λ N i = 1 N J f ( x i ) F 2 + γ E [ J f ( x ) J f ( x + ε ) F 2 ] .

3. CAE-H Based on Nuclear Norm Constraint

Part 2 introduces the shortcomings of the Frobenius norm constrained optimization model in CAE. The advantage of the nuclear norm is that it has a sparse low-dimensional manifold direction, and the nuclear norm is similar to the L 1 norm of singular value vectors. L 1 norm has good sparsity, the model is easy to be interpreted, and the local low dimension of the characterizing manifold is relatively good, which can completely retain the geometric characteristics of the original data, so that the information of the original data is not lost, and the resistance to noise is strong. Therefore, this section proposes a CAE-HNC.

3.1. Definition of Nuclear Norm and Its Jacobian Approximation

The nuclear norm function F , also known as trace norm, is defined as:
F ( A ) = A * = T r ( A T A ) = T r ( B ) ,
where   B = A T A . The nuclear norm is a convex function which can be optimized effectively and is the best convex approximation of the rank function on the matrix identity sphere with the norm less than 1. When matrix variables are symmetric and positive semidefinite, this heuristic is equivalent to the tracking heuristic often used in control systems. The nuclear norm heuristic has been observed in practice to produce very low-rank solutions, but the theoretical representation of when it produces the minimum rank solution has not been obtained before.
Theorem 1.
The Jacobian matrix of the nuclear norm F(A) defined in Equation (17) with respect to A can be approximated as:
J F ( A ) = D T [ B ; B ] A T + D [ B ; B ] A T ,
where D [ B ; B ] = ( B T B ) 1 , is Kronecker sum. If A n × n , B m × m , Kronecker sum is defined as:
A B = A I m + I n B ,
And
B = V D 1 2 V T .
Proof of Theorem 1.
Since the differential of the nuclear normal function can be expressed as:
d F = T r ( D [ B ; B ] d B ) = T r ( D [ B ; B ] ( d A T A + A T d A ) ) = T r ( D [ B ; B ] d A T A ) + T r ( D [ B ; B ] A T d A ) = T r ( d A T A D [ B ; B ] ) + T r ( D [ B ; B ] A T d A ) = T r ( D T [ B ; B ] A T d A ) + T r ( D [ B ; B ] A T d A )
Therefore, the Jacobian approximation of the nuclear norm function is:
J F ( A ) = D T [ B ; B ] A T + D [ B ; B ] A T
where D [ B ; B ] = ( B T B ) 1 and is Kronecker sum. If A n × n , B m × m , then Kronecker sum is defined as:
A B = A I m + I n B
And
B = V D 1 2 V T
 □

3.2. The Robust CAE-H with Nuclear Norm Constraints

Nuclear norm with sparse is the advantage of low dimensional manifold directions, and nuclear norm is similar to the singular value vector of L 1 norm, and L 1 norm has good sparse, the model is easy to explain, and characterization of the manifold local low dimensional relatively well, which can keep complete geometric characteristics of original data, the original data information is not lost, the stronger noise resistance. CAE-HNC is still a regular auto-encoder, whose main purpose is to suppress the disturbance of the training sample data in all directions, and to better achieve the local space contraction effect by adding a Jacobian nuclear norm constraint penalty term on the objective function, as follows:
J f ( x ) * = T r ( A T A ) = T r ( B ) .
The Frobenius norm of Hessian can be used to describe the geometric structure of the manifold with smoother data. Combining Equation (21), the objective function of the high-order contraction auto-encoder with nuclear norm constraint can be obtained as follows:
J C A E H N C ( W , b ) = i = 1 N y i x i 2 2 + λ N i = 1 N J f ( x i ) * + γ N i = 1 N H f ( x i ) F 2 .
The high-order contraction auto-encoder with nuclear norm constraint is transformed into the following optimization problem:
min W , b { J C A E H N C ( W , b ) = i = 1 N y i x i 2 2 + λ N i = 1 N J f ( x i ) * + γ N i = 1 N H f ( x i ) F 2 } .
The learning parameters of solution are:
W , b = a r g min W , b { J C A E H N C ( W , b ) = i = 1 N y i x i 2 2 + λ N i = 1 N J f ( x i ) * + γ N i = 1 N H f ( x i ) F 2 } .

3.3. Solution Algorithm

Aiming at the specific problem of feature representation of wetland hyperspectral remote sensing image, this part proposes a feature representation method of CAE-HNC. The regular terms of the Jacobian matrix with nuclear norm constraint are designed, and the sparse representation of the Jacobian matrix is realized, the model is easier to understand, and the local low dimension of manifolds is easier to describe. The parameter updating process in the solution algorithm of method CAE-HNC is as follows Algorithm 1:
Algorithm 1: High order contractive auto-encoder with nuclear norm constrain
Input: X ;
Output: Θ K = { W K , b h K , b r K , U K } .
1: Initialization: Θ = { W 0 , b h 0 , b r 0 , U 0 } ;
2:  For t = 0 , 1 , 2 , , K ;
3:    Update W W t + 1 = W t η w φ ( { W t , b h t , b r t , U t } ) ;
4:    Update b h     b h t + 1 = b h t η b h φ ( { W t + 1 , b h t , b r t , U t } ) ;
5:    Update b r     b r t + 1 = b r t η b r φ ( { W t + 1 , b h t + 1 , b r t , U t } ) ;
6:     Update   U     U t + 1 = P ψ * β ( β J f T ( X ; W t + 1 , b h t + 1 , b r t + 1 ) + U t ) ;
7: End

4. Experimental Results and Analysis

Three groups of hyperspectral remote sensing images were selected in this paper. In order to verify the effectiveness of the proposed algorithm, experimental simulation analysis was carried out on three groups of hyperspectral images, and feature extraction and representation of different algorithms were performed using hyperspectral data such as CAE, CAE-H, SCAE, DCAE, and CAE-HNC, then classification comparison was conducted to verify the robustness of the proposed method. The experiment was run on a Windows (Inter(R) Celeron(R) G4900 [email protected] 3.10 GHz 8GB RAM) 64-bit operating system, using Matlab2016 and ENVI5.3 for simulation verification.

4.1. The Robust CAE-H

In this paper, simulation experiments are carried out on three groups of data, all of which are obtained by spatial and spectral degradation simulation using hyperspectral images, as shown in Figure 1.
The first set of data was collected by an AVIRIS sensor over the Indian Pine Proving Ground in northwestern Indiana. The image size was 145 × 145, containing 21,025 pixels. Out of the original 224 spectral bands, 24 water absorption bands (104–108,150–163 and 220) were removed, and a total of 200 bands participated in the experiment. With a wavelength range of 0.4 to 2.5 μm and a spatial resolution of 20 m, most areas of the image represent fields with a variety of crops, while the rest represent forests and dense vegetation. The map contains 16 different types of crops, including corn, grass, soybean, forest and so on.
The second set of data is a hyperspectral image acquired by ROSIS sensor over the University of Pavia, Italy. The image size is 610 × 340 and contains 220,740 pixels. There are 103 spectral bands with a wavelength range of 0.4–0.9 μm and a spatial resolution of 1.3 m after removing 12 bands affected by noise. The corresponding map of real features contains 9 types of features, including trees, Asphalt roads, Bricks, Meadows, etc.
The third set of data, also obtained by AVIRIS sensor, is located in the Salinas Valley region of California. Image size is 512 × 217, contains 111,104 pixels, removing water absorption band to participate in a total of 204 regions of the spectrum experiment, the wavelength range of 0.4–2.5 microns, the spatial resolution of 3.7 m, the corresponding real line map contains 16 class features, including Fallow, Celery, etc.
The experimental data have been radiometric correction and geometric registration. The number of available samples for each group of data is shown in Table 1.

4.2. Classification Results of Hyperspectral Image Features

In order to verify the effectiveness of CAE-HNC feature representation method based on the kernel norm constraint proposed in this paper, this paper uses the CAE-HNC algorithm to extract the feature representation information of hyperspectral images, and uses the SVM classifier to realize the feature classification. The kernel function of SVM adopts gaussian RBF kernel function, the kernel function parameter = 0.3, the penalty factor C = 50. In order to test the robustness of the proposed algorithm, 5% and 10% of the hyperspectral data were selected as training samples, and CAE-HNC method was used for data feature extraction. Then 40% of the extracted feature data were used as training samples and 60% as test samples for ground object classification in SVM.
The proposed CAE-HNC method is compared with different methods, including Contractive Auto-encoder (CAE), High-order Contractive auto-encoder (CAE-H), Stacked Contractive Auto-encoder (SCAE) and Denoising Contractive Auto-encoder (DCAE).The experimental results were measured by the Overall Accuracies (OA) A and Kappa coefficient (R) of ground object classification in hyperspectral images. For each group of experiments, the experiment was repeated 10 times, and then the average value of the 10 results was selected for comparison. The experimental results are shown in Table 2, where the results shown in bold are the relatively highest accuracy and Kappa coefficient.
Table 2 shows that the proposed CAE-HNC method can well describe the low-dimensional manifold structure of local features of hyperspectral images, with good robustness and satisfactory classification results.
Figure 2 is the local region magnification of several different feature representation methods on the University of Pavia (UP) data set after SVM classification results, in which the training sample is 5%. As can be seen from Figure 2, SCAE method (Figure 2d), DCAE method (Figure 2e) and CAE-HNC feature representation method (Figure 2f) are superior to CAE (Figure 2b) and CAE-H method (Figure 2c) in overall classification effect. Figure 2d–f can better eliminate salt and pepper noise in classification mapping, so that the image becomes clearer and smoother. In fact, this is the result of effectively improving the phenomena of “different objects with the same spectrum” and “the same object with different spectra”. For ground objects with similar spectrum, such as grass and trees, buildings composed of gravel and bricks, there is a certain degree of misclassification. The CAE-HNC method proposed in this paper has more accurate classification results in the red area of Figure 2. The contours of trees and asphalt buildings are clearer and more in line with the real ground features.

5. Conclusions

Hyperspectral data has a special nonlinear structure in high dimensional space, and this nonlinear structure is also the area where hyperspectral data are distributed and concentrated in high density. This kind of nonlinear structure relation is not effectively described in the dimensionless representation of hyperspectral data, resulting in poor robustness of the feature representation of hyperspectral image. Therefore, this article focus on the study of the characteristics of wetland hyperspectral image said learning method, aimed at effectively depict hyperspectral data in low dimensional space nonlinear manifold structure, analyzes the Jacobian matrix of CAE Frobenius norm approximation, CAE geometric interpretation and CAE-H Frobenius norm Hession matrix approximation, is proposed based on nuclear norm constraint of high order contractive auto-encoder (CAE-HNC) feature representation. By introducing Jacobian matrix in the CAE of the nuclear norm constraint, the sparse feature is enhanced, the low-dimensional manifold structure of the data is characterized effectively, the Frobenius norm second-order penalty term of the Hessian matrix is added, and the smoother low-dimensional manifold geometry is encouraged. Experiments on hyperspectral images show that CAE-HNC is a compact and robust feature representation method, which provides effective help for the ground object classification and target recognition of hyperspectral images.

Author Contributions

Conceptualization, R.D.; Data curation, X.L.; Methodology, X.Y.; Validation, X.L. and J.S.; Writing—Original draft, X.Y.; Writing—Review and editing, X.Y. and R.D. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Basic Scientific Research Operating Expenses Project of Provincial Universities in Heilongjiang Province (No.2020-KYYWF-358) and Doctor Initiation Fund Project of Harbin Normal University (No.XKB202113).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, C.J.; Li, G.D.; Lei, R.M.; Du, S.H.; Wu, Z.F. Deep feature aggregation network for hyperspectral remote sensing image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 5314–5325. [Google Scholar] [CrossRef]
  2. Sui, C.H.; Tian, Y.; Xu, Y.P.; Xie, Y. Weighted Spectral-Spatial Classification of Hyperspectral Images via Class-Specific Band Contribution. IEEE Trans. Geosci. Remote 2017, 55, 7003–7017. [Google Scholar] [CrossRef]
  3. Zhang, B.; Zhao, L.; Zhang, X.L. Three-dimensional convolutional neural network model for tree species classification using airborne hyperspectral images. Remote Sens. Environ. 2020, 247, 111938. [Google Scholar] [CrossRef]
  4. Zhang, S.Z.; Li, S.T. Spectral-spatial classification of hyperspectral images via multiscale superpixels based sparse representation. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium, Beijing, China, 10–15 July 2016; p. 7729625. [Google Scholar]
  5. Li, Q.T.; Wang, C.Z.; Zhang, B.; Lu, L.L. Object-Based Crop Classification with Landsat-MODIS Enhanced Time-Series Data. Remote Sens. 2015, 7, 16091–16107. [Google Scholar] [CrossRef] [Green Version]
  6. Calders, K.; Adams, J.S.; Armston, J.; Bartholomeus, H.; Verbeeck, H. Terrestrial laser scanning in forest ecology: Expanding the horizon. Remote Sens. Environ. 2020, 251, 112102. [Google Scholar] [CrossRef]
  7. Khan, S.; Hussain, S.; Yang, S.K. Contrast Enhancement of Low-Contrast Medical Images Using Modified Contrast Limited Adaptive Histogram Equalization. J. Med. Imaging Health Inform. 2020, 10, 1795–1803. [Google Scholar] [CrossRef]
  8. Lefkimmiatis, S.; Bourquard, A.; Unser, M. Hessian-based Norm Regularization for Image Restoration with Biomedical Applications. IEEE Trans. Image Process. 2011, 21, 983–995. [Google Scholar] [CrossRef] [PubMed]
  9. Prashnani, M.; Chekuri, R.S. Identification of military vehicles in hyper spectral imagery through spatio-spectral filtering. In Proceedings of the 2013 IEEE Second International Conference on Image Information Processing, Shimla, India, 9–11 December 2013; pp. 527–532. [Google Scholar]
  10. Bioucas-Dias, J.M.; Plaza, A.; Camps-Valls, G.; Scheunders, P.; Nasrabadi, N.M.; Chanussot, J. Hyperspectral Remote Sensing Data Analysis and Future Challenges. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–36. [Google Scholar] [CrossRef] [Green Version]
  11. Li, J.; Marpu, P.R.; Plaza, A.; Bioucas-Dias, J.M.; Benediktsson, J.A. Generalized Composite Kernel Framework for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote 2013, 51, 4816–4829. [Google Scholar] [CrossRef]
  12. Zhou, Y.C.; Peng, J.T.; Chen, C.L.P. Dimension Reduction Using Spatial and Spectral Regularized Local Discriminant Embedding for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote 2015, 53, 1082–1095. [Google Scholar] [CrossRef]
  13. Garcia-Salgado, B.P.; Ponomaryov, V.I.; Sadovnychiy, S.; Reyes-Reyes, R. Efficient dimension reduction of hyperspectral images for big data remote sensing applications. J. Appl. Remote Sens. 2020, 14, 032611. [Google Scholar] [CrossRef]
  14. Deng, Y.J.; Li, H.C.; Pan, L.; Shao, L.Y.; Du, Q.; Emery, W. Modified Tensor Locality Preserving Projection for Dimensionality Reduction of Hyperspectral Images. IEEE Geosci. Remote Sens. 2018, 15, 277–281. [Google Scholar] [CrossRef]
  15. Gu, Y.F.; Liu, T.Z.; Jia, X.P.; Benediktsson, J.A.; Chanussot, J. Nonlinear Multiple Kernel Learning with Multiple-Structure-Element Extended Morphological Profiles for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote 2016, 54, 3235–3247. [Google Scholar] [CrossRef]
  16. Zhang, P.; He, H.X.; Gao, L.R. A Nonlinear and Explicit Framework of Supervised Manifold-Feature Extraction for Hyperspectral Image Classification. Neurocomputing 2019, 337, 315–324. [Google Scholar] [CrossRef]
  17. Wen, J.H.; Yan, W.D.; Lin, W. Supervised linear manifold learning feature extraction for hyperspectral image classification. In Proceedings of the 2014 IEEE Geoscience and Remote Sensing Symposium, Quebec City, QC, Canada, 13–18 July 2014; pp. 3710–3713. [Google Scholar]
  18. Wang, L.L.; Li, Z.Y.; Sun, J.X. Improved ISOMAP Algorithm for Anomaly Detection in Hyperspectral Images. In Proceedings of the Fourth International Conference on Machine Vision, Singapore, 9–10 December 2011; p. 834902. [Google Scholar]
  19. Wang, B.J.; Huang, M.; Zhu, Q.B.; Wang, S. UVE-LLE Classification of Apple Mealiness Based on Hyperspectral Scattering Image. Acta Photonica Sin. 2011, 40, 1132–1136. [Google Scholar] [CrossRef]
  20. Yang, Q.W.; Li, Y.; Sun, F.C. Independent Component Analysis: Embedded LTSA. In Foundations and Applications of Intelligent Systems; Springer: Berlin, Germany, 2014; Volume 213, pp. 711–722. [Google Scholar]
  21. Huang, H. Classification of Hyperspectral Remote-sensing Images Based on Sparse Manifold Learning. J. Appl. Remote Sens. 2013, 7, 464–477. [Google Scholar] [CrossRef]
  22. Pan, X.; Ruan, Q.Q. Palmprint Recognition with Improved Two-dimensional Locality Preserving Projections. Image Vision Comput. 2008, 26, 1261–1268. [Google Scholar] [CrossRef]
  23. Rifai, S.; Vincent, P.; Muller, X.; Glorot, X.; Bengio, Y. Contractive Auto-encoders: Explicit Invariance During Feature Extraction. In Proceedings of the 28th International Conference on International Conference on Machine Learning, Bellevue, WA, USA, 28 June–2 July 2011; pp. 836–844. [Google Scholar]
  24. Rifai, S.; Mesnil, G.; Vincent, P.; Muller, X.; Glorot, X. Higher Order Contractive Auto-Encoder. In Proceedings of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, Athens, Greece, 5–9 September 2011; pp. 645–660. [Google Scholar]
  25. Yu, Y.Z.; Hui, J. A Study on Text Classification based on Stacked Contractive Auto-Encoder. In Proceedings of the International Conference on Electronics Instrumentation and Information Systems, Harbin, China, 3–5 June 2017; pp. 1–6. [Google Scholar]
  26. Aamir, M.; Nawi, N.M.; Wahid, F.; Mahdin, H. A Deep Contractive Autoencoder for Solving Multiclass Classification Problems. Evol. Intell. 2020, 14, 1619–1633. [Google Scholar] [CrossRef]
  27. Ng, P.C.; She, J. Denoising-Contractive Autoencoder for Robust Device-free Occupancy Detection. IEEE Internet Things 2019, 6, 9572–9582. [Google Scholar] [CrossRef]
  28. Zhang, Y.Y.; Li, X.Y.; Gao, L.; Chen, W.; Li, P.G. Ensemble Deep Contractive Auto-encoders for Intelligent Fault Diagnosis of Machines under Noisy Environment. Knowl.-Based Syst. 2020, 196, 105764. [Google Scholar] [CrossRef]
Figure 1. Experimental data sets of hyperspectral images. (a) Hyperspectral and ground truth images of Indian Pine; (b) Hyperspectral and ground truth images over University of Pavia; (c) Hyperspectral and ground truth images of Salinas.
Figure 1. Experimental data sets of hyperspectral images. (a) Hyperspectral and ground truth images of Indian Pine; (b) Hyperspectral and ground truth images over University of Pavia; (c) Hyperspectral and ground truth images of Salinas.
Electronics 10 02667 g001
Figure 2. Local region enlarge image of Classification results by different algorithms.
Figure 2. Local region enlarge image of Classification results by different algorithms.
Electronics 10 02667 g002
Table 1. Number of available samples for each group of data.
Table 1. Number of available samples for each group of data.
Indian PineUniversity of PaviaSalinas
TypeNumber of samplesTypeNumber of samplesTypeNumber of samples
Alfalfa46Asphalt6631Brocoli_green_weeds_12009
Corn-notill1428Meadows18,649Brocoli_green_weeds_23726
Corn-min830Gravel2099Fallow1976
Corn237Trees3064Fallow_rough_plow1394
Grass/Pasture483Painted metal sheets1345Fallow_smooth2678
Grass/Trees730Bare soil5029Stubble3959
Grass/Pasture-mowed28Bitumen1330Celery3579
Hay-windrowed478Self-Blocking Bricks3682Grapes_untrained11,271
Oats20Shadows947Soil_vinyard_develop6203
Soybeans-notill972 Corn_senesced_green_weeds3278
Soybeans-min2455 Lettuce_romaine_4wk1068
Soybeans-clean693 Lettuce_romaine_5wk1927
Wheat205 Lettuce_romaine_6wk916
Woods1265 Lettuce_romaine_7wk1070
Bldg-Grass-Tree-Drives386 Vinyard_untrained7268
Stone-steel towers93 Vinyard_vertical_trellis1807
Table 2. Number of available samples in each experimental data set.
Table 2. Number of available samples in each experimental data set.
Data Set NameTraining Sample EvaluationCAE [23]CAE-H [24] SCAE [25]DCAE [27]Proposed CAE-HNC
Indian Pine (IP)5%OA (%)74.3980.3989.0691.8393.39
Kappa0.72500.78310.86500.90140.9278
10%OA (%)75.5181.1690.3391.8893.87
Kappa0.74030.77070.84030.89110.9283
University of Pavia (UP)5%OA (%)83.9689.1796.0696.8198.04
Kappa0.82560.88070.95050.95990.9701
10%OA (%)84.9290.3596.7997.4498.13
Kappa0.83880.89050.95150.96340.9703
Salinas (SV)5%OA (%)76.3382.7792.7593.7895.36
Kappa0.74330.81230.91430.92340.9406
10%OA (%)76.9983.1494.0994.9595.88
Kappa0.75270.82120.93020.94010.9472
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yu, X.; Ding, R.; Shao, J.; Li, X. Hyperspectral Remote Sensing Image Feature Representation Method Based on CAE-H with Nuclear Norm Constraint. Electronics 2021, 10, 2667. https://doi.org/10.3390/electronics10212667

AMA Style

Yu X, Ding R, Shao J, Li X. Hyperspectral Remote Sensing Image Feature Representation Method Based on CAE-H with Nuclear Norm Constraint. Electronics. 2021; 10(21):2667. https://doi.org/10.3390/electronics10212667

Chicago/Turabian Style

Yu, Xiaodong, Rui Ding, Jingbo Shao, and Xiaohui Li. 2021. "Hyperspectral Remote Sensing Image Feature Representation Method Based on CAE-H with Nuclear Norm Constraint" Electronics 10, no. 21: 2667. https://doi.org/10.3390/electronics10212667

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop