Next Article in Journal
Performance Enhancement for Full-Duplex Relaying with Time-Switching-Based SWIPT in Wireless Sensors Networks
Next Article in Special Issue
TaijiGNN: A New Cycle-Consistent Generative Neural Network for High-Quality Bidirectional Transformation between RGB and Multispectral Domains
Previous Article in Journal
Wearables for Industrial Work Safety: A Survey
Previous Article in Special Issue
In Vitro and In Vivo Multispectral Photoacoustic Imaging for the Evaluation of Chromophore Concentration
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scale Superpixels Dimension Reduction Hyperspectral Image Classification Algorithm Based on Low Rank Sparse Representation Joint Hierarchical Recursive Filtering

1
School of Software, Henan University, Kaifeng 475001, China
2
Institute of Intelligence Networks System, Henan University, Kaifeng 475001, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(11), 3846; https://doi.org/10.3390/s21113846
Submission received: 27 April 2021 / Revised: 29 May 2021 / Accepted: 30 May 2021 / Published: 2 June 2021
(This article belongs to the Special Issue Recent Advances in Multi- and Hyperspectral Image Analysis)

Abstract

:
The original Hyperspectral image (HSI) has different degrees of Hughes phenomenon and mixed noise, leading to the decline of classification accuracy. To make full use of the spatial-spectral joint information of HSI and improve the classification accuracy, a novel dual feature extraction framework joint transform domain-spatial domain filtering based on multi-scale-superpixel-dimensionality reduction (LRS-HRFMSuperPCA) is proposed. Our framework uses the low-rank structure and sparse representation of HSI to repair the unobserved part of the original HSI caused by noise and then denoises it through a block-matching 3D algorithm. Next, the dimension of the reconstructed HSI is reduced by principal component analysis (PCA), and the dimensions of the reduced images are segmented by multi-scale entropy rate superpixels. All the principal component images with superpixels are projected into the reconstructed HSI in parallel. Secondly, PCA is once again used to reduce the dimension of all HSIs with superpixels in scale with hyperpixels. Moreover, hierarchical domain transform recursive filtering is utilized to obtain the feature images; ultimately, the decision fusion strategy based on a support vector machine (SVM) is used for classification. According to the Overall Accuracy (OA), Average Accuracy (AA) and Kappa coefficient on the three datasets (Indian Pines, University of Pavia and Salinas), the experimental results have shown that our proposed method outperforms other state-of-the-art methods. The conclusion is that LRS-HRFMSuperPCA can denoise and reconstruct the original HSI and then extract the space-spectrum joint information fully.

1. Introduction

HSI uses numerous continuous narrow-band electromagnetic wave bands to image the surface species and obtain rich joint information of the space-spectrum. With the rapid development of HSI processing and analysis in recent years, HSI classification technology is extensively utilized in agriculture [1], environmental detection [2], marine monitoring [3], and other fields; however, considering the effects of imaging sensor breakdown [4], environmental pollution [5], and other factors, the obtained HSI has mixed noise [6,7,8,9] that reduces the classification accuracy. For example, Tu et al. [7] proposed a kernel entropy component analysis (KECA)-based method for noisy label detection that can remove noisy labels for the PaviaU with 50 true samples and 10 noisy labels per class, KECA obtains 81.11% OA, with 50 true samples and 30 noisy labels per class, the OA is 71%. Xu et al. [8] proposed a spectral-spatial classification of HSI based on low-rank decomposition, by removing the sparse part, the LRD-NWFE-GC [8] obtains 92.3% on Indian Pines even if the training sample is small. On the contrary, the NWFE-GC [8] does not remove the sparse noise, and the classification accuracy is reduced to 74.53%. So, the HSI denoising needs to be solved.
The common noises in HSI are gaussian, impulse, and stripe noises [10]. The methods of noise removal are divided into transform domain filtering and spatial domain filtering [11]. Original spatial domain filtering is easy to perform; however, it will cause a blur of the image and losing some edge and texture details. At present, the edge-preserving filtering method can eliminate the image noise and keep the boundary between various species clear. Typical edge-preserving filtering methods based on spatial domain include a guided filter presented by He et al. [12] and a domain transform recursive filter proposed by Gastal et al. [13]. The non-local mean method is also a type of spatial domain filtering. The related non-local mean algorithms are the shape adaptive non-local mean algorithm [14] and the block matching 3D (BM3D) algorithm [15]. BM3D can preserve the image edge and texture information and obtain a high signal-to-noise ratio [15]. The transform domain filtering denoising transforms the image from the spatial domain to the transform domain through a set of orthogonal transforms. It separates the signal and noise via the different features of noise and signal in the transform domain. In recent years, the denoising algorithm based on sparse representation is extensively used in the field of HSI processing. Low-rank sparse representation is a technique to decompose the observation matrix into the low-rank matrix and sparse matrix, using the low-rank attribute of HSI. This technology can restore the low-rank component robustly when the image is destroyed by mixed noise. In this regard, Zhang et al. [16] proposed an HSI restoration method based on low-rank matrix restoration. A fast denoising algorithm based on low-rank sparse representation was proposed by Lina et al. [17], taking full advantage of the HSIs’ low-rank structure and self-similarity to further improve the denoising effect.
HSI can present rich spectral information by increasing the dimension of HSI data. The classification accuracy will first increase and then decrease, leading to the dimension disaster [18]. The dimension reduction algorithms of HSI are divided into band selection [19,20] and feature extraction. A feature extraction algorithm is divided into supervised and unsupervised. Among the traditional unsupervised algorithms is PCA [21]. PCA does not need label information, it can be reduced to any dimension and is easy to implement. Later, some efficient unsupervised algorithms were proposed, such as subspace feature learning [22] and potential subclass learning [23]. A supervised dimensionality reduction algorithm utilizes supervised information known as labels, to learn the feature space after dimensionality reduction. The representative work includes linear discriminant analysis (LDA) [24] and local Fisher discriminant analysis (LFDA) [25]. LDA cannot only reduce the dimension, but also distinguish different kinds of samples. On the other hand, many HSIs contain complex types of surface features. LDA needs a lot of manpower and costs much time to acquire prior knowledge, and when the ratio between the number of training samples and the number of features is very small, the LDA will achieve lower classification accuracy [24]. In recent years, the spectral features of each pixel are used in most feature extraction algorithms ignoring the spatial features, in order to make full use of spatial features. Kang et al. [26] proposed a spatial-spectral joint classification algorithm based on image fusion (IFRF). An HSI classification algorithm was proposed by Tu et al. [27] combining the correlation coefficient and the joint sparse representation (CCJSR). It can make full use of spectral similarity information and spatial context information at the same time. With the development of deep learning technology, a convolutional neural network (CNN) is extensively utilized in image processing [28]. The relevant CNN classification algorithms based on the spatial-spectral combination include 2D-CNN [29], SSRN [30], CDL-MLR [31] and SSFC [32] to obtain a robust result. Although deep learning methods have great advantages in classification, the architecture of CNNs often has feature redundancy. Therefore, Arijitet et al. [33] proposed a CNN structure based on the GhoMR module to reduce the number of parameters and form a lightweight feature extraction module. Compared with the original 2D-CNN algorithm, the OA of the GhoMR algorithm is improved by 18%, on Indian Pines with 10% training set. Besides, different species often exist in various regions of HSI, resulting in different spectral features. However, most of the extraction algorithms often construct a unified projection space for the feature image after dimensionality reduction [21,26,33], making it impossible to make full use of the different spectral features of various species. To solve this problem, segmentation technology divides the observed image into several specific regions. Jakub et al. [34]. proposed an end-to-end approach to segment hyperspectral images in a fully unsupervised way. Jiang et al. [35]. proposed a feature extraction algorithm based on multi-scale superpixel PCA(MSuperPCA). A superpixel representation based on K-Nearest Neighbor was proposed by Tu et al. [36] for HSI classification (KNNRS). The authors of [34,35,36] fully extracted the spatial-spectral features of HSI.
Compared with the basic feature extraction algorithm PCA, in recent years, there are many advanced feature extraction algorithms, for example, LFDA [25] and 2D-CNN [29] improve the OA only by at least 3%, IFRF [26], CCJSR [27], SSRN [30], CDL-MLR [31], GhoMR [33], MSuperPCA [35], and KNNRS [36] improve the OA by at least 15% on the Indian Pines with 10% training sets; however, only the spatial domain denoising algorithm is selected for feature extraction [26,36]. The noise in HSI is often mixed noise. Different noise has various imaging reasons and denoising strategies. To reduce the mixed noise of HSI and further enhance the subsequent classification accuracy, transformation domain filtering algorithms based on low-rank sparse representations can robustly recover the low-rank component when the images are destroyed by mixed noise. Nevertheless, HSI classification algorithms only selecting low-rank sparse representations cannot acquire sufficiently spatial structural information, leading to decreased classification accuracy on datasets that are large and complex in texture and structure [27]. The original HSI is not denoised by the HSI feature extraction algorithm based on multi-scale superpixel PCA. Hence, the classification accuracy improvement is not obvious when the training set is large [35]. The problem of feature redundancy of traditional convolutional neural networks is solved by convolutional neural network structure based on the GhoMR module achieving high classification accuracy. However, GhoMR performs PCA on the original HSI and constructs a single projection space, making it impossible to make full use of the spectral features of different species [33]. We propose a multi-scale superpixel unsupervised dimension reduction HSI classification algorithm based on hierarchical recursive filtering and low-rank sparse representation (LRS-HRFMSuperPCA), the schematic diagram is shown in Figure 1. Firstly, the low-rank sparse representation algorithm is used to predict the subspace of HSI. The fully observed HSI is obtained according to the subspace. Next, BM3D is utilized to denoise HSI, and PCA is used to reduce the dimension of HSI after restoration and denoising. The principal component images are segmented by multi-scale entropy rate superpixels. To construct multiple projection spaces, the segmented principal component images with superpixels are projected parallel to the denoised HSIs. The dimension of HSI with superpixels is reduced by PCA again, and each feature image in the multi-scale, after dimension reduction, is transformed by hierarchical recursive filtering. Ultimately, the decision fusion algorithm based on an SVM is used to classify the multi-scale feature images after hierarchical recursive filtering and obtain the final classification result.
The main contributions of LRS-HRFMSuperPCA can be summarized as follows:
  • Compared to other HSI classification algorithms, we propose to integrate the new joint feature extraction framework of spatial domain filtering and transform domain filtering(LRS-HRF) into the MSuperPCA. Firstly, the sparse feature and low-rank structure of HSI are utilized to realize transform domain filtering to separate the mixed noise in the image and improve the PSNR of HSI. Then, the separated feature images are transformed and filtered recursively in a multi-level domain based on the spatial domain. To enhance the spatial-spectral features of the feature image, the strong edges of different objects in the image are preserved and the fine texture structure is removed.
  • The HSI with restoration and denoising is obtained by filtering in the transform domain segmented by multi-scale superpixels, and the low-dimension characteristics of different species in different regions are fully obtained. Then, PCA is performed for each HSI with superpixels in multi-scale, and the spatial information of each HSI is fully utilized.
  • Compared to the state-of-the-art HSI classification algorithms, the LRS-HRFMSuperPCA algorithm achieves the highest OA, AA, and Kappa coefficient on the three real datasets.

2. Methodology

2.1. Estimation of Noise in Hyperspectral Image

Suppose an original HSI Y L × Q [ y 1 , y 2 , , y Q ] R L , Q is the number of pixels in each band, L is the number of bands. The high correlation between neighboring spectral bands and the spatial correlation within that band are the reasons underlying the good performance of the linear regression theory in hyperspectral applications [37,38].
Calculate the autocorrelation matrix R ^ : = ( Y Y T ) and the inverse matrix R : = R ^ 1 . Traversing all bands, we assume that z i is explained by a linear combination of the remaining L 1 bands, z i contains the data read by the hyperspectral sensor at the i th band for all image pixels, Z = Y T , so we have [39]:
z i = Z i β i + ξ i
where Z i = [ z 1 , , z i 1 , z i + 1 , , z L ] denotes data from other band image pixels after removing the i -th band, β i is the regression vector of size ( L 1 ) × 1 , and ξ i is the noise vector of size N × 1 , the least squares estimator of the regression vector β i is given by [39]:
β ^ i = ( Z i T Z i ) 1 Z i T z i
In order to reduce the computational complexity, define the pseudoinverse Z i # = ( Z i T Z i ) 1 Z i T , of size ( L 1 ) × ( L 1 ) . Let L × L be the size of symmetric and positive definite matrices R and R 1 that are partitioned into block matrices as follows:
R = [ A b b T c ]   R 1 = [ A b b T c ]
β ^ i = Z i # z i
where A and A are ( L 1 × L 1 ) matrices, b and b are ( L 1 × 1 ) vectors, and c and c are scalars. Because they are symmetric and positive, we find:
A 1 = A b b T / c
So, let A = [ R ^ ] i , i ,   A = [ R ] i , i , [ R ] i , i is the inverse of the autocorrelation matrix obtained by deleting row i and column i , and we have [ R ^ ] i , i 1 = [ R ] i , i [ R ] i , i [ R ] i , i / [ R ] i , i , the following is the assignment operation, Z i # : = [ R ^ ] i , i 1 , z i : = [ R ^ ] i , i ,   and   Equation   ( 4 ) , thus:
β ^ i : = ( [ R ] i , i [ R ] i , i [ R ] i , i / [ R ] i , i ) [ R ^ ] i , i
ξ i ^ : = z i Z i β ^ is calculated for the regression vector β ^ of each band, the end traversal output ξ ^ : = [ ξ ^ 1 , ξ ^ 2 , , ξ ^ L ] yields the estimated noise in the HSI.

2.2. Subspace Prediction of the Signal from the Hyperspectral Image

This section performs a subspace prediction of the signal from the HSI. First, calculate the autocorrelation matrix R ^ y ( Y Y T ) Q and the autocorrelation matrix R ^ n : = i ( ξ ^ i , ξ ^ i T ) Q of the noise according to the noise vector ξ ^ i . Then, the autocorrelation matrix R ^ x : = i ( ( y i ξ ^ i ) ( y i ξ ^ i ) T ) Q is computed for the predicted signals. The eigenvectors E : = [ e 1 , , e L ] for R ^ x are computed. Given a permutation π = { i 1 , , i L } of indices i = 1 , , L , let us decompose the space L into two orthogonal subspaces. The two orthogonal subspaces are: the k -D subspace E k [ e i 1 , , e i k ] and E 1 k [ e i k + 1 , , e i L ] , E 1 k is the orthogonal complement of the subspace E k . Let Uk be the projection matrix onto E k and x ^ k U k y be the projection of the observed spectral vector y onto the subspace E k , we introduce the concept of minimum mean square error between x and x ^ k as a criterion of subspace [40].
m s e ( k ) = t r ( U 1 k R y ) + 2 t r ( U k R ^ n ) + c
where c is an irrelevant constant, with respect to all the permutations π = { i 1 , , i L } of size L and to k . So, we can construct the setting item K as follows:
m s e ( k ) = t r ( U 1 k R y ) + 2 t r ( U k R ^ n ) + c
Let k ^ be the number of negative values of δ = t r ( U 1 k R ^ y ) + 2 t r ( U k R ^ y ) , where δ : = [ δ 1 , , δ L ] . Ultimately, the proper subset of feature vector E : = [ e 1 , , e L ] corresponding to negative δ i is found from δ : = [ δ 1 , , δ L ] as the prediction. The value S ^ = [ e i ^ 1 , , e i ^ k ^ ] of the HSI subspace is retrieved from the signal subspace from k ^ and π .

2.3. Domain Transform Recursive Filtering (DTRF)

The principle of DTRF is to find a transform t : R 2 R for the input signal I , in the new domain R . It is stated that the Euclidean distance between neighboring samples in the new domain R must equal the l 1 distance between them in the original domain R 2 [13].
| t ( x i , I ( x i ) ) ( x j , I ( x i ) ) | = ( x i , I ( x i ) ) ( x j , I ( x i ) )   x i , x j S  
where S = { x 0 , x 1 , , x n } represents the sample set on the signal. Set up c t ( x ) = t ( x , I ( x ) ) ,   h is the sampling interval, and convert Equation (8) to:
c t ( x + h ) c t ( x ) = h + | I ( x + h ) I ( x ) |
In general, to avoid calculating the absolute value, c t is defined as a simple increasing function set up c t ( 0 ) = 0 . We have:
c t ( u ) = 0 u 1 + | I ( x ) | d x , u Ω
In the actual filtering process, two adjustable parameters are introduced to adjust the filter, the spatial standard deviation σ s and the spectral standard deviation σ r , as:
c t ( u ) = 0 u 1 + σ s σ r | I ( x ) | d x , u Ω
where u is the transform domain signal, I ( x ) represents the derivative of input one-dimensional signal I ( x ) . Then, the input signal I is processed by recursive filtering as:
K i = ( 1 a b ) I i + a b K j 1
where a = exp ( 2 δ s ) ϵ   [ 0 , 1 ] refers to the feedback coefficient, b is the distance between two adjacent samples in the transform domain, K i represents the filtered result. By increasing b , a b tends to be zero, which stops the propagation chain to preserve sharp edges in the signal.

3. Proposed LRS-HRFMSuperPCA Method

The HSIs are inevitably contaminated with various noises due to equipment limitation and atmospheric environment impact. On the other hand, the computational complexity will grow exponentially as the dimension increases, which will cause the “curse of dimensionality”; two problems that limit the subsequent classification. We proposed a novel transformation domain-space domain feature extraction framework, called LRS-HRFMSuperPCA, which can effectively improve the classification accuracy and is effective for HSI denoising.
The framework of this joint feature extraction algorithm includes four parts: low-rank sparse representation denoising, multi-scale superpixel dimensionality reduction, and hierarchical domain transform recursive filtering, SVM classification. We detail all parts and provide a pseudo code in the following.

3.1. Low-Rank Sparse Representation Denoising Based on MSuperPCA

Transform domain filtering denoising algorithm is used to remove the mixed noise in the original HSI. Following Jiang et al. [35] and Lina et al. [17], we propose the low-rank sparse representation denoising algorithm to remove noise from the original HSI, as the input of the MSuperPCA algorithm. According to Section 2.2, we can obtain the subspace S ^ of the original HSI Y L × Q
x : = vec ( X ) = vec ( S ^ Z ) = ( I S ^ ) z
where represents Kronecker product and z : = v e c ( Z ) , X is the clean image without noise and I is the identity matrix, Z represents the correlation coefficient matrix. The denoising and inpainting problem is formulated as
min z 1 2 M ( I S ^ ) z y 2 + λ ϕ ( z )
where ϕ ( z ) is the regularization function, λ is the regularization coefficient, y : = v e c ( Y ) . Let the mask M i   acting on the unobserved pixel i and HSI y o , i affected by noise, because x i = S ^ z i , we have
y o , i = M i S ^ z i + n o , i
In addition, n o , i is the noise part, according to the literature [40], we can obtain z ^ i = ( S ^ T M i T M i S ^ ) 1 S ^ T M i T y o , i . By calculating y ^ i = S ^ z ^ i , we can obtain the fully observed HSI Y ^ = [ y ^ 1 , y ^ 2 , , y ^ L ] . After recovering the unobserved component, we should solve the denoising problem
min Z 1 2 S ^ Z Y ^ F 2 + λ ϕ ( Z )
For the optimization problem of Equation (17), we choose BM3D for denoising [15], because the image obtained by the BM3D algorithm keeps a high PSNR. Then, the correlation coefficient matrix Z is obtained. The HSI X ^ is solved by X ^ = S ^ Z .
The dimension of HSI X ^ is reduced by PCA, the first principal component images I f are segmented into N principal component images with various superpixel blocks utilizing the entropy rate superpixel algorithm [41], where N = 2 C + 1 , C is a user-defined number of scales. The number of superpixels per image in the scale is S c = ( 2 ) c S f , where c = [ C , C + 1 , , 0 , , C 1 , C ] , c denotes the rank of each picture with superpixels, the segmented superpixel blocks of each scale are parallel projected onto the repaired and denoised HSI X ^ . SuperPCA was performed on X ^ [35], therefore, we have
F i = S u p e r P C A ( J m , X ^ , n u m _ P C )
where n u m   _   P C is the dimension after PCA, J m represents the principal component image after superpixel segmentation.

3.2. Hierarchical Domain Transform Recursive Filtering (HDTRF)

Inspired by the idea of deep learning [42,43], we set up a hierarchical structure based on DTRF [13], as shown in Figure 2, the output feature set of the upper level DTRF unit is used as the input of the next level DTRF unit. The spatial information and spectral information of hyperspectral data are fully mined. The obtained spatial-spectral features preserve the global similarity structure, local geometry structure and rich spatial information of the hyperspectral dataset.

3.3. Decision Fusion Strategy Based on Support Vector Machine

We use multi-scale entropy rate superpixel segmentation to obtain multiple principal component images. We classify the feature image sets after hierarchical recursive filtering by using SVM. SVM is a supervised learning model, and it is mainly oriented to the linear separable case.
In the case of linear inseparability, the linear inseparable samples in the low-dimensional input space are transformed into a high-dimensional feature space by using a nonlinear transformation algorithm. The nonlinear transformation is to map the sample information in the input space into a high-dimensional space through a properly defined inner product function, and then find the optimal linear classification surface in the new space. The optimal classification surface is obtained when the maximum interval classification line is expanded to the n -dimensional space. The optimal classification line cannot only correctly separate some individuals of different categories, but also maximize the overall classification interval. The general principle of SVM is shown in Figure 3.
In the face of the feature classification in HSIs, the problem is always nonlinear. Therefore, we should use some nonlinear feature transformation to map the sample information in the original input space to a high-dimensional feature space, and then find the optimal classification surface in the new space. The objective function is:
min φ ( ω , τ ) = 1 2 ω 2 + c 1 i = 1 l τ i   s . t .   y i [ w · x i + b ] 1 + τ i 0 ,   τ i 0   , i = 1 , 2 , , n
where φ ( ω , τ ) represents the spatial variation function, τ i represents the relaxation factor, c 1 represents the regularization parameter. Facing different classification tasks, we need to construct different kernel functions. The RBF kernel function has the characteristic of unique best approximation. As a kernel function, RBF can map input samples to high-dimensional feature space in SVM. We use the RBF kernel function in the LIBSVM toolbox.
As a feature image can obtain one classification accuracy, we can get 2 C + 1 classification accuracies from different feature images. The decision fusion algorithm is used to improve classification accuracy. We leverage the majority voting-based decision fusion strategy due to its insensitivity to inaccurate estimates of posterior probabilities, the majority vote is given by [44]:
v = a r g m a x i { 1 , 2 C } N ( i ) ,   where   N ( i ) = j = 1 2 C + 1 α j I 1 ( v j = i )
where I 1 is the indicator function, v is the class label from one of the C 1 possible classes for the test pixel, j is the classifier index, 2 C + 1 is the number of classifiers, N ( i ) is the number of times class i was detected in the bank of classifiers, and α j denotes the voting strength of the j -th classifier. In our method, we use the equal voting strength, α j = 1 2 C + 1 ( j = 1 , 2 , , 2 C + 1 ) .
The detailed procedure for the LRS-HRFMSuperPCA classification is described in Algorithm 1.
Algorithm 1 LRS-HRFMSuperPCA
Inputs: Insert original HSI Y L × Q ,   Y _ m i n is the minimum spectral value of all bands traversed;
Step 1 Noise prediction of HSI
(1) Y N = Y L × Q Y _ m i n / ( Y _ m a x Y _ m i n ) ,   Y N [ y 1 , y 2 , , y Q ] , Y N [ 0 , 1 ] ;
(2) R ^ : = ( Y N Y N T ) ,   R : = R ^ 1 ;
(3) for i :=1 to L do
(4) β ^ i : = ( [ R ] i , i [ R ] i , i [ R ] i , i / [ R ] i , i ) [ R ^ ] i , i ; i = 1 , , i 1 , i + 1 , , L
(5) ξ ^ i : = z i Z i β ^ i
(6) end for
(7) Output ξ ^ : = [ ξ ^ 1 , ξ ^ 2 , , ξ ^ L ] ;
Step 2 Subspace prediction of HSI
(1) R ^ y ( Y N Y N T ) Q , R ^ n : = i ( ξ ^ i , ξ ^ i T ) Q , R ^ x : = i ( ( y i ξ ^ i ) ( y i ξ ^ i ) T ) Q ;
(2) Calculate the eigenvector of R ^ x , E : = [ e 1 , , e L ] ;
(3) Construction setting item K   ,   K = ( k ^ , π ^ ) = a r g   m i n k , π { t r ( U 1 k R ^ y ) + 2 t r ( U k R ^ y ) }   , U k = E k E k T ;
(4) Set δ = t r ( U 1 k R ^ y ) + 2 t r ( U k R ^ n ) , δ : = [ δ 1 , , δ L ] ; ( δ ^ , π ^ ) : = s o r t ( δ ) by ascending order, save the permutation π ^ ;
(5) k ^ : = number of terms δ ^ i < 0   , find subspace prediction S ^ = [ e i ^ 1 , , e i ^ k ^ ]
Step 3 HSI restoration and denoising
(1) Let the mask M i acting on the unobserved pixel i and HSI y o , i affected by noise;
(2) The correlation coefficient of the subspace z ^ i is calculated by z ^ i = ( S ^ T M i T M i S ^ ) 1 S ^ T M i T y o , i ;
(3) Determine the fully observed HSI Y ^ = [ y ^ 1 , y ^ 2 , , y ^ L ] , y ^ i = S ^ z ^ i ;
(4) Solve the optimal value problem of min Z 1 2 S ^ Z Y ^ F 2 + λ ϕ ( Z ) based on the BM3D algorithm; the denoised HSI X ^ is obtained by X ^ = S ^ Z ;
Step 4 Dimension reduction algorithm based on multi-scale superpixels segmentation
(1) Insert scale number C   , N = 2 C + 1 , C ϵ N * ; Input initial number of pixels S ;
(2) Let p = 1 , P = p c a ( X ^ , p ) ;   [ h e i g h t , w i d t h ] = s i z e ( P ) ;   S c = ( 2 ) c S , c = 0 , ± 1 , ± 2 , , ± C
(3) for m : = 1 to 2 C + 1 do
(4) J m = s e g 2 b m a p ( m e x _ e r s ( P , ( 2 ) c S ) , w i d t h ,   h e i g h t ) ;
(5) end for
(6) for k : = 1 to N do
(7) F k = S u p e r P C A ( J k , X ^ , n u m _ P C )
(8) end for
(9) Output F k = [ F 1 , F 2 , , F N ]
Step 5 Decision fusion classification algorithm based on hierarchical recursive filtering
(1) for l   : = =1 to N do
(2) D e n o R S l = H R F ( F l , σ s , σ r , M )
(3) end for
(4) A C l = S V M ( D e n o R S l )
(5) v = a r g   m a x i { 1 , 2 C } N ( i ) ,   where   N ( i ) = j = 1 2 C + 1 α j I 1 ( v j = i )

4. Experiment and Analysis

4.1. Experimental Datasets

Indian Pines is a 145 × 145 pixel HSI collected by the AVIRIS hyperspectral spectrometer sensor at the Indiana agricultural test site. It contains 220 bands of 16 types of ground objects, and 20 bands with noise removed. The remaining 200 bands are utilized as the experimental dataset, and the total number of labeled pixels is 10,249, Figure 4 shows Indian Pines’ ground true value map, specific species and false color map.
PaviaU is a 610 × 340 pixel HSI collected by the ROSIS hyperspectral spectrometer sensor over Pavia University. A total of 115 bands of 9 types of ground objects are included, and 12 bands with noise are eliminated. The remaining 103 bands are utilized as the experimental dataset, and the total number of labeled pixels is 42,776, Figure 5 shows PaviaU’s ground value map, specific species and false color map.
Salinas Scene is a 512 × 217 pixel HSI collected by the AVIRIS hyperspectral spectrometer sensor over the Salinas Valley, California. A total of 224 bands of 16 types of ground objects are included, and 20 bands with noise are eliminated. The remaining 204 bands are employed as the experimental dataset, Figure 6 shows Salinas’ ground value map, specific species and false color map. The total number of labeled pixels is 54,129, Table 1 shows Number of samples in the Indian Pines, PaviaU, and Salinas.

4.2. Parameter Analysis

The commonly used HSI classification performance indicators are as follows: OA, AA and Kappa coefficient. Taking the hyperspectral dataset with C 1 kinds of surface features as an example, by comparing the classification results with the real surface feature distribution data, the confusion matrix M 1 C 1 × C 1 is obtained, in the matrix, m i j represents the number of samples that belong to class i and are divided into class j . Therefore, the diagonal element m i i in the confusion matrix represents the number of correctly classified samples in each feature category. Next, each evaluation index can be calculated separately.
OA: It refers to the proportion of all correctly classified samples in the total number of samples in the hyperspectral dataset. The calculation formula is:
OA = i = 1 C 1 m i i i = 1 C 1 i = 1 C 1 m i j
AA: According to the OA, the AA is:
AA = i = 1 C 1 m i i j = 1 C 1 m i j C
Kappa coefficient: In order to make full use of the information of the confusion matrix to measure the consistency between the classification results and the real surface features distribution map, researchers put forward the concept of Kappa, which is defined as follows:
Kappa = N 1 ( i = 1 C 1 m i i ) i = 1 C 1 ( j = 1 C 1 m i j j = 1 C 1 m j i ) N 1 2 i = 1 C 1 ( j = 1 C 1 m i j j = 1 C 1 m j i )
where, N 1 = i = 1 C 1 j = 1 C 1 m i j . It represents the total number of samples in the hyperspectral dataset. There are three types of parameters in our paper.

4.2.1. Scale Number C and Initial Superpixel Number S

As seen in Figure 7, when C is 0, the model degenerates to a single-scale superpixel dimension reduction. As a result, it is unable to make full use of the spatial information of multiple images, leading to the decline of classification accuracy. When the C value is too large, an over-segmentation will occur. Hence, a single-pixel block is obtained not being visible and the correct label cannot be identified. As shown in Figure 7a, when C is 7, OA will be maximized on the Indian Pines. Figure 7b reveals that, when C is 10, the maximum value is obtained on PaviaU, since the datasets with complex structures and textures and large sizes need to be segmented into more hyper pixels (for instance, PaviaU). Fewer hyper pixels are segmented when the dataset size is smaller (for example, the Indian Pines).
For the number of initial superpixels S , according to Figure 8, the trend of classification accuracy for the three datasets increases first and then decreases with increasing number of initial hyperpixels. Too few or too many initial superpixels will lead to poor classification accuracy. The reason is that when the initial number of superpixel blocks is too small, the number of superpixel blocks calculated for each image in the scale is also small, and the pixel blocks contain various ground objects, which cannot make full use of the spectral characteristics of different species. When the number of superpixels is too large, a single-pixel block will be smaller, and a single-pixel block will contain fewer pixels and will not be able to utilize all samples belonging to a uniform area. As seen in Figure 8, when S is 40, the maximum OA can be obtained for the three datasets. This indicates that hyperpixel segmentation of the reduced-dimension image can capture the spatial structure of the HSI and improve the OA of the subsequent process.

4.2.2. The Spatial and Range Standard Deviations of the Filter

As seen in Figure 9, on the Indian Pines, when σ s is 30 and σ r is 0.8, the method can obtain the optimal classification accuracy. This will not be repeated for the Salinas and PaviaU. For the recursive layer number L , when the recursion level is 0, the LRS-HRFMSuperPCA method degenerates to the LRS-MSuperPCA method leading to the failure to guarantee similar eigenvalues for the adjacent pixels on the same side of the edge and the classification accuracy is minimum, when the recursion level is 1, the fine texture structure cannot be completely eliminated by single-layer recursive filtering. Hence, it eliminates the fine texture structure and preserves the edges. This can be inferred from the SuperPCA part in schematic Figure 1. According to Figure 9, Figure 10 and Figure 11, it is indicated that the proposed HDTRF can enhance the classification accuracy of HSI.

4.2.3. Kernel Function Parameter G and Regularization Parameter c 1 in SVM

The SVM classifier used in the LRS-HRFMSuperPCA technique utilizes the LIBSVM toolbox in MATLAB. Moreover, the comparison algorithm of the SVM classifier also uses the LIBSVM toolbox in MATLAB containing the kernel parameter gamma G and regularization parameter c 1 . Figure 12 and Figure 13 represent the effect of different kernel parameters G and regularization parameter c 1 on classification accuracy in three datasets, respectively. It is observed that for the three datasets, the maximum classification accuracy can be achieved when the kernel parameter G is 300, 40 and 400, respectively, and the different regularization parameter c 1 is set to 106, 106 and 108. In our classifier, a higher gamma value G leads to a higher dimension of the mapping, and the training result is better. We choose the higher c 1 , as the higher c 1 aims to classify all training samples correctly by giving the model the freedom to choose more samples as support vectors. Since the Salinas training sample is larger than the Indian Pines and PaviaU, we can see that the value of c 1 will increase with the increase of training set samples.
The experimental parameter settings of the LRS-HRFMSuperPCA method in subsequent experiments are shown in Table 2.

4.3. Experimental Results and Analysis

The HSI restored by the transform domain filtering algorithm based on low-rank sparse representation is better than the original image. In our paper, the peak-signal-to-noise ratio (PSNR) was utilized to compare the quality of the denoised HSI and the original HSI, as follows:
PSNR = 20 · l o g 10 ( M A X I M S E )
M S E = 1 m n i = 0 m 1 j = 0 n 1 [ I ( i , j ) K ( i , j ) ] 2
where M A X I is the maximum value of the color of the image point, M S E represents the mean square error between two monochromatic images I and K of m∗n size. In our method, the original HSI with additive Gaussian noise with a noise variance σ 2 of 0.01 is used as the reference image to determine the PSNR. In the experiment, the original HSI and the denoised HSI obtained by low-rank sparse representation are considered as the processed images. Figure 14 represents the PSNR values of all bands of the original HSI and the HSI denoised by low-rank sparse representation on the Indian Pines. Table 3 denotes the average PSNR values of all bands of the two HSIs after summation.
When comparing the PSNR values, the reference image used in our method is a noisy HSI with additive Gaussian noise. Thus, the greater the MSE between the processed image and the reference image, the greater the difference between the processed image and the reference image. Moreover, the smaller the corresponding PSNR value, in our method, the smaller the difference between the processed image and the reference image damaged by additive noise. The larger corresponding PSNR value indicates that the quality of the processed image is poor. As seen in Figure 14, the PSNR values of the reconstructed HSI obtained by our method are lower compared to the original HSI in all bands. The third row of Table 3 reveals that the OA of the LRS-HRFMSuperPCA is better than that of the HRFMSuperPCA without low-rank sparse representation denoising on the 10% training set. Therefore, the transform domain filtering denoising algorithm based on low-rank sparse representation of HSI can enhance the quality of the original HSI and improve the OA.
The LRS-HRFMSuperPCA is compared with several HSI classification models, including original classification algorithm, SVM, PCA and the recently proposed IFRF [26], CCJSR [27], MSuperPCA [35], SSRN [30], KNNRS [36] and GhoMR [33]. To avoid the chance of the experimental results, the average value of five runs is considered as the final result, including our algorithm and all other comparison algorithms. At the same time, for a fair comparison, the parameters of the comparison algorithm are set as the default optimal parameters based on the relevant literature. Table 4 represents the specific experimental parameters of the comparison algorithm. The experimental environment of the LRS-HRFMSuperPCA method and other comparative algorithms is a notebook with 12 GB memory, Intel Core i5 2.2 GHz CPU and the development environment in MATLAB R2018a is used. GhoMR and SSRN are performed utilizing PyTorch 1.6.0 with CUDA 10.1 in the open-source GPU environment of Google Colaboratory. The code for the LRS-HRFMSuperPCA is available at https://gitee.com/lxxds123/my-dem (accessed on 25 April 2021).
In order to fully compare with the training set in the references corresponding to the comparison algorithm, the contrast experiments are carried out on the training set which accounts for t 1 = 77 (10% Training set) and t 1 = 173 (20% Training set) for per class on Indian Pines and t 2 = 475 (10% Training set) and t 2 = 1139 (20% Training set) for PaviaU, t 3 = 102 (3% Training set) and t 2 = 170 (5% Training set) on Salinas, leaving the rest samples to form the testing samples. For 10% training data on Indian Pines and PaviaU, 3% training data on Salinas, from Table 5, our method obtains 99.32%, 99.87% and 99.73% OA values on the three datasets. For 20% training data on Indian Pines and PaviaU, 5% training data on Salinas, from Table 6, our method obtains 99.63%, 99.91% and 99.90% OA values on the three datasets.
To compare 10%, 20%, 3% and 5% of the total randomly selected training samples, the eight advanced comparison algorithms in Table 4 and the LRS-HRFMSuperPCA method are used. In this section, only the comparison results on the Indian Pines are discussed. Table 7 shows that, when the training set is 10%, the OA of our method is improved by 17.65% and 20.27%, respectively, compared with the original SVM and PCA on the Indian Pines. Compared to the IFRF, according to Table 3, the low-rank sparse representation denoising algorithm can enhance the quality of the original HSI, thus improving the OA of 0.9% on 10% of the Indian Pines. Compared to the CCJSR, the OA of our method is enhanced by 3.32% and 2.78% over the two training sets on Indian Pines. In comparison to the MSuperPCA, when the training set is 20%, the OA of our method on the Indian Pines dataset is improved by 1.78%. The reason is that our method combines the spatial domain filtering algorithm and the transform domain filtering algorithm. The mixed noise can be removed by the transform domain filtering denoising algorithm based on low-rank sparse representation within the original HSI. At the same time, the small-scale structure of the HSI can be eliminated by the multi-level spatial domain recursive filtering algorithm fully obtaining the spatial-spectral joint information of the HSI and improving the OA. Compared to KNNRS, the OA of our method is always higher than that of the LRS-HRFMSuperPCA on the three datasets (from Table 7 and Table 8). Compared to GhoMR, the single projection space constructed by the GhoMR makes it impossible to make full use of the spectral features of different species. Our method can randomly control the scale of multi-scale superpixel segmentation, and each obtained superpixel block is a set of pixels with similar characteristics. Subsequently, a hierarchical recursive filtering algorithm is used to extract the spatial-spectral joint information in each pixel block region to improve the consistency and similarity within the region. At the same time, our method runs in a non-GPU environment. Compared to the GhoMR running on GPU, we use CPU to reduce the experimental equipment cost.
Figure 15, Figure 16 and Figure 17 represent the random feature classification map obtained by our method and different comparison algorithms and the OA of the corresponding feature classification map. As seen in Figure 15, the OA of the SVM and PCA is not ideal since they do not combine the joint information of spatial-spectrum, and the “salt and pepper” noise phenomenon appears in the classification result graph. In contrast, IFRF and KNNRS combined with the spatial-spectrum can effectively eliminate the noise pixels, and the “salt and pepper” noise phenomenon is greatly reduced. Compared to PCA and SVM, the OA of IFRF is improved by 18.33% and 15.92%, respectively, and that of KNNRS is improved by 17.3% and 19.6%, respectively. According to the surface features classification map of Indian Pines, the OA of our method is 1.43%, 2.47%, and 0.16% higher than that of IFRF, CCJSR, and KNNRS. Compared to the advanced MSuperPCA, SSRN and GhoMR algorithms, the OA of our algorithm on the Indian Pines is improved by 1.58%, 0.61% and 0.68%, respectively. Improvements or comparable results are obtained on PaviaU and Salinas as well that are not repeated here.
This section analyzes the effects of various training and test sets on the OA of our method and the other methods. Figure 18 represents the classification results of the LRS-HRFMSuperPCA method. The number of training samples of three datasets increased from 1% to 6%. According to Figure 18, with different training sample numbers, the OA can be improved significantly by our method by increasing the number of training samples. Hence, we can conclude that our method can achieve the best OA for the three datasets compared to other state-of-the-art comparison algorithms.
The computational time of the LRS-HRFMSuperPCA algorithm and other comparison algorithms are represented in Table 9. The running environment of the LRS-HRFMSuperPCA and all the comparison algorithms is a notebook computer with 12 GB memory and a 2.2 GHz CPU processor. Except for GhoMR and SSRN, the running environment of GhoMR and SSRN are pytorch 1.6.0 and CUDA10.1 in the open-source GPU environment of Google lab, the running environment of other algorithms is MATLAB2018a. Table 8 shows the number of training samples as a percentage of the total sample number in the measurement run time to experiment on the three datasets. According to the analysis of Table 9, the LRS-HRFMSuperPCA algorithm takes a long time on the Indian Pines. The main reason is that this technique utilizes multi-level domain transform recursive filtering and multi-scale superpixel segmentation. As seen in Figure 18a,b, under different training sets of Indian Pines, the OA of our method is enhanced compared to other comparison algorithms. On the PaviaU and Salinas, the CCJSR took the most time since the computation of the correlation coefficient of the training set increased by increasing the training set, followed by the KNNRS. The main computation of the KNNRS is caused by the operation of KNNs. The operation time of the MSuperPCA is mainly caused by multiscale hyperpixel segmentation of HSIs. The run time of our method on the PaviaU is lower compared to the KNNRS, MSuperPCA, and CCJSR. The OA index of this method is also improved compared to other algorithms on PaviaU, for the GhoMR. Although the time cost of our algorithm on Indian Pines and PaviaU is higher than that of the GhoMR, the classification accuracy of our algorithm is better than that of the GhoMR in various training sets. Moreover, the time cost on Salinas is lower than that of GhoMR. Thus, the experimental results show that the run time of LRS-HRFMSuperPCA is acceptable.

5. Discussion

We reveal some interesting points about the LRS-HRFMSuperPCA methods by conducting several comparison experiments and parameter discussion on three HSIs.
  • In general, we notice that the mixed noise in HSI will reduce the classification accuracy [9]. Considering the effects of imaging sensors breakdown, environmental pollution, and other factors, these degradation factors account for a small proportion in HSI. We usually build them as sparse noise models here. A clear natural image (specific to our article, it is a HSI) is often a low-rank structure; therefore, when dealing with the mixed noise problem of natural image, we usually use the low-rank sparse representation algorithm to solve it. In order to remove the small-scale textures in HSI, domain transform recursive filtering is used to solve this problem. Another knowledge base is that after the stable low-rank feature natural image (specific to our article, it is a HSI) is extracted (specific to our article, it is LRS-MSuperPCA in Section 3.1), the joint information of the spatial-spectrum of the feature image can be obtained by hierarchical filtering (specific to our article, it is HDTRF in Section 3.2). In order to facilitate understanding, we apply hierarchical recursive filtering to IFRF (Figure 19) for experimental verification.
    Compared with the LRS-HRFMSuperPCA algorithm, using hierarchical filtering in Figure 9c, with the increase of the level, the OA of the three kinds of classification in Figure 19 decreases significantly. The experimental results show that our knowledge base is correct.
  • Based on the two knowledge bases proposed in the previous point, denoising based on low-rank sparse representation (let it set to function A ) belongs to the transform domain filtering algorithm, while denoising based on hierarchical recursive filtering (let it set to function B ) belongs to the spatial domain filtering algorithm. Let us set the penalty factor ρ that controls the proportion between function A and function B , ρ [ 0 , 1 ] , the optimal filtering of the finding F can be expressed as a function, which can be used for future research to find an optimal framework:
    F = ρ × A + ( 1 ρ ) B
  • According to the data in Table 7 and Table 8, LRS-HRFMSuperPCA is superior to the feature extraction algorithm using a single filter. Compared with IFRF and KNNRS, our method uses low-rank sparse representation algorithm to denoise the original HSI. It can be seen from Table 3 that the OA of LRS-HRFMSuperPCA is 0.23% higher than that of HRFMSuperPCA, and PSNR has also improved. This supports the correctness of LRS [17] that can remove mixed noise in the original HSI. Compared with LRS-MSuperPCA, the small-scale textures of the HSI can be eliminated by the multi-level spatial domain recursive filtering algorithm, fully obtaining the spatial-spectral joint information of the HSI and improving the OA. It can be seen from Figure 9c, Figure 10c and Figure 11c that the OA of LRS-HRFMSuperPCA is always higher than that of LRS-MSuperPCA on the three datasets. The comparative experiments verify that the DTRF [13] and IFRF [26] can remove texture structure and improve classification accuracy. These results support earlier studies.
  • Our finding is a novel feature extraction framework with joint low-rank sparse representation-hierarchical domain transform recursive filtering. The framework agrees with the existence of mixed noise in the original HSI [9] and the spectral vectors live in a low-dimensional subspace [45]. The proposed framework is applied to HSI feature extraction (MSuperPCA), so the algorithm LRS-HRFMSuperPCA is obtained. As can be seen from Figure 1, we use hierarchical recursive filtering to extract features images from HSI. According to the experimental results in Table 7 and Table 8, it improves the OA, AA and Kappa. In the field of remote sensing science, there are two expectations for remote sensing image classification [46]: the first is feature mining, and the second is to design a classifier with high classification accuracy. Our LRS-HRFMSuperPCA algorithm achieves the two expectations.
  • Compared with the HSI classification algorithm based on non-deep learning [26,27,35,36] and deep learning [30,33], the advantage of the LRS-HRFMSuperPCA algorithm is the use of the dual filtering framework. Specifically, compared with IFRF [26] and KNNRS [36] which only use spatial domain filtering, LRS-HRFMSuperPCA uses spatial-transform dual domain filtering, LRS-HRFMSuperPCA reconstructs the original HSI by recovering the pixels not observed in some known bands. We conduct an experiment to prove the anti- noise performance of the algorithm, in which 0-means additive Gaussian noise is added to the original HSI. According to Figure 20, with the increase of Gaussian noise intensity, our algorithm still maintains an OA of more than 98%, accounting for 10% of the Indian Pines samples on the training sets. Compared with CCJSR [27], which only uses a sparse representation, LRS-HRFMSuperPCA uses hierarchical domain transformation recursive filtering to make full use of spatial context information. Compared with SSRN [30] and GhoMR [33], the two space-spectrum joint classification algorithms based on deep learning use a single projection space. They cannot make full use of the low-dimensional characteristics of each species. Multi-scale superpixel segmentation can learn the low-dimensional features of different regions in HSIs by inputting the number of scales. At the same time, the architectures of SSRN [30] and GhoMR [33] need expensive GPU hardware to train and store. LRS-HRFMSuperPCA uses CPU to reduce the consumption of the experimental equipment. However, we find that most feature extraction algorithms (such as PCA in LRS-HRFMSuperPCA) can compress the information transformation of the original data into the low-dimensional feature space and this may lose the physical meaning of the original feature. In future research, we will suggest to focus on the method of unsupervised band selection algorithm instead of SuperPCA. Band selection selects a desired band subset from the original bands and preserves well the spectral meaning of spectral channels. In practice, prior knowledge about the surface features are difficult to obtain, so it is necessary to develop an unsupervised feature extraction algorithm. Specifically, for the reconstructed HSI with superpixels X ^ , hybrid scheme-based methods implement multiple schemes to select appropriate bands from X ^ . We can try to integrate both clustering and ranking based techniques to search for the best combination of bands with higher classification accuracy. Another major drawback of our method is that its performance is still affected by the number of superpixel segmentation and the shape of superpixel in our method is irregular. Specifically, once pixels of different classes exist in the same superpixel region, the pixels in these superpixels cannot be effectively classified. In future research, for different categories of pixels in the same superpixel, we will segment all kinds of pixels again, according to the weight of each pixel in the superpixel. The weight can be measured by relative entropy or Euclidean distance, etc. On the other hand, a gradient-based superpixel method has the advantage of regular shape and controllable number of generated superpixels, which is the direction to solve the problem of irregular shape of superpixels in the future.

6. Conclusions

In this paper, a simple and effective HSI classification algorithm is proposed combining transform domain filtering and spatial domain filtering. The objective of our method is to combine the PCA based on multi-scale entropy rate superpixel dimension reduction within two types of joint filtering denoising framework. To remove the mixed noise in the original HSI, the denoising algorithm based on a low-rank sparse representation is used to denoise and restore the original HSI, improve the peak-signal-to-noise ratio of the original HSI, and the subsequent classification accuracy. A multi-scale entropy ratio-based superpixel segmentation algorithm is utilized to segment the repaired image. By dividing the whole HSI into several regions, the same superpixel block has similar reflection characteristics, which is convenient for subsequent dimension reduction processing and finding the low-dimensional features of the HSI. Secondly, hierarchical recursive filtering is applied to the feature images after SuperPCA is making full use of the spatial information contained in HSI. Experiments on three real HSIs reveal that the proposed LRS-HRFMSuperPCA algorithm enhances the classification accuracy compared with the original MSuperPCA algorithm and the recently proposed HSI classification algorithm. However, the disadvantage of the LRS-HRFMSuperPCA algorithm is that the parameters of the algorithm are manually selected, thereby reducing the operability of the algorithm. Therefore, in future research, we will focus on how to automatically determine the optimal parameters of our algorithm.

Author Contributions

All authors made significant contributions to the manuscript. X.L. and S.Q. put forward the idea; S.Q. and X.L. conceived and designed the experiments; S.Q. and X.L. presented tools and carried out the data analysis; X.L. and S.Q. wrote the paper. S.Q. and S.L. guided and revised the paper; S.Q. provided the funding; All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Henan science and technology development plan project (No. 212102210538), Postgraduate education innovation and quality improvement project of Henan University (No. SYL20040121).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

The authors would like to sincere thank all the editors and anonymous reviewers gave valuable comments on our research.

Conflicts of Interest

The authors declare that there are no conflict of interest.

References

  1. Park, B.; Lu, R. Hyperspectral Imaging Technology in Food and Agriculture; Springer: Berlin, Germany, 2015. [Google Scholar]
  2. Pontius, J.; Martin, M.; Plourde, L.; Hallett, R. Ash decline assessment in emerald ash borer-infested regions: A test of tree-level, hyperspectral technologies. Remote Sens. Environ. 2008, 112, 2665–2676. [Google Scholar] [CrossRef]
  3. Ruitenbeek, F.; Debba, P.; Meer, F.; Cudahy, T.; Meijde, M.; Hale, M. Mapping white micas and their absorption wavelengths using hyperspectral band ratios. Remote Sens. Environ. 2006, 102, 211–222. [Google Scholar] [CrossRef]
  4. Majumdar, A.; Ansari, N.; Aggarwal, H.; Biyani, P. Impulse denoising for hyper-spectral images: A blind compressed sensing approach. Signal Process. 2016, 119, 136–141. [Google Scholar] [CrossRef] [Green Version]
  5. Hu, X.; Gu, X.; Yu, T.; Zhang, Z.; Li, J.; Luan, H. Effects of aerosol optical thickness on the optical remote sensing imaging quality. Spectrosc. Spect. Anal. 2014, 34, 735–740. [Google Scholar]
  6. Acito, N.; Diani, M.; Corsini, G. Subspace-based striping noise reduction in hyperspectral images. IEEE Trans. Geosci. Remote Sens. 2011, 49, 1325–1342. [Google Scholar] [CrossRef]
  7. Tu, B.; Zhou, C.; Peng, J.; He, W.; Ou, X.; Xu, Z. Kernel Entropy Component Analysis-Based Robust Hyperspectral Image Supervised Classification. Remote Sens. 2019, 11, 2823. [Google Scholar] [CrossRef] [Green Version]
  8. Xu, Y.; Wu, Z.; Wei, Z. Spectral–spatial classification of hyperspectral image based on low-rank decomposition. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2370–2380. [Google Scholar] [CrossRef]
  9. Rasti, B.; Ghamisi, P.; Chanussot, J. Mixed Noise Reduction in Hyperspectral Imagery. In Proceedings of the 2019 10th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing, Amsterdam, The Netherlands, 24–26 September 2019; pp. 1–4. [Google Scholar]
  10. Chen, Y.; Guo, Y.; Wang, Y.; Dong, W.; Chong, P.; He, G. Denoising of Hyperspectral Images Using Nonconvex Low Rank Matrix Approximation. IEEE Trans. Geosci. Remote Sens. 2017, 55, 5366–5380. [Google Scholar] [CrossRef]
  11. Aggarwal, H.K.; Majumdar, A. Hyperspectral Unmixing in the Presence of Mixed Noise Using Joint-Sparsity and Total Variation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 4257–4266. [Google Scholar] [CrossRef]
  12. He, K.; Sun, J.; Tang, X. Guided Image Filtering. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1397–1409. [Google Scholar] [CrossRef]
  13. Palumbo, N.F.; Harrison, G.A.; Blauwkamp, R.A.; Marquart, J.K. Guidance filter fundamentals. Johns Hopkins APL Tech. Digest 2010, 29, 60–70. [Google Scholar]
  14. Deledalle, C.A.; Duval, V.; Salmon, J. Non-local methods with shape-adaptive patches (nlm-sap). J. Math. Imaging Vision 2012, 43, 103–120. [Google Scholar] [CrossRef] [Green Version]
  15. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image Denoising by Sparse 3-D Transform-Domain Collaborative Filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef]
  16. Zhang, H.; He, W.; Zhang, L.; Shen, H.; Yuan, Q. Hyperspectral image restoration using low-rank matrix recovery. IEEE Trans. Geosci. Remote Sens. 2014, 52, 4729–4743. [Google Scholar] [CrossRef]
  17. Zhuang, L.; Bioucas-Dias, J.M. Fast Hyperspectral image Denoising based on low rank and sparse representations. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium, Beijing, China, 10–15 July 2016; pp. 1847–1850. [Google Scholar]
  18. Hughes, G. On the mean accuracy of statistical pattern recognizers. IEEE Trans. Inf. Theory 1968, 14, 55–63. [Google Scholar] [CrossRef] [Green Version]
  19. Sun, W.; Du, Q. Hyperspectral Band Selection: A Review. IEEE Geosci. Remote Sens. Mag. 2019, 7, 118–139. [Google Scholar] [CrossRef]
  20. Ribalta, L.P.; Tulczyjew, L.; Marcinkiewicz, M.; Nalepa, J. Hyperspectral band selection using attention-based convolutional neural networks. IEEE Access 2020, 8, 42384–42403. [Google Scholar] [CrossRef]
  21. Prasad, S.; Bruce, L.M. Limitations of Principal Components Analysis for Hyperspectral Target Recognition. IEEE Geosci. Remote Sens. Lett. 2008, 5, 625–629. [Google Scholar] [CrossRef]
  22. Slavkovikj, V.; Verstockt, S.; De, N.W.; Van, H.S. Unsupervised spectral sub-feature learning for hyperspectral image classification. Int. J. Remote Sens. 2016, 37, 309–326. [Google Scholar] [CrossRef] [Green Version]
  23. Wei, W.; Zhang, Y.; Tian, C. Latent subclass learning-based unsupervised ensemble feature extraction method for hyperspectral image classification. Remote Sens. Lett. 2015, 6, 257–266. [Google Scholar] [CrossRef]
  24. Bandos, T.V.; Bruzzone, L.; Camps, V.G. Classification of Hyperspectral Images with Regularized Linear Discriminant Analysis. IEEE Trans. Geosci. Remote Sens. 2009, 47, 862–873. [Google Scholar] [CrossRef]
  25. Li, W. Locality-Preserving Dimensionality Reduction and Classification for Hyperspectral Image Analysis. IEEE Trans. Geosci. Remote Sens. 2012, 50, 1185–1198. [Google Scholar] [CrossRef] [Green Version]
  26. Kang, X.; Li, S.; Benediktsson, J.A. Feature extraction of hyperspectral images with image fusion and recursive filtering. IEEE Trans. Geosci. Remote Sens. 2014, 52, 3742–3752. [Google Scholar] [CrossRef]
  27. Tu, B.; Zhang, X.; Kang, X.; Zhang, G.; Wang, J.; Wu, J. Hyperspectral Image Classification via Fusing Correlation Coefficient and Joint Sparse Representation. IEEE Geosci. Remote Sens. Lett. 2018, 15, 340–344. [Google Scholar] [CrossRef]
  28. Zheng, Q.; Yang, M.; Yang, J.; Zhang, Q.; Zhang, X. Improvement of Generalization Ability of Deep CNN via Implicit Regularization in Two-Stage Training Process. IEEE Access 2018, 6, 15844–15869. [Google Scholar] [CrossRef]
  29. Makantasis, K.; Karantzalos, K.; Doulamis, A.; Doulamis, N. Deep supervised learning for hyperspectral data classification through convolutional neural networks. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, IGARSS 2015, Milan, Italy, 26–31 July 2015; pp. 4959–4962. [Google Scholar]
  30. Zhong, Z.; Li, J.; Luo, Z.; Michael, C. Spectral–Spatial Residual Network for Hyperspectral Image Classification: A 3-D Deep Learning Framework. IEEE Trans. Geosci. Remote Sens. 2018, 56, 847–858. [Google Scholar] [CrossRef]
  31. Ma, X.; Geng, J.; Wang, H. Hyperspectral image classification via contextual deep learning. EURASIP J. Image Video Process. 2015, 20, 1–12. [Google Scholar] [CrossRef] [Green Version]
  32. Zhao, W.; Du, S. Spectral-Spatial Feature Extraction for Hyperspectral Image Classification: A Dimension Reduction and Deep Learning Approach. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4544–4554. [Google Scholar] [CrossRef]
  33. Das, A.; Saha, I.; Scherer, R. GhoMR: Multi-Receptive Lightweight Residual Modules for Hyperspectral Classification. Sensors 2020, 20, 6823. [Google Scholar] [CrossRef]
  34. Nalepa, J.; Myller, M.; Imai, Y.; Honda, K.I.; Takeda, T.; Antoniak, M. Unsupervised Segmentation of Hyperspectral Images Using 3-D Convolutional Autoencoders. IEEE Geosci. Remote Sens. Lett. 2019, 17, 1948–1952. [Google Scholar] [CrossRef]
  35. Jiang, J.; Ma, J.; Chen, C.; Wang, Z.; Cai, Z.; Wang, L. SuperPCA: A Superpixelwise PCA Approach for Unsupervised Feature Extraction of Hyperspectral Imagery. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4581–4593. [Google Scholar] [CrossRef] [Green Version]
  36. Tu, B.; Wang, J.; Kang, X.; Zhang, G.; Ou, X. KNN-Based representation of superpixels for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 4032–4047. [Google Scholar] [CrossRef]
  37. Roger, R.E.; Arnold, J.F. Reliably estimating the noise in AVIRIS hyperspectral images. Int. J. Remote Sens. 1996, 17, 1951–1962. [Google Scholar] [CrossRef]
  38. Chang, C.I.; Du, Q. Estimation of number of spectrally distinct signal sources in hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2004, 42, 608–619. [Google Scholar] [CrossRef] [Green Version]
  39. Kim, S.-J.; Koh, K.; Lustig, M.; Boyd, S.; Gorinevsky, D. An Interior-Point Method for Large-Scale -Regularized Least Squares. IEEE J. Sel. Top. Signal. Process. 2007, 1, 606–617. [Google Scholar] [CrossRef]
  40. Bioucas-Dias, J.M.; Nascimento, J.M.P. Hyperspectral subspace identification. IEEE Trans. Geosci. Remote Sens. 2008, 46, 2435–2445. [Google Scholar] [CrossRef] [Green Version]
  41. Liu, M.Y.; Tuzel, O.; Ramalingam, S.; Chellappa, R. Entropy rate superpixel segmentation. In Proceedings of the 24th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; pp. 2097–2104. [Google Scholar]
  42. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep Learning-Based Classification of Hyperspectral Data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 7, 2094–2107. [Google Scholar] [CrossRef]
  43. Zhou, Y.; Wei, Y. Learning Hierarchical Spectral-Spatial Features for Hyperspectral Image Classification. IEEE Trans. Cybern. 2017, 46, 1667–1678. [Google Scholar] [CrossRef]
  44. Prasad, S.; Bruce, L.M. Decision Fusion With Confidence-Based Weight Assignment for Hyperspectral Target Recognition. IEEE Trans. Geosci. Remote Sens. 2008, 46, 1448–1456. [Google Scholar] [CrossRef]
  45. Bioucas-Dias, J.M.; Plaza, A.; Dobigeon, N.; Parente, M.; Du, Q.; Gader, P.; Chanussot, J. Hyperspectral Unmixing Overview: Geometrical, Statistical, and Sparse Regression-Based Approaches. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2012, 5, 354–379. [Google Scholar] [CrossRef] [Green Version]
  46. Zhang, B. Advancement of hyperspectral image processing and information extraction. J. Remote Sens. 2016, 20, 1062–1090. [Google Scholar] [CrossRef]
Figure 1. The schematic of the proposed LRS-HRFMSuperPCA method.
Figure 1. The schematic of the proposed LRS-HRFMSuperPCA method.
Sensors 21 03846 g001
Figure 2. The schematic of the proposed HDTRF method.
Figure 2. The schematic of the proposed HDTRF method.
Sensors 21 03846 g002
Figure 3. Linear SVM.
Figure 3. Linear SVM.
Sensors 21 03846 g003
Figure 4. (a) Indian Pines ground truth; (b) Indian Pines land-cover category; (c) Indian Pines false-color image.
Figure 4. (a) Indian Pines ground truth; (b) Indian Pines land-cover category; (c) Indian Pines false-color image.
Sensors 21 03846 g004
Figure 5. (a) PaviaU ground truth; (b) PaviaU land-cover category; (c) PaviaU false-color image.
Figure 5. (a) PaviaU ground truth; (b) PaviaU land-cover category; (c) PaviaU false-color image.
Sensors 21 03846 g005
Figure 6. (a) Salinas ground truth; (b) Salinas land-cover category; (c) Salinas false-color image.
Figure 6. (a) Salinas ground truth; (b) Salinas land-cover category; (c) Salinas false-color image.
Sensors 21 03846 g006
Figure 7. The effects of different scales on classification accuracy of HSIs; (a) Indian Pines; (b) PaviaU; (c) Salinas.
Figure 7. The effects of different scales on classification accuracy of HSIs; (a) Indian Pines; (b) PaviaU; (c) Salinas.
Sensors 21 03846 g007
Figure 8. The effects of superpixels on classification accuracy of HSIs; (a) Indian Pines; (b) PaviaU; (c) Salinas.
Figure 8. The effects of superpixels on classification accuracy of HSIs; (a) Indian Pines; (b) PaviaU; (c) Salinas.
Sensors 21 03846 g008aSensors 21 03846 g008b
Figure 9. The influence of σ s , σ r and L on the OA of Indian Pines. (a)   σ s ; (b)   σ r ; (c)   L .
Figure 9. The influence of σ s , σ r and L on the OA of Indian Pines. (a)   σ s ; (b)   σ r ; (c)   L .
Sensors 21 03846 g009
Figure 10. The influence of σ s , σ r and L on the OA of PaviaU. (a)   σ s ; (b)   σ r ; (c)   L .
Figure 10. The influence of σ s , σ r and L on the OA of PaviaU. (a)   σ s ; (b)   σ r ; (c)   L .
Sensors 21 03846 g010aSensors 21 03846 g010b
Figure 11. The influence of σ s ,   σ r and L on the OA of Salinas. (a)   σ s ; (b)   σ r ; (c)   L .
Figure 11. The influence of σ s ,   σ r and L on the OA of Salinas. (a)   σ s ; (b)   σ r ; (c)   L .
Sensors 21 03846 g011
Figure 12. The effect of the different kernel function G in SVM on classification accuracy; (a) Indian Pines; (b) PaviaU; (c) Salinas.
Figure 12. The effect of the different kernel function G in SVM on classification accuracy; (a) Indian Pines; (b) PaviaU; (c) Salinas.
Sensors 21 03846 g012
Figure 13. The effect of the different regularization parameter c 1 in SVM on OA; (a) Indian Pines; (b) PaviaU; (c) Salinas.
Figure 13. The effect of the different regularization parameter c 1 in SVM on OA; (a) Indian Pines; (b) PaviaU; (c) Salinas.
Sensors 21 03846 g013
Figure 14. The PSNR values of all bands of original HSI and denoised reconstructed HSI on the Indian Pines.
Figure 14. The PSNR values of all bands of original HSI and denoised reconstructed HSI on the Indian Pines.
Sensors 21 03846 g014
Figure 15. The classification maps obtained by different methods on the Indian Pines when training samples account for 10% of the reference. (a) GroundT. (b) SVM (OA = 81.99%). (c) PCA (OA = 79.58%). (d) IFRF (OA = 97.91%). (e) CCJSR (OA = 96.87%). (f) MSuperPCA (OA = 97.76%). (g) SSRN (OA = 98.73%). (h) KNNRS (OA = 99.18%). (i) GhoMR (OA = 98.66%). (j) LRS-HRFMSuperPCA (OA = 99.34%).
Figure 15. The classification maps obtained by different methods on the Indian Pines when training samples account for 10% of the reference. (a) GroundT. (b) SVM (OA = 81.99%). (c) PCA (OA = 79.58%). (d) IFRF (OA = 97.91%). (e) CCJSR (OA = 96.87%). (f) MSuperPCA (OA = 97.76%). (g) SSRN (OA = 98.73%). (h) KNNRS (OA = 99.18%). (i) GhoMR (OA = 98.66%). (j) LRS-HRFMSuperPCA (OA = 99.34%).
Sensors 21 03846 g015
Figure 16. The classification maps obtained by different methods on the PaviaU when training samples account for 10% of the reference. (a) GroundT. (b) SVM (OA = 94.17%). (c) PCA (OA = 90.50%). (d) IFRF (OA = 99.21%). (e) CCJSR (OA = 93.73%). (f) MSuperPCA (OA = 99.40%). (g) SSRN (OA = 99.47%). (h) KNNRS (OA = 99.45%). (i) GhoMR (OA = 99.54%). (j) LRS-HRFMSuperPCA (OA = 99.88%).
Figure 16. The classification maps obtained by different methods on the PaviaU when training samples account for 10% of the reference. (a) GroundT. (b) SVM (OA = 94.17%). (c) PCA (OA = 90.50%). (d) IFRF (OA = 99.21%). (e) CCJSR (OA = 93.73%). (f) MSuperPCA (OA = 99.40%). (g) SSRN (OA = 99.47%). (h) KNNRS (OA = 99.45%). (i) GhoMR (OA = 99.54%). (j) LRS-HRFMSuperPCA (OA = 99.88%).
Sensors 21 03846 g016
Figure 17. The classification maps obtained by different methods on the Salinas dataset when training samples account for 3% of the reference. (a) GroundT. (b) SVM (OA = 93.05%). (c) PCA (OA = 92.94%). (d) IFRF (OA = 99.42%). (e) CCJSR (OA = 95.37%). (f) MSuperPCA (OA = 99.53%). (g) SSRN (OA = 99.34%). (h) KNNRS (OA = 99.68%). (i) GhoMR (OA = 99.30%). (j) LRS-HRFMSuperPCA (OA = 99.78%).
Figure 17. The classification maps obtained by different methods on the Salinas dataset when training samples account for 3% of the reference. (a) GroundT. (b) SVM (OA = 93.05%). (c) PCA (OA = 92.94%). (d) IFRF (OA = 99.42%). (e) CCJSR (OA = 95.37%). (f) MSuperPCA (OA = 99.53%). (g) SSRN (OA = 99.34%). (h) KNNRS (OA = 99.68%). (i) GhoMR (OA = 99.30%). (j) LRS-HRFMSuperPCA (OA = 99.78%).
Sensors 21 03846 g017
Figure 18. OA of the proposed LRS-HRFMSuperPCA method with different numbers of training samples on different images. (a,b) Indian Pines; (c,d) PaviaU; (e,f) Salinas.
Figure 18. OA of the proposed LRS-HRFMSuperPCA method with different numbers of training samples on different images. (a,b) Indian Pines; (c,d) PaviaU; (e,f) Salinas.
Sensors 21 03846 g018
Figure 19. The accuracy of image fusion and recursive filtering algorithm (IF gets non-low-rank feature image) under adding the influence of hierarchical filtering on Indian Pines.
Figure 19. The accuracy of image fusion and recursive filtering algorithm (IF gets non-low-rank feature image) under adding the influence of hierarchical filtering on Indian Pines.
Sensors 21 03846 g019
Figure 20. Under different additive Gaussian noise environments, the OA of LRS-HRFMSuperPCA algorithm.
Figure 20. Under different additive Gaussian noise environments, the OA of LRS-HRFMSuperPCA algorithm.
Sensors 21 03846 g020
Table 1. Number of samples in the Indian Pines, PaviaU, and Salinas.
Table 1. Number of samples in the Indian Pines, PaviaU, and Salinas.
Indian PinesPaviaUSalinas
Class NamesNumbersClass NamesNumbersClass NamesNumbers
Alfalfa46Asphalt6631Weeds_12009
Corn-N1428Meadows18,649Weeds_23726
Corn-M830Gravel2099Fallow1976
Corn237Trees3064Fallow_P1394
Grass-M483Sheets1345Fallow_S2678
Grass-T730Soil5029Stubble3959
Grass-P28Bitumen1330Celery3579
Hay_W478Bricks3682Grapes11,271
Oats20Shadows947Soil6203
Soybean_N972 Corn3278
Soybean_M2455 Lettuce_4wk1068
Soybean_C593 Lettuce_5wk1927
Wheat205 Lettuce_6wk916
Woods1265 Lettuce_7wk1070
Buildings386 Vinyard_U7268
Stone93 Vinyard_T1807
Total Number10,249Total Number42,776Total Number54,129
Table 2. The experimental parameters settings in LRS-HRFMSuperPCA.
Table 2. The experimental parameters settings in LRS-HRFMSuperPCA.
Datasets C S σ s σ r L G c 1
Indian Pines740300.810300106
PaviaU1040300.2440106
Salinas5401200.96400108
Table 3. The average PSNR of the original Indian Pines and the reconstructed Indian Pines in all bands and the OA of the LRS-HRFMSuperPCA method and HRFMSuperPCA method are compared.
Table 3. The average PSNR of the original Indian Pines and the reconstructed Indian Pines in all bands and the OA of the LRS-HRFMSuperPCA method and HRFMSuperPCA method are compared.
Original Indian PinesReconstructed Indian Pines
Average value of PSNR21.26 dB20.96 dB
OA(10% Train set)99.04%(HRFMSuperPCA)99.27%(LRS-HRFMSuperPCA)
Table 4. The application of experimental parameters in the comparison algorithm.
Table 4. The application of experimental parameters in the comparison algorithm.
MethodParameters or Structure Used in the Comparison Algorithm
SVMPenalty parameter c = 100 , Kernel function parameter G = 0.125 .
PCA(SVM)Feature dimension of Indian Pines after dimension reduction D = 100 , the D of PaviaU and Salinas are 70 and 120, Penalty parameter c = 3000 , Kernel function parameter G = 2 7 .
IFRF(SVM)The parameters of recursive filtering are set to δ s =200 and δ r = 0.3 . The number of features k is set 20, penalty parameter c = 100 , kernel function parameter G = 4 .
CCJSR The   relationship   among   OA ,   the   joint   neighboring   scale   S e , and   the   sparsity   level   S l , the   number   of   nearest   neighbors   N , and   the   regularized   parameter   λ . When   S e is   set   to   6   and   S l is   set   to   2 ,   λ equals   0.6   and   N is set to 4. The highest classification accuracy is obtained by the proposed method.
MSuperPCA(SVM)The scale number of C for Indian Pines, PaviaU, and Salinas is set to 4, 6, and 4, respectively. The number of superpixels is set to the initial optimal value,   S = 100,   S = 20, and 100 for Indian Pines, PaviaU, and Salinas, for SVM, penalty parameter c = 10 5 , Kernel function parameter G = 500 .
SSRNThe SSRN includes two convolutional layers and two spectral residual blocks as a spectral feature learning section, a 3-D convolutional layer and two spatial residual blocks as a spatial feature learning section, an average pooling layer, and a fully connected layer.
KNNRSThe parameters of recursive filtering are set to δ s =204 and δ r = 0.6 . The parameters of KNN are set to k 1 = 1 and k 2 = 9 . The number of superpixels N for Indian Pines, PaviaU, and Salinas is set to 3300, 8000, and 9000.
GhoMRThe GhoMR have two hyperparameters—number of Ghost transformations ( T ) and spatial size of ghost filters ( K T ), for Indian Pines T = 2 and K T = 3 , for PaviaU T = 4 and K T = 7 , for Salinas T = 2 and K T = 3
Table 5. The data distribution along with class-wise accuracies, OA, AA, Kappa on Indian Pines, PaviaU, and Salinas datasets, respectively, for randomly chosen 10% training data on Indian Pines and PaviaU, 3% training data on Salinas, leaving the rest samples to form the test set.
Table 5. The data distribution along with class-wise accuracies, OA, AA, Kappa on Indian Pines, PaviaU, and Salinas datasets, respectively, for randomly chosen 10% training data on Indian Pines and PaviaU, 3% training data on Salinas, leaving the rest samples to form the test set.
Indian Pines (10% Training Data)PaviaU (10% Training Data)Salinas (3% Training Data)
NameTrainingTestAccuracyNameTrainingTestAccuracyNameTrainingTestAccuracy
Alfalfa2323100Asphalt475615699.69Weeds_11021907100
Corn-N77135198.67Meadows47518,17499.92Weeds_21023624100
Corn-M77753100Gravel475162499.88Fallow1021874100
Corn77160100Trees475258999.61Fallow_P1021292100
Grass-M7740699.75Sheets475870100Fallow_S102257699.42
Grass-T77653100Soil475455499.98Stubble102385799.95
Grass-P1414100Bitumen47585599.77Celery102347799.91
Hay_W77401100Bricks475320799.91Grapes10211,16999.91
Oats1010100Shadows475472100Soil1026101100
Soybean_N77895100 Corn102317899.93
Soybean_M77237898.70 Lettuce_4wk102966100
Soybean_C7751697.87 Lettuce_5wk102182599.45
Wheat7712899.22 Lettuce_6wk10281499.16
Woods771188100 Lettuce_7wk102968100
Buildings7730999.68 Vinyard_U102716699.21
Stone4746100 Vinyard_T102170599
OA(%) 99.32OA(%) 99.87OA(%) 99.73
AA(%) 99.62AA(%) 99.86AA(%) 99.75
Kappa 0.9921Kappa 0.9982Kappa 0.9970
Table 6. The data distribution along with class-wise accuracies, OA, AA, Kappa on Indian Pines, PaviaU, and Salinas datasets, respectively, for randomly chosen 20% training data on Indian Pines and PaviaU, 5% training data on Salinas, leaving the rest samples to form the test set.
Table 6. The data distribution along with class-wise accuracies, OA, AA, Kappa on Indian Pines, PaviaU, and Salinas datasets, respectively, for randomly chosen 20% training data on Indian Pines and PaviaU, 5% training data on Salinas, leaving the rest samples to form the test set.
Indian Pines (20% Training Data)PaviaU (20% Training Data)Salinas (5% Training Data)
NameTrainingTestAccuracy (%)NameTrainingTestAccuracy (%)NameTrainingTestAccuracy (%)
Alfalfa2323100Asphalt1139549299.84Weeds_11701839100
Corn-N173125599.60Meadows11391751099.94Weeds_21703556100
Corn-M173657100Gravel1050104999.90Fallow1701806100
Corn119118100Trees1139192599.58Fallow_P1701214100
Grass-M173310100Sheets673672100Fallow_S170250899.80
Grass-T173557100Soil1139389099.97Stubble170378999.84
Grass-P1414100Bitumen66566599.85Celery170340999.94
Hay_W173305100Bricks1139254399.96Grapes1701110199.93
Oats1010100Shadows47447399.79Soil1706033100
Soybean_N173799100 Corn170310899.97
Soybean_M173228299.12 Lettuce_4wk170898100
Soybean_C17342099.29 Lettuce_5wk170175799.94
Wheat10310299.02 Lettuce_6wk17074699.87
Woods1731092100 Lettuce_7wk170900100
Buildings17321399.53 Vinyard_U170709899.59
Stone4746100 Vinyard_T1701637100
OA(%) 99.63OA(%) 99.91OA(%) 99.90
AA(%) 99.79AA(%) 99.87AA(%) 99.93
Kappa 0.9957Kappa 0.9986Kappa 0.9988
Table 7. OA, AA, and Kappa using the LRS-HRFMSuperPCA and other state-of-the-art methods on 10% and 20% randomly chosen training samples.
Table 7. OA, AA, and Kappa using the LRS-HRFMSuperPCA and other state-of-the-art methods on 10% and 20% randomly chosen training samples.
TrainMethodsIndian PinesPercentage of OA Value Increase (%)PaviaUPercentage of OA Value Increase (%)
OA (%)AA (%)KappaOA (%)AA (%)Kappa
10%SVM81.6779.840.787617.6591.0893.390.87918.79
PCA79.0577.020.760220.2790.7488.050.87639.13
IFRF98.4297.800.98250.9099.1498.730.99060.73
CCJSR96.0095.440.94783.3293.6690.800.91606.21
MSuperPCA96.3797.960.95822.9599.3799.420.99140.50
SSRN98.6598.530.98170.6799.7999.660.99720.08
KNNRS99.2496.990.99130.0899.3299.110.99050.55
GhoMR98.6498.000.98450.6899.7599.330.99670.12
LRS-HRFMSuperPCA99.3299.620.9921/99.8799.860.9982/
20%SVM86.2483.150.842713.3995.2093.600.93634.72
PCA82.6480.520.794716.9991.2188.420.88298.71
IFRF98.8598.620.98660.7899.4199.160.99110.51
CCJSR96.8596.220.96412.7896.2893.670.95083.64
MSuperPCA97.8597.960.97491.7899.6199.600.99440.31
SSRN99.4599.520.99370.1899.8799.810.99890.05
KNNRS99.3196.620.99220.3299.5399.180.99360.39
GhoMR99.5499.300.99470.0999.9099.820.99860.02
LRS-HRFMSuperPCA99.6399.790.9957/99.9299.880.9988/
Table 8. OA, AA, and Kappa using the LRS-HRFMSuperPCA and other state-of-the-art methods on 3% and 5% randomly chosen training samples on Salinas.
Table 8. OA, AA, and Kappa using the LRS-HRFMSuperPCA and other state-of-the-art methods on 3% and 5% randomly chosen training samples on Salinas.
TrainMethodsSalinasPercentage of OA Value Increase (%)
OA (%)AA (%)Kappa
3%SVM85.9689.870.842913.77
PCA91.7495.250.90977.99
IFRF99.2899.200.99160.45
CCJSR95.7497.000.95543.99
MSuperPCA99.5399.370.99480.20
SSRN99.4199.360.99330.32
KNNRS99.7299.740.99690.01
GhoMR99.3799.610.99320.36
LRS-HRFMSuperPCA99.7399.750.9970/
5%SVM87.4891.360.860212.42
PCA92.3695.860.914817.54
IFRF99.7599.720.99690.15
CCJSR97.8597.860.97652.05
MSuperPCA99.6999.590.99650.21
SSRN99.6099.580.99530.30
KNNRS99.7999.750.99770.11
GhoMR99.7799.810.99740.13
LRS-HRFMSuperPCA99.9099.930.9988/
Table 9. The computational time (in seconds) of different methods (10% training set).
Table 9. The computational time (in seconds) of different methods (10% training set).
Dataset (10%)SVMPCAIFRFCCJSRMSuperPCASSRNKNNRSGhoMRLRS-HRFMSuperPCA
Indian Pines203184901417469178347
PaviaU55111175977166110616846801618
Salinas1392382105827073324525860762
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Qu, S.; Liu, X.; Liang, S. Multi-Scale Superpixels Dimension Reduction Hyperspectral Image Classification Algorithm Based on Low Rank Sparse Representation Joint Hierarchical Recursive Filtering. Sensors 2021, 21, 3846. https://doi.org/10.3390/s21113846

AMA Style

Qu S, Liu X, Liang S. Multi-Scale Superpixels Dimension Reduction Hyperspectral Image Classification Algorithm Based on Low Rank Sparse Representation Joint Hierarchical Recursive Filtering. Sensors. 2021; 21(11):3846. https://doi.org/10.3390/s21113846

Chicago/Turabian Style

Qu, Shenming, Xuan Liu, and Shengbin Liang. 2021. "Multi-Scale Superpixels Dimension Reduction Hyperspectral Image Classification Algorithm Based on Low Rank Sparse Representation Joint Hierarchical Recursive Filtering" Sensors 21, no. 11: 3846. https://doi.org/10.3390/s21113846

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop