Next Article in Journal
Demodulation of EM Telemetry Data Using Fuzzy Wavelet Neural Network with Logistic Response
Next Article in Special Issue
Spatial Analysis for Landscape Changes
Previous Article in Journal
Plasmon Resonant Two-Photon Luminescence Inducing Photosensitization and Nonlinear Optical Microscopy In Vivo by Near-Infrared Excitation of Au Nanopeanuts
Previous Article in Special Issue
Spatial Analysis for Landscape Changes: A Bibliometric Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Superpixel-Based Singular Spectrum Analysis for Effective Spatial-Spectral Feature Extraction

by
Subhashree Subudhi
1,
Ramnarayan Patro 
1,
Pradyut Kumar Biswal
1,* and
Fabio Dell’Acqua
2
1
Department of Electronics and Communication Engineering, IIIT, Bhubaneswar 751003, India
2
Department of Electrical, Computer and Biomedical Engineering, University of Pavia, 27100 Pavia, Italy
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(22), 10876; https://doi.org/10.3390/app112210876
Submission received: 29 September 2021 / Revised: 1 November 2021 / Accepted: 2 November 2021 / Published: 17 November 2021
(This article belongs to the Special Issue Spatial Analysis for Landscape Changes)

Abstract

:
In the processing of remotely sensed data, classification may be preceded by feature extraction, which helps in making the most informative parts of the data emerge. Effective feature extraction may boost the efficiency and accuracy of the following classification, and hence various methods have been proposed to perform it. Recently, Singular Spectrum Analysis (SSA) and its 2-D variation (2D-SSA) have emerged as popular, cutting-edge technologies for effective feature extraction in Hyperspectral Images (HSI). Using 2D-SSA, each band image of an HSI is initially decomposed into various components, and then the image is reconstructed using the most significant eigen-tuples relative to their eigen-values, which represent strong spatial features for the classification task. However, instead of performing reconstruction on the whole image, it may be more effective to apply reconstruction to object-specific spatial regions, which is the proposed objective of this research. As an HSI may cover a large area, multiple objects are generally present within a single scene. Hence, spatial information can be highlighted accurately by specializing the reconstruction based on the local context. The local context may be defined by the so-called superpixels, i.e., finite sets of pixels that constitute a homogeneous set. Each superpixel may undergo tailored reconstruction, with a process expected to perform better than non-spatially-adaptive approaches. In this paper, a Superpixel-based SSA (SP-SSA) method is proposed where the image is first segmented into multiple regions using a superpixel segmentation approach. Next, each segment is individually reconstructed using 2D-SSA. In doing so, the spatial contextual information is preserved, leading to better classifier performance. The performance of the reconstructed features is evaluated using an SVM classifier. Experiments on four popular benchmark datasets reveal that, in terms of the classification accuracy, the proposed approach overperforms the standard SSA technique and various common spatio-spectral classification methods.

1. Introduction

Recent advancements in hyperspectral sensors resulted in the increased availability of Hyperspectral Images (HSI) and a boost in their circulation among the remote sensing community. HSI data enables the discrimination of objects even with minor differences as it contains several contiguous spectral bands acquired from the visible to the infrared region [1] so that every small spectral difference can, in principle, be captured. The information is available in the form of a 3-D structure that contains a 2-D spatial scene along with a 1-D spectral signature. These unique characteristics of HSI have made them popular in several application areas, such as agriculture [2], mineralogy [3], land cover classification [4], target detection [5], and others. However, effective classification of HSI is still an open challenge.
Several classification techniques, such as K-nearest neighbor (KNN) [6], support vector machine (SVM) [7], multinomial logistic regression (MLR) [8], Extreme Learning Machine (ELM) [9], and Sparse Representation Classifier (SRC) [10] have been proposed in the past decades. The richness in spectral information attracted research efforts on pixel-based processing and classification. SVM is the most popular and widespread classifier due to its lower generalization error rate that makes it capable of identifying even minor changes in spectral signatures. Due to the high spectral dimensionality compared to a generally limited number of class-specific training samples, it is quite difficult to properly estimate the model parameters. Hence, there is a need to adopt effective spatial-spectral feature extraction approaches to overcome the aforementioned challenges.
To deal with the issue of higher spectral dimensionality, several linear (e.g., Principal Component Analysis (PCA) [11], Independent Component Analysis (ICA) [12], and Linear Discriminant Analysis (LDA) [13]) and non-linear (manifold learning [14]) dimensionality reduction (DR) methods have been introduced. Band selection approaches may also be utilized to select the most informative bands out of several available bands [15,16]. Spectral features alone, however, may not be sufficient to score very high accuracy values. To improve performance, it is necessary to incorporate spatio-spectral features that help increasing the separability of classes.
Recently, various spatial feature extraction techniques have been proposed. Mathematical Morphology is one of the most popular approaches that is extensively utilized by researchers. The concept of Extended Morphological Profiles (EMP) for FE in HSI was first proposed by Benediktsson et al. [17]. This technique utilizes morphological opening and closing transformations to extract spatial geometrical information. Later, Dalla Mura et al. [18] proposed Morphological Attribute Filters (MAP) for the spatial FE. From that point onwards, several variations of Attribute Profiles (AP) were created. Ghamisi et al. [19] conducted a comprehensive survey on the evolution in Attribute Profiles.
Texture Descriptors, including Wavelet transform [20], Gray-Level Co-occurrence Matrix (GLCM) [21], Local Binary Patterns (LBP) [22], and Gabor filters [23], have also been used in the literature for spatial FE. Filtering, i.e., moving-window-based processing, is another approach to extract spatial-spectral features. Various edge-preserving filters, such as Bilateral Filters [24], Trilateral Filters [25], Guidance filters [26], and Domain Transform Recursive Filters [27] have been introduced for spatial FE in the literature. The texture and noise variations are minimized by performing smoothing operations; however, important details, such as edges and lines, are well preserved by these filters [28].
In addition to these techniques, 2D-SSA is another interesting approach for spatial feature extraction. Using this approach, each band image of HSI is initially decomposed into varying trends, oscillations, and noise. Later, HSI is reconstructed using the selected oscillations and trends [29,30]. In 2D-SSA, the spatial structural information is extracted by utilizing the characteristics of the surrounding pixels in a specific embedding window. It can withstand high levels of noise and generally achieves good data classification results.
2D-SSA suffers, however, from several limitations, such as reduced utilization of the abundant spectral information available in data, and over-smoothing or under-smoothing of classification results because of the fixed embedding window size. To overcome the challenge of selecting the optimal embedding window size, recently, multi-scale 2D-SSA has been proposed for the effective extraction of discriminative spatial features under different noise conditions [31,32].
Superpixel segmentation techniques have gained popularity in recent years due to their capability of exploiting spatial structural information adaptively in an image. In [33], a survey on superpixel segmentation as a preprocessing step in HSI analysis was presented. A superpixel-based classification via multiple kernels (SCMK) approach was proposed in [34]. In [35], a region-based relaxed multiple kernel (R2MK) method was proposed that combines the multiscale spectral and spatial features using a kernel collaborative representation classification technique.
To obtain superior classification performance and solve the problem of optimal superpixel number selection, an adjacent superpixel-based multiscale spatial-spectral kernel (ASMGSSK) was proposed in [36]. In [37], a multiscale segmentation-based SuperPCA model (MSuperPCA) was developed, which can effectively integrate multiscale spatial information to obtain the optimal classification result by decision fusion.
Recently, deep learning techniques have become quite popular in the classification of HSI data due to their ability to extract discriminant and abstract features by using a series of hierarchical layers. The initial layers usually extract texture and edge information, whereas deeper layers highlight more complicated features. Some of the most popular deep learning frameworks include stacked autoencoders (SAE) [38], Deep Belief Networks (DBN) [39], Convolution Neural Networks (CNN) [40], Recurrent Neural Networks (RNN) [41], Generative Adversarial Networks (GAN) [42], etc.
Although deep learning approaches have several advantages, they also pose significant challenges in HSI applications. First of all, to achieve better classification result, often deep learning techniques demand large volumes of training samples. Moreover, a large number of hyper-parameters (like the kernel sizes, learning rate, etc.) are involved in training complex deep learning networks mainly designed for feature extraction and classification. Hence, the process becomes computationally expensive.
The disadvantages of combining SSA with structured approaches to incorporating spatial information may be overcome by using more flexible ways to spatially partition the dataset. In line with this consideration, in this work, a superpixel-based SSA (SP-SSA) algorithm was proposed as a means to increase the classifier performance. Instead of performing direct reconstruction, an object-specific reconstruction is performed to accurately preserve the local contextual information. Superpixel segmentation is first applied on the input HSI to generate a segmented HSI where each sub-region carries similar characteristic features, and its shape and size is adjusted according to the local image structure information. Next, 2D-SSA is individually applied on each segmented region to produce the reconstructed HSI. Lastly, the final classification map is generated by using the popular SVM classifier. The major novel contributions of this work are highlighted in the following list:
  • Direct reconstruction is usually performed in standard 2D-SSA algorithms, where the full image is reconstructed. In HSIs, however, object-specific reconstruction is always better than direct reconstruction, as, in this way, local contextual information can be captured accurately. In this work, a novel SP-SSA approach is proposed that performs object-specific reconstruction.
  • Superpixel segmentation and 2D-SSA are combined together for the first time for accurate spatial-spectral feature extraction. Using SP-SSA, each superpixel, i.e., object-specific spatial region is reconstructed.
  • Superior classifier performance is achieved with the proposed method in comparison to other state-of-the-art methods, even with a comparatively small number of training samples.
The remainder of this paper is organized as follows. A detailed description of the proposed method is presented in Section 2. The experimental setup, results, and analysis are described in Section 3. Finally, some conclusions and future work are discussed in Section 4.

2. The Proposed Methodology

The proposed SP-SSA method includes three stages as described in Figure 1. In stage 1, superpixel segmentation is applied on the input HSI to obtain the segmented HSI. In stage 2, each segmented region is reconstructed using 2D-SSA to obtain the reconstructed HSI. In the final stage, an SVM classifier is applied on the reconstructed HSI to build the final classification map. A detailed description of each of these stages is presented in the subsections below.

2.1. Superpixel Segmentation

Superpixel segmentation approaches have gained popularity in recent years as these approaches have several benefits. Using superpixels, the computational complexity can be drastically reduced by computing features on more meaningful regions rather than acting on each individual pixel in HSI [43]. Simple Linear Iterative Clustering (SLIC) [44] is one of the most popular gradient-ascent-based superpixel segmentation approaches, where an initially defined tentative set of cluster points are iteratively refined using a gradient-ascent method until some convergence criteria are met. This algorithm has lower computational complexity as it applies the k-means method locally. The algorithm includes four key steps that can be summarized as follows.
The first step is cluster center initialization. Let the input HSI be denoted as H b { h 1 b , h 2 b , , h N b } with N pixels, where { h i b } represents the value at the i th pixel for the b th spectral band and i = 1 , 2 , N ; b = 1 , 2 , B . B is the total number of spectral bands. Each pixel can be labeled as A i = [ h i , r i , u i ] , where h i T = [ h 1 , h 2 , , h B ] T is the spectral vector and [ r i , u i ] T is the position vector. The K number of initial cluster centers C j = [ h j , r j , u j ] T are sampled on a regular Q × Q ( Q = N K ) grid and are, thus, equally spaced apart [45].
The next step is the cluster assignment step, where each pixel is assigned to the nearby cluster center based on the computed distance measure D. Distance is computed within a 2 Q × 2 Q window around the cluster center. The distance between the cluster center C j and pixel A i is calculated as follows (Equation (1)):
D = D s p e c t r a l + w Q D s p a t i a l
where w is the weighting factor between spectral and spatial features. The spectral and spatial distance between pixel i and j are represented as in Equations (2) and (3) below.
D s p e c t r a l = b = 1 B h i b h j b 2
where D s p e c t r a l is the measure of homogeneity within the superpixels.
D s p a t i a l = ( r i r j ) 2 + ( u i u j ) 2
where ( r i , u i ) denotes the location of pixel i in the superpixels. The spatial distance D s p a t i a l ensures regularity and compactness in the generated superpixels.
In the third step, the cluster centers are updated with the mean value of all pixels belonging to the same cluster. The second and third steps are iteratively repeated until convergence is achieved.
In the final step, post-processing is performed to enforce connectivity by reassigning disjoint pixels to nearby superpixels.

2.2. 2D-SSA

SSA is capable of decomposing a series into multiple independent components or subseries, where each extracted eigenvalue represents an individual component of the original series. The SSA can be applied to the respective spectral bands of the hypercube, thereby, decomposing the 2-D scene, and then reconstructing it using the respective main components while removing the noise contribution. As a data cube is decomposed in this way, the local structure and main spatial trends are typically found in the first component. Hence, when all images within the hyperspectral cube are decomposed and only the first components are selected to individually reconstruct each of them, a resulting cube with minimum noise is generated. The SSA can be implemented using the following four steps:

2.2.1. Embedding

Imagine a HSI dataset H, with a size of N x × N y × B , where N x , N y indicates the band image size and B represents the total number of available bands. Each band image H b ( b B ) can be expressed as follows:
H b = H 1 , 1 b H 1 , 2 b H 1 , N y b H 2 , 1 b H 2 , 2 b H 2 , N y b H N x , 1 b H N x , 2 b H N x , N y b N x × N y
Next, a 2D window Q b is defined, whose dimensions are M x × M y .
Q b = H i , j b H i , j + 1 b H i , j + M y 1 b H i + 1 , j b H i + 1 , j + 1 b H i + 1 , j + M y 1 b H i + M x 1 , j b H i + M x 1 , j + 1 b H i + M x 1 , j + M y 1 b M x × M y
where 1 M x N x , 1 M y N y , and 1 < M x M y < N x N y . Each pixel is spatially positioned by ( i , j ) within the image H b . The pixels in a window Q b can be rearranged into a column vector C i , j b R M x M y according to the reference position ( i , j ) as follows:
C i , j b = [ H i , j b , H i , j + 1 b , , H i , j + M y 1 b , H i + 1 , j b , H i + 1 , j + 1 b , , H i + M x 1 , j + M y 1 b ] T
To scan the whole image H b , this 2-D window is slid across it from top left to bottom right until it has visited every position on the entire image (see also Figure 2 for a graphical explanation).
As a result, the trajectory matrix Z b of all feasible 2-D windows of image H b of size M x M y × ( N x M x + 1 ) ( N y M y + 1 ) can be obtained as follows:
Z b = ( C 1 , 1 b ) T , ( C 1 , 2 b ) T , , ( C 1 , N y M y + 1 b ) T , ( C 2 , 1 b ) T , , ( C N x M x + 1 , N y M y + 1 b ) T M x M y × ( N x M x + 1 ) ( N y M y + 1 )
Note that the trajectory matrix Z b has a structure of Hankel–block–Hankel (HbH). Z b can be expressed as follows:
Z b = P 1 b P 2 b P N x M x + 1 b P 2 b P 3 b P N x M x + 2 b P M x b P M x + 1 b P N x b M x × ( N x M x + 1 )
Each of the submatrices P i b corresponds to a Hankel structure as follows:
P i b = H i , 1 b H i , 2 b H i , N y M y + 1 b H i , 2 b H i , 3 b H i , N y M y + 1 b H i , M y b H i , M y + 1 b H i , N y b M y × ( N y M y + 1 )

2.2.2. Singular Value Decomposition (SVD)

After obtaining the trajectory matrix Z b , SVD is applied to determine the eigenvalues λ 1 λ 2 λ M x M y , and the corresponding eigenvectors U 1 , U 2 , , U M x M y of Z b Z b T . It is possible to rewrite Z b as follows:
Z b = Z 1 b + Z 2 b + + Z M x M y b
where the i th elementary matrix is Z i b = λ i U i V i T and its Principal Components (PCs) are defined as V i = ( Z b ) T U i λ i

2.2.3. Grouping

A subsequent operation is eigenvalue grouping, during which the total set of M x M y individual components in (10) are divided into m subsets, designated as S = [ S 1 , S 2 , , S m ] . By selecting one or more elementary matrices Z i b from each subset, it is possible to derive the main information contained in an image without being disturbed by high noise levels. As a result, the trajectory matrix Z b can be represented as follows:
Z b = Z S 1 b + Z S 2 b + + Z S m b
The reconstruction of a single band scene of HSI using various numbers of components ( Z i b ) is compared in Figure 3. In general, the component with the highest eigenvalue is the most informative one, containing key features with the lowest noise contribution. With the inclusion of additional components, the reconstructed scene begins to resemble the actual scene. The reconstructed image obtained by grouping the 1st–5th components and 1st–10th components are very similar with marginal differences (Figure 3c,d). Hence, a small number of key components are sufficient to reconstruct the scene satisfactorily.

2.2.4. Diagonal Averaging

Z b , in this case, does not necessarily belong to the HbH matrix type. It is projected into a 2D-signal by applying the Hankelization process in two steps; first inside every block (9) and next block-to-block (8) by averaging the anti-diagonal elements in the matrix. Thus, it is possible to obtain a reconstructed image that contains the distinctive spatial features based on the local contextual information present in a 2D window defined by the user.

2.3. Novelty of the Proposed SP-SSA Method

The proposed approach integrates SSA and superpixel segmentation for the first time to extract improved the spatio-spectral features from HSI. Reconstruction of object-specific spatial sections, rather than the entire image, may be more effective. Hence, in the proposed work, 2D-SSA is applied individually to each superpixel segmented region to extract the local contextual information accurately. The pseudo-code for the proposed SP-SSA algorithm is outlined in Algorithm 1.
Algorithm 1: Proposed SP-SSA algorithm for HSI classification.
Input: HS image, H R n × b
   Ground Truth G T
   number of superpixels: K
   Embedding Window Size: M x × M y
   Eigen Value Grouping: E V
Output: Classification Map c l s m a p generated by SVM.
  1: for b = 1 to B do
  2:  Perform SLIC superpixel segmentation to obtain segmentation map L from h b
     containing K superpixel segments
     L = SLIC ( h b ,K)    (as outlined in Section 2.1)
  3:  for k = 1 to K do
  4:     h k b = r e c o n s t r u c t 2 D S S A (L, M x , M y , E V )    (as outlined in Section 2.2)
  5:  end for
  6: end for
  7: Obtained Reconstructed HSI H R n × b
  8:  c l s m a p = SVM ( H , GT)
For each superpixel, the reconstruction ( r e c o n s t r u c t 2 D S S A (Algorithm 1)) is applied to the rectangular Region of Interest (ROI) surrounding the superpixel (Figure 4). The ROI is created based on the location information of the pixels available in that particular segment. Only the reconstructed pixels specific to those pixels in the selected superpixel are stored as spatial features, while the remaining reconstructed pixels in the ROI are discarded as they do not belong to the superpixel under test. The same procedure is applied to all other superpixels, and the HS image is reconstructed using the proposed SP-SSA approach. This procedure collects local object-specific superpixel-based spatial features for each band in the image.

2.4. Classification

The selection of an appropriate classifier is critical in assessing the performance of the above-mentioned features, especially in hyperspectral images with a limited number of training samples. SVM is the most widely used supervised statistical learning framework among pixel-wise classifiers. With the help of a kernel function, data can be mapped to a higher-dimensional space via a nonlinear transformation, aiming to determine the best hyperplane for separating samples belonging to different classes. The performance of SVM in HSI classification is outstanding despite the variation of the data dimensions [46,47]. Hence, in this work, the SVM classifier is utilized to evaluate the performance of the reconstructed features.

3. Results and Discussion

This section reports the outcome of testing the proposed approach on some of the most popular benchmark datasets and compares it with other, state-of-art classification approaches.

3.1. Dataset Description

In this subsection, the datasets used for testing the proposed approach are presented and described.

3.1.1. Indian Pines

The first dataset, named “Indian Pines” (IP), was collected over Northwestern Indiana, USA, with the airborne AVIRIS sensor; it includes a total of 220 bands covering wavelengths from 0.4 to 2.5 μ m. About 70% of the imaged area is agricultural land, while the remaining portions are forests. Due to the comparatively low spatial resolution (20 m/pixel) of the sensor, this dataset is challenging as it contains highly mixed pixels. The number of samples obtained per class is also unbalanced, which further complicates classification. The size of the scene is 145 × 145 pixels, and its Ground Truth (GT) data defines 16 different classes. The pseudo-color image, the GT map, and the class names for the dataset are all included in Figure 5.

3.1.2. Pavia University

The ROSIS sensor was instrumental to the collection of this dataset over the University of Pavia, Italy. The dataset is called “Pavia University” (PU). It has a spatial resolution of 1.3 m and originally comprises 115 spectral bands covering wavelength ranges from 0.43 to 0.86 μ m. In the final analysis, 103 bands are used after the elimination of noisy channels. The image has a size of 610 × 340 pixels, and it has nine challenging classes with nearly similar spectral reflectances. Detailed information about the false-color image, Ground Truth, and class names is displayed in Figure 6.

3.1.3. Salinas Dataset

The “Salinas” (SAL) dataset was captured over the Salinas Valley, California, USA, using the AVIRIS Sensor. The sensor has 224 channels with spectral range varying from 0.43 μ m to 2.5 μ m. This scene has a size of 512 × 217 pixels and spatial resolution of 3.7 m per pixel. The number of bands reduces to 204 after discarding 20 water absorption bands: [108–112], [154–167], 224. The scene is mainly an agricultural area, with 16 classes in its Ground Truth. A false color representation, the Ground Truth, and the class names for the Salinas dataset are shown in Figure 7.

3.1.4. Houston 2018

The 2018 IEEE GRSS Data Fusion Contest (DFC) triggered public dissemination of this rich dataset, which was included in our tests to increase their statistical significance. The image of the Houston campus and its surrounding area was captured by the IRTES CASI-1500 sensor at a GSD of 1 m over Houston, Texas, USA. It has 601 × 2384 pixels and 50 spectral bands with wavelengths ranging from 380 to 1050 nm sampled at 10 nm intervals. The scene contains 20 urban landcover classes. The false-color composite image, ground truth image, and class names for the Houston 2018 dataset are provided in Figure 8.

3.2. Experimental Setup

Our proposed approach was evaluated by comparing its performance with eight state-of-the-art approaches for HSI feature extraction (Algorithm 2, see Section 3.4.5). These include SVM [7], Edge Preserving Filter (EPF) [26], superpixel-based classification via multiple kernels (SCMK) [34], region-based relaxed multiple kernel (R2MK) [35], adjacent superpixel-based multiscale generalized spatial-spectral kernel (ASMGSSK) [36], Multiscale superpixel-based PCA (MsuperPCA) [37], 2D Singular Spectrum Analysis (2D-SSA) [29], and 2D Multiscale Singular Spectrum Analysis (2D-MSSA) [31]. A common way to measure the efficiency of feature extraction is through the accuracy of the classifier scored by the experiments. As a result, the classification setup must be appropriate with the current state-of-the-art. In light of this, SVMs have demonstrated themselves to be robust and efficient in multi-class classification applications.
The LIBSVM toolbox [48] is used to implement SVM as the default classifier for all of the involved methods. A Gaussian RBF kernel is utilized for SVM implementation, and a grid search is applied to tune both key parameters of RBF-SVM; the penalty c and the gamma γ . The SVM parameters are kept constant across all competitive experiments for a fair comparison. To avoid systematic errors and reduce random discrepancies, all experiments were independently carried out ten times each with different training and testing subsets, with no overlap between each training and the corresponding testing subset. This was intended to ensure good statistical significance for our experiments.
Stratified sampling was used to randomly obtain the training and testing subsets. For training, 3%, 2%, 1%, and 0.2% samples per class were selected for the IP, PU, SAL, and Houston 2018 datasets, respectively. Additionally, four objective quality indices are utilized to evaluate image classification results: namely the OA, the average accuracy (AA), the kappa coefficient, and class-by-class accuracy. All experiments were conducted using MATLAB R2018b software, installed on a personal computer with an Intel core i5-6200 CPU clocked at 2.30 GHz, and 16 GB RAM.

3.3. Parameter Sensitivity Analysis

Table 1 displays the best parameter settings for the competing algorithms, found by experimentation. For the proposed SP-SSA algorithm, the size of the 2-D embedding window was set to 5 × 5 pixels for the IP and Salinas dataset; whereas, for the PU and Houston 2018 dataset, the window size was set at 3 × 3 pixels. For the IP and SAL datasets, superpixels were set at 100. However, the amount of superpixels in the PU and Houston 2018 datasets were set to 150 and 500, respectively. The effect of window size variation for different number of superpixels on the classification performance for the experimental datasets is provided in Figure 9.
As each superpixel is reconstructed individually, smaller window sizes are preferred since they lead to better image reconstructions. Using a large window may smooth the results too much and result in mixing errors. A 2D-SSA algorithm was presented in [29] for feature extraction in HSI, where various window sizes, such as 5 × 5, 10 × 10, 20 × 20, 40 × 40, and 60 × 60, were examined. The IP and SAL datasets produced the best classification accuracy when the window size was set at 10 × 10. When analyzing the PU and Houston dataset, the window sizes of 5 × 5 showed the best classification results. Since the optimal window size may vary depending on the dataset, ref. [31] adopts a multiscale strategy to improve the generalization ability.

3.4. Experimental Result and Analysis

In this section, the four HSI data sets outlined in Section 3.1 are utilized, and several experiments are performed to examine the efficacy of the proposed SP-SSA method. Figure 10 compares the classification results obtained with varying numbers of training samples on four datasets. It can be noted that better classification performance is evident when larger numbers of labeled samples are utilized for training; after passing the percentages used in this work; however, the accuracy level mostly plateaus, and no further significant improvement is observed. Our proposed approach attains the best classification accuracy in almost all cases, regardless of the number of samples, proving its robustness. Classification results from all four data sets are provided in Table 2, Table 3, Table 4 and Table 5 and quantitatively support the dominance of the proposed method. Individual classification maps generated by the proposed SP-SSA method and all the compared approaches are displayed in Figure 11, Figure 12, Figure 13 and Figure 14.

3.4.1. Results from the Indian Pines Dataset

Based on the results shown in Table 2, the proposed method achieves the best values across three metrics, and its accuracy exceeded 89% on almost all classes. In the tables, the best results in each row are highlighted in bold font. When comparing SP-SSA with raw HSI data, the OA improved substantially from 76.42% to 98.15%. In addition, comparisons between SVM and other methods indicated that the incorporation of spatial features can enhance the classification performance compared to considering spectral features alone.
Superpixel-based methods, such as SCMK, R2MK, ASMGSSK, and MsuperPCA techniques, yield higher classification accuracy as compared to non-superpixel based techniques (EPF, 2D-SSA, and 2D-MSSA); by grouping spectrally identical regions, superpixels offer a powerful way to exploiting spatial/contextual information. It can also be noted that methods considering multi-scale windows (ASMGSSK, MsuperPCA, and 2D-MSSA) perform better with respect to fixed-window methods. Due to the different window sizes, unique local spatial features can be exploited, which allows better covering of different sizes of land cover classes and different scales of spatial features. On the downside, the use of multiscale approaches involves heavier processing burdens. As the proposed method reconstructs each superpixel individually, better classification results are obtained.
Figure 11 displays the classification maps produced by various approaches for the Indian Pines dataset. For the SVM approach, the classification map appears very noisy if spatial features are not considered. Through the use of neighborhood spatial information, the EPF and 2D-SSA techniques can suppress spot-wise misclassification to a large extent, but these methods do not preserve the detailed structures of the HSI well enough.
However, by adopting superpixel-based approaches, the generated classification map becomes much smoother, and more accurate estimates are obtained in the detailed region. With the utilization of multi-scale approaches (like ASMGSSK, MSuperPCA, and 2D-MSSA), the amount of misclassification is further reduced. Still, even with multi-scale approaches, landcover boundaries are frequently misplaced. As can be observed from Figure 11, the proposed approach effectively solved the above-mentioned problems due to its considerate utilization of spectral and spatial features.

3.4.2. Results from the Pavia University Dataset

Quantitative results are presented in Table 3. The proposed SP-SSA method still achieved higher classification accuracy and ranked first among all the compared methods, closely followed by the ASMGSSK algorithm. Also, in comparison to EPF, SCMK, R2MK, 2D-SSA, MSuperPCA, and 2D-MSSA techniques, the average improvement of the proposed approach is over 4.41%, 3.64%, 2.09%, 2.37%, 1.3%, and 1.48%, respectively. For comparison, the top results in the tables are boldfaced. In Figure 12, different classification maps are shown, based on various testing methods applied to the PU dataset.
According to Figure 12, the classification map for SVM still continues to remain noisy. Both EPF and 2D-SSA can generate a relatively smooth result; however, some significant regions remain undetected (e.g., the detailed areas). The superpixel-based methods (SCMK, R2MK, ASMGSSK, and MSuperPCA) and SSA-based approach (2D-SSA and 2D-MSSA) offer significantly improved performance, but the proposed 2D-SSA method remains the most promising approach as it outperforms all the compared algorithms.

3.4.3. Results from the Salinas Dataset

The visual classification maps and quantitative results obtained by various classifiers on the Salinas dataset are shown in Figure 13 and Table 4, respectively. In the table, the best results are shown in bold. Based on the visual quality as well as objective metrics, it can be observed that the proposed SP-SSA method outperformed other competing approaches. In addition, compared with the 2D-SSA method that globally reconstructs the image using fixed-size embedded windows, the SP-SSA method considers the local spatial information by reconstructing each superpixel individually, which helps in further reducing the disturbances and improving the class assignment.

3.4.4. Results from the Houston 2018 Dataset

The quantitative results for the Houston 2018 dataset with 0.2% training samples from each class are presented in Table 5. The corresponding classification map is shown in Figure 14. The best results from the tables are displayed in bold font for comparison. As observed from Table 5, the proposed methods are robust and achieve good classification results even for challenging scenes. The proposed approach improves accuracy from 68.19% to 83.57% for the SVM method. In this case also, the superpixel-based approaches (SCMK, R2MK) display superior performance as compared to non-superpixel based methods (EPF, 2DSSA). Here also, multi-scale window approaches (ASMGSSK, MsuperPCA, and 2D-MSSA) outperform fixed-window based methods as different scales of spatial features are incorporated into the analysis. Figure 14 also highlights the superiority of the proposed method. The salt and pepper noise is reduced by a greater extent, and a smoother classification map is produced with the proposed method.

3.4.5. Statistical Evaluation

The effectiveness of the proposed method was statistically evaluated using McNemar’s test. The classification results for all the test cases were compared using this test. The McNemar’s test is defined as in Equation (12), where it is assumed that two generic algorithms, named Algorithm 1 and Algorithm 2 are compared.
Z = f 12 f 21 f 12 + f 21
In the equation above, f 12 indicates the number of samples correctly classified by Algorithm 1 and incorrectly classified by Algorithm 2, and f 12 indicates the number of samples for the opposite case. The performance of Algorithm 1 is better than Algorithm 2 if Z > 0 . The differences between Algorithm 1 and Algorithm 2 are statistically significant if Z > 1.96 . In our case, Algorithm 1 is the algorithm proposed in our manuscript, and Algorithm 2 is —sequentially— each one from the list of standard algorithms: SVM, EPF, SCMK, R2MK, ASMGSSK, MsuperPCA, 2DSSA, 2DMSSA.
McNemar’s test between the proposed SP-SSA algorithm and the algorithms listed above for the Indian Pines, Pavia University, Salinas, and Houston 2018 datasets are provided in Table 6. The test result clearly reveals that the classification results for the proposed method were significantly better—in a McNemar’s statistical sense—compared with other approaches.

3.5. Advantage of Proposed Method over 2D-SSA

3.5.1. Applying SP-SSA on General Images

In the proposed approach, 2D-SSA is applied on each and every superpixel segmented region. Hence, it can be considered as a local 2D-SSA approach that can extract accurate spatial information on each single object. In the case of global 2D-SSA, features are over-smoothed, and features are not prominent for specific classes. In local 2D-SSA instead, object-specific texture information can be highlighted. In Figure 15, the popular cameraman image and an artificial test image are used to demonstrate the effectiveness of the proposed approach over the 2D-SSA approach.
When the cameraman image is reconstructed using the 2D-SSA method, the Mean Square Error (MSE) comes out to 115.8865; however, when the same image is reconstructed using the proposed SP-SSA approach, the MSE reduces to 93.0468. A similar behavior is also observed with the test image. With the proposed SP-SSA method, the MSE reduces to 237.1038 from 287.5323. This signifies that the proposed method can reconstruct an image with minimum error and can effectively integrate local information during the reconstruction process.

3.5.2. Applying SP-SSA on HSI

The HSI is composed of a stack of 2D images carrying valuable information about each spectral band. To demonstrate the effectiveness of the proposed method, a randomly selected spectral band at 667 nm was considered for our analysis. Figure 16b,c contains the scene as reconstructed by 2 D - S S A and S P - S S A , respectively. Since the HSI was acquired over a large area, it includes multiple objects with different textural information. This is a typical case where object-specific reconstruction works better than direct reconstruction. Textural information can be highlighted accurately by using local reconstruction as opposed to global reconstruction. The error in SP-SSA-based reconstruction is indeed lower as compared to 2D-SSA-based reconstruction. The same conclusion can also be drawn from Figure 16.
In the case of 2D-SSA-based reconstruction, the Mean Square Error (MSE) is 612.4349, while, in the case of SP-SSA-based reconstruction, the MSE is 504.5685. Figure 16d,e contains the difference image for 2D-SSA-based reconstruction and SP-SSA-based methods. It can be clearly observed that edge information is preserved with the proposed method. The SP-SSA-based reconstruction is applied to all spectral bands and generates a modified hypercube with preserved local structure information and minimum noise level. These latter features generally lead to better classification performance.

4. Conclusions and Future Scope

Feature extraction is one of the most crucial steps in HSI classification. It is essential to capture comprehensive spatial and spectral information for accurate feature extraction. For image reconstruction, the conventional 2D-SSA algorithm usually extracts spatial features directly by applying the embedding window to the entire image. However, HSI scenes frequently encompass a broader area and contain several items. As a result, spatial information pertaining to local objects must be recovered. To solve this problem, in the proposed method, a superpixel-based SSA technique was presented, which can capture the object specific spatio-spectral information accurately.
In this work, the original HSI was first divided into various semantic sub-regions by the superpixel segmentation algorithm. Next, each segment was reconstructed individually by applying 2D-SSA. The generated reconstructed HSI was then classified using the SVM classifier, and the final classification map was produced. Local characteristics may be collected effectively in the suggested method since 2D-SSA is applied at the superpixel level. However, two parameters must be adjusted: the amount of superpixels and the embedding window size. Future developments will aim at finding the optimal criteria to determine the parameters of the procedure and to investigate relationships between the characteristics of the HSI and quality of the results.

Author Contributions

Conceptualization, P.K.B. and R.P.; methodology, S.S.; software, R.P. and S.S.; validation, S.S., R.P. and P.K.B.; formal analysis, S.S.; investigation, S.S. and R.P.; resources, F.D. and P.K.B.; data curation, S.S.; writing—original draft preparation, S.S.; writing—review and editing, F.D.; visualization, S.S. and F.D.; supervision, F.D.; project administration, F.D. and P.K.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets have been utilized for conducting the experiments. Details are provided in the text of the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Imani, M.; Ghassemian, H. An overview on spectral and spatial information fusion for hyperspectral image classification: Current trends and challenges. Inf. Fusion 2020, 59, 59–83. [Google Scholar] [CrossRef]
  2. Fabiyi, S.D.; Vu, H.; Tachtatzis, C.; Murray, P.; Harle, D.; Dao, T.K.; Andonovic, I.; Ren, J.; Marshall, S. Varietal classification of rice seeds using RGB and hyperspectral images. IEEE Access 2020, 8, 22493–22505. [Google Scholar] [CrossRef]
  3. Fox, N.; Parbhakar-Fox, A.; Moltzen, J.; Feig, S.; Goemann, K.; Huntington, J. Applications of hyperspectral mineralogy for geoenvironmental characterisation. Miner. Eng. 2017, 107, 63–77. [Google Scholar] [CrossRef]
  4. Chen, Y.; Zhu, K.; Zhu, L.; He, X.; Ghamisi, P.; Benediktsson, J.A. Automatic design of convolutional neural network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 7048–7066. [Google Scholar] [CrossRef]
  5. Bitar, A.W.; Cheong, L.F.; Ovarlez, J.P. Sparse and low-rank matrix decomposition for automatic target detection in hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5239–5251. [Google Scholar] [CrossRef] [Green Version]
  6. Tu, B.; Wang, J.; Kang, X.; Zhang, G.; Ou, X.; Guo, L. KNN-based representation of superpixels for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 4032–4047. [Google Scholar] [CrossRef]
  7. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  8. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Spectral–spatial hyperspectral image segmentation using subspace multinomial logistic regression and Markov random fields. IEEE Trans. Geosci. Remote Sens. 2011, 50, 809–823. [Google Scholar] [CrossRef]
  9. Lv, Q.; Niu, X.; Dou, Y.; Xu, J.; Lei, Y. Classification of hyperspectral remote sensing image using hierarchical local-receptive-field-based extreme learning machine. IEEE Geosci. Remote Sens. Lett. 2016, 13, 434–438. [Google Scholar] [CrossRef]
  10. Chen, Y.; Nasrabadi, N.M.; Tran, T.D. Hyperspectral image classification using dictionary-based sparse representation. IEEE Trans. Geosci. Remote Sens. 2011, 49, 3973–3985. [Google Scholar] [CrossRef]
  11. Rodarmel, C.; Shan, J. Principal component analysis for hyperspectral image classification. Surv. Land Inf. Sci. 2002, 62, 115–122. [Google Scholar]
  12. Wang, J.; Chang, C.I. Independent component analysis-based dimensionality reduction with applications in hyperspectral image analysis. IEEE Trans. Geosci. Remote Sens. 2006, 44, 1586–1600. [Google Scholar] [CrossRef]
  13. Bandos, T.V.; Bruzzone, L.; Camps-Valls, G. Classification of hyperspectral images with regularized linear discriminant analysis. IEEE Trans. Geosci. Remote Sens. 2009, 47, 862–873. [Google Scholar] [CrossRef]
  14. Ma, L.; Crawford, M.M.; Tian, J. Local manifold learning-based k-nearest-neighbor for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2010, 48, 4099–4109. [Google Scholar] [CrossRef]
  15. Patro, R.N.; Subudhi, S.; Biswal, P.K.; Dell’Acqua, F. A Review on Unsupervised Band Selection Techniques: Land Cover Classification for Hyperspectral Earth Observation Data. IEEE Geosci. Remote Sens. Mag. 2021, 9, 72–111. [Google Scholar] [CrossRef]
  16. Patro, R.N.; Subudhi, S.; Biswal, P.K. Spectral clustering and spatial Frobenius norm-based Jaya optimisation for BS of hyperspectral images. IET Image Process. 2019, 13, 307–315. [Google Scholar] [CrossRef]
  17. Benediktsson, J.A.; Palmason, J.A.; Sveinsson, J.R. Classification of hyperspectral data from urban areas based on extended morphological profiles. IEEE Trans. Geosci. Remote Sens. 2005, 43, 480–491. [Google Scholar] [CrossRef]
  18. Dalla Mura, M.; Benediktsson, J.A.; Waske, B.; Bruzzone, L. Morphological attribute profiles for the analysis of very high resolution images. IEEE Trans. Geosci. Remote Sens. 2010, 48, 3747–3762. [Google Scholar] [CrossRef]
  19. Ghamisi, P.; Dalla Mura, M.; Benediktsson, J.A. A survey on spectral–spatial classification techniques based on attribute profiles. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2335–2353. [Google Scholar] [CrossRef]
  20. Guo, X.; Huang, X.; Zhang, L. Three-dimensional wavelet texture feature extraction and classification for multi/hyperspectral imagery. IEEE Geosci. Remote Sens. Lett. 2014, 11, 2183–2187. [Google Scholar]
  21. Huang, X.; Liu, X.; Zhang, L. A multichannel gray level co-occurrence matrix for multi/hyperspectral image texture representation. Remote Sens. 2014, 6, 8424–8445. [Google Scholar] [CrossRef] [Green Version]
  22. Ye, Z.; Fowler, J.E.; Bai, L. Spatial–spectral hyperspectral classification using local binary patterns and Markov random fields. J. Appl. Remote Sens. 2017, 11, 035002. [Google Scholar] [CrossRef]
  23. Jia, S.; Shen, L.; Li, Q. Gabor feature-based collaborative representation for hyperspectral imagery classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 1118–1129. [Google Scholar]
  24. Tomasi, C.; Manduchi, R. Bilateral filtering for gray and color images. In Proceedings of the Sixth International Conference onComputer Vision, Bombay, India, 4–7 January 1998; pp. 839–846. [Google Scholar]
  25. Sun, W.; Messinger, D.W. Trilateral filter on multispectral imagery for classification and segmentation. In Algorithms and Technologies for Multispectral, Hyperspectral, and Ultraspectral Imagery XVII; International Society for Optics and Photonics: Bellingham, WA, USA, 2011; Volume 8048, p. 80480Y. [Google Scholar]
  26. Kang, X.; Li, S.; Benediktsson, J.A. Spectral–spatial hyperspectral image classification with edge-preserving filtering. IEEE Trans. Geosci. Remote Sens. 2014, 52, 2666–2677. [Google Scholar] [CrossRef]
  27. Kang, X.; Li, S.; Benediktsson, J.A. Feature extraction of hyperspectral images with image fusion and recursive filtering. IEEE Trans. Geosci. Remote Sens. 2014, 52, 3742–3752. [Google Scholar] [CrossRef]
  28. Subudhi, S.; Patro, R.; Biswal, P.K. Bilateral kernel extraction from pca for classification of hyperspectral images. PFG–J. Photogramm. Remote Sens. Geoinf. Sci. 2019, 87, 205–221. [Google Scholar] [CrossRef]
  29. Zabalza, J.; Ren, J.; Zheng, J.; Han, J.; Zhao, H.; Li, S.; Marshall, S. Novel two-dimensional singular spectrum analysis for effective feature extraction and data classification in hyperspectral imaging. IEEE Trans. Geosci. Remote Sens. 2015, 53, 4418–4433. [Google Scholar] [CrossRef] [Green Version]
  30. Zabalza, J.; Qing, C.; Yuen, P.; Sun, G.; Zhao, H.; Ren, J. Fast implementation of two-dimensional singular spectrum analysis for effective data classification in hyperspectral imaging. J. Frankl. Inst. 2018, 355, 1733–1751. [Google Scholar] [CrossRef] [Green Version]
  31. Ma, P.; Ren, J.; Zhao, H.; Genyun, S.; Murray, P.; Zheng, J. Multiscale 2D singular spectrum analysis and principal component analysis for spatial-spectral noise-robust feature extraction and classification of hyperspectral images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 14, 1233–1245. [Google Scholar] [CrossRef]
  32. Fu, H.; Sun, G.; Ren, J.; Zhang, A.; Jia, X. Fusion of PCA and Segmented-PCA Domain Multiscale 2-D-SSA for Effective Spectral-Spatial Feature Extraction and Data Classification in Hyperspectral Imagery. IEEE Trans. Geosci. Remote Sens. 2020. [Google Scholar] [CrossRef]
  33. Subudhi, S.; Patro, R.N.; Biswal, P.K.; Dell’Acqua, F. A Survey on Superpixel Segmentation as a Preprocessing Step in Hyperspectral Image Analysis. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 5015–5035. [Google Scholar] [CrossRef]
  34. Fang, L.; Li, S.; Duan, W.; Ren, J.; Benediktsson, J.A. Classification of hyperspectral images by exploiting spectral–spatial information of superpixel via multiple kernels. IEEE Trans. Geosci. Remote Sens. 2015, 53, 6663–6674. [Google Scholar] [CrossRef] [Green Version]
  35. Liu, J.; Wu, Z.; Xiao, Z.; Yang, J. Region-based relaxed multiple kernel collaborative representation for hyperspectral image classification. IEEE Access 2017, 5, 20921–20933. [Google Scholar] [CrossRef]
  36. Sun, L.; Ma, C.; Chen, Y.; Shim, H.J.; Wu, Z.; Jeon, B. Adjacent superpixel-based multiscale spatial-spectral kernel for hyperspectral classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 1905–1919. [Google Scholar] [CrossRef]
  37. Jiang, J.; Ma, J.; Chen, C.; Wang, Z.; Cai, Z.; Wang, L. SuperPCA: A superpixelwise PCA approach for unsupervised feature extraction of hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4581–4593. [Google Scholar] [CrossRef] [Green Version]
  38. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep learning-based classification of hyperspectral data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
  39. Chen, Y.; Zhao, X.; Jia, X. Spectral–spatial classification of hyperspectral data based on deep belief network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2381–2392. [Google Scholar] [CrossRef]
  40. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  41. Mou, L.; Ghamisi, P.; Zhu, X.X. Deep recurrent neural networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3639–3655. [Google Scholar] [CrossRef] [Green Version]
  42. Zhu, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Generative adversarial networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5046–5063. [Google Scholar] [CrossRef]
  43. Stutz, D.; Hermans, A.; Leibe, B. Superpixels: An evaluation of the state-of-the-art. Comput. Vis. Image Underst. 2018, 166, 1–27. [Google Scholar] [CrossRef] [Green Version]
  44. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Süsstrunk, S. SLIC superpixels compared to state-of-the-art superpixel methods. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2274–2282. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Xue, Z.; Zhou, S.; Zhao, P. Active learning improved by neighborhoods and superpixels for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2018, 15, 469–473. [Google Scholar] [CrossRef]
  46. Tarabalka, Y.; Fauvel, M.; Chanussot, J.; Benediktsson, J.A. SVM-and MRF-based method for accurate classification of hyperspectral images. IEEE Geosci. Remote Sens. Lett. 2010, 7, 736–740. [Google Scholar] [CrossRef] [Green Version]
  47. Chi, M.; Bruzzone, L. Semisupervised classification of hyperspectral images by SVMs optimized in the primal. IEEE Trans. Geosci. Remote Sens. 2007, 45, 1870–1880. [Google Scholar] [CrossRef]
  48. Chang, C.C.; Lin, C.J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 1–27. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the proposed method.
Figure 1. Flowchart of the proposed method.
Applsci 11 10876 g001
Figure 2. Moving window across the image H b to create the trajectory matrix Z b .
Figure 2. Moving window across the image H b to create the trajectory matrix Z b .
Applsci 11 10876 g002
Figure 3. Implementation of 2D-SSA on a HSI scene (a) Original scene at 667 nm. (b) 1st component grouping. (c) 1–5th component grouping. (d) 1–10th component grouping, where M x = 5, and M y = 5.
Figure 3. Implementation of 2D-SSA on a HSI scene (a) Original scene at 667 nm. (b) 1st component grouping. (c) 1–5th component grouping. (d) 1–10th component grouping, where M x = 5, and M y = 5.
Applsci 11 10876 g003
Figure 4. Possible Region of Interest (ROI) around the superpixel segment. ( r , u ) denotes the location of pixel i in superpixels. r m i n , r m a x represents the minimum and maximum row index, and u m i n , u m a x are the corresponding min and max column indices.
Figure 4. Possible Region of Interest (ROI) around the superpixel segment. ( r , u ) denotes the location of pixel i in superpixels. r m i n , r m a x represents the minimum and maximum row index, and u m i n , u m a x are the corresponding min and max column indices.
Applsci 11 10876 g004
Figure 5. (a) False Color Composite Image, (b) Ground Truth Image and (c) Class names for the Indian Pines Dataset.
Figure 5. (a) False Color Composite Image, (b) Ground Truth Image and (c) Class names for the Indian Pines Dataset.
Applsci 11 10876 g005
Figure 6. (a) False Color Composite Image, (b) Ground Truth Image and (c) Class names for the Pavia University Dataset.
Figure 6. (a) False Color Composite Image, (b) Ground Truth Image and (c) Class names for the Pavia University Dataset.
Applsci 11 10876 g006
Figure 7. (a) False Color Composite Image, (b) Ground Truth Image, and (c) Class names for the Salinas Dataset.
Figure 7. (a) False Color Composite Image, (b) Ground Truth Image, and (c) Class names for the Salinas Dataset.
Applsci 11 10876 g007
Figure 8. (a) False-Color Composite Image, (b) Ground Truth Image, and (c) Class names for the Houston-2018 Dataset.
Figure 8. (a) False-Color Composite Image, (b) Ground Truth Image, and (c) Class names for the Houston-2018 Dataset.
Applsci 11 10876 g008
Figure 9. Effect of window size variation for different number of superpixels on the classification performance for the (a) Indian Pines (b) Pavia University, and (c) Salinas, and (d) Houston 2018 datasets.
Figure 9. Effect of window size variation for different number of superpixels on the classification performance for the (a) Indian Pines (b) Pavia University, and (c) Salinas, and (d) Houston 2018 datasets.
Applsci 11 10876 g009
Figure 10. Effect of training sample variation on the classification performance for the (a) Indian Pines (b) Pavia University, (c) Salinas, and (d) Houston 2018 datasets.
Figure 10. Effect of training sample variation on the classification performance for the (a) Indian Pines (b) Pavia University, (c) Salinas, and (d) Houston 2018 datasets.
Applsci 11 10876 g010
Figure 11. (a) Ground Truth Image, Classification Maps of (b) SVM (c) EPF (d) SCMK (e) R2MK (f) ASMGSSK (g) MsuperPCA (h) 2D-SSA (i) 2D-MSSA (j) SP-SSA for Indian Pines dataset.
Figure 11. (a) Ground Truth Image, Classification Maps of (b) SVM (c) EPF (d) SCMK (e) R2MK (f) ASMGSSK (g) MsuperPCA (h) 2D-SSA (i) 2D-MSSA (j) SP-SSA for Indian Pines dataset.
Applsci 11 10876 g011
Figure 12. (a) Ground Truth Image, Classification Maps of (b) SVM (c) EPF (d) SCMK (e) R2MK (f) ASMGSSK (g) MsuperPCA (h) 2D-SSA (i) 2D-MSSA (j) SP-SSA for the Indian Pines dataset.
Figure 12. (a) Ground Truth Image, Classification Maps of (b) SVM (c) EPF (d) SCMK (e) R2MK (f) ASMGSSK (g) MsuperPCA (h) 2D-SSA (i) 2D-MSSA (j) SP-SSA for the Indian Pines dataset.
Applsci 11 10876 g012
Figure 13. (a) Ground Truth Image, Classification Maps of (b) SVM (c) EPF (d) SCMK (e) R2MK (f) ASMGSSK (g) MsuperPCA (h) 2D-SSA (i) 2D-MSSA (j) SP-SSA for Salinas dataset.
Figure 13. (a) Ground Truth Image, Classification Maps of (b) SVM (c) EPF (d) SCMK (e) R2MK (f) ASMGSSK (g) MsuperPCA (h) 2D-SSA (i) 2D-MSSA (j) SP-SSA for Salinas dataset.
Applsci 11 10876 g013
Figure 14. (a) Ground Truth Image, Classification Maps of (b) SVM (c) EPF (d) SCMK (e) R2MK (f) ASMGSSK (g) MsuperPCA (h) 2D-SSA (i) 2D-MSSA (j) SP-SSA for Houston 2018 dataset.
Figure 14. (a) Ground Truth Image, Classification Maps of (b) SVM (c) EPF (d) SCMK (e) R2MK (f) ASMGSSK (g) MsuperPCA (h) 2D-SSA (i) 2D-MSSA (j) SP-SSA for Houston 2018 dataset.
Applsci 11 10876 g014
Figure 15. (a) Cameraman image (b) 2D-SSA Reconstructed image [MSE = 115.8865] (c) SP-SSA reconstructed image [MSE = 93.0468] (d) Test image (e) 2D-SSA Reconstructed image [MSE = 287.5323] (f) SP-SSA reconstructed image [MSE = 237.1038].
Figure 15. (a) Cameraman image (b) 2D-SSA Reconstructed image [MSE = 115.8865] (c) SP-SSA reconstructed image [MSE = 93.0468] (d) Test image (e) 2D-SSA Reconstructed image [MSE = 287.5323] (f) SP-SSA reconstructed image [MSE = 237.1038].
Applsci 11 10876 g015
Figure 16. (a) Original scene at band 667 nm (b) Reconstructed scene by 2D-SSA (c) Reconstructed scene by SP-SSA (d) Difference image for the 2D-SSA reconstructed scene (e) Difference image for the SP-SSA reconstructed scene.
Figure 16. (a) Original scene at band 667 nm (b) Reconstructed scene by 2D-SSA (c) Reconstructed scene by SP-SSA (d) Difference image for the 2D-SSA reconstructed scene (e) Difference image for the SP-SSA reconstructed scene.
Applsci 11 10876 g016
Table 1. Parameter settings for different algorithms in the Indian Pines, Pavia University, Salinas, and Houston 2018 Datasets.
Table 1. Parameter settings for different algorithms in the Indian Pines, Pavia University, Salinas, and Houston 2018 Datasets.
MethodIndian Pines, SalinasPavia UniversityHouston 2018
SVMNANANA
EPF δ s = 3, δ r = 0.2, r = 3, and ϵ = 0.01 δ s = 3, δ r = 0.2, r = 3, and ϵ = 0.01 δ s = 5, δ r = 0.1, r = 3, and ϵ = 0.02
SCMK σ = 2 6 ; spnum = 600; μ 1 = 0.2, μ 2 = 0.2 σ = 2 4 ; spnum = 900; μ 1 = 0.2, μ 2 = 0.4 σ = 2 5 ; spnum = 1600; μ 1 = 0.1, μ 2 = 0.3
R2MK σ = 2 6 ; spnum = [20,50,100,200,400,800]; μ = 0.2 σ = 2 5 ; spnum = [50,100,200,400,800,1600]; μ = 0.3 σ = 2 4 ; spnum = [50,100,200,400,800,1600,3200]; μ = 0.1
ASMGSSK r 0 = 0.1; σ = 2 7 ; spnum = [100,200,400,800,1600,3200] r 0 = 0.1; σ = 2 5 ; spnum = [200,400,800,1600,3200,6400] r 0 =0.2; σ = 2 4 ; spnum = [200,400,800,1600,3200,6400]
MsuperPCAFundamental spnum = 100; scale no = 4Fundamental spnum = 20; scale no = 6Fundamental spnum = 100; scale no = 8
2DSSAWindow Size: 10 × 10 ; EVG = 1 st Window Size: 5 × 5 ; EVG = 1 2 nd Window Size: 5 × 5 ; EVG = 1 2 nd
2DMSSAWindow Size: 5 × 5 , 10 × 10 , 20 × 20 , 40 × 40 , 60 × 60 Window Size: 5 × 5 , 10 × 10 , 20 × 20 , 40 × 40 , 60 × 60 Window Size: 5 × 5 , 10 × 10 , 20 × 20 , 40 × 40 , 60 × 60
SP-SSAspnum: 100; Window Size: 5 × 5 ; EVG = 1spnum: 150; Window Size: 3 × 3 ; EVG = 1 2 nd spnum: 500; Window Size: 3 × 3 ; EVG = 1 2 nd
Table 2. Classification results for the Indian Pines Dataset with 3% training for SVM, EPF, SCMK, R2MK, ASMGSSK, MsuperPCA, 2D-SSA, 2D-MSSA, and SP-SSA algorithms.
Table 2. Classification results for the Indian Pines Dataset with 3% training for SVM, EPF, SCMK, R2MK, ASMGSSK, MsuperPCA, 2D-SSA, 2D-MSSA, and SP-SSA algorithms.
ClassSamplesSVMEPFSCMKR2MKASMGSSKMsuperPCA2DSSA2DMSSASP-SSA
Alfalfa4612.6522.2262.7992.869587.854.5587.892.11
Corn-notill142875.1892.8592.5594.0694.8597.0991.8195.2598.25
Corn-mintill83082.4683.449093.3297.3790.1188.9696.6597.99
Corn23747.8679.3985.5995.4196.1499.5380.8989.296.98
Grass-pasture48369.2568.0489.6797.399.0595.8190.6192.4197.04
Grass-trees73080.1996.7299.5698.3699.2198.6193.2399.5499.51
Grass-pasture-mowed2888.8988.4669.23801009692.5992100
Hay-windrowed47893.461009898.1898.3295.0798.4699.5499
Oats2036.1247.3710052.6347.0694.1242.1183.33100
Soybean-notill97276.7482.9688.2990.1696.5795.2787.1296.9193.63
Soybean-mintill245579.5197.0395.5497.4898.596.6693.197.1999.03
Soybean-clean5935086.9993.1895.7997.2995.2790.9696.0798.6
Wheat20599.0196.9492.2310097.2198.999.4898.91100
Woods126588.7499.2698.9199.4899.2799.0296.4299.8299.62
Buildings-Grass-Trees-Drives38665.1887.8799.4594.3796.7388.9590.1694.2498.46
Stone-Steel-Towers9341.3297.7572.7392.9497.593.997.7598.8189.74
OA::76.4291.2593.8295.9897.5496.0491.9596.8398.15
AA::67.9182.9689.2392.0294.3895.1386.7694.8697.5
K::73.0589.9892.9595.3297.295.4990.8296.3997.89
Table 3. Classification results for the Pavia University Dataset with 2% training for SVM, EPF, SCMK, R2MK, ASMGSSK, MsuperPCA, 2D-SSA, 2D-MSSA, and SP-SSA algorithms.
Table 3. Classification results for the Pavia University Dataset with 2% training for SVM, EPF, SCMK, R2MK, ASMGSSK, MsuperPCA, 2D-SSA, 2D-MSSA, and SP-SSA algorithms.
ClassSamplesSVMEPFSCMKR2MKASMGSSKMsuperPCA2DSSA2DMSSASP-SSA
Asphalt663190.7996.3394.0195.4399.2397.1997.1297.6198.94
Meadows18,64999.5498.0699.1499.7899.7699.8399.5299.4599.85
Gravel209953.1280.8683.6990.8191.8589.7389.3590.5394.17
Trees306481.983.8187.3493.898.2295.5396.2495.7699
Painted metal sheets134593.0410099.1710099.1710099.4799.46100
Bare Soil502989.6494.1596.4696.8199.698.0195.7698.5299.74
Bitumen133055.4889.7486.8691.7199.4294.5485.9797.2198.85
Self-Blocking Bricks368287.9291.4790.9790.9794.6993.0589.1192.6495.62
Shadows94790.7590.1997.1298.1798.5998.8995.6988.6798.76
OA::90.7194.5395.396.8598.6797.6496.5797.4698.94
AA::82.4791.6292.7595.2897.8496.3194.2595.5498.33
K::87.5792.7393.7695.8298.2396.8795.4496.6398.6
Table 4. Classification Results for the Salinas Dataset with 1% training for SVM, EPF, SCMK, R2MK, ASMGSSK, MsuperPCA, 2D-SSA, 2D-MSSA, and SP-SSA algorithms.
Table 4. Classification Results for the Salinas Dataset with 1% training for SVM, EPF, SCMK, R2MK, ASMGSSK, MsuperPCA, 2D-SSA, 2D-MSSA, and SP-SSA algorithms.
ClassSamplesSVMEPFSCMKR2MKASMGSSKMsuperPCA2DSSA2DMSSASP-SSA
Brocoli-green-weeds-1200999.7498.4399.4499.8599.3699.3998.8310099.64
Brocoli-green-weeds-2372610099.9210098.6210099.8210010099.96
Fallow197693.3499.3199.910099.7199.8999.6499.64100
Fallow-rough-plow139496.3797.2898.529898.2698.5697.3897.5498.67
Fallow-smooth267891.1999.0697.7799.4298.7299.7598.4299.2599.63
Stubble395910099.9298.5499.92100100100100100
Celery357999.1899.6599.1499.5499.899.7810099.8899.96
Grapes-untrained11,27191.3892.4994.695.296.7795.769596.3197.91
Soil-vinyard-develop620397.3499.8899.6599.9899.8299.9699.6699.95100
Corn-senesced-green-weeds327893.9697.3798.9697.4298.4798.2496.7398.398.82
Lettuce-romaine-4wk106881.6696.5596.8197.7897.9997.9297.2497.8699.87
Lettuce-romaine-5wk192798.03100100100100100100100100
Lettuce-romaine-6wk91698.8597.0197.4199.5599.6999.7699.7499.6999.69
Lettuce-romaine-7wk107092.9298.2399.1398.1799.0798.8699.7899.4799.73
Vinyard-untrained726882.6990.4191.3292.2895.8992.9894.2195.5897.7
Vinyard-vertical-trellis180795.6399.1390.399.8399.699.8298.5799.2999.92
OA::93.6496.5696.8297.598.4397.8997.6398.3399.1
AA::94.5297.7997.5998.4798.9598.7898.4598.9299.47
K::92.9196.1796.4697.2198.2597.6597.3698.1499
Table 5. Classification results for the Houston Dataset with 0.2% training for SVM, EPF, SCMK, R2MK, ASMGSSK, MsuperPCA, 2D-SSA, 2D-MSSA, and SP-SSA algorithms.
Table 5. Classification results for the Houston Dataset with 0.2% training for SVM, EPF, SCMK, R2MK, ASMGSSK, MsuperPCA, 2D-SSA, 2D-MSSA, and SP-SSA algorithms.
ClassSamplesSVMEPFSCMKR2MKASMGSSKMsuperPCA2DSSA2DMSSASP-SSA
Healthy grass979962.8465.717374.4981.4676.4476.6484.5179.28
Stressed grass32,50284.485.0883.8386.4289.8888.6783.790.9991.33
Artificial turf68410098.8310099.41100100100100100
Evergreen trees13,58883.7273.9980.1584.7687.7782.4380.4391.1387.37
Deciduous trees504843.9836.7550.5451.7463.8255.8946.3858.2673.94
Bare earth451679.3282.6786.2888.4694.0191.180.6489.0496.21
Water26666.4267.861.8968.0683.0868.1874.6266.9285.71
Residential buildings39,76268.2677.2977.5279.284.378274.7678.0787.39
Non-residential buildings223,68482.5785.5187.388.3191.6489.1886.0590.6392.73
Roads45,81040.4143.746.7947.3358.8651.3942.7553.2563.89
Sidewalks34,00231.3635.337.5837.5143.7439.2533.6551.3549.71
Crosswalks15164.969.028.639.1213.669.267.625.214.21
Major thoroughfares46,35850.9755.9159.4163.0472.5365.158.1760.8375.48
Highways984949.760.3561.1963.9373.8370.7859.5568.9278.56
Railways693779.9285.839289.5295.2988.0780.2295.8497.29
Paved parking lots11,47554.9163.274.0269.9984.5874.6364.974.1884.33
Unpaved parking lots14958.7883.1183.6777.0381.6385.7181.0881.7693.06
Cars657843.994749.3359.6665.0953.7854.5162.8570.22
Trains536540.8340.5458.9251.377.461.1452.6679.1979.96
Stadium seats682486.4293.2587.7893.8696.6294.1883.5398.7198.43
OA::68.1971.6474.175.4881.3277.272.1179.0883.57
AA::60.6964.5467.9969.1676.9671.3666.0974.0879.95
K::59.1363.4166.5268.2875.8270.5563.9472.6178.73
Table 6. Statistics of the McNemar Test for the Indian Pines, Pavia University, Salinas, and Houston 2018 datasets.
Table 6. Statistics of the McNemar Test for the Indian Pines, Pavia University, Salinas, and Houston 2018 datasets.
ZIndian PinesPavia UniversitySalinasHouston 2018
Proposed Method (SP-SSA)
SVM42.53456.34237.21561.512
EPF39.15248.18625.11357.084
SCMK27.46737.42821.42143.115
R2MK23.61528.52118.47239.721
ASMGSSK10.62416.8328.35118.521
MsuperPCA21.52419.44111.48635.431
2DSSA34.32131.08415.32151.322
2DMSSA16.81923.2179.09127.634
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Subudhi, S.; Patro , R.; Biswal, P.K.; Dell’Acqua, F. Superpixel-Based Singular Spectrum Analysis for Effective Spatial-Spectral Feature Extraction. Appl. Sci. 2021, 11, 10876. https://doi.org/10.3390/app112210876

AMA Style

Subudhi S, Patro  R, Biswal PK, Dell’Acqua F. Superpixel-Based Singular Spectrum Analysis for Effective Spatial-Spectral Feature Extraction. Applied Sciences. 2021; 11(22):10876. https://doi.org/10.3390/app112210876

Chicago/Turabian Style

Subudhi, Subhashree, Ramnarayan Patro , Pradyut Kumar Biswal, and Fabio Dell’Acqua. 2021. "Superpixel-Based Singular Spectrum Analysis for Effective Spatial-Spectral Feature Extraction" Applied Sciences 11, no. 22: 10876. https://doi.org/10.3390/app112210876

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop