Next Article in Journal
Optimising Sampling Strategies in Coral Reefs Using Large-Area Mosaics
Previous Article in Journal
Mapping at 30 m Resolution of Soil Attributes at Multiple Depths in Midwest Brazil
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spectral DWT Multilevel Decomposition with Spatial Filtering Enhancement Preprocessing-Based Approaches for Hyperspectral Imagery Classification

The State Key Laboratory of Information Engineering in Surveying, Mapping, and Remote Sensing, Wuhan University, Wuhan 430079, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2019, 11(24), 2906; https://doi.org/10.3390/rs11242906
Submission received: 9 November 2019 / Revised: 27 November 2019 / Accepted: 3 December 2019 / Published: 5 December 2019
(This article belongs to the Special Issue Spatial Enhancement of Spectral Data and Applications)

Abstract

:
In this paper, spectral–spatial preprocessing using discrete wavelet transform (DWT) multilevel decomposition and spatial filtering is proposed for improving the accuracy of hyperspectral imagery classification. Specifically, spectral DWT multilevel decomposition (SDWT) is performed on the hyperspectral image to separate the approximation coefficients from the detail coefficients. For each level of decomposition, only the detail coefficients are spatially filtered instead of being discarded, as is often adopted by the wavelet-based approaches. Thus, three different spatial filters are explored, including two-dimensional DWT (2D-DWT), adaptive Wiener filter (AWF), and two-dimensional discrete cosine transform (2D-DCT). After the enhancement of the spectral information by performing the spatial filter on the detail coefficients, DWT reconstruction is carried out on both the approximation and the filtered detail coefficients. The final preprocessed image is fed into a linear support vector machine (SVM) classifier. Evaluation results on three widely used real hyperspectral datasets show that the proposed framework using spectral DWT multilevel decomposition with 2D-DCT filter (SDWT-2DCT_SVM) exhibits a significant performance and outperforms many state-of-the-art methods in terms of classification accuracy, even under the constraint of small training sample size, and execution time.

Graphical Abstract

1. Introduction

Hyperspectral imagery contains rich discriminative spectral and spatial characteristics about surface materials [1], which makes it a prominent source of information for a varied range of applications in areas such as agriculture, environmental planning, surveillance, target detection, medicine [2,3,4], etc. Most of these hyperspectral applications rely on the supervised classification task. However, supervised classification faces challenges from the Hughes phenomenon, caused by the unbalance between high dimensionality and limited availability of training samples. Furthermore, embedded noises from both the sensors and the environment decrease the classification accuracy.
Thus, to address these problems, intensive work has been devoted to provide accurate classifiers for hyperspectral images, including [5] support vector machines (SVMs) [6], random forests [7], neural networks [8,9], active learning [10,11], and sparse representation [12,13,14] methods. Particularly, the SVM classifier has shown high performance with regards to the classification accuracy [6]. Although these aforementioned classifiers can fully make use of the spectral information in hyperspectral imagery, they do not incorporate spatial information. Thus, the resulting classification maps are often corrupted with salt-and-pepper noise [15].
Recent advances in hyperspectral image analysis have shown that the integration of spectral and spatial information in the classification approach can abolish salt-and-pepper noise in classification maps and noticeably improve image classification accuracy [16,17,18]. Many spectral–spatial hyperspectral image classification approaches have been proposed to improve the accuracy further based on various methodologies including structural filtering [19,20,21], morphological profile (MP) [22,23], random field [24], deep learning-based approaches [25,26], sparse representation-based approaches (SRC) [27,28], and segmentation-based approaches [29,30].
The structural filtering methodology is a widely used approach for spatial preprocessing of hyperspectral images [31]. In contrast to other spatial methodologies, spatial filtering is simple and easy to implement [32]. This advantage makes it particularly suitable for practical applications. A great deal of ongoing research is focused on structural filtering to obtain the contextual features where the simplest way is to extract the spatial information based on moment criteria [33,34]. However, recent work demonstrates that transform-based filters can retain texture features, edges, and details while discarding noise [35]. At the same time, much attention has been paid to wavelet transform-based filtering [36,37]. For instance, in [19], a spectral–spatial classification framework for hyperspectral images was proposed based on using one-dimensional discrete wavelet transform (1D-DWT) to create the extended morphological profile (EMP) and two-dimensional DWT (2D-DWT) for image denoising before stacking the original data to the EMP prior to the support vector machine (SVM) classification. Another spectral–spatial classification method based on a combination of a 1D-DWT with kernel minimum noise fraction (KMNF) in the spectral domain and a 2D-DWT in the spatial domain was proposed in [38]. In this work, the combination of 1D-DWT, KMNF, and 2D-DWT is suggested to create spatial spectral Schroedinger eigenmap (SSSE) features. Similar to the approach proposed in [38], a classification framework based on the combination of 1D-DWT, kernel maximum autocorrelation factor (KMAF), and 2D-DWT with hidden Markov random fields as classification post processing was proposed in [39].
Three-dimensional filtering using three-dimensional DWT (3D-DWT) for hyperspectral imagery classification has received great attention and produced successful spectral–spatial classification approaches by considering hyperspectral image data as a 3D cube where spatial and spectral features are treated as a whole [40]. Specifically, the extracted spectral–spatial 3D-DWT coefficients are combined with other techniques to enhance the classification performance further [37]. For instance, in [40], the 3D-DWT texture extraction was combined with the sparse logistic regression for hyperspectral image classification. The 3D-DWT features have also been integrated with Markov random field in [41]. Moreover, following the recent trend of using convolutional neural networks (CNN) and deep learning-based approaches in hyperspectral image classification, in [42], the low-frequency sub-bands of the 3D redundant wavelet transform (3D-RDWT) were used to create multiple views that are integrated into a multi-view active learning approach. In [43], a framework based on combining CNN with 3D multi-resolution wavelets was proposed.
Although the existing wavelet-based spectral and spatial hyperspectral classification methods have been largely successful, there are still some issues to be solved. The main issue with wavelet-based approaches is the substitution of the original spectral signatures by only the corresponding approximation coefficients. Subsequently, when increasing the level of decomposition, the original spectral signatures get smoothed more, and they will be almost similar to each other, which affects the performance of the classifier. Nevertheless, the detail coefficients emphasizing differences between the spectral signatures must be further exploited to enhance the effectiveness of the classifier.
There is another issue related to the 3D-DWT-based approaches. Indeed, a simple observation of a hyperspectral data cube discloses that the degree of irregularity is higher in the spectral dimension than in the spatial dimensions [44]. Hence, the variation of the noise variance in the spectral dimension is, in general, more drastic than that in the spatial dimensions. However, the 3D wavelet filtering process implicitly considers that the noise variance is the same in the three dimensions and ignores the dissimilarity between the spatial and the spectral dimensions. Furthermore, for better image representation, 3D feature extraction methods create a large number of spectral–spatial features, leading to another feature selection issue [40].
To handle these issues and fully exploit the spectral and spatial information in the original hyperspectral imagery, we propose three effective spectral–spatial hyperspectral classification frameworks based on performing the spatial filtering process in the spectral wavelet domain. The proposed approaches are based on a hybrid spectral and spatial filtering preprocessing to benefit from the signal nature dissimilarity in the spectral and the spatial dimensions. Thus, we perform a spectral multilevel DWT filter on the original dataset to separate the approximation part from the detail part that will be further filtered using a spatial filter for each level of decomposition, instead of being discarded, as it is usually adopted by wavelet-based approaches. Furthermore, the spatial filtering process is performed only on the detail part in the frequency domain to extract the remaining meaningful information effectively. Three different spatial filters are explored to filter the obtained detail information including 2D-DWT, two-dimensional discrete cosine transform (2D-DCT), and two-dimensional adaptive Wiener filter (2D-AWF). For each approach, a DWT reconstruction was performed on both the approximation and the filtered detail parts. Finally, the output of the DWT reconstruction is fed to the linear SVM classifier. As we fully exploit the spectral and spatial information, the SVM classifier is adopted in our frameworks for its robustness to the Hughes phenomenon [6] as well as for its capacity for generalization [6].
The main contribution of the paper is the application of the spatial filter in the transform domain on the detail coefficients for each level of decomposition rather than discarding this part of information, as usually adopted by the feature extraction-based approaches. The proposed frameworks make full use of the filtered spectral and spatial information of the hyperspectral image. Another important contribution of this paper concerns the simplicity of the tuning configuration. Indeed, recently, the proposed classifiers have reached very high classification accuracy where the competitive difference between them is more related to their tuning configuration simplicity. We provide effective classification frameworks involving the tuning at maximum a couple of parameters for each applied filter. The proposed frameworks fully exploit the spectral–spatial features in reasonable execution time, which is helpful to improve the classification performance. The rest of this paper is structured as follows: The principles of multilevel DWT decomposition (ML-DWT) and 2D-DWT are introduced in Section 2. The proposed frameworks are described in Section 3. Datasets and evaluation process are presented in Section 4. Section 5 gives the experimental setup and reports the results with the related analysis. In Section 6, some conclusions are given.

2. Materials and Methods

2.1. Multilevel Discrete Wavelet TransformDecomposition

Wavelet transform (WT) is constructed from small waves of varying frequency with limited duration. These waves are obtained from basic wavelet function by dilation and translation [16], where DWT is any wavelet transform for which the wavelets are discretely sampled. DWT provides a multiresolution wavelet analysis that represents and analyzes the signals at more than one resolution. Thus, features undetected at one level are easy to be detected at another level [17]. Therefore, the multilevel decomposition characteristic of DWT provides the features of an image at different resolutions.
For the application of hyperspectral data, the multilevel decomposition provides a simple hierarchical framework for interpreting the spectral information. Especially, the local spectral variation of a spectrum in different spectral bands at each scale can be detected and can provide useful information for hyperspectral image classification.

2.1.1. One-Dimensional Discrete Wavelet Transform (1D-DWT)

In the context of hyperspectral data, 1D-DWT is usually used for data reduction in the spectral dimension where the transform is applied on each pixel spectral curve x considered as a discrete signal represented by a vector x = [x0, x1, x2,…, xN−1] in N-dimensional spectral band space (N is the number of the spectral bands).
In DWT, the signal is passed through the low-pass and high-pass filters [28]. The output of the low-pass filter is the low-resolution version of the original signal, called approximation coefficients, where the output of the high-pass filter embeds the residual information of the original signal, known as detail coefficients. The most common approach to the multilevel discrete wavelet transform involves further decomposition of only the approximation coefficient at each subsequent level [45].
The best way to describe DWT is through an analysis filter bank in which the input signal x[n] is decomposed into two sub-band signals. As shown in the Figure 1a, we represent the low-pass filter by g(n) and the high-pass filter by h(n). The output of each filter is then down-sampled by two to achieve the two sub-band signals. Thus, the low-pass filter g(n) produces the approximation coefficient x1,L[n], which is the informative band, while high-pass filter h(n) produce the detail coefficient x1,H[n]. The original signal can be reconstructed by synthesis filters which take the upsampled x1,L[n] and x1,H[n], as shown in Figure 1b.
The multilevel DWT decomposition is obtained by the recursive application of the analysis filter bank on the low-frequency bands (the approximation coefficients) as shown in Figure 2.
In our approach, the multilevel 1D-DWT decomposition is performed separately on the spectral curve of each pixel to split the approximation coefficients from the detail coefficients.

2.1.2. Two-Dimensional Discrete Wavelet Transform (2D-DWT) Filter Bank

The 2D-DWT can also be implemented using digital filters and down-samplers, similar to the 1D-DWT. For an input image x[n,m], we simply take the 1D-DWT of the rows followed by the 1D-DWT of the resulting columns to obtain four frequency parts (x1,LL[n,m], x1,LH[n,m], x1,HL[n,m], and x1,HH[n,m]), as described in Figure 3. x1,LL[n,m] is the low frequency sub-band containing the approximation of the original image. x1,LH[n,m] is the high-frequency sub-band contain the vertical details of the image. x1,HL[n,m] is the high-frequency sub-band that contains the horizontal details of the image. x1,HH[n,m] is the high-frequency sub-band of the diagonal image.
In 2D-DWT multilevel decomposition, only the approximation coefficients are further decomposed. Figure 4 shows two levels of 2D-DWT decomposition.
In our approach, the 2D-DWT is explored as a spatial filter in the transform domain to extract the rest of useful information embedded in the detail part resulted from the application of spectral DWT.
Regarding the threshold technique, we opted the universal threshold of Donoho and Johnstone [46], scaled by a robust estimate of the variance. For a matrix M(L,C) of size [L,C], the universal threshold can be represented by the following formula:
Thresh = σ 2 ln ( N ) ,
where σ is the average variance of the noise and N is the signal length (L*C).
σ is calculated using the median estimate method, as follows:
σ = Median ( | M | ) 0.6745 .

3. The Proposed Approach

The proposed classification frameworks are based on using spectral DWT multilevel decomposition followed by spatial filtering of the detail coefficients for each level of decomposition. Thus, in our proposed approaches, we perform a spectral DWT multilevel decomposition on the original dataset to separate the most significant information concentrated in the approximation coefficients from the detailed and noisy information embedded in the detail coefficients. For each decomposition level, a spatial filter is carried out on the detail coefficients to filter the remaining meaningful information further. In the spatial filtering step, three spatial filters are explored, including 2D-DWT, 2D-AWF, and 2D-DCT. Therefore, regarding the type of the spatial filter, there are three different proposed frameworks: spectral DWT multilevel decomposition (SDWT)/2D-DWT, SDWT/2D-AWF, and SDWT/2D-DCT. The inverse SDWT is performed on all coefficients comprising the filtered ones to get the final preprocessed dataset that will be fed to the linear SVM classifier. The detailed flowchart of the proposed framework is illustrated in Figure 5, where the pipeline of the proposed spectral multilevel DWT decomposition-based classification approaches can be summarized in Algorithm 1.
Algorithm 1
Input: Hyperspectral imagery dataset X ∈ R MxLxP(M rows, L columns, and P bands).
1: Flatting X to obtain a matrix E ∈ R NxP where N = M × L.
2: Performing spectral multilevel DWT decomposition on the matrix E (on the P dimension) to obtain two matrices:
-
F ∈ R NxK for the K approximation coefficients.
-
R∈R Nx P-K forthe P – K detail coefficients.
3: Unflatting the matrix R to obtain the data cube Rc ∈ RMxLxP-K ofthe detail coefficients.
4: Performing a spatial filter on each matrix Rm ∈ RMxL in the cube Rc, using one of the following filters for each approach (there are P-K matrices):
-
2D-DWT
-
2D-AWF
-
2D-DCT
5: Flatting the filtered data cube Rc to return back to the matrix R ∈ RNx P-K.
6: Stack the two matrices F and R to obtain the matrix D ∈ R NxP, with the same size as the matrix E in step 1.
7: Apply an inverse DWT on the matrix D to obtain the preprocessed matrix E.
8: Perform a linear SVM classifier on the resulting matrix E.
Output: Classification map.

4. Data and Evaluation Process

4.1. Data

Three hyperspectral scenes with variant spatial resolution downloaded at http://www.ehu.eus/ccwintco/index.php?title=Hyperspectral_Remote_Sensing_Scenes were used for the experiments including the Indian Pines, Salinas, and Pavia University datasets. The selected datasets were captured from two different sensors, including Airborne Visible Infrared Imaging Spectrometer (AVIRIS) and Reflective Optics System Imaging Spectrometer (ROSIS), capturing rural and urban areas.
The first dataset concerns the Indian Pines scene. This scene was captured over northwest Indiana’s Indian Pines area by AVIRIS at a low spatial resolution of 20 m. This dataset has 220 bands of size 145 × 145 pixels where the bands affected by noise and water absorption were discarded, leaving 200 bands. The ground reference data contains 16 classes; among them, four classes have similar spectral signatures and are very difficult to discriminate. Besides the spectral similarity between the classes, the low spatial resolution makes this scene a challenging classification scenario [10]. Table 1 lists the related classes and describes the number of samples, the training, and the test points in this dataset.
Figure 6 shows the original band and the ground truth data for the Indian Pines dataset.
The second test dataset was the Pavia University scene. This scene was captured by the ROSIS sensor over the Pavia area, northern Italy. It has 610 × 610 pixels with 1.3-m spatial resolution and contains 103 spectral bands. For classification purposes, the scene is divided into nine regions listed in Table 2 with the number of samples that belong to each region and the numbers of training and testing points.
Figure 7 illustrates the Pavia University image and its related ground truth maps.
The third test dataset was the Salinas scene. This dataset captured the area over Salinas Valley in California by AVIRIS at 3.7-m spatial resolution. This image comprises 220 bands of size 512 × 217 pixels where the bands affected by noise and water absorption were discarded, leaving 204 bands. There are 16 classes representing a variety of crops in the ground reference data. Table 3 lists these classes with the related number of samples, training, and testing points in this dataset. Figure 8illustrates the Salinas image and its ground truth categorization maps.

4.2. Evaluation Process

Several state-of-the-art methods were considered for comparison to validate the effectiveness of our proposed spectral DWT multilevel decomposition with 2D-DWT filter (SDWT-2DWT_SVM), spectral DWT multilevel decomposition with Wiener filter (SDWT-WF_SVM), and spectral DWT multilevel decomposition with 2D-DCT filter (SDWT-2DCT_SVM) frameworks based on the spectral DWT multilevel decomposition. These included two pixel-wise classifiers represented by the standard SVM [6], and Principal Component Analysis followed by SVM (PCA_SVM) [47]. Four closely related spectral–spatial classification methods based on the structural filtering methodology were used for comparison, including 3D-DWT (3D_SVM) [41], 3D-DWT with graph cut (3DG_SVM) [41]. edge-preserving filtering (EPF) [48], and the image fusion and recursive filtering (IFRF) [49]. Moreover, two well-known denoising methods, block-matching 4D filtering (BM4D_SVM) [50] and parallel factor analysis-based approach (PARAFAC_SVM) [51], were included for comparison.
The parameter settings opted for in the compared techniques were the default settings used in the corresponding research works; the related source codes were provided by the respective authors. Specifically, for a fair comparison, linear SVM is adopted as a pixel-wise classifier for all the compared methods. The SVM classifier was implemented using MATLAB libsvm library [52]. Regarding the PCA_SVM method, the number of the principal components (PCs) fed to the linear SVM classifier was calculated using HySime criterion for the intrinsic dimension estimation for each dataset [53].
The compared methods were assessed numerically using five criteria: Average Accuracy (AA), Overall Accuracy (OA), statistical kappa coefficient (κ) [54], the accuracy of each class, and the computational time. Specifically, AA is the average of class classification accuracies, OA is calculated by dividing the number of correctly classified samples by the total number of test samples, and kappa coefficient (κ) considers errors of both omission and commission. In order to calculate the three criteria, the classification confusion matrix should be obtained first, which is defined as follows:
M ( m 11 m 1 k m k 1 m k k ) ,
where mij shows the number of pixels that should belong to class i but are assigned to class j, and K is the number of classification classes. Then, according to the confusion matrix M, OA, AA, and kappa coefficient (κ) are calculated as follows:
O A = 1 N t e s t i = 1 K m i i ,
where N t e s t is the total number of test samples. Then, the AA is calculated as:
C A i = m i i N i ,   i = 1 , 2 , , K ,
A A = 1 K i = 1 K C A i ,
where C A i is the accuracy of class i and N i is the total test samples in class i. Finally, the kappa coefficient (κ) is calculated by
κ = O A P e 1 P e
where P e = i = 1 i = K P i . P . i is the expected agreement, P i . = R i N t e s t , P . i = C i N t e s t .
R i and C i represent the sums of ith row and ith column in the confusion matrix, respectively. For all the three criteria, a larger value indicates a better classification performance. The experiments were repeated 20 times, and the average classification accuracies (kappa, OA, AA) were retained for comparison.
In order to verify the classification robustness of our proposed frameworks in the case of a small sample size problem, we evaluated the compared methods with insufficient and sufficient training samples; 2%, 4%, 6%, 8%, and 10% training samples were randomly selected from each class to form the training set, and the rest were used as test samples. The experiments were repeated 20 times, and the average results were reported to evaluate our approaches.
To compare the computational time of the tested approaches, we calculated the time consumed during the whole classification process for an average of 20 times. The experiments were performed on a computer with an Intel(R) Xeon(R) with CPU W3680 of 3.33 GHz, and 6 GB memory; all the methods were implemented with MATLAB®.
The parameter settings of the proposed approaches will be described in the next section, followed by the experimental results and analysis.

5. Experimental Results and Analysis

In this section, the parameter estimation for our proposed approaches are described, and the experimental results are analyzed to evaluate the performance of our proposed approaches relative to the other closely related hyperspectral classification methods.

5.1. Parameter Settings

Two types of parameters related to the spectral and spatial filters must be independently fixed for each of our proposed approaches:
SDWT-2DWT_SVM framework:
-
For both the spectral and the spatial filters, we need to fix the wavelet mother function type and its level of decomposition.
-
Default global threshold is used for the spatial filter [46].
SDWT-WF_SVM framework:
-
For the spectral filter, we need to fix the wavelet mother function type and its level of decomposition.
-
For the spatial filter, we have to fix the Wiener filter patch size.
SDWT-2DCT_SVM framework:
-
For the spectral filter, we need to fix the wavelet mother function type and its level of decomposition
-
For the spatial filter, we need to fix the threshold of the 2D-DCT filter.
Regarding the Wavelet mother function type, we evaluated for each proposed approach the most common wavelet families, including 30 wavelet functions, as mentioned in Table 4.
The estimation of the required parameters for each proposed approach is explained in the following three subsections independently.

5.1.1. The SDWT-2DWT_SVM Framework Parameter Estimation

The spectral filter parameters in this approach are the wavelet mother function type and its level of decomposition. These two parameters were set by comparing the classification accuracy OA with different types of wavelets for the first level of decomposition for each dataset as listed in Table 5. Then two functions with the highest accuracies were selected and further decomposed to fix the best function with its level of decomposition as listed in Table 6. The level of decomposition depends on the wavelet function type and the dimension of the selected dataset. The type of wavelet mother function and level of decomposition correspond to the most accurate classification OA for each dataset.
As shown in Table 5, for Indian Pines dataset, ‘db7’ and ‘db9’ functions have the highest classification accuracies (84.84% OA and 84.82% OA, respectively). For Pavia University dataset, ‘coif2’and ‘bior1.3’ functions have the highest classification accuracies (98.13% OA and 98.12% OA, respectively). For Salinas dataset, the highest accuracies correspond to ‘coif3’ with an OA of 95.01% and ‘sym4’ with an OA of 94.90%. For each dataset, the aforementioned functions corresponding to the first two highest accuracies are further decomposed to fix only one function with its best level of decomposition. Table 6 lists the obtained classification results using the best two functions for each dataset in the proposed SDWT-2DWT approach.
As depicted from Table 6, the highest classification accuracy (84.70% OA) was obtained using ‘db9’ function at the third level of decomposition for Indian Pines dataset. For Pavia University dataset, the highest accuracy (98.21% OA) was obtained using ‘bior1.3’ function at the fourth level of decomposition. For Salinas dataset, the highest accuracy (94.53% OA) was obtained using sym4 function at its fourth level of decomposition. Thus, for the SDWT-2DWT approach, we opted the following spectral parameters (Dataset, SDWT, Wavelet function, level of decomposition):
-
(Indian Pines dataset, SDWT, ‘db9’, 3);
-
(Pavia University dataset, SDWT, ‘bior1.3’, 4);
-
(Salinas dataset, SDWT, ‘sym4’, 4).
The spatial filter parameters in the proposed SDWT-2DWT approach are the wavelet mother function type and its level of decomposition. These two parameters were set using the same logic in the last subsection by comparing the classification accuracy OA with different types of wavelets for the second level of decomposition for each dataset. Then, we use the first two functions having the highest classification accuracies with more levels of decomposition to fix only one function with the best level of decomposition having the highest obtained classification accuracy. Table 7 lists the classification accuracy OA with different types of wavelets for the first level of decomposition on the Indian Pines, Pavia University, and Salinas datasets.
As can be seen in Table 7, for the Indian Pines dataset, ‘haar’ and ‘bior2.2’ functions have the highest classification accuracies (81.67% OA and 81.82% OA, respectively). For Pavia University dataset, ‘haar’ and ‘coif1’ functions have the highest classification accuracies (95.43% OA and 95.51%OA, respectively). For the Salinas dataset, the highest accuracies correspond to ‘haar’ with an OA of 90.70% and ‘coif1’ with an OA of 90.66%. For each dataset, the aforementioned functions corresponding to the first two highest accuracies are further decomposed to fix only one function to its best level of decomposition. Table 8 lists the obtained classification results using the best two functions for the 2D-DWT for each dataset in the proposed SDWT-2DWT_SVM approach.
As shown in Table 8, for the three datasets, the highest classification accuracy was obtained using ‘haar’ function. Specifically, the highest accuracy was obtained at the fifth level of decomposition for the Indian Pines dataset (85.66% OA), at the eighth level of decomposition for the Pavia University dataset, and at the seventh level of decomposition for the Salinas dataset (95.90% OA). Thus, for the SDWT-2DWT approach, we adopted the following spatial parameters (dataset, 2D-DWT function, and level of decomposition):
-
(Indian Pines dataset, 2D-DWT, ‘haar’, 5);
-
(Pavia University dataset, 2D-DWT, ‘haar’, 8);
-
(Salinas dataset, 2D-DWT, ‘haar’, 7).
For the spatial filtering threshold, we used the default global threshold (universal threshold) developed by Donoho and Johnstone, as it provides easy, fast, and automatic thresholding [46].

5.1.2. The SDWT-WF_SVM Framework Parameter Estimation

The spectral filter parameters in this approach are the wavelet mother function type and its level of decomposition. These two parameters were set by comparing the classification accuracy OA with different types of wavelets for the first level of decomposition for each dataset as listed in Table 9 for Indian Pines, Pavia University, and Salinas datasets. For each dataset, two functions with the highest accuracies were selected and further decomposed to fix the best function with its level of decomposition, as listed in Table 10.
As can be seen in Table 9, for the Indian Pines dataset, ‘bior2.4’and ‘sym3’ functions have the highest classification accuracies (90.70% OA and 90.52% OA, respectively). For the Pavia University dataset, ‘coif2’, and ‘sym4’functions have the highest classification accuracies (95.33% OA and 94.91%OA, respectively). For the Salinas dataset, the highest accuracies correspond to ‘db4’ with an OA of 96.78% and ‘coif2’ with an OA of 96.65%. For each dataset, the aforementioned functions corresponding to the first two highest accuracies are further decomposed to fix only one function with its best level of decomposition. Table 10 lists the obtained classification results using the best two functions for each dataset in the proposed SDWT-WF_SVM approach.
As depicted from Table 10, the highest classification accuracy (92.37% OA) was obtained using ‘sym3’ function at the fifth level of decomposition for the Indian Pines dataset. For the Pavia University dataset, the highest accuracy (97.34% OA) was obtained using ‘coif2’ function at the third level of decomposition. For the Salinas dataset, the highest accuracy (97.53% OA) was obtained using ‘db4’ function at its fourth level of decomposition. Thus, for the SDWT-WF_SVM approach, we adopted the following spectral parameters (dataset, SDWT, wavelet function, level of decomposition):
-
(Indian Pines dataset, SDWT, ‘sym3’, 5);
-
(Pavia University dataset, SDWT, ‘coif2’, 3);
-
(Salinas dataset, SDWT, ‘db4’, 4).
Wiener filter patch size represents the spatial parameter in the SDWT-WF_SVM approach. This parameter was set by comparing the classification accuracy OA with varying Wiener filter patch sizes where the opted patch size corresponds to the best classification OA for each dataset. Figure 9 depicts the variation of the classification accuracy OA with varying filter patch sizes, ranging from 3 × 3 to 63 × 63 pixels, on the Indian Pines, Pavia University, and Salinas datasets.
As can be seen in Figure 9, classification accuracy (OA) increased with a larger patch size value for the three datasets. Thus, accuracies tend to become stable at around a 41 × 41 patch size for the Indian Pines dataset, at around 27 × 27 patch size for the Pavia University dataset, and at around a 47 × 47 patch size for the Salinas dataset. Therefore, these patch sizes corresponding to each dataset were selected for our experiments.

5.1.3. The SDWT-2DCT_SVM Framework Parameter Estimation

The spectral filter parameters in the proposed SDWT-2DCT_SVM approach are the wavelet mother function type and its level of decomposition. These two parameters were set by comparing the classification accuracy OA with different types of wavelets for the first level of decomposition for each dataset as listed in Table 11 for the Indian Pines, Pavia University, and Salinas datasets. Then, for each dataset, two functions with the highest accuracies were selected and further decomposed to fix the best function with the level of decomposition as listed in Table 12.
As can be seen in Table 11, for the Indian Pines dataset, ‘db7’and ‘db9’functions have the highest classification accuracies (85.64% OA and 85.47% OA, respectively). For Pavia University dataset, ‘bior1.3’and ‘coif2’ functions have the highest classification accuracies (98.36% OA and 98.13%OA, respectively). For the Salinas dataset, the highest accuracies correspond to ‘bior3.7’ with an OA of 98.70% and ‘bior4.4’ with an OA of 98.67%. For each dataset, the aforementioned functions corresponding to the first two highest accuracies are further decomposed to fix only one function with its best level of decomposition. Table 12 lists the obtained classification results using the best two functions for each dataset in the proposed SDWT-2DCT_SVM approach.
As shown in Table 12, the highest classification accuracy (85.86% OA) was obtained using ‘db7’ function at the second level of decomposition for the Indian Pines dataset. For the Pavia University dataset, the highest accuracy (98.51% OA) was obtained using ‘bior1.3’ function at the fourth level of decomposition. For the Salinas dataset, the highest accuracy (98.65% OA) was obtained using ‘bior3.7’ function at its second level of decomposition. Thus, for the SDWT-2DCT_SVM approach, we opted the following spectral parameters (dataset, SDWT, wavelet function, level of decomposition):
-
(Indian Pines dataset, SDWT, ‘db7’, 2);
-
(Pavia University dataset, SDWT, ‘bior1.3’, 4);
-
(Salinas dataset, SDWT, ‘bior3.7’, 2).
The spatial filter parameter in the SDWT-2DCT_SVM approach was the threshold for 2D-DCT. This parameter was set by comparing the classification accuracy OA with varying threshold values where the opted threshold corresponds to the best classification OA for each dataset. Figure 10 depicts the variation of the classification accuracy OA with varying threshold values, ranging from 50 to 1000 with the step of 50, on the Indian Pines, Pavia University, and Salinas datasets.
As can be seen in Figure 10, the OA increased with a higher threshold value for all datasets and tended to be stable around the threshold of 1400 for Indian Pines, while this degree of stability was obtained at around a threshold of 1050 for the Pavia University dataset and around 850 for the Salinas dataset. Therefore, these threshold values corresponding to each dataset were selected for our experiments.

5.2. Classification of Indian Pines Dataset

In this subsection, we compare the classification effectiveness of the proposed frameworks SDWT-2DWT_SVM, SDWT-WF_SVM, and SDWT-2DCT_SVM with the other classification methods on the Indian Pines dataset. For this dataset, the parameter tunings for the three proposed approaches are recapitulated in Table 13. The number of PCs obtained using HySime criterion was 18 and therefore was used in PCA_SVM method.
The first experiment assessed the validity of our proposed approaches relative to the other methods. We randomly chose 100 samples for each class from the reference data as training samples, and the rest were used as test samples as given in Table 1. The results reported in Table 14 represent the averages of the individual classification accuracies (%), κ statistic, standard deviation, AA, OA, and computational time in seconds. Classification maps for all the approaches are shown in Figure 11. From Table 14 and Figure 11, several observations can be made.
The approaches incorporating only the spectral information, including SVM (70.40% OA) and PCA_SVM (64.40% OA) attain lower classification results when compared with the other spectral–spatial techniques. This can be confirmed in Figure 11, where it is clear that the classification maps resulting from these pixel-wise techniques are degraded by salt-and-pepper noise.
The methods considering both the spectral and spatial information, including 3D_SVM, 3DG_SVM, EPF, IFRF, BM4D_SVM, PARAFAC_SVM, and our three proposed approaches are most accurate. Specifically, the two proposed approaches combining two different filters and taking advantage of both of them, including SDWT-2DCT_SVM and SDWT-WF_SVM, achieved high classification results. The proposed SDWT-2DCT_SVM approach achieved the highest values for the three classification criteria (91.47% κ, 92.59% OA, and 95.46% AA), followed by the proposed SDWT-WF_SVM with the second-best performance (90.76% κ, 91.97% OA, and 94.89% AA). Furthermore, the proposed SDWT-2DCT_SVM and SDWT-WF_SVM approaches achieve 2.27%–28.19% and 1.65%–27.57% advantages in OA, respectively, over the other state-of-the-art methods. The classification maps produced from both the SDWT-2DCT_SVM and SDWT-WF_SVM approaches are noiseless and the closest to the ground truth map, as shown in Figure 11. These results demonstrate that the classification performance was improved by exploring the DWT multilevel decomposition in the spectral domain when spatially filtering the details coefficients using another different filter.
The proposed SDWT-2DWT_SVM approach using DWT in both the spectral and spatial filtering steps achieved worse classification results than the 3DG_SVM method and better results than the 3D_SVM (85.46% OA, 90.32% OA, and 85.04% OA, respectively). While the edge-preserving filtering (EPF)-based approach achieved the fourth-best classification result with an OA of 86.61%, the second EPF-based approach, which is the IFRF with an OA of 74.30%, achieved worse classification results than the two denoising-based approaches, namely the including BM4D_SVM and PARAFAC_SVM approaches (83.08% OA and 81.62% OA, respectively).
For the computational cost, Table 14 shows that the shortest classification time was 1.86 s achieved with the IFRF, which is a feature extraction-based method. Nevertheless, it had low classification accuracy (74.30% OA). Denoising based methods are time-consuming with low classification accuracies where CDCT-BM4D_SVM achieves an OA of 83.08% in 351.37 s, and the PARAFAC_SVM achieves an OA of 81.62% in 297.39 s.
The proposed SDWT_2DCT_SVM and SDWT_WF_SVM approaches achieved the two highest classification accuracies within a short execution time (less than one minute). For example, SDWT-2DCT_SVM achieved an OA of 92.59% in 32.03 s, while SDWT-WF_SVM achieved an OA of 91.97% in 47.74 s. The third highest classification accuracy, achieved by the 3DG_SVM (90.32% OA) method within 210.16 s, represents the longest execution time among all the compared methods.
In the second experiment, the performance of the proposed approaches was evaluated against state-of-the-art approaches under different training conditions, in which 2%, 4%, 6%, 8%, and 10% labeled samples of each class are randomly selected as training samples, while the rest are used as test samples. Each experiment was performed 20 times, and the average results in terms of the OA values for each method are plotted in Figure 12.
Figure 12 illustrates that the two proposed approaches SDWT-2DCT_SVM and SDWT-WF_SVM can significantly improve classification results when varying the training sample size, as the size of training samples changes from 2% to 10%, these two proposed approaches always achieved the first and the second highest OA. Meanwhile, with 2% and 4% as a training sample, the proposed CDCT-WF_SVM approach outperforms the SDWT-2DCT_SVM approach that takes back the best performance when using more than 6% training samples per class. In this scenario, our proposed SDWT-2DWT_SVM outperforms the 3DG_SVM method when using only 2% training sample per class. Meanwhile, using 4% and above, the 3DG_SVM method is more accurate than the proposed SDWT-2DWT_SVM approach.

5.3. Classification of Pavia University Dataset

Regarding the Pavia University dataset experiments, the parameter tunings of the three proposed approaches are recapitulated in Table 15. In the PCA_SVM method, the number of PCs obtained using HySime criterion on this dataset was 45.
In the first experiment, the effectiveness of our proposed frameworks against other methods was evaluated. We randomly chose 300 samples for each class from the reference data as training samples, and the rest were used as test samples, as summarized in Table 2. The quantitative classification results, including the individual classification accuracies (%), κ statistic, standard deviation, OA, AA, and computational time in seconds are reported in Table 16. The classification maps of the compared approaches on the Pavia University dataset are shown in Figure 13.
As can be seen in Table 16, the highest classification accuracies are always obtained by the spectral–spatial classification techniques, especially the structural filtering based approaches. The proposed method SDWT-2DCT_SVM achieved the highest values for the three classification criteria (98.51% κ, 98.90% OA, and 98.94% AA), followed by the proposed SDWT-2DWT_SVM (97.95% κ, 98.49% OA, and 98.37% AA). The proposed SDWT-2DCT_SVM and SDWT-2DWT_SVM approaches achieve 0.83%–18.41% and 0.42%–18% advantages in OA, respectively, over the other state-of-the-art methods. However, for this dataset, 3DG_SVM method (98.07% OA) achieved better classification results than the proposed SDWT-2DCT_SVM (97.29% OA) and the 3D_SVM method (96.95% OA). The BM4D_SVM, a denoising-based method, achieved classification results comparable with the EPF method (87.23% OA and 86.81% OA, respectively). Meanwhile, PARAFAC_SVM and IFRF were the least accurate among the spectral–spatial classifiers with OAs of 83.99% and 83.14%, respectively.
As can be visually seen from Figure 13, the classification maps produced from methods combining spectral and spatial information are smooth. In particular, wavelet-based approaches have smoother classification maps than the other approaches. The classification maps obtained from our proposed SDWT-2DCT_SVM and SDWT-2DWT_SVM approaches closely match the ground-truth map.
Although wavelet-based approaches have the highest classification accuracies, the processing time of our proposed frameworks is faster than 3DG_SVM and 3D_SVM methods. For example, the proposed SDWT-2DCT_SVM achieve an OA of 98.90% in 426.28 s, while 3DG_SVM achieves an OA of 98.07% in 2142.40 s, which is five times slower than our proposed SDWT-2DCT_SVM approach. The shortest processing time was achieved by the IFRF method within 18.14 s; however, this method delivered the lowest classification accuracy among the spectral–spatial approaches with an OA of 83.14%. In addition, denoising-based approaches are time-consuming with low classification accuracies. For example, the BM4D_SVM achieved an OA of 87.23% in 1535.13 s and the PARAFAC_SVM achieved an OA of 83.99% in 1275.06 s.
In the second experiment, we evaluated the performance of our proposed approaches on the Pavia University dataset against state-of-the-art methods with different training conditions, in which 2%, 4%, 6%, 8%, and 10% labeled samples of each class were randomly selected as training samples, and the rest were used for testing. The average results in terms of OA values are plotted in Figure 14 for each compared method.
Figure 14 illustrates that all wavelet-based approaches outperform the other compared methods when varying the training sample size. Specifically, the proposed SDWT-2DCT_SVM and SDWT-2DWT_SVM approaches always significantly improved the classification accuracy significantly with a different number of training samples, and as the size of training samples changed from 2% to 10%, these two proposed approaches always achieved the two highest OA values, followed by 3DG_SVM and the proposed SDWT-WF_SVM methods. Additionally, BM4D_SVM provided higher classification accuracy than the IFRF and EPF methods even when varying the training sample size.

5.4. Classification of Salinas Dataset

Parameter tuning the three proposed approaches for Salinas dataset is recapitulated in Table 17. For PCA_SVM method, the number of PCs obtained using HySime criterion for this dataset was 20.
In the first experiment, to compare our proposed approaches and other methods, we randomly chose 100 training samples for each class from the reference data, and the rest were used as test samples, as summarized in Table 3. Table 18 reports the quantitative classification results, including the individual classification accuracies (%), κ statistic, standard deviation, AA, OA, and computational time in seconds. The classification maps for all the compared approaches on the Salinas image are illustrated in Figure 15.
From Table 18, we can see that our three proposed approaches provide the highest three classification accuracies in terms of all criteria (AA, OA, and kappa) with stable performance for the proposed SDWT-2DCT_SVM (0.21 as the lowest standard deviation value for OA). Hence, the proposed SDWT-2DCT_SVM approach achieved the highest values for the three classification criteria (98.91% κ, 99.02% OA, and 99.19% AA), followed by the proposed SDWT-WF_SVM with the second-best performance (97.46% κ, 97.72% OA, and 98.61% AA). The third-best performance was achieved by our proposed SDWT-2DWT_SVM (96.13% κ, 96.71% OA, and 97.97% AA). Furthermore, the proposed approaches achieve 2.61%–17.4%, 1.31%–16.1%, and 0.3%–15.09% OA advantage over the other state-of-the-art methods in the SDWT-2DCT_SVM, the SDWT-WF_SVM, and the SDWT-2DWT_SVM approaches, respectively.
In addition, the classification maps produced from the three proposed approaches are smooth and close to the ground truth map, as shown in Figure 15. Specifically, the classification map resulting from the proposed SDWT-2DWT_SVM is noiseless and very similar to the ground truth map.
In this scenario, the IFRF method with an OA of 96.41% outperforms the 3DG_SVM (95.02% OA), with a smoother classification map. The EPF, 3D_SVM, and BM4D_SVM methods provided close classification results (92.95% OA, 92.88% OA, and 91.96% OA, respectively), resulting in similar classification maps as shown in Figure 15. The PARAFAC_SVM approach delivered the lowest classification accuracy (81.62% OA) with a noisy produced classification map.
As for the computational time, Table 18 shows that feature extraction-based methods achieved the shortest times, using 4.77 s for the IFRF method and 13.18 s for the PCA_SVM method. In general, classifiers incorporating spatial information are slower than the pixel-wise methods. Our proposed approaches, including SDWT-2DCT_SVM, SDWT-WF_SVM, and SDWT-2DWT_SVM, had the fourth, the fifth, and the sixth shortest execution times, among all the tested spectral–spatial methods. Specifically, the highest classification accuracy (99.09% OA) was obtained by our proposed SDWT-2DCT_SVM approach using 169.79 s and was less than three minutes. The 3DG_SVM and BM4D_SVM methods were time-consuming, using 1142.52 and 1854.07 s, respectively. Spectral DWT is performed on each pixel, and the spatial filters performed on each band in the three proposed approaches, which means that they can be parallelized. Thus, the computational time could be greatly improved.
The robustness of the proposed approaches regarding the state-of-the-art methods with different training conditions was evaluated in the second experiment, in which 2%, 4%, 6%, 8%, and 10% labeled samples of each class were randomly selected as training samples. The rest were used as testing samples. The average results of the OA values of 20 trials are plotted in Figure 16.
Figure 16 shows that the three proposed approaches can improve the classification accuracy significantly with different training sample sizes, as the number of training samples varies from 2% to 10%. Our proposed approaches (SDWT-2DCT_SVM and SDWT-WF_SVM) always achieved the first- and the second-highest OA. However, the IFRF method outperforms our proposed SDWT-2DWT_SVM approach when using a smaller number of training samples (2%–6%). When the training sample size was greater than 8%, the proposed SDWT-2DWT_SVM approach tends to be more accurate than the IFRF method.
In general, several observations can be made from the experiments reported in this section on the three widely used hyperspectral datasets:
-
The proposed approaches outperform the other considered methods; especially, our proposed SDWT-2DCT_SVM framework in terms of all classification criteria (AA, OA, and κ). It delivers higher accuracy on the three datasets with smoother classification maps than the other compared methods.
-
The proposed approaches deliver higher classification accuracies under the constraint of the small size of training samples.
-
The proposed frameworks are computationally efficient with a reasonable tradeoff between accuracy and computational time. Subsequently, they will be quite useful for applications requiring a fast response.
-
The proposed methods can deal with different spatial resolutions (20 m, 3.7 m, and 1.7 m). Particularly, for the Indian Pines scene at a spatial resolution of 20 m, our proposed SDWT-2DCT_SVM and SDWT-WF_SVM approaches achieved the first- and the second-highest classification accuracies. Our proposed frameworks are effective for the classification of low spatial resolution images.
Indeed, the potential strength of our proposed strategies consists of the following advantages:
-
The spectral and spatial dimensions are filtered separately as the noise variance is more drastic in the spectral dimension than in the spatial dimension.
-
The performed spectral filters can potentially separate the valuable information from the details using the wavelet multi-resolution analysis.
-
The spatial filter is performed in the transform domain, where the separation of the significant information from the detail is easy.
-
The spatial filter in the transform domain is performed only on the detail coefficients and does not alter the significant information resulting from the spectral filter.
These characteristics made this filtering process very effective on the three tested datasets as the highest classification accuracies were always achieved by our proposed approaches.

6. Conclusions

In this paper, three effective classification approaches based on spectral–spatial preprocessing using DWT multilevel decomposition and spatial filtering are proposed. In these approaches, we perform spectral DWT multilevel decomposition on the original hyperspectral image to separate the approximation coefficients that represent the most significant information from the detail coefficients. Then, for each level of decomposition, only the details coefficients are spatially filtered in the transform domain to extract the rest of the meaningful spatial information and further enhance the spectral filtered part. After the spatial filtering step using three different filters, namely 2D-DWT, 2D-DCT, and AWF, corresponding to the three proposed frameworks, the DWT reconstruction was carried out on both the approximate and the spatially filtered detail coefficients. Finally, the output-preprocessed image is fed into a simple linear SVM classifier.
Different performance indicators were used to assess our experiments, including the classification accuracy, execution time, the classification maps, and the tests with different training sample sizes. Our experimental results, conducted with three hyperspectral images, confirm that the proposed approaches yield good classification results that were comparable to or better than those obtained by other state-of-the-art classifiers. In particular, the proposed SDWT-2DCT_SVM approach exhibited a significantly higher performance and outperformed all the comparable methods in terms of classification accuracy, even with small training sample sizes, and within a reasonable execution time on all the three datasets. Furthermore, in the three proposed approaches, spectral DWT multilevel decomposition is performed on each pixel separately, and the spatial filters are performed on each band, so they could be parallelized easily for significantly reduced computational times.

Author Contributions

All authors have made great contributions to the work. Conceptualization, R.B. and K.B.; Software, R.B. and K.B.; validation, H.W., R.B. and K.B.; formal analysis, R.B.; supervision, H.W., Writing—original draft, R.B.; and Writing—review and editing, H.W. and K.B.

Funding

This research received no external funding.

Acknowledgments

The authors are grateful for the valuable comments and propositions from the reviewer of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tong, Q.; Xue, Y.; Zhang, L. Progress in Hyperspectral Remote Sensing Science and Technology in China Over the Past Three Decades. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 70–91. [Google Scholar] [CrossRef]
  2. Adão, T.; Hruška, J.; Pádua, L.; Bessa, J.; Peres, E.; Morais, R.; Sousa, J.J. Hyperspectral Imaging: A Review on UAV-Based Sensors, Data Processing and Applications for Agriculture and Forestry. Remote Sens. 2017, 9, 1110. [Google Scholar] [CrossRef] [Green Version]
  3. Yokoya, N.; Chan, J.C.-W.; Segl, K. Potential of Resolution-Enhanced Hyperspectral Data for Mineral Mapping Using Simulated EnMAP and Sentinel-2 Images. Remote Sens. 2016, 8, 172. [Google Scholar] [CrossRef] [Green Version]
  4. He, J.; He, Y.; Zhang, C. Determination and Visualization of Peimine and Peiminine Content in Fritillaria thunbergii Bulbi Treated by Sulfur Fumigation Using Hyperspectral Imaging with Chemometrics. Molecules 2017, 22, 1402. [Google Scholar] [CrossRef]
  5. Ghamisi, P.; Plaza, J.; Chen, Y.; Li, J.; Plaza, A.J. Advanced spectral classifiers for hyperspectral images: A review. IEEE Geosci. Remote Sens. Mag. 2017, 5, 8–32. [Google Scholar] [CrossRef] [Green Version]
  6. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  7. Ham, J.; Chen, Y.; Crawford, M.M.; Ghosh, J. Investigation of the random forest framework for classification of hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2005, 43, 492. [Google Scholar] [CrossRef] [Green Version]
  8. Ratle, F.; Camps-Valls, G.; Weston, J. Semisupervised neural networks for efficient hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2010, 48, 2271–2282. [Google Scholar] [CrossRef]
  9. Zhong, Y.; Zhang, L. An adaptive artificial immune network for supervised classification of multi-/hyperspectral remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2012, 50, 894–909. [Google Scholar] [CrossRef]
  10. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Spectral--spatial classification of hyperspectral data using loopy belief propagation and active learning. IEEE Trans. Geosci. Remote Sens. 2013, 51, 844–856. [Google Scholar] [CrossRef]
  11. Di, W.; Crawford, M.M. View generation for multiview maximum disagreement based active learning for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2012, 50, 1942–1954. [Google Scholar] [CrossRef]
  12. Iordache, M.-D.; Bioucas-Dias, J.M.; Plaza, A. Total variation regulatization in sparse hyperspectral unmixing. In Proceedings of the 2011 3rd Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Lisbon, Portugal, 6–9 June 2011; pp. 1–4. [Google Scholar]
  13. Chen, Y.; Nasrabadi, N.M.; Tran, T.D. Hyperspectral Image Classification via Kernel Sparse Representation. IEEE Trans. Geosci. Remote Sens. 2013, 51, 217–231. [Google Scholar] [CrossRef] [Green Version]
  14. Castrodad, A.; Xing, Z.; Greer, J.B.; Bosch, E.; Carin, L.; Sapiro, G. Learning Discriminative Sparse Representations for Modeling, Source Separation, and Mapping of Hyperspectral Imagery. IEEE Trans. Geosci. Remote Sens. 2011, 49, 4263. [Google Scholar] [CrossRef]
  15. Wang, Y.; Duan, H. Classification of Hyperspectral Images by SVM Using a Composite Kernel by Employing Spectral, Spatial and Hierarchical Structure Information. Remote Sens. 2018, 10, 441. [Google Scholar] [CrossRef]
  16. Camps-Valls, G.; Tuia, D.; Bruzzone, L.; Benediktsson, J.A. Advances in hyperspectral image classification: Earth monitoring with statistical learning methods. IEEE Signal Process. Mag. 2014, 31, 45–54. [Google Scholar] [CrossRef] [Green Version]
  17. Plaza, A.; Benediktsson, J.A.; Boardman, J.W.; Brazile, J.; Bruzzone, L.; Camps-Valls, G.; Chanussot, J.; Fauvel, M.; Gamba, P.; Gualtieri, A.; et al. Recent advances in techniques for hyperspectral image processing. Remote Sens. Environ. 2009, 113, S110–S122. [Google Scholar] [CrossRef]
  18. Casalino, G.; Gillis, N. Sequential dimensionality reduction for extracting localized features. Pattern Recognit. 2017, 63, 15–29. [Google Scholar] [CrossRef] [Green Version]
  19. Quesada-Barriuso, P.; Arguello, F.; Heras, D.B. Spectral--spatial classification of hyperspectral images using wavelets and extended morphological profiles. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 1177–1185. [Google Scholar] [CrossRef]
  20. Jia, S.; Wu, K.; Zhu, J.; Jia, X. Spectral-Spatial Gabor Surface Feature Fusion Approach for Hyperspectral Imagery Classification. IEEE Trans. Geosci. Remote Sens. 2018, 1–13. [Google Scholar] [CrossRef]
  21. Tang, Y.Y.; Lu, Y.; Yuan, H. Hyperspectral image classification based on three-dimensional scattering wavelet transform. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2467–2480. [Google Scholar] [CrossRef]
  22. Fauvel, M.; Benediktsson, J.A.; Chanussot, J.; Sveinsson, J.R. Spectral and spatial classification of hyperspectral data using SVMs and morphological profiles. IEEE Trans. Geosci. Remote Sens. 2008, 46, 3804–3814. [Google Scholar] [CrossRef] [Green Version]
  23. Ghamisi, P.; Benediktsson, J.A.; Cavallaro, G.; Plaza, A. Automatic Framework for Spectral–Spatial Classification Based on Supervised Feature Extraction and Morphological Attribute Profiles. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2147. [Google Scholar] [CrossRef]
  24. Sun, S.; Zhong, P.; Xiao, H.; Wang, R. An MRF model-based active learning framework for the spectral-spatial classification of hyperspectral imagery. IEEE J. Sel. Top. Signal Process. 2015, 9, 1074–1088. [Google Scholar] [CrossRef]
  25. Pan, B.; Shi, Z.; Xu, X. MugNet: Deep learning for hyperspectral image classification using limited samples. ISPRS J. Photogramm. Remote Sens. 2017, 145, 108–119. [Google Scholar] [CrossRef]
  26. Liu, B.; Yu, X.; Yu, A.; Zhang, P.; Wan, G. Spectral-spatial classification of hyperspectral imagery based on recurrent neural networks. Remote Sens. Lett. 2018, 9, 1118–1127. [Google Scholar] [CrossRef]
  27. Song, B.; Li, J.; Dalla Mura, M.; Li, P.; Plaza, A.; Bioucas-Dias, J.M.; Benediktsson, J.A.; Chanussot, J. Remotely sensed image classification using sparse representations of morphological attribute profiles. IEEE Trans. Geosci. Remote Sens. 2014, 52, 5122. [Google Scholar] [CrossRef] [Green Version]
  28. Ni, D.; Ma, H. Hyperspectral image classification via sparse code histogram. IEEE Geosci. Remote Sens. Lett. 2015, 12, 1843–1847. [Google Scholar]
  29. Wang, Y.; Song, H.; Zhang, Y. Spectral-spatial classification of hyperspectral images using joint bilateral filter and graph cut based model. Remote Sens. 2016, 8, 748. [Google Scholar] [CrossRef] [Green Version]
  30. Fang, L.; Li, S.; Kang, X.; Benediktsson, J.A. Spectral--spatial classification of hyperspectral images with a superpixel-based discriminative sparse model. IEEE Trans. Geosci. Remote Sens. 2015, 53, 4186–4201. [Google Scholar] [CrossRef]
  31. He, L.; Li, J.; Liu, C.; Li, S. Recent advances on spectral--spatial hyperspectral image classification: An overview and new guidelines. IEEE Trans. Geosci. Remote Sens. 2018, 56, 1579–1597. [Google Scholar] [CrossRef]
  32. Cao, X.; Ji, B.; Ji, Y.; Wang, L.; Jiao, L. Hyperspectral image classification based on filtering: A comparative study. J. Appl. Remote Sens. 2017, 11, 35007. [Google Scholar] [CrossRef]
  33. Liu, K.-H.; Lin, Y.-Y.; Chen, C.-S. Linear spectral mixture analysis via multiple-kernel learning for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2254–2269. [Google Scholar] [CrossRef] [Green Version]
  34. Zhou, Y.; Peng, J.; Chen, C.L.P. Extreme learning machine with composite kernels for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2351–2360. [Google Scholar] [CrossRef]
  35. Bazine, R.; Wu, H.; Boukhechba, K. Spatial Filtering in DCT Domain-Based Frameworks for Hyperspectral Imagery Classification. Remote Sens. 2019, 11, 1405. [Google Scholar] [CrossRef] [Green Version]
  36. Oktem, R.; Ponomarenko, N.N. Image filtering based on discrete cosine transform. Telecommun. Radio Eng. 2007, 66. [Google Scholar] [CrossRef]
  37. Guo, X.; Huang, X.; Zhang, L. Three-Dimensional Wavelet Texture Feature Extraction and Classification for Multi/Hyperspectral Imagery. IEEE Geosci. Remote Sens. Lett. 2014, 11, 2183. [Google Scholar] [CrossRef]
  38. Kordi Ghasrodashti, E.; Helfroush, M.S.; Danyali, H. A wavelet-based classification of hyperspectral images using Schroedinger eigenmaps. Int. J. Remote Sens. 2017, 38, 3608–3634. [Google Scholar] [CrossRef]
  39. Kordi Ghasrodashti, E.; Helfroush, M.S.; Danyali, H. Spectral-spatial classification of hyperspectral images using wavelet transform and hidden Markov random fields. Geocarto Int. 2018, 33, 771–790. [Google Scholar] [CrossRef]
  40. Qian, Y.; Ye, M.; Zhou, J. Hyperspectral image classification based on structured sparse logistic regression and three-dimensional wavelet texture features. IEEE Trans. Geosci. Remote Sens. 2013, 51, 2276–2291. [Google Scholar] [CrossRef] [Green Version]
  41. Cao, X.; Xu, L.; Meng, D.; Zhao, Q.; Xu, Z. Integration of 3-dimensional discrete wavelet transform and Markov random field for hyperspectral image classification. Neurocomputing 2017, 226, 90–100. [Google Scholar] [CrossRef]
  42. Zhou, X.; Prasad, S.; Crawford, M.M. Wavelet-domain multiview active learning for spatial-spectral hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 4047–4059. [Google Scholar] [CrossRef]
  43. Shi, C.; Pun, C.-M. 3D multi-resolution wavelet convolutional neural networks for hyperspectral image classification. Inf. Sci. (Ny.) 2017, 420, 49–65. [Google Scholar] [CrossRef]
  44. Othman, H.; Qian, S.-E. Noise reduction of hyperspectral imagery using hybrid spatial-spectral derivative-domain wavelet shrinkage. IEEE Trans. Geosci. Remote Sens. 2006, 44, 397–408. [Google Scholar] [CrossRef]
  45. Mallat, S.G. A theory for multiresolution signal decomposition: The wavelet representation. IEEE Trans. Pattern Anal. Mach. Intell. 1989, 11, 674. [Google Scholar] [CrossRef] [Green Version]
  46. Donoho, D.L. Denoising by softthresholding. IEEE Trans. Inf. Theory 1995, 41, 613–627. [Google Scholar] [CrossRef] [Green Version]
  47. Prasad, S.; Bruce, L.M. Limitations of principal components analysis for hyperspectral target recognition. IEEE Geosci. Remote Sens. Lett. 2008, 5, 625–629. [Google Scholar] [CrossRef]
  48. Kang, X.; Li, S.; Benediktsson, J.A. Spectral-Spatial Hyperspectral Image Classification with Edge-Preserving Filtering. IEEE Trans. Geosci. Remote Sens. 2014, 52, 2666–2677. [Google Scholar] [CrossRef]
  49. Kang, X.; Li, S.; Benediktsson, J.A. Feature extraction of hyperspectral images with image fusion and recursive filtering. IEEE Trans. Geosci. Remote Sens. 2014, 52, 3742–3752. [Google Scholar] [CrossRef]
  50. Maggioni, M.; Foi, A. Nonlocal transform-domain denoising of volumetric data with groupwise adaptive variance estimation. In Proceedings of the IS&T/SPIE Electronic Imaging 2012, Burlingame, CA, USA, 22–26 January 2012; Volume 8296, p. 82960O. [Google Scholar]
  51. Liu, X.; Bourennane, S.; Fossati, C. Denoising of hyperspectral images using the PARAFAC model and statistical performance analysis. IEEE Trans. Geosci. Remote Sens. 2012, 50, 3717–3724. [Google Scholar] [CrossRef]
  52. Chang, C.-C.; Lin, C.-J. {LIBSVM}: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 27:1–27:27. [Google Scholar] [CrossRef]
  53. Bioucas-Dias, J.M.; Nascimento, J.M.P. Estimation of signal subspace on hyperspectral data. In Proceedings of the SPIE Remote Sensing, Bruges, Belgium, 19–22 September 2005; Volume 5982, p. 59820L. [Google Scholar]
  54. Story, M.; Congalton, R.G. Accuracy assessment: A user’s perspective. Photogramm. Eng. Remote Sens. 1986, 52, 397–399. [Google Scholar]
Figure 1. The principle of 1D-DWT. (a). DWT analysis filter bank; (b). DWT synthesis filter bank.
Figure 1. The principle of 1D-DWT. (a). DWT analysis filter bank; (b). DWT synthesis filter bank.
Remotesensing 11 02906 g001
Figure 2. The principle of multilevel 1D-DWT decomposition.
Figure 2. The principle of multilevel 1D-DWT decomposition.
Remotesensing 11 02906 g002
Figure 3. The principle of two-dimensional DWT (2D-DWT) analysis filter bank.
Figure 3. The principle of two-dimensional DWT (2D-DWT) analysis filter bank.
Remotesensing 11 02906 g003
Figure 4. Two levels of 2D-DWT decomposition.
Figure 4. Two levels of 2D-DWT decomposition.
Remotesensing 11 02906 g004
Figure 5. Detailed flow chart of the proposed spectral DWT multilevel decomposition-based approaches.
Figure 5. Detailed flow chart of the proposed spectral DWT multilevel decomposition-based approaches.
Remotesensing 11 02906 g005
Figure 6. The ground truth information of the Indian Pines dataset: (a) original dataset; (b) ground truth map; (c) training map; (d) test map.
Figure 6. The ground truth information of the Indian Pines dataset: (a) original dataset; (b) ground truth map; (c) training map; (d) test map.
Remotesensing 11 02906 g006
Figure 7. The ground truth information of the Pavia University dataset: (a) original dataset; (b) ground truth map; (c) training map; (d) testing map.
Figure 7. The ground truth information of the Pavia University dataset: (a) original dataset; (b) ground truth map; (c) training map; (d) testing map.
Remotesensing 11 02906 g007
Figure 8. The ground truth information map of Salinas dataset: (a)original dataset; (b) ground truth map; (c) training map; (d) test map.
Figure 8. The ground truth information map of Salinas dataset: (a)original dataset; (b) ground truth map; (c) training map; (d) test map.
Remotesensing 11 02906 g008
Figure 9. Classification accuracy (overall accuracy (OA)%) with varying Wiener filter patch size in SDWT-WF_SVM on the Indian Pines, Pavia University, and Salinas datasets.
Figure 9. Classification accuracy (overall accuracy (OA)%) with varying Wiener filter patch size in SDWT-WF_SVM on the Indian Pines, Pavia University, and Salinas datasets.
Remotesensing 11 02906 g009
Figure 10. Classification accuracy (OA %) with varying 2D-DCT thresholds in SDWT-2DCT_SVM on the Indian Pines, Pavia University, and Salinas datasets.
Figure 10. Classification accuracy (OA %) with varying 2D-DCT thresholds in SDWT-2DCT_SVM on the Indian Pines, Pavia University, and Salinas datasets.
Remotesensing 11 02906 g010
Figure 11. Classification maps of different methods on the Indian Pines dataset.
Figure 11. Classification maps of different methods on the Indian Pines dataset.
Remotesensing 11 02906 g011
Figure 12. Classification accuracy (OA, %) achieved by the compared methods using different training sample sizes on the Indian Pine dataset.
Figure 12. Classification accuracy (OA, %) achieved by the compared methods using different training sample sizes on the Indian Pine dataset.
Remotesensing 11 02906 g012
Figure 13. Classification maps of different methods on the Pavia University dataset.
Figure 13. Classification maps of different methods on the Pavia University dataset.
Remotesensing 11 02906 g013
Figure 14. Classification accuracy (OA, %) achieved by the compared methods using different training sample sizes on the Pavia University dataset.
Figure 14. Classification accuracy (OA, %) achieved by the compared methods using different training sample sizes on the Pavia University dataset.
Remotesensing 11 02906 g014
Figure 15. Classification maps of different methods on the Salinas dataset.
Figure 15. Classification maps of different methods on the Salinas dataset.
Remotesensing 11 02906 g015
Figure 16. Classification accuracy (OA, %) achieved by the compared methods using different training sample sizes on the Salinas dataset.
Figure 16. Classification accuracy (OA, %) achieved by the compared methods using different training sample sizes on the Salinas dataset.
Remotesensing 11 02906 g016
Table 1. Classes of the Indian Pines scene with the number of testing and training samples.
Table 1. Classes of the Indian Pines scene with the number of testing and training samples.
ClassTypeSamplesTrainingTesting
1Alfalfa462323
2Corn-notill14281001328
3Corn-mintill830100730
4Corn237100137
5Grass-pasture483100383
6Grass-trees730100630
7Grass-pasture-mowed281414
8Hay-windrowed478100378
9Oats201010
10Soybean-notill972100872
11Soybean-mintill24551002355
12Soybean-clean593100493
13Wheat205100105
14Woods12651001165
15Buildings-grass-trees-drives386100286
16Stone-steel-towers934746
Table 2. Classes of Pavia University scene with the related numbers of testing and training samples.
Table 2. Classes of Pavia University scene with the related numbers of testing and training samples.
ClassTypeSamplesTrainingTesting
1Asphalt66313006331
2Meadows18,64930018,349
3Gravel20993001799
4Trees30643002764
5Painted metal sheets13453001045
6Bare soil50293004729
7Bitumen13303001030
8Self-blocking bricks36823003382
9Shadows947300647
Table 3. Classes of the Salinas scene with the related number of test and training samples.
Table 3. Classes of the Salinas scene with the related number of test and training samples.
ClassTypeSamplesTrainingTesting
1Brocoligreenweeds120091001909
2Brocoligreenweeds237261003626
3Fallow19761001876
4Fallowroughplow13941001294
5Fallowsmooth26781002578
6Stubble39591003859
7Celery35791003479
8Grapesuntrained11,27110011,171
9Soilvinyarddevelop62031006103
10Cornsenescedgreenweeds32781003178
11Lettuceromaine4wk1068100968
12Lettuceromaine5wk19271001827
13Lettuceromaine6wk916100816
14Lettuceromaine7wk1070100970
15Vinyarduntrained72681007168
16Vinyardverticaltrellis18071001707
Table 4. The wavelet families and function types investigated for parameter tuning.
Table 4. The wavelet families and function types investigated for parameter tuning.
Wavelet FamilyMother Function
Daubechieshaar, db2, …, db9
Biorthogonalbior1.1, bior1.3, bior1.5, bior2.2, bior2.4, bior2.6, bior3.1, bior3.3, bior3.5, bior3.7, bior4.4, bior5.5
Coifletscoif1, …, coif5
Symletssym2, …, sym6
Table 5. Classification accuracies with respect to the spectral DWT mother function on all datasets for the SDWT-2DWT_SVM approach.
Table 5. Classification accuracies with respect to the spectral DWT mother function on all datasets for the SDWT-2DWT_SVM approach.
WaveletIndian PinesPavia UniversitySalinasWaveletIndian PinesPavia UniversitySalinas
haar82.7198.1694.52bior3.381.6697.8294.45
db283.4897.8694.10bior3.582.5797.7594.41
db383.4797.6594.96bior3.782.7997.9594.77
db483.0198.0394.42bior4.483.5697.7094.44
db584.2497.6394.69bior5.583.2298.0393.94
db683.6797.8694.17coif182.7597.9494.68
db784.8497.8293.97coif282.5398.1394.17
db883.4097.9393.19coif383.1198.0695.01
db984.8297.7494.02coif482.7897.8593.88
bior1.383.5598.1294.22coif583.2198.0494.45
bior1.582.8798.0994.12sym283.3697.7194.52
bior2.282.2997.8694.49sym383.7897.6894.91
bior2.483.0397.8194.63sym482.8397.7994.90
bior2.683.3297.9694.06sym583.4197.9894.27
bior3.179.2097.8194.37sym683.5497.9494.14
Table 6. Classification accuracies with respect to the spectral DWT level of decomposition on all datasets for the SDWT-2DWT_SVM approach.
Table 6. Classification accuracies with respect to the spectral DWT level of decomposition on all datasets for the SDWT-2DWT_SVM approach.
Decomposition LevelIndian PinesPavia UniversitySalinas
Db7Db9Coif2Bior1.3Coif3Sym4
180.3479.8391.9895.2691.6991.13
282.3782.9996.6696.9893.0493.08
384.1284.7098.1198.1194.4794.24
4 98.21 94.53
Table 7. Classification accuracies with respect to the 2D-DWT mother function on all datasets for the SDWT-2DWT_SVM.
Table 7. Classification accuracies with respect to the 2D-DWT mother function on all datasets for the SDWT-2DWT_SVM.
WaveletIndian PinesPavia UniversitySalinasWaveletIndian PinesPavia UniversitySalinas
haar81.6795.4390.70bior3.378.0893.6890.12
db281.0195.1790.57bior3.579.0494.4289.54
db379.9395.3990.14bior3.778.8394.3489.92
db480.5894.9690.15bior4.480.0895.1590.58
db580.0694.7990.53bior5.580.2095.1789.87
db680.1795.2890.20coif180.7595.5190.66
db780.3195.1090.19coif281.3895.3489.92
db880.1995.1590.24coif380.9795.2590.19
db980.2794.9790.07coif480.5295.1390.16
bior1.379.6194.8990.18coif580.9295.2590.15
bior1.580.1194.2889.70sym280.9095.2090.61
bior2.281.8295.2189.21sym380.3595.3290.18
bior2.479.7095.0690.44sym480.9895.3389.99
bior2.680.0995.0690.12sym581.1995.2390.18
bior3.176.7989.5388.35sym680.4694.9890.08
Table 8. Classification accuracies with respect to 2D-DWT level of decomposition on all datasets for the SDWT-2DWT_SVM approach.
Table 8. Classification accuracies with respect to 2D-DWT level of decomposition on all datasets for the SDWT-2DWT_SVM approach.
Decomposition LevelIndian PinesPavia UniversitySalinas
HaarBior2.2HaarCoif1HaarCoif1
177.2776.8890.3990.6388.8689.08
281.6781.8295.4395.5190.7090.66
384.9182.9597.4697.3492.9992.57
485.6885.3398.3198.0393.9194.35
585.66 98.2798.1595.4194.85
685.54 98.2998.2895.24
785.26 98.33 95.90
8 98.44
Table 9. Classification accuracies with respect to the DWT mother function on all datasets for the SDWT-WF_SVM approach.
Table 9. Classification accuracies with respect to the DWT mother function on all datasets for the SDWT-WF_SVM approach.
WaveletIndian PinesPavia UniversitySalinasWaveletIndian PinesPavia UniversitySalinas
haar90.02--bior3.389.0493.6195.76
db289.0394.5096.57bior3.588.7493.9195.86
db390.0594.4895.83bior3.789.9793.8296.04
db489.4494.6696.78bior4.490.3894.2796.06
db590.2494.7596.62bior5.590.2394.4596.44
db689.5394.6495.69coif190.0594.5296.58
db790.0894.2696.26coif289.6595.3396.65
db890.0493.8995.97coif389.5994.1896.15
db990.2294.1995.81coif489.4193.9296.43
bior1.3---coif589.8094.2595.56
bior1.590.41--sym289.2894.4896.40
bior2.289.5594.1596.30sym390.5294.4096.17
bior2.490.7094.6995.66sym490.4794.9195.97
bior2.689.3894.6095.96sym589.4794.3896.49
bior3.187.6193.4395.09sym690.3394.5996.29
Table 10. Classification accuracies with respect to the DWT level of decomposition on all datasets for the SDWT-WF_SVM approach.
Table 10. Classification accuracies with respect to the DWT level of decomposition on all datasets for the SDWT-WF_SVM approach.
Decomposition LevelIndian PinesPavia UniversitySalinas
BIOR2.4Sym3Coif2Sym4Db4Coif2
184.7284.8790.3191.0493.0993.33
287.3888.6595.7295.6194.8594.48
390.2690.2097.3497.0396.7896.45
491.0190.68 97.5397.18
5 92.37
Table 11. Classification accuracies with respect to the DWT mother function on all datasets for the SDWT-2DCT approach.
Table 11. Classification accuracies with respect to the DWT mother function on all datasets for the SDWT-2DCT approach.
WaveletIndian PinesPavia UniversitySalinasWaveletIndian PinesPavia UniversitySalinas
haar84.3398.2898.28bior3.382.3895.7998.48
db284.4698.1998.44bior3.581.3396.9498.67
db384.5798.0498.31bior3.783.3595.3498.70
db483.0598.3598.20bior4.484.6397.6098.67
db582.5896.9998.41bior5.584.5397.6098.64
db685.3798.1298.56coif183.9598.1398.46
db785.6497.7098.02coif284.8398.3698.50
db884.4097.8798.38coif384.4197.5098.10
db985.4797.4698.41coif484.5997.9698.16
bior1.384.9698.6297.90coif584.3597.5398.31
bior1.583.6898.2998.27sym284.5198.0698.36
bior2.284.2098.0198.54sym384.2197.9098.65
bior2.484.4597.0098.49sym483.7897.6298.47
bior2.683.9497.7098.58sym584.3297.7598.12
bior3.178.9596.3798.58sym684.7597.5698.29
Table 12. Classification accuracies with respect to the DWT level of decomposition on all datasets for the spectral DWT multilevel decomposition with 2D-DCT filter (SDWT-2DCT_SVM) approach.
Table 12. Classification accuracies with respect to the DWT level of decomposition on all datasets for the spectral DWT multilevel decomposition with 2D-DCT filter (SDWT-2DCT_SVM) approach.
Decomposition LevelIndian PinesPavia UniversitySalinas
Db7Db8Bior1.5Coif2Bior3.7Bior4.4
183.1681.9796.1787.5398.2698.15
285.8685.4297.8696.3798.6598.52
385.4284.8398.4998.4098.6598.64
4 98.51 98.59
Table 13. The parameter tuning for the proposed approach in the Indian Pines dataset.
Table 13. The parameter tuning for the proposed approach in the Indian Pines dataset.
ApproachSDWT
(Function, Level)
2D-DWT
(Function, Level)
Wiener Filter
Patch Size
2D-DCT
Threshold
SDWT-2DWT_SVM(‘db9’, 3)(‘haar’, 5)//
SDWT-WF_SVM(‘sym4’, 4)/41/
SDWT-2DCT_SVM(‘db7’, 2)//1400
Table 14. Classification results (%) for the Indian Pines dataset with standard deviation (in parentheses).
Table 14. Classification results (%) for the Indian Pines dataset with standard deviation (in parentheses).
ClassSVMPCA_SVM3D_SVM3DG_SVMEPFIFRFBM4D_SVMPARAFAC_SVMSDWT-2DWT_SVMSDWT-WF_SVMSDWT-2DCT_SVM
188.7078.2698.7098.7010071.7091.7493.4894.3597.8395.22
272.7156.4476.4287.3185.7174.2583.2577.6482.4890.5692.25
370.1159.4081.5990.9290.7949.7480.7873.9087.9789.5293.81
480.9575.2698.3299.6465.0763.1591.6896.5090.9597.2395.91
592.6988.9896.8997.1096.0180.5195.4892.8794.3396.1696.45
695.2291.7097.2299.5699.4689.6096.6097.0898.0298.8799.17
792.1478.5795.0097.141000087.8687.8694.2992.1493.57
896.7595.7910010010010098.9999.6898.6599.9599.39
978.0066.0010094.001000089.0094.0095.0097.0097.00
1072.9955.0980.9189.9171.3461.4882.2678.4186.4091.8991.09
1156.2446.8275.2383.7291.6994.1968.7470.2872.2785.8284.52
1272.7251.8991.0393.4166.3942.1183.8780.3483.5987.7794.71
1398.7696.8699.1499.7110081.7699.1499.3399.1499.5299.33
1485.9586.4795.7296.2199.3497.8992.4691.7995.5998.5299.06
1571.4760.2495.1099.9378.4779.0988.4392.6992.4898.0898.08
1697.3997.6198.9194.8393.6797.5096.7499.3598.4897.3997.83
κ70.4059.6382.8488.8784.6871.0680.6678.9583.3690.7691.47
(0.79)(1.24)(0.78)(1.18)(1.96)(1.66)(0.64)(1.03)(0.69)0.73(0.74)
OA73.9764.4085.0490.3286.6174.3083.0881.6285.4691.9792.59
(0.71)(1.13)(0.70)(1.04)(1.72)(1.46)(0.56)(0.92)(0.62)0.64(0.65)
AA82.6774.0992.5194.4189.8767.6989.1989.0891.5094.8995.46
(0.97)(1.41)(0.72)(0.70)(1.60)(1.80)(0.87)(1.00)(0.69)0.93(0.55)
Time (s)3.1042.7253.41210.166.801.86351.37297.3940.0447.7432.03
The bold values indicate critical values.
Table 15. The parameter tuning for the proposed approaches in the Pavia University dataset.
Table 15. The parameter tuning for the proposed approaches in the Pavia University dataset.
ApproachSDWT
(Function, Level)
2D-DWT
(Function, Level)
Wiener Filter
Patch Size
2D-DCT
Threshold
SDWT-2DWT_SVM(‘bior1.3’, 4)(‘haar’, 8)//
SDWT-WF_SVM(‘coif2’, 3)/27/
SDWT-2DCT_SVM(‘bior1.3’, 4)//1050
Table 16. Classification results (%) for the Pavia University dataset with standard deviation (in parentheses).
Table 16. Classification results (%) for the Pavia University dataset with standard deviation (in parentheses).
ClassSVMPCA_SVM3D_SVM3DG_SVMEPFIFRFBM4D_SVMPARAFAC_SVMSDWT_2DWT_SVMSDWT_WF_SVMSDWT_2DCT_SVM
171.7172.1396.8998.8898.6876.1880.0777.9096.4296.0097.64
282.9182.7997.6498.9197.2097.9089.4985.3798.9798.6699.43
378.8978.2388.1590.0892.5152.3982.6170.8198.2895.2998.69
491.8991.0099.0199.4267.8184.9994.1593.8597.6797.2597.91
599.7099.7010010099.8998.8599.6699.7899.7799.6799.88
684.3977.9496.6899.1664.3192.7489.7684.8399.1098.4499.29
776.9677.5798.9199.5377.2564.2285.5977.8999.1599.4199.65
869.4471.0794.6496.0884.4053.2475.8874.3096.0789.8998.04
999.9199.8699.6399.7296.1743.6399.8899.9299.9599.9499.94
κ75.2474.2595.8897.8282.7977.6383.0080.9197.9596.3398.51
(1.43)(2.16)(0.30)(0.12)(1.03)(0.25)(1.40)(0.47)(0.31)(0.54)(0.30)
OA81.1880.4996.9598.0786.8183.1487.2383.9998.4997.2998.90
(1.15)(1.75)(0.22)(0.09)(1.68)(0.19)(1.10)(0.40)(0.23)(0.40)(0.22)
AA83.9883.3796.8497.9886.4773.7988.5784.9698.3797.1798.94
(0.69)(1.68)(0.14)(0.12)(1.01)(0.17)(0.79)(0.40)(0.21)(0.36)(0.14)
Time (s)80.0578.68485.92142.4052.8418.141535.131275.06432.03349.64426.28
The bold values indicate critical values.
Table 17. The parameter tuning for the proposed approaches in the Salinas dataset.
Table 17. The parameter tuning for the proposed approaches in the Salinas dataset.
ApproachSDWT
(Function, Level)
2D-DWT
(Function, Level)
Wiener Filter
Patch Size
2D-DCT
Threshold
SDWT-2DWT_SVM(‘sym4’, 4)(‘haar’, 7)//
SDWT-WF_SVM(‘db4’, 4)/47/
SDWT-2DCT_SVM(‘bior3.7’, 2)//850
Table 18. Classification results (%) for the Salinas dataset with standard deviation (in parentheses).
Table 18. Classification results (%) for the Salinas dataset with standard deviation (in parentheses).
ClassSVMPCA_SVM3D_SVM3DG_SVMEPFIFRFBM4D_SVMPARAFAC_SVMSDWT-2DWT_SVMSDWT-WF_SVMSDWT-2DCT_SVM
199.1399.3699.1999.7010098.9999.2593.4899.7299.8199.54
299.5199.5299.3599.6899.9710099.6377.6499.6499.6999.82
399.5899.0997.4898.9596.0799.8499.7773.9099.5099.4799.82
499.3499.3599.3699.4398.4390.0599.3096.5099.2399.6399.30
598.7497.7898.9899.4699.8099.9898.5592.8798.0698.3098.94
699.6899.6799.8699.9599.9910099.6797.0899.6899.6299.59
799.5799.5699.6699.6799.9399.7999.5587.8699.5899.6099.76
867.9872.4086.2788.6785.4099.5378.0099.6891.0593.4198.54
999.0097.5098.4399.1898.7399.9899.4794.0098.9699.8699.56
1095.0194.5193.7995.6093.6099.7296.3678.4196.0998.0197.03
1198.9098.7599.6899.9397.5899.0298.9770.2898.9498.3499.20
1299.8299.7099.9799.8799.6798.8299.8480.3499.9599.9199.91
1399.5899.5599.0298.9799.9397.7099.7299.3399.7599.4699.96
1497.7697.9298.0198.1997.8997.5898.4391.7997.9596.9698.79
1570.0470.5175.9285.5277.6883.0679.7292.6990.8097.0598.50
1698.8698.8898.8398.7099.7110098.4899.3598.7198.5998.79
κ87.0487.8692.0594.4492.1396.0191.0478.9596.1397.4698.91
(0.62)(1.07)(0.41)(0.89)(0.90)(0.27)(0.69)(1.03)(0.61)(0.35)(0.23)
OA88.3689.1092.8895.0292.9596.4191.9681.6296.7197.7299.02
(0.57)(0.97)(0.37)(0.79)(0.81)(0.25)(0.62)(0.92)(0.55)(0.32)(0.21)
AA95.1695.2596.4997.5996.5297.7596.5489.0897.9798.6199.19
(0.20)(0.41)(0.15)(0.36)(0.31)(0.29)(0.21)(1.00)(0.21)(0.16)(0.08)
Time (s)99.1313.18183.51142.5219.294.771854.07297.39228.56273.15169.79
The bold values indicate critical values.

Share and Cite

MDPI and ACS Style

Bazine, R.; Wu, H.; Boukhechba, K. Spectral DWT Multilevel Decomposition with Spatial Filtering Enhancement Preprocessing-Based Approaches for Hyperspectral Imagery Classification. Remote Sens. 2019, 11, 2906. https://doi.org/10.3390/rs11242906

AMA Style

Bazine R, Wu H, Boukhechba K. Spectral DWT Multilevel Decomposition with Spatial Filtering Enhancement Preprocessing-Based Approaches for Hyperspectral Imagery Classification. Remote Sensing. 2019; 11(24):2906. https://doi.org/10.3390/rs11242906

Chicago/Turabian Style

Bazine, Razika, Huayi Wu, and Kamel Boukhechba. 2019. "Spectral DWT Multilevel Decomposition with Spatial Filtering Enhancement Preprocessing-Based Approaches for Hyperspectral Imagery Classification" Remote Sensing 11, no. 24: 2906. https://doi.org/10.3390/rs11242906

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop