Next Article in Journal
Impacts of Urban Green Space on Land Surface Temperature from Urban Block Perspectives
Next Article in Special Issue
A Hybrid Classification of Imbalanced Hyperspectral Images Using ADASYN and Enhanced Deep Subsampled Multi-Grained Cascaded Forest
Previous Article in Journal
Retrieval of Aerosol Optical Properties over Land Using an Optimized Retrieval Algorithm Based on the Directional Polarimetric Camera
Previous Article in Special Issue
A Comparative Analysis of Swarm Intelligence and Evolutionary Algorithms for Feature Selection in SVM-Based Hyperspectral Image Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dimensionality Reduction and Classification of Hyperspectral Remote Sensing Image Feature Extraction

College of Information and Technology, Jilin Agricultural University, Changchun 130118, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(18), 4579; https://doi.org/10.3390/rs14184579
Submission received: 23 August 2022 / Revised: 7 September 2022 / Accepted: 9 September 2022 / Published: 13 September 2022

Abstract

:
Terrain classification is an important research direction in the field of remote sensing. Hyperspectral remote sensing image data contain a large amount of rich ground object information. However, such data have the characteristics of high spatial dimensions of features, strong data correlation, high data redundancy, and long operation time, which lead to difficulty in image data classification. A data dimensionality reduction algorithm can transform the data into low-dimensional data with strong features and then classify the dimensionally reduced data. However, most classification methods cannot effectively extract dimensionality-reduced data features. In this paper, different dimensionality reduction and machine learning supervised classification algorithms are explored to determine a suitable combination method of dimensionality reduction and classification for hyperspectral images. Soft and hard classification methods are adopted to achieve the classification of pixels according to diversity. The results show that the data after dimensionality reduction retain the data features with high overall feature correlation, and the data dimension is drastically reduced. The dimensionality reduction method of unified manifold approximation and projection and the classification method of support vector machine achieve the best terrain classification with 99.57% classification accuracy. High-precision fitting of neural networks for soft classification of hyperspectral images with a model fitting correlation coefficient (R2) of up to 0.979 solves the problem of mixed pixel decomposition.

Graphical Abstract

1. Introduction

Since the 20th century, the widespread application of remote sensing technology reflects various surface information such as that related to agriculture, forestry, water, soil, minerals, energy, and oceans [1,2,3]. Hyperspectral remote sensing imaging promotes the development of precision agriculture [4]. Hyperspectral remote sensing images usually contain hundreds of bands with small wavelength intervals between each band. These images provide a large amount of extremely rich data for the target area [5], as well as facilitate considerable finer ground object information [6]. However, in some specific practical applications, a higher amount of data does not necessarily result in an increased amount of information because excessive information redundancy exists in hyperspectral images [7]. To extract all characteristic terrain data from hyperspectral data, dimensionality reduction must be performed according to data characteristics and band types. Hyperspectral image downscaling methods focus on feature extraction based on the linear and nonlinear, supervised and unsupervised, spatial and spectral features of the images. Hyperspectral images are processed by downscaling and classification. Deep learning has been a powerful feature extraction tool widely used for hyperspectral image classification, and convolutional neural networks (CNN) are capable of extracting nonlinear features [8,9].
Current hyperspectral dimensionality reduction methods are classified into feature extraction [10] and band selection [11]. Hyperspectral feature extraction refers to the recombination and optimization of the original spectral–spatial features to extract new features that are most suitable for the application requirements. Feature extraction is usually performed in terms of feature vectors [12] and feature kernels [13] of hyperspectral data for dimensionality reduction. Dimensionality reduction attempts to retain local and global features. The global feature transforms the entire remote sensing data, and the local feature extracts the part of the data with obvious data features. Many methods of hyperspectral image data classification have been developed; for example, hyperspectral data classification based on deep learning [14], hyperspectral remote sensing image classification using the support vector machine (SVM) [15], semi-supervised learning image classification [16], and dimensionality reduction methods for image classification [17]. The most popular application is deep learning for remote sensing image classification; deep-learning-based hyperspectral image classification methods [18] are based on two main directions. The first is basic spectral classification methods such as SVMs [19,20], k-nearest neighbor (KNN) [21,22], decision trees [23] and other classification algorithms based on spectral features and appropriate feature transformations. The second is to link image contextual features for classification. For example, spatial context and spectral correlation are utilized to enhance hyperspectral image classification [24], and a residual network [25] is employed to obtain features at each level and fuse deep features to the classification of the network [26]. In most dimensionality reduction classification methods, the data features retained by data dimensionality reduction are indistinguishable in classification, producing poor classification results. Determining the most suitable model for dimensionality reduction and classification of hyperspectral remote sensing data is crucial because the resolution of each pixel of hyperspectral remote sensing data is extremely high. Each pixel represents a large real area and mixes multiple terrains. If hard classification alone is too decisive, the model may fail to classify what proportion of terrain a single pixel occupies.
To solve these problems and improve the combination method of data dimensionality reduction and classification, this paper classifies hyperspectral urban terrain data that are subjected to various data dimensionality reduction algorithms. The supervised machine learning-based classification algorithm is used to observe which feature data after dimensionality reduction is sensitive to produce a strong classification effect. In addition, this study evaluates the overall accuracy of the combination of hyperspectral image dimensionality reduction and classification methods. Despite redundant data of hyperspectral remote sensing, the proportion of hyperspectral data features is preserved by different dimensionality reduction methods. The dimensionality reduction data retain the difference of the original features, and the supervised machine learning-based classification method is used to determine which data classification effect is better. The classification process is generally divided into hard and soft classification methods. This study also determines which dimensionality reduction and classification methods for hyperspectral remote sensing data are suitable for soft and hard classification.
The method determined in this study can quickly and accurately determine the terrain information of remote sensing images. Furthermore, the dimensionality of the overall data can be greatly reduced and redundant bands can be filtered, thereby increasing the calculation speed of remote sensing terrain analysis and retaining its maximum characteristics. Using lower-dimensional data to effectively express original high-dimensional features is conducive to the rapid extraction of information while compressing the amount of data. Soft classification and hard classification can be used for distinguishing pixel singularity for hyperspectral remote sensing terrain classification. In summary, this paper has two targets: (1) For the high-dimensional characteristics of hyperspectral data, various dimensionality reduction methods are applied to retain all the data features; then, it is verified whether unified manifold approximation and projection (UMAP) dimensionality reduction and SVM classification can both remove redundant data and maintain good classification accuracy. (2) For large-resolution hyperspectral images leading to hard pixel classification, a neural network model is applied to obtain the best fit of the pixel class decomposition.

2. Hyperspectral Dimensionality Reduction

2.1. Feature Extraction Dimensionality Reduction

Hyperspectral remote sensing images contain spectral information of all bands as well as considerable spatial and band redundancy. Hyperspectral feature extraction refers to the recombination and optimization of the original spectral–spatial features to extract new features that are most suitable for current application requirements. Hyperspectral feature extraction is based on the principle of reducing the data dimension to a lower level to achieve improved feature selection performance.
Based on the above, traditional machine learning methods can be applied for dimensionality reduction of hyperspectral image data because of the high number of spectra they contain and the high dimensionality of the data. Graph learning can effectively reveal the intrinsic relationships of data and is currently widely applied to hyperspectral imaging. For example, a hybrid graph learning method is applied to separate the intra-class compactness and inter-class separability of sample data and obtain the optimal spatial transformation for dimensionality reduction [27]. Graph learning considers only the individual information of each sample in certain features, and discriminant analysis-based hyperspectral image classification for dimensionality reduction [28] adds domain, tangential, and statistical features of samples to achieve feature complementarity and improve the classification performance of hyperspectral images. The difficulty of applying graph embedding methods to dimensionality reduction of hyperspectral images lies in choosing the appropriate neighborhood construction to explore spatial information. Spatial spectral manifold reconstruction preserving embedding [29] utilizes a new spatial and spectral combination distance (SSCD) to fuse spatial structure and spectral information to extract discriminative features. While deep learning convolutional neural networks can reduce the dimensionality of spectral and spatial data, the adoption of autoencoders in convolutional neural networks reduces the dimensionality [30], using pooled hyperspectral images [31].
This paper mainly focuses on the method of feature extraction based on machine learning for hyperspectral images. The existing feature extraction methods mainly include principal component analysis (PCA) [32], linear discriminant analysis (LDA) [33], factor analysis (FA) [34], singular value decomposition (SVD) [35], independent component analysis (ICA) [36], local linear embedding (LLE) [37], t-distributed stochastic neighbor embedding (t-SNE) [38], UMAP [39], and other dimensionality reduction methods. The data features retained by each method are different, and based on these differences, the feature extraction methods can be divided into linear dimensionality reduction and nonlinear dimensionality reduction. Nonlinear dimensionality reduction can be divided into global feature and local feature methods.

2.2. Linear Dimensionality Reduction

PCA is a linear transformation unsupervised dimensionality reduction algorithm. It maintains data information and simplifies the dimension of hyperspectral remote sensing data by transforming data information of all bands into a new coordinate system. In this method, the eigenvalue with the largest variance contribution is selected, which reduces the dimension of the dataset without much classification accuracy loss and achieves faster calculation. In view of the high-dimensional characteristics of hyperspectral data, the PCA algorithm selects multiple eigenvectors for feature extraction. The eigenvector selection rule based on experience is that when the number of principal components is greater than 1, the data with an eigenvalue greater than 1 and a variance ratio greater than 85% are selected.
LDA is a supervised linear dimensionality reduction algorithm; it projects high-dimensional pattern samples into the optimal discriminant vector space to extract classification information and compress the dimension of feature space. After projection, it is ensured that the pattern samples have the largest inter-class distance and the shortest intra-class distance in the new subspace; that is, the pattern has the best separability in this space. Classification aims to achieve the farthest possible distance between different categories and the shortest distance between the same category. Compared with the PCA dimensionality reduction algorithm, LDA is more concerned with classification rather than maintaining data information. After dimensionality reduction, the intra-class variance of the data is the smallest, and the inter-class variance is the largest. The retained dimension is smaller than the classification category. When the number of layers of dimension reduction is three, the test dataset retains 99.9% of the original data features.
SVD dimensionality reduction is another feature extraction method similar to PCA. The method of considering the largest eigenvalue for Eigen decomposition is only suitable for the decomposition of square matrices but not for non-square matrices. Singular values, such as an m × n matrix, construct an m × m square matrix on the left, an n × n square matrix on the right, and an m × n singular matrix with non-zero diagonals in the middle; then, the values of the singular matrix are sorted by size for feature extraction. In the process of singular value decomposition, the singular value is reduced particularly fast, and fewer dimensions can be selected to retain more feature data.
ICA is another type of linear dimensionality reduction algorithm. It is suitable for the dimensionality reduction of non-Gaussian data because the signal obtained after the Gaussian distribution is mixed is Gaussian, and the mixed data show no difference. ICA is suitable for the separation of such mixed signals. For hyperspectral data, the mixed multiple spectra can be separated by ICA, extracting mutually independent attributes and thus reducing the dimension. The purpose of ICA is to extract mutually independent attributes and reduce dimensionality. It considers the data after dimensionality reduction to be a linear combination of several statistically independent components. It only focuses on independence and not the size of the data and the variance between them; therefore, it is more conducive to data separation.

2.3. Nonlinear Dimensionality Reduction

LLE is a nonlinear dimensionality reduction algorithm that preserves the local features of the data, allowing the dimensionality-reduced data to maintain a good manifold structure; hence, it is a manifold learning technology. LLE first measures the degree of linear correlation between each training instance and its nearest neighbors; then, the low-dimensional vector to represent the features of the training set is determined such that the local relationships are best preserved. When the LLE algorithm is used to reduce the dimensionality of the dataset to three-dimensional data, the dimensionality reduction error is 5.5 × 10 18 .
t-SNE is a nonlinear dimensionality reduction method. It usually pays more attention to maintaining similarity such that the distance between similar points in the low-dimensional space is smaller. The t-distribution has the characteristics of long tail; that is, when there are outliers, the entire distribution will not be separated from most of the original data because of the outliers. While t-distribution mapping is used for low-dimensional data, normal-distribution mapping is used for high-dimensional data. The t-SNE algorithm uses Gaussian distribution in a high-dimensional space and t-distribution in a low-dimensional space. The t-distribution satisfies the requirement for t-SNE that the distances between points in the same cluster are similar, whereas those between points in different clusters are large. t-SNE converts the similarity between data points into probabilities.

2.4. UMAP

UMAP is a nonlinear dimensionality reduction algorithm. It is based on the principle of manifold and projection technology to achieve dimensionality reduction. First, the distances between points in a high-dimensional space are calculated and projected to a low-dimensional space. Then, the distances between points in that low-dimensional space are calculated. Subsequently, stochastic gradient descent is used to minimize the difference between these distances. The most prominent feature of the UMAP output is the balance between local and global structure; UMAP tends to preserve the global structure in the final projection. It is used to analyze high-dimensional data of any data type, providing fast running time and high repeatability.
The UMAP dimensionality reduction algorithm mainly involves two steps—learning the manifold structure in the high-dimensional space and determining the low-dimensional representation method of the manifold.
Step 1: For a given original hyperspectral image dataset X = ( X 1 , X 2 , X 3 X n ) , dimension reduction is initialized to obtain a low-dimensional dataset Y = ( Y 1 , Y 2 , Y 3 Y n ) ~ N (0, 10 4 × I n ). Then, p i j and initial q i j are calculated.
The conditional probability of i for j is given by
p i | j = e d ( x i , x j ) σ i
The symmetric formula of the similarity matrix P for X is as follows:
p i j = p i | j + p j | i p i | j · p j | i
The similarity matrix Q of Y is
q i j = [ 1 + a ( y i y j ) 2 b ] 1
where a ≈ 1.93 and b ≈ 0.79 for default UMAP hyperparameters.
p i j measures the similarity between X i and X j ; q i j measures the similarity between Y i and Y j .
Step 2: The cost loss is calculated using binary cross entropy (CE):
C E ( p , q ) = i j [ p i j · l o g p i j q i j + ( 1 p i j ) · log ( 1 p i j 1 q i j ) ]
Step 3: The parameters are optimized, and the number of iterations t, learning speed v, and momentum a are set. The target result is a low-dimensional data representation Y = Y 1 , Y 2 , Y 3 Y n .
Step 4: Optimization is started: Y t = Y t 1 + v · d C d Y + a · ( Y t 1 Y t 2 ) , where d C d Y   is the gradient vector of the loss function with respect to Y; d C d Y = ( C y i ) 1 × n .
In the field of remote sensing, PCA is commonly used for feature dimensionality reduction and machine learning for classification. Considering the diversity of different regions of a superpixel-level PCA for dimensionality reduction [40], a novel collapsed PCA was developed [41], in which the spectral vector is collapsed into a matrix to effectively determine the covariance matrix for better dimensionality reduction of hyperspectral image data. The results of different types of PCA and their linear variants were compared, and it was found that the segmented PCA has better accuracy than the collapsed PCA, which has a lower spatial complexity [42]. Because PCA projection looks for the direction that maximizes the variance, it usually ignores the local structure and the variance in other directions. UMAP in machine learning, on the other hand, differs from PCA in that not only local features but also global features are preserved when clustering.
Remote sensing data contains hundreds of bands, and the combination of each band constitutes the pixel point category of image data. The input image is categorized according to the type of each pixel point, and the UMAP algorithm is applied to reduce the dimensionality of hyperspectral linear data. UMAP maps the original band data to the low-dimensional space, for which it first learns the popular structure of hyperspectral nonlinearity to determine the low-dimensional representation of the stream shape. The optimal low-dimensional data are found by utilizing the least cost function, which results in dimensionality reduction. Unlike other dimensionality reduction algorithms, such as PCA and t-SNE, which only consider the local structure and lose many features, UMAP considers the global structure.

3. Hyperspectral Image Classification Methods

The current difficulties to be overcome by feature extraction methods in the field of remote sensing image research are the large spatial feature span of hyperspectral images and the limited number of samples. CNNs are widely applied for hyperspectral image classification [43], and the insufficient number of samples of hyperspectral images is a bottleneck that limits the deep learning training, for example, utilizing knowledge migration [44] and domain generalization [45] approaches to address the characteristics of the insufficient number of samples. Hyperspectral image classification using deep neural networks with inadequate samples and feature learning [46]. A new method was proposed for hyperspectral image classification based on multi-view deep neural networks, which fuses spectral and spatial features, employing only a small number of labeled samples [47]. Multi-level discontinuous features are extracted from remote sensing images [48]. Moreover, hyperspectral images are classified using a regularized subspace of Manhattan distances [49] by introducing a deep hybrid multi-view neural network that implements information interaction, making excellent utilization of different graph filters [50].
In the face of hyperspectral image classification, both hard classification and soft classification are applied. Hard classification mainly involves deep dimensionality reduction of hyperspectral images first, followed by supervised classification by machine learning. Soft classification is the probability obtained from image classification using logistic regression and neural networks and represents the decomposition of hyperspectral image elements.

3.1. Hard and Soft Classification for Hyperspectral Images

Hard classification and soft classification are common strategies for hyperspectral remote sensing image classification. Figure 1 and Figure 2 show the flowcharts of hard and soft classification, respectively. Hard classification assigns each pixel in a remote sensing image a single category, and the classification is based on the similarity of pixel features, spectral features, texture features, or a mixture of multiple features with known statistical features of each category.
In soft classification, the surface area corresponding to the pixel is often composed of multiple categories of ground objects according to the actual situation, therefore it is assumed that each pixel belongs to multiple categories or is composed of multiple categories. Then, the relationship between the pixel and each category is calculated according to a specific algorithm. The classification output is the probability fuzzy classification that the pixel belongs to each category or the proportion of each category of objects in the pixel.
Soft classification and hard classification have their own advantages in the classification of hyperspectral remote sensing images. Hard classification focuses only on the maximum value of the corresponding classification probability of each pixel, ignoring the relationship between other surrounding pixels and the corresponding relationship. Soft classification can divide a single pixel into the proportion of target objects. The hard classification method is absolute; therefore, it determines the category of an object according to the category of its adjacent objects.
The dimensionality reduction operation is performed on the basis of the distribution characteristics of hyperspectral data. In this study, the data are obtained from public datasets with data labels. The terrain classification methods of hyperspectral images mainly include supervised machine learning, KNN, Bayesian classification, logistic regression, neural network, SVM, and decision tree classification methods. Bayesian and logistic regression usually have better classification effect on linear features. The KNN algorithm determines the object category according to the distance of adjacent nearest points and provides a high classification speed and good prediction effect. The decision tree algorithm can process multiple continuous fields and can effectively classify according to the intensity of hyperspectral bands. The SVM can solve the nonlinear problem of hyperspectral light intensity, providing relatively strong generalization ability and handling high-dimensional data effectively. The neural network can be fitted by multi-dimensional parameters and has strong learning ability and high classification accuracy.
Soft classification of hyperspectral images is a probabilistic model in which the reduced-dimensional data features are classified by a logistic regression or neural network model, yielding a probability that a single pixel on the image can be simultaneously classified into multiple categories; this yields a probability that the pixel is classified into each category, which can be decomposed by the probability for image elements. The hard classification of hyperspectral images is a non-probabilistic model, i.e., the results are derived by a decision classification function, and the exact classification of each pixel point is learned by a supervised classifier of machine learning for image distribution features.

3.2. Neural Networks

Feedforward neural networks are also known as multi-layer perceptrons, in which different neurons belong to different layers, namely an input layer, a hidden layer, and an output layer, which are fully connected to each other. Figure 3 illustrates a neural network structure. The neural network contains functions such as activation and loss functions. For the classification of hyperspectral images, the activation function is selected as tanh, the Adam gradient descent function is used, the number of nodes in the hidden layer is selected as 50, and the number of iterations is selected as 5000 for training. The neuron is directly connected to the following formula:
y i = j w i j x j + b
where W is the weight and b is bias.
The tanh activation function is obtained as
tanh ( x ) = e x e x e x + e x

3.3. Support Vector Machine

SVM is one of the classification algorithms in supervised learning, whose core idea is to determine a divided hyperplane that separates sample points in space; the plane with the largest distance to the nearest point among the samples of different categories with the best generalization ability is selected as the divided hyperplane.
The point-to-plane distance γ is obtained as
γ = | w T x i + b | | | w | |
where w is the weight and b is the bias.
The point with the smallest sample interval is found to have the largest distance to the plane.
The nonlinear data of hyperspectral images are not in the same low-dimensional space and need to be mapped to the high-dimensional space by applying kernel functions so that linearly indistinguishable sample points can be made linearly distinguishable.

4. Experimental Results and Analysis

4.1. Hyperspectral Image Description

The dataset used in this study is a public dataset comprising hyperspectral images of a city obtained by the HYDICE sensor. The image size is 307 × 307 pixels. The urban dataset originally has a total of 210 bands; after removing noise and water absorption bands, 162 bands remain for subsequent dimensionality reduction processing and analysis. There are four types of ground objects: asphalt, roof, grass, and tree. The training set and test set are in the ratio of 8:2. The test set is (75,399, 18,850). The number of pixels in each test set and training set is shown in Table 1. An original hyperspectral remote sensing image and its label image showing different terrains are shown in Figure 4.
Hyperspectral images have high resolution and rich spectral information; however, the band interval between each spectrum is small, a nonlinear relationship exists between each spectrum, and the large amount of data results in abundant redundant data, making identification difficult. Through dimensionality reduction of hyperspectral remote sensing images, data that retains the overall feature correlation according to the distribution characteristics of the data can be extracted, thereby drastically reducing the data dimension.

4.2. Results and Analysis

4.2.1. PCA

PCA is applied to the urban hyperspectral image data, with 162 bands and features, to reduce the dimension, and the ratio of the features is sorted from the largest to the smallest. The data dimension retained by PCA dimensionality-reduced data and the variance ratio of the occupied features are shown in Figure 5.
The figure indicates that the first and second vectors account for more than 60% and 30% of the total proportion. A PCA dimensionality reduction dimension of the largest three features accounts for 98% of the total variance ratio. For each additional dimension, the proportion of features occupied by data increases slowly. When the classification dimension is four, the proportion of features reaches 99%. Thus, the purpose of reducing the amount of data and preserving the overall characteristics is met.

4.2.2. KNN Proximity Classification

The choice of the k value of the KNN will have a considerable effect on the algorithm classification. When k = 1, the test instance is related to the closest sample, and the training error is small. By contrast, when the test sample has noise, the test error will be large. While a small k value results in overfitting, when the k value is large, it is equivalent to training with a large range of data, and the test result is the class with the most instances in the range, which will cause underfitting. Thus, selecting an appropriate k value is important. As the test sample, we consider the overall data of the test set without dimensionality reduction. The k value and the segmentation accuracy are selected as shown in Table 2.
When k = 4, the feature ratio reaches 99%, and the overall classification accuracy is somewhat improved compared with that when k = 3. When the k value increases, the calculation amount increases and the classification accuracy decreases. In the KNN proximity algorithm, four nearby points are selected as the criteria for their classification.

4.2.3. Gaussian Maximum Likelihood Classifier

The KNN proximity algorithm only considers the distance from the sample to be classified to the center of each sample, ignoring the overall distribution of the sample. By contrast, the Gaussian maximum likelihood classifier additionally considers the distribution characteristics of known classes as well. The classifier is run on the hyperspectral image dataset to train the reduced principal components and classify them according to the training data. The classified images are shown in Figure 6. The classification accuracy of each terrain is shown in Table 3.
As the figure shows, in the classification result of the Gaussian maximum likelihood classifier after dimensionality reduction, the black pixels are points that are classified incorrectly. Overall, a reasonable classification effect is achieved. The table indicates that the best classification accuracy of Gaussian maximum likelihood estimation for each terrain is 91.39%. However, the accuracy required for terrain segmentation of hyperspectral remote sensing data is insufficient, and the Gaussian maximum likelihood classifier cannot distinguish the spectral features retained by PCA dimensionality reduction well.

4.2.4. Dimensionality Reduction Method Combined with Classification

Next, we determine an effective classification algorithm to identify dimensionality-reduced data with multiple features retained by different dimensionality reduction methods. After dimensionality reduction, the data features include linear features, local features, manifold features, and non-popular features. In addition, the classifiers are divided into linear and nonlinear classifiers. A suitable method for dimensionality reduction and classification of hyperspectral images can be determined through experiments. The classification results of various dimensionality reduction and classification methods are shown in Table 4, and the classification accuracy is shown in Figure 7.
We analyzed the adaptation of different classification algorithms to the features retained by the dimensionality reduction methods. For the KNN proximity algorithm, the data features retained by UMAP dimensionality reduction can be well captured. Comparison of the classification accuracy of FA and SVD dimensionality reduction with the classification accuracy of logistic regression indicates that they are comparable to unreduced dimensionality; therefore, it can be concluded that the dimensionality reduction not only reduces the amount of data but also achieves good classification accuracy. The Naive Bayes algorithm and SVD have the best classification accuracy of hyperspectral data compared with other dimensionality reduction methods. The classification accuracy of SVM and UMAP dimensionality reduction is the best among all dimensionality reduction algorithms. The multi-layer perceptron algorithm is not sensitive to LLE dimensionality reduction. The combination of multi-spectral classification with SVD dimensionality reduction has the best effect compared with other combination methods. The combination method of decision tree classification and UMAP dimensionality reduction shows superior results. The accuracy of classification with the dimensionality reduction algorithm and that without dimensionality reduction were statistically analyzed, and the p-value was less than 0.05, indicating a significant difference.

4.2.5. Accuracy of Various Dimensionality Reduction Methods for Each Terrain Classification

The accuracy of each terrain classification obtained by the dimensionality reduction method and the classification method is shown in Figure 8.
Clearly, the Naive Bayes data classification effect on dimensionality reduction is not as good as that of other classification algorithms. Compared with the classification accuracy of algorithms without dimensionality reduction, the dimensionality reduction algorithms achieve the expected goal of not only reducing the amount of calculation but also preserving the original classification effect.
The dimensionality reduction mode of UMAP preserves the integrity of the data features. In the case of the Naive Bayes algorithm, the classification accuracy based on t-SNE dimensionality reduction for roof terrain is 0, because the remote sensing spectra are nonlinear, independent, and discrete. According to the results, the Bayesian classifier is not highly effective for handling the nonlinear problem of hyperspectral data. The comparison chart of terrain classification accuracy by UMAP dimensionality reduction is shown Figure 9.
The nonlinear local features retained by UMAP are not sensitive to the classification effect of linear logistic regression for some linear classifiers such as the Bayesian classifier. However, the classification accuracy obtained by UMAP dimensionality reduction with other classifiers can reach more than 99%. UMAP has the ability to infer local and global structures while maintaining the relative global distance in the low-dimensional space. UMAP not only has high classification accuracy but also reduces the running time. With its high accuracy, UMAP provides favorable technical support for the classification of hyperspectral remote sensing terrains.

4.3. Soft Classification of Hyperspectral Images

The classification of hyperspectral remote sensing images is restricted by its spatial resolution. The spatial resolution refers to the size of the smallest target object that the sensor can distinguish; it is the ground range corresponding to one pixel in the actual satellite observation image. Each pixel of the image represents the real area of land, and the image will face various terrains on its pixels. Most of the classification prioritizes the type of land, ignoring the texture and structure of the remote sensing image feature. The topography of pixel elements is decomposed into probabilities for soft classification of hyperspectral images.
The following two methods are used to fit the correct classification model. First, the soft classification model of pixels is trained according to the overall spectral data, and then the multiple logistic regression and neural network models will be fitted. One pixel of the image contains the range of the square of the real object, and the side length of the square depends on the image resolution at the time of capturing the image. For a pixel mixed with multiple terrains, the proportion of each type of terrain in a single pixel is obtained. The topographic ratio of the pixel classification determines the local image classification.
Using a linear logistic regression function with hyperspectral remote sensing data as input, the targets are classified into terrain types. Multi-variate logistic regression is performed, and the output is the probability ratio of each class. Logistic regression and neural network training are performed for 162 bands of remote sensing data as input parameters. In the original dataset, each pixel is divided into proportions of target objects, and the error calculation and fitting are performed between the probability result of logistic regression and the true proportion. The fitting results and errors obtained are shown in Table 5.
Table 5 indicates that the fitting degree of the neural network is smaller than that of the linear logistic regression. The multi-layer neurons of the neural network can fit the nonlinear data characteristics between the spectra well, and the pixel classification of the hyperspectral image can obtain the proportion of terrain categories through the soft classification probability of the output of the neural network. The spatial characteristics of hyperspectral images can be enhanced according to the adjacent pixel types, thus providing support for efficient classification accuracy improvement in hyperspectral images.
The statistical significance test comparing the classification results of different methods also suggests that linear regression downscaling is not relevant for hyperspectral downscaling analysis, and the linear regression fit is insufficient because of the nonlinearity of the spectra. However, the neural network has the best fit and correlation effect.

5. Discussion

The UMAP dimensionality reduction and SVM classification methods for hyperspectral images can retain and extract the main features. Statistical significance tests are performed for the results. The computational effort of UMAP dimensionality reduction and SVM classification of hyperspectral images is low and their accuracy of classification is improved. Considering the high pixel resolution of hyperspectral images and the complexity of image elements, the pixel decomposition of the image and the fitting of the neural network model can solve the problem of excessive classification errors of image pixels due to a high image resolution.
The nearest regularized subspace (NRS) classification method is applied to determine the residuals between the approximation and the corresponding pixels, and the Tikhonov matrix of each class and the pixels to be classified into the corresponding classes. In determining the best distance metric [51], this method achieves a maximum accuracy of 96% but fails to reduce the number of parameters computed for hyperspectral images. When the combined hyperspectral image and LiDAR data are used, LiDAR data help to better characterize the elevation information of the same measurement area, thereby improving the classification performance [52]. The fusion of CNN networks and LiDAR data to fuse feature- and decision-level classification achieved good accuracy in the test set [53]. Thus, this fusion strategy can be applied for pixel classification of hyperspectral images. Hyperspectral dimensionality reduction classification using QPCA [54] and deep learning networks combined with meta-learning, as well as population intelligence and evolutionary algorithms (SIEA), can be used to solve the feature selection problem for hyperspectral images [55]. Although all the above-mentioned methods preserve the main features of hyperspectral data, they are inferior to UMAP, which preserves the global features of the images.
Here, we selected supervised classification methods that are highly dependent on the quality of the sample labels of the dataset and require a small sample size and training feature classification for fewer samples. All the dimensionality reduction methods involve feature transformation and extraction of data by ignoring the connection between spectra. Deep learning networks can help both dimensionality reduction and image classification of hyperspectral images. In our future work, we will consider combining deep learning for hyperspectral information classification.

6. Conclusions

In this study, the combination of feature extraction dimension reduction and classification of hyperspectral remote sensing images was investigated. The analysis results indicate that the UMAP data dimension reduction algorithm and the SVM classification algorithm achieve superior performance in the terrain classification of hyperspectral images. A classification accuracy of 99.57% could be achieved. The SVM algorithm effectively captures the classification features after dimensionality reduction, and it not only reduces the amount of calculation but also improves the classification accuracy. For the single-pixel classification problem of hyperspectral images, it is advisable to not rely solely on the maximum probability. The neural network fitting classification model was found to be the most effective for fitting the classification probability of pixels according to the spectral features; it achieved a fitting correlation coefficient (R2) of 0.979. It can provide a method for solving the single-pixel classification problem. To prepare for the classification and recognition of hyperspectral data and ground objects, the topographic and edge features of ground objects can be identified on the basis of the results of soft classification, which improves identification and classification accuracy. This study emphasizes that deep learning neural networks for image classification are expected to be widely applied to hyperspectral images.

Author Contributions

Conceptualization, H.L. and L.C.; methodology, L.C. and X.Z.; software, H.L. and J.C.; validation, L.C., Y.H. and H.L.; writing—original draft preparation, H.L.; writing—review and editing, L.C. and J.C. All authors have read and agreed to the published version of the manuscript.

Funding

This paper was supported by the National Natural Science Foundation of China (Grant No. U19A2061), the 2021 Jilin Provincial Budget Construction Fund (Innovation Capacity Development) (Grant No. 2021C044-4), and the 2021 Science and Technology Research Project of Jilin Provincial Department of Education (Grant No. JJKH20210337KJ).

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no competing interests.

References

  1. Huang, Q.Q.; Bao-Qing, H.U.; Luo, C. Research into Remote Sensing Technology Application in Geological Hazard Analysis. J. Guangxi Teach. Educ. Univ. (Nat. Sci. Ed.) 2016, 33, 130–134. [Google Scholar]
  2. Jian, W.U. Advances in researches on hyperspectral remote sensing forestry information-extracting technology. Spectrosc. Spectr. Anal. 2011, 31, 2305. [Google Scholar]
  3. Ge, Y.; Thomasson, J.A.; Sui, R. Remote sensing of soil properties in precision agriculture: A review. Front. Earth Sci. 2011, 5, 229–238. [Google Scholar] [CrossRef]
  4. Lu, B.; Dao, P.D.; Liu, J.; He, Y.; Shang, J. Recent advances of hyperspectral imaging technology and applications in agriculture. Remote Sens. 2020, 12, 2659. [Google Scholar] [CrossRef]
  5. Bioucas-Dias, J.M.; Plaza, A.; Camps-Valls, G.; Scheunders, P.; Nasrabadi, N.; Chanussot, J. Hyperspectral Remote Sensing Data Analysis and Future Challenges. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–36. [Google Scholar] [CrossRef]
  6. Zhang, L.; Zhang, L.; Tao, D.; Huang, X. On Combining Multiple Features for Hyperspectral Remote Sensing Image Classification. IEEE Trans. Geosci. Remote Sens. 2012, 50, 879–893. [Google Scholar] [CrossRef]
  7. Haut, J.M.; Paoletti, M.E.; Plaza, J.; Plaza, A. Fast dimensionality reduction and classification of hyperspectral images with extreme learning machines. J. Real-Time Image Process. 2018, 15, 439–462. [Google Scholar] [CrossRef]
  8. Mou, L.; Lu, X.; Li, X.; Zhu, X.X. Nonlocal graph convolutional networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 8246–8257. [Google Scholar] [CrossRef]
  9. Hang, R.; Li, Z.; Liu, Q.; Ghamisi, P.; Bhattacharyya, S.S. Hyperspectral image classification with attention-aided CNNs. IEEE Trans. Geosci. Remote Sens. 2020, 59, 2281–2293. [Google Scholar] [CrossRef]
  10. Borges, J.S.; Marcal, A. Evaluation of feature extraction and reduction methods for hyperspectral images. New Dev. Chall. Remote Sens. 2007, 29, 265–274. [Google Scholar]
  11. Yuan, Y.; Zhu, G.; Wang, Q. Hyperspectral Band Selection by Multitask Sparsity Pursuit. IEEE Trans. Geosci. Remote Sens. 2014, 53, 631–644. [Google Scholar] [CrossRef]
  12. Fauvel, M.; Chanussot, J.; Benediktsson, J.A. Kernel Principal Component Analysis for the Classification of Hyperspectral Remote Sensing Data over Urban Areas. EURASIP J. Adv. Signal Process. 2009, 2009, 783194. [Google Scholar] [CrossRef]
  13. Kuo, B.C.; Ho, H.H.; Li, C.H.; Hung, C.C. A Kernel-Based Feature Selection Method for SVM with RBF Kernel for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2013, 7, 317–326. [Google Scholar] [CrossRef]
  14. Chen, Y.; Lin, Z.; Xing, Z.; Gang, W.; Gu, Y. Deep Learning-Based Classification of Hyperspectral Data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 7, 2094–2107. [Google Scholar] [CrossRef]
  15. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  16. Camps-Valls, G.; Marsheva, T.; Zhou, D. Semi-Supervised Graph-Based Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3044–3054. [Google Scholar] [CrossRef]
  17. Damodaran, B.B.; Nidamanuri, R.R. Assessment of the impact of dimensionality reduction methods on information classes and classifiers for hyperspectral image classification by multiple classifier system. Adv. Space Res. 2014, 53, 1720–1734. [Google Scholar] [CrossRef]
  18. Li, S.; Song, W.; Fang, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Deep learning for hyperspectral image classification: An overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef]
  19. Okwuashi, O.; Ndehedehe, C.E. Deep support vector machine for hyperspectral image classification. Pattern Recognit. 2020, 103, 107298. [Google Scholar] [CrossRef]
  20. Jain, D.K.; Dubey, S.B.; Choubey, R.K.; Sinhal, A.; Arjaria, S.K.; Jain, A.; Wang, H. An approach for hyperspectral image classification by optimizing SVM using self organizing map. J. Comput. Sci. 2018, 25, 252–259. [Google Scholar] [CrossRef]
  21. Tu, B.; Wang, J.; Kang, X.; Zhang, G.; Ou, X.; Guo, L. KNN-based representation of superpixels for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 4032–4047. [Google Scholar] [CrossRef]
  22. Bo, C.; Lu, H.; Wang, D. Spectral-spatial K-Nearest Neighbor approach for hyperspectral image classification. Multimed. Tools Appl. 2018, 77, 10419–10436. [Google Scholar] [CrossRef]
  23. Xu, S.; Liu, S.; Wang, H.; Chen, W.; Zhang, F.; Xiao, Z. A hyperspectral image classification approach based on feature fusion and multi-layered gradient boosting decision trees. Entropy 2020, 23, 20. [Google Scholar] [CrossRef] [PubMed]
  24. Yang, X.; Ye, Y.; Li, X.; Lau, R.Y.; Zhang, X.; Huang, X. Hyperspectral image classification with deep learning models. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5408–5423. [Google Scholar] [CrossRef]
  25. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 30 June 2016; pp. 770–778. [Google Scholar]
  26. Song, W.; Li, S.; Fang, L.; Lu, T. Hyperspectral image classification with deep feature fusion network. IEEE Trans. Geosci. Remote Sens. 2018, 56, 3173–3184. [Google Scholar] [CrossRef]
  27. Luo, F.; Zhang, L.; Du, B.; Zhang, L. Dimensionality reduction with enhanced hybrid-graph discriminant learning for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 5336–5353. [Google Scholar] [CrossRef]
  28. Li, W.; Feng, F.; Li, H.; Du, Q. Discriminant analysis-based dimension reduction for hyperspectral image classification: A survey of the most recent advances and an experimental comparison of different techniques. IEEE Geosci. Remote Sens. Mag. 2018, 6, 15–34. [Google Scholar] [CrossRef]
  29. Huang, H.; Shi, G.; He, H.; Duan, Y.; Luo, F. Dimensionality reduction of hyperspectral imagery based on spatial–spectral manifold learning. IEEE Trans. Cybern. 2019, 50, 2604–2616. [Google Scholar] [CrossRef]
  30. Ramamurthy, M.; Robinson, Y.H.; Vimal, S.; Suresh, A. Auto encoder based dimensionality reduction and classification using convolutional neural networks for hyperspectral images. Microprocess. Microsyst. 2020, 79, 103280. [Google Scholar] [CrossRef]
  31. Paul, A.; Chaki, N. Dimensionality reduction of hyperspectral images using pooling. Pattern Recognit. Image Anal. 2019, 29, 72–78. [Google Scholar] [CrossRef]
  32. Shlens, J. A tutorial on principal component analysis: Derivation, discussion and singular value decomposition. arXiv 2003, arXiv:1404.1100. [Google Scholar]
  33. Blei, D.M.; Ng, A.Y.; Jordan, M.I. Latent Dirichlet Allocation. J. Mach. Learn. Res. 2003, 3, 993–1022. [Google Scholar]
  34. Mooi, E.; Sarstedt, M. Factor Analysis. In A Concise Guide to Market Research: The Process, Data, and Methods Using IBM SPSS Statistics; Springer: Berlin/Heidelberg, Germany, 2011; pp. 201–236. [Google Scholar]
  35. Golub, G.H. Singular value decomposition and least squares solutions. Numer. Math. 1970, 14, 403–420. [Google Scholar] [CrossRef]
  36. Prasad, P.S. Independent Component Analysis; Cambridge University Press: Cambridge, UK, 2001. [Google Scholar]
  37. Roweis, S.; Saul, L. Nonlinear Dimensionality Reduction by Locally Linear Embedding. Science 2000, 290, 2323–2326. [Google Scholar] [CrossRef]
  38. Hinton, P.G. Visualizing Data using t-SNE Laurens van der Maaten MICC-IKAT. J. Mach. Learn. Res. 2017, 9, 2579–2605. [Google Scholar]
  39. McInnes, L.; Healy, J.; Melville, J. Umap: Uniform manifold approximation and projection for dimension reduction. arXiv 2018, arXiv:1802.03426. [Google Scholar]
  40. Jiang, J.; Ma, J.; Chen, C.; Wang, Z.; Cai, Z.; Wang, L. SuperPCA: A superpixelwise PCA approach for unsupervised feature extraction of hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4581–4593. [Google Scholar] [CrossRef] [Green Version]
  41. Zabalza, J.; Ren, J.; Yang, M.; Zhang, Y.; Wang, J.; Marshall, S.; Han, J. Novel folded-PCA for improved feature extraction and data reduction with hyperspectral imaging and SAR in remote sensing. ISPRS J. Photogramm. Remote Sens. 2014, 93, 112–122. [Google Scholar] [CrossRef]
  42. Uddin, M.; Mamun, M.; Hossain, M. Feature extraction for hyperspectral image classification. In Proceedings of the 2017 IEEE Region 10 Humanitarian Technology Conference (R10-HTC), Dhaka, Bangladesh, 21–23 December 2017; IEEE: New York, NY, USA, 2017; pp. 379–382. [Google Scholar]
  43. Ahmad, M.; Shabbir, S.; Raza, R.A.; Mazzara, M.; Distefano, S.; Khan, A.M. Hyperspectral image classification: Artifacts of dimension reduction on hybrid CNN. arXiv 2021, arXiv:2101.10532. [Google Scholar]
  44. Zhao, X.; Liang, Y.; Guo, A.J.; Zhu, F. Classification of small-scale hyperspectral images with multi-source deep transfer learning. Remote Sens. Lett. 2020, 11, 303–312. [Google Scholar] [CrossRef]
  45. Peng, J.; Sun, W.; Ma, L.; Du, Q. Discriminative transfer joint matching for domain adaptation in hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2019, 16, 972–976. [Google Scholar] [CrossRef]
  46. Wambugu, N.; Chen, Y.; Xiao, Z.; Tan, K.; Wei, M.; Liu, X.; Li, J. Hyperspectral image classification on insufficient-sample and feature learning using deep neural networks: A review. Int. J. Appl. Earth Obs. Geoinf. 2021, 105, 102603. [Google Scholar] [CrossRef]
  47. Sellami, A.; Tabbone, S. Deep neural networks-based relevant latent representation learning for hyperspectral image classification. Pattern Recognit. 2022, 121, 108224. [Google Scholar] [CrossRef]
  48. Zhou, B.; Duan, X.; Ye, D.; Wei, W.; Damaševičius, R. Multi-Level Features Extraction for Discontinuous Target Tracking in Remote Sensing Image Monitoring. Sensors 2019, 19, 4855. [Google Scholar] [CrossRef]
  49. Khan, S.S.; Ran, Q.; Khan, M.; Zhang, M. Hyperspectral image classification using nearest regularized subspace with Manhattan distance. J. Appl. Remote Sens. 2019, 14, 1. [Google Scholar] [CrossRef]
  50. Yao, D.; Zhi-Li, Z.; Xiao-Feng, Z.; Wei, C.; Fang, H.; Yao-Ming, C.; Cai, W.-W. Deep hybrid: Multi-graph neural network collaboration for hyperspectral image classification. Def. Technol. 2022. [Google Scholar] [CrossRef]
  51. Khan, S.S.; Khan, M.; Haider, S.; Damaševičius, R. Hyperspectral image classification using NRS with different distance measurement techniques. Multimedia Tools Appl. 2022, 81, 24869–24885. [Google Scholar] [CrossRef]
  52. Roy, S.K.; Deria, A.; Hong, D.; Ahmad, M.; Plaza, A.; Chanussot, J. Hyperspectral and LiDAR Data Classification Using Joint CNNs and Morphological Feature Learning. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar]
  53. Hang, R.; Li, Z.; Ghamisi, P.; Hong, D.; Xia, G.; Liu, Q. Classification of hyperspectral and LiDAR data using coupled CNNs. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4939–4950. [Google Scholar] [CrossRef] [Green Version]
  54. L-Alimi, D.A.; Al-qaness, M.A.; Cai, Z.; Dahou, A.; Shao, Y.; Issaka, S. Meta-Learner Hybrid Models to Classify Hyperspectral Images. Remote Sens. 2022, 14, 1038. [Google Scholar] [CrossRef]
  55. Shang, Y.; Zheng, X.; Li, J.; Liu, D.; Wang, P. A Comparative Analysis of Swarm Intelligence and Evolutionary Algorithms for Feature Selection in SVM-Based Hyperspectral Image Classification. Remote Sens. 2022, 14, 3019. [Google Scholar] [CrossRef]
Figure 1. Flowchart of hard classification.
Figure 1. Flowchart of hard classification.
Remotesensing 14 04579 g001
Figure 2. Flowchart of soft classification.
Figure 2. Flowchart of soft classification.
Remotesensing 14 04579 g002
Figure 3. Neural network structure diagram.
Figure 3. Neural network structure diagram.
Remotesensing 14 04579 g003
Figure 4. (a) Hyperspectral remote sensing image; (b) label image showing different terrains.
Figure 4. (a) Hyperspectral remote sensing image; (b) label image showing different terrains.
Remotesensing 14 04579 g004
Figure 5. PCA cumulative explained variance ratio.
Figure 5. PCA cumulative explained variance ratio.
Remotesensing 14 04579 g005
Figure 6. (a) Classified image of Gaussian maximum likelihood; (b) Result of comparison with the label image.
Figure 6. (a) Classified image of Gaussian maximum likelihood; (b) Result of comparison with the label image.
Remotesensing 14 04579 g006
Figure 7. Comparison chart of overall classification accuracy of different dimensionality reduction methods.
Figure 7. Comparison chart of overall classification accuracy of different dimensionality reduction methods.
Remotesensing 14 04579 g007
Figure 8. Comparison chart of terrain classification accuracy by different dimensionality reduction methods.
Figure 8. Comparison chart of terrain classification accuracy by different dimensionality reduction methods.
Remotesensing 14 04579 g008
Figure 9. Comparison chart of terrain classification accuracy by UMAP dimensionality reduction.
Figure 9. Comparison chart of terrain classification accuracy by UMAP dimensionality reduction.
Remotesensing 14 04579 g009
Table 1. Train and test sets of the dataset.
Table 1. Train and test sets of the dataset.
AsphaltGrassTreeRoof
Number29,95432,32824,8057162
Train set23,96325,91019,7665760
Test set5991641850391402
Table 2. Selection of KNN classification k.
Table 2. Selection of KNN classification k.
KNNk = 3 k = 4 k = 5
AA0.95870.97540.9751
OA0.95640.97700.9762
KAPPA0.93820.96750.9664
RECALL0.96000.97550.9752
F1-SCORE0.95930.97540.9751
Table 3. Accuracy of Gaussian Maximum Likelihood Classification.
Table 3. Accuracy of Gaussian Maximum Likelihood Classification.
AsphaltGrassTreeRoof
PA0.85020.90290.89680.9139
OA0.92340.88140.91370.7108
F1-Score0.88530.89200.90510.7996
Table 4. Classification accuracy of hyperspectral images with different dimensionality reduction classification methods.
Table 4. Classification accuracy of hyperspectral images with different dimensionality reduction classification methods.
None Dimensionality ReductionPCALDALLET-SNESVDICAFAUMAP
k-Nearest NeighborKappa0.96120.96740.91260.94600.93220.96590.94010.94790.9938
Recall0.96900.97550.93140.96420.94680.97450.96170.96080.9957
AA0.97360.97540.92970.96310.94990.97500.96040.95930.9987
F1-score0.97120.97540.93060.96360.94830.97470.96100.96000.9957
OA0.97290.97700.93820.96180.95210.97590.95770.96310.9956
P 0.14452 × 10−50.0440.00050.24710.02050.01144 × 10−6
Naive Bayesian ClassifierKappa0.52690.82000.81810.52370.49310.81900.47510.73920.2816
Recall0.61280.87020.88590.63460.52710.86920.63460.78430.3961
AA0.63580.84410.82530.69230.49010.84360.66610.79020.3972
F1-score0.60700.85510.84100.64050.50350.85440.59980.78330.3369
OA0.66170.87210.86920.67080.64940.97140.62850.81650.5167
P 9 × 10−61 × 10−60.54110.07676 × 10−50.84660.00010.0011
Support Vector MachineKappa0.97900.97960.92280.48500.91270.98190.78030.97630.9937
Recall0.98520.98550.93710.50920.93120.98620.74920.98390.9958
AA0.98450.98320.93750.75520.93280.98550.89080.98220.9957
F1-score0.98480.98430.93730.45420.93190.98590.77950.98310.9957
OA0.98510.98560.94550.65440.93840.98720.84850.98330.9955
P 0.87801 × 10−60.00012 × 10−60.29530.00020.32261 × 10−5
Decision TreeKappa0.93740.95620.88550.93000.91160.94830.92080.94880.9890
Recall0.95400.96820.90750.95460.93460.96270.94770.93160.9923
AA0.95200.96560.90490.95040.93310.95960.94640.95890.9927
F1-score0.95290.96690.90620.95250.93380.96120.94690.96030.9925
OA0.95580.96910.91910.95060.93760.96350.94410.96390.9922
P 0.00659 × 10−50.62850.00780.08060.16880.74302 × 10−6
Logistic RegressionKappa0.94910.94510.91670.44800.46630.94590.54270.94680.5121
Recall0.95970.95690.93260.48040.54160.95690.53720.95710.5308
AA0.95880.95760.93250.62620.57540.95740.58750.95820.4935
F1-score0.95920.95720.93260.39460.54830.95720.49790.95760.5106
OA0.96410.96130.94120.63080.62830.96180.69230.96240.6650
P 0.50740.00042 × 10−53 × 10−70.53533 × 10−60.63501 × 10−8
Multi-layer PerceptronKappa0.97010.96660.91970.82740.89240.97840.91540.97490.9748
Recall0.98140.97720.93620.80420.91420.98540.92810.98320.9786
AA0.97510.97430.93300.90020.91940.98350.94410.98190.9809
F1-score0.97820.97570.93460.83220.91680.98440.93560.98260.9797
OA0.97890.97640.94320.87980.92410.98470.94040.98230.9822
P 0.35168 × 10−69 × 10−55 × 10−60.02224 × 10−50.12490.3116
Table 5. Soft classification fitting accuracy.
Table 5. Soft classification fitting accuracy.
RMSER2p-Value
Linear Logistic Regression0.1030.7090.869
Neural Networks0.0420.9790.071
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, H.; Cui, J.; Zhang, X.; Han, Y.; Cao, L. Dimensionality Reduction and Classification of Hyperspectral Remote Sensing Image Feature Extraction. Remote Sens. 2022, 14, 4579. https://doi.org/10.3390/rs14184579

AMA Style

Li H, Cui J, Zhang X, Han Y, Cao L. Dimensionality Reduction and Classification of Hyperspectral Remote Sensing Image Feature Extraction. Remote Sensing. 2022; 14(18):4579. https://doi.org/10.3390/rs14184579

Chicago/Turabian Style

Li, Hongda, Jian Cui, Xinle Zhang, Yongqi Han, and Liying Cao. 2022. "Dimensionality Reduction and Classification of Hyperspectral Remote Sensing Image Feature Extraction" Remote Sensing 14, no. 18: 4579. https://doi.org/10.3390/rs14184579

APA Style

Li, H., Cui, J., Zhang, X., Han, Y., & Cao, L. (2022). Dimensionality Reduction and Classification of Hyperspectral Remote Sensing Image Feature Extraction. Remote Sensing, 14(18), 4579. https://doi.org/10.3390/rs14184579

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop