Next Article in Journal
Inspecting Pond Fabric Using Unmanned Aerial Vehicle-Assisted Modeling, Smartphone Augmented Reality, and a Gaming Engine
Next Article in Special Issue
Evaluation of Leaf Chlorophyll Content from Acousto-Optic Hyperspectral Data: A Multi-Crop Study
Previous Article in Journal
A Satellite-Drone Image Cross-View Geolocalization Method Based on Multi-Scale Information and Dual-Channel Attention Mechanism
Previous Article in Special Issue
Exploration of Data Scene Characterization and 3D ROC Evaluation for Hyperspectral Anomaly Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Band Selection via Band Density Prominence Clustering for Hyperspectral Image Classification

1
Center for Hyperspectral Imaging in Remote Sensing (CHIRS), Information and Technology College, Dalian Maritime University, Dalian 116026, China
2
Remote Sensing Signal and Image Processing Laboratory, Department of Computer Science and Electrical Engineering, University of Maryland, Baltimore County (UMBC), MD 21250, USA
3
Department of Electrical Engineering, National Cheng Kung University, Tainan 70101, Taiwan
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(6), 942; https://doi.org/10.3390/rs16060942
Submission received: 21 December 2023 / Revised: 21 February 2024 / Accepted: 6 March 2024 / Published: 7 March 2024
(This article belongs to the Special Issue Advances in Hyperspectral Data Exploitation II)

Abstract

:
Band clustering has been widely used for hyperspectral band selection (BS). However, selecting an appropriate band to represent a band cluster is a key issue. Density peak clustering (DPC) provides an effective means for this purpose, referred to as DPC-based BS (DPC-BS). It uses two indicators, cluster density and cluster distance, to rank all bands for BS. This paper reinterprets cluster density and cluster distance as band local density (BLD) and band distance (BD) and also introduces a new concept called band prominence value (BPV) as a third indicator. Combining BLD and BD with BPV derives new band prioritization criteria for BS, which can extend the currently used DPC-BS to a new DPC-BS method referred to as band density prominence clustering (BDPC). By taking advantage of the three key indicators of BDPC, i.e., cut-off band distance bc, k nearest neighboring-band local density, and BPV, two versions of BDPC can be derived called bc-BDPC and k-BDPC, both of which are quite different from existing DPC-based BS methods in three aspects. One is that the parameter bc of bc-BDPC and the parameter k of k-BDPC can be automatically determined by the number of clusters and virtual dimensionality (VD), respectively. Another is that instead of using Euclidean distance, a spectral discrimination measure is used to calculate BD as well as inter-band correlation. The most important and significant aspect is a novel idea that combines BPV with BLD and BD to derive new band prioritization criteria for BS. Extensive experiments demonstrate that BDPC generally performs better than DPC-BS as well as many current state-of-the art BS methods.

1. Introduction

With advances in remote sensing technology, hyperspectral imaging has become an emerging technique in recent years to solve many issues that cannot be resolved by multispectral imaging, such as subpixel detection, mixed pixel classification, spectral unmixing, endmember finding, etc. [1]. However, such advantages also come with unnecessary abundant spectral redundancy that can be removed without compromising data exploitation. One effective means is data spectral dimensionality reduction (DR) [2] (chapter 6). Although many DR techniques have been proposed over the past years, band selection (BS) remains one of most widely used DR methods for hyperspectral data reduction due to the fact that it can remove inter-band correlation resulting from very fine spectral resolution while retaining data integrity provided by selected bands.

1.1. Band Priorirtization for BS

Many BS methods have been developed in the past. Generally speaking, BS can be performed by band prioritization (BP), which ranks all bands according to a BP criterion. As a result, BP can be considered as unsupervised BS without requiring prior knowledge, selecting bands in accordance with their priority scores with no reference to a specific application, such as variance, entropy, information divergence, maximum variance principal component analysis (MVPCA) [3], mutual information [4,5], constrained band selection (CBS) [1,6], minimum estimated abundance covariance (MEAC) [7], etc. This type of BP is easy to implement and has been extensively used for BS.

1.2. Band Selection via Search Strategies

As an alternative to BP, another type of BS involves selecting desired bands according to a band search strategy, usually determined by a particular application, such as detection, classification, spectral unmixing, etc. Consequently, such BS is generally supervised and carried out by a specifically designed band search algorithm, such as sequential forward selection (SFS) algorithm [8], sequential backward selection (SBS) algorithm, multitask sparsity pursuit (MTSP) [9], multigraph determinantal point process (MDPP) [10], dominant set extraction BS (DSEBS) in [11], band subset (BSS) [12,13,14] for anomaly detection, linearly constrained minimum variance-based BSS for classification [15], and also evolution-based algorithms [16], using a variety of evolutionary computing strategies; for example, particle swarm optimization (PSO) [17,18], firefly [19], and colony algorithm [20,21].

1.3. Band Clustering/Grouping for BS

In addition to the above-mentioned BP-based and BS strategy-based BS methods, band clustering/grouping-based BS methods have also received considerable interest in the past. For example, Wang et al. [22] developed an adaptive subspace partition strategy (ASPS)-based band selection method by partitioning a hyperspectral image cube into multiple sub-cubes by maximizing the ratio of interclass distance to intra-class distance. In a follow-up work, Wang et al. [23] further developed a fast neighborhood grouping method for hyperspectral band selection (FNGBS), which partitions a hyperspectral image cube into several groups using a coarse-to-fine strategy, so that bands with the maximum product of local density and information entropy in groups are selected as a band subset. Most recently, Wang et al. [24] also developed an optimal neighborhood reconstruction (ONR)-based BSS based on correlated neighborhood property (CNP) to exploit strong correlations between neighboring bands. Interestingly, Baisantry et al. [25] developed a feature extraction-based clustering BS method that uses sparse subspace clustering to cluster similar bands into groups and prioritizes bands by a new metric called the combined divergence-correlation index to select the most discriminative as well as least correlated bands as cluster representatives.

1.4. Density Peak Clustering for BS

As noted above, band clustering/grouping-based methods cluster or group bands according to band neighboring correlation without taking data spatial distribution into account. To mitigate this dilemma, a new concept of density peak clustering (DPC) was derived in [26], which introduces two indicators, local density that characterizes the spatial distribution surrounding each data point and cluster distance of a data point from clusters with densities higher than its density. By virtue of DPC, we not only can calculate the density of each data point but also its distance to higher density clusters. Now, if we consider data points as spectral bands, then for each spectral band, say Bi (with its expanded band vector, denoted by bi), the data density can be interpreted as band local density (BLD), denoted by ρi, which is determined by the cut-off band distance (BD), denoted by bc, and the cluster distance as BD, denoted by δi of bi, from band images with their BLD higher than ρi. Then, a DPC score is calculated by multiplying ρi with δi, referred to as γ i = ρ i × δ i , to be used as a BP criterion to rank all bands for BS, referred to as DPC-BS.

1.5. Motivations

Although the above-mentioned DPC-BS methods have shown some success in BS, three issues still remain. One is how to automatically determine the cut-off band distance bc, which is crucial to calculate the BLD. It must be determined in advance or empirically. Another is the used Euclidean distance, which generally measures the spherical distribution of the data so that a data point is always assigned to the nearest center. However, as for the bands of hyperspectral data, a better BD indicator should be a spectral discrimination measure that captures spectral characteristics better than distance. Last but not least, there is a lack of criteria for measuring inter-band correlation.
The motivation of this paper is to investigate an approach that can resolve the above three issues. More specifically, we develop a novel DPC-BS method called band density prominence clustering (BDPC), which introduces a new concept of band prominence value (BPV) for each band, say BPV ( b i ) , as a third indicator to take care of inter-band correlation. Despite the fact that this idea is inspired from the band prominent peaks of a curve calculated by self-mutual information (SMI) [5], it does not calculate prominent peaks. Instead, it calculates the BPV for each band bi and then integrates the BPV ( b i ) with the DPC score γ i as a curve to produce a BDPC score, given by η i = γ i × BPV ( b i ) , which will be used to prioritize all bands for BS similarly to how γi plays the same role for DPC-BS. Using this BDPC score, two versions of BDPC-BS can be further developed, to be called bc-BDPC and k-BDPC. bc-BDPC is developed by fixing the cut-off band distance, bc at a pre-determined value, while k-BDPC is developed by fixing the number of nearest neighbors k at a pre-determined value. Particularly, bc-BDPC extends currently used DPC-BS methods from using two indicators (BD, BLD) to three indicators (BD, BLD, and BPV ( b i ) ). On the other hand, k-BDPC is new and developed based on a given number of k nearest neighboring bands to calculate the three indictors (BD, BLD, and BPV ( b i ) ). Unlike DPC-BS, which requires a pre-specified value for bc, the proposed bc-BDPC can automatically determine the value of bc by the number of clusters, nclusters. Specifically, for hyperspectral image classification (HSIC), nclusters can be determined by the number of classes, nclasses. As for k-BDPC, the value of k can be determined by virtual dimensionality (VD) [27,28].
Finally, to make BDPC more effective for hyperspectral BS, the Euclidean distance, commonly used by DPC, is replaced by a spectral discrimination measure such as spectral angle mapper (SAM) [27], spectral information divergence (SID) [29], or a joint SID with SAM called SIDAM [30], etc., which can be used to calculate BD and also inter-band correlation.

1.6. Contributions

Several contributions are summarized as follows:
  • A new concept of BPV is introduced into DPC as a third indicator to extend the commonly used two-indicator DPC-BS, (BLD,BD)-DPC-BS, to a new three-indicator BDPC-BS, (BLD,BD,BPV)-BDPC-BS.
  • Using (BLD,BD,BPV)-BDPC-BS, two versions of BDPC-BS, bc-BDPC and k-BDPC are developed. Specifically, bc-BDPC can be considered as an extension to DPC-BS.
  • Automatic rules are particularly derived to determine the value of the cut-off band distance, bc, for bc-BDPC and the number of nearest neighboring bands, k, for k-BDPC.
  • Spectral discrimination measures are used in BDPC-BS to replace Euclidean distance in DPC-BS to better capture BD and spectral inter-band correlation.
The remainder of this paper is organized as follows: Section 2.1 reviews DPC-related works, while Section 2.2 discusses the most recent DPC-BS-based methods. Section 2.3 develops new variants of DPC-BS methods. Section 2.4 introduces the concept of BPV to explore inter-band correlation to extend DPC-BS to BDPC-BS, where two versions of BDPC BS methods, bc-BDPC and k-BDPC, are derived in great detail. In particular, the automatic rules for determining the cut-off band distance, bc, and the k nearest neighbors are also derived. Section 3.1 describes three hyperspectral images used for classification experiments. Section 3.2 conducts extensive experiments for HSIC along with their discussions. Section 4 describes the novelties of this paper and draws some conclusions.

2. Methods

DPC was originally proposed by Rodriguez and Laio for data point clustering and has received considerable interest in pattern recognition [26]. In [31] Wei et al. presented an overview on DPC to first analyze the theory of DPC and its performance advantages and disadvantages and then summarizes the improvement of DPC in recent years through the improvement effect via experiments and new ideas for improving DPC algorithm in the future. Most recently, Tobin and Zhang [32] provided theoretical properties that characterize DPC and further develops a clustering algorithm, called Component-wise Peak-Finding (CPF) to deal with detection of erroneous points with high density and large distance to points of higher density as well as incoherent cluster assignment caused by noise.

2.1. DPC

The idea of DPC assumes that a cluster center is surrounded by its neighboring data points, with local density measured by the number of its neighbors, and each cluster has a relatively large distance from other clusters. More specifically, DPC designs two indicators to cluster data points. One is local density, defined by:
ρ i = j i χ d c ( d c d i j )
where dc is the predetermined cut-off distance, dij is the Euclidean distance between two data points xi and xj, and
χ d c ( d c d i j ) = 1 ;   if   d c d i j > 0 0 ;   otherwise
The other is the cluster distance defined by:
δ i = min ρ j > ρ i d i j
which is calculated by minimizing the distance between data points xi and xj with local density ρj higher than ρi.
For the data point with the highest local density, δ i max , it is defined as:
δ i max = max j d i j
Thus, a point with relatively high ρ i and large δ i can be considered a cluster center. These two quantities are then multiplied together to yield a quantity for a data point xi, given by:
γ i = ρ i × δ i
which can be used to define a DPC score of xi. Those points that have larger γ i are more likely to be distinguished as cluster centers. With this in mind, (5) can be used to rank the cluster centers.
On many occasions, the discrete value of the local density in (1) is generally replaced by a continuous value, as follows:
ρ i = d i j < d c exp d i j d c α
where α is an adjusting factor.
One crucial issue arising from DPC is the determination of dc. Xu et al. [33] introduced density–distance clustering (DDC) to develop an improved automatic density peaks clustering (ADPC) algorithm that could automatically select the suitable cut-off distance, dc, and acquires the optimal number of clusters without additional parameters. Also, DPC generally makes an assumption that cluster centers are often surrounded by data points with lower local density and are far away from other data points with higher local density. However, this assumption is not necessarily true. To address this issue, Wang et al. [34] developed a variational density peak clustering (VDPC) algorithm, which is designed to systematically and autonomously perform the clustering task on datasets with various types of density distributions.

2.1.1. k-Nearest Neighbors-Based DPC

It is interesting to note that the concept of local density, ρi is actually very closely related to k-nearest neighbors (kNNs). Du et al. [35] utilized kNNs to calculate ρi to derive DPC-KNN. Since then, many works along this line have been reported [36,37,38,39]. Specifically, local density is very closely related to kNNs. For each data point xi, k N N ( x i ) is defined as the set of k data points that are the nearest neighbors close to xi. Three kNN-based DPC methods are of particular interest.

2.1.2. DPC-KNN

Due to the fact that the global data structure may lead DPC to miss many clusters, Du et al. [34] proposed a DPC based on kNNs (DPC-KNN) by including kNNs into DPC to address this issue. Additionally, in high-dimensional data spaces, DPC is likely to generate an incorrect number of clusters. To cope with this problem, DPC-KNN also includes principal component analysis (PCA) into DPC-KNN to derive DPC-KNN-PCA. However, two issues that have significant impacts on DPC performance were not addressed, that is: how to determine the value of k and the number of principal components, both of which were determined empirically. However, for multi-density data scenarios, one parameter cannot satisfy all data sets. So, clustering often cannot achieve good results. To resolve this issue, Yin et al. [40] extended the DPC algorithm to deal with multi-density data. The cut-off distance dc was selected using KNN to sort the neighbor distances of each data point to draw a line graph of the KNN distance and found the global bifurcation point to divide the data with different densities.

2.1.3. G-DPC-KNN

Compared to DPC-KNN, which empirically determined the value of k, Jiang et al. [41] developed a method called G-DPC-KNN to calculate the cutoff distance dc based on the Gini coefficient and then used kNNs to find cluster centers. Most recently, Anandarao and Chellasamy [42] also addressed the issue of the random selection of the cut-off distance parameter, dc, by using the Gini index or Gaussian function to make a valid guess of dc. Unfortunately, the same issue arising from DPC-KNN in how to determine an appropriate value of k for G-DPC-KNN remains challenging. Wang et al. [43] developed a novel density peak clustering algorithm for the automatic selection of clustering centers based on K-nearest neighbors (AKDPC) to remedy manual determination of cluster centers and poor performance on complex datasets with varying densities.

2.1.4. Shared Nearest Neighbors (SNNs)

Since kNNs generally run into an issue where two data points may share certain nearest neighbors (SNNs), it results in double accounts for these shared nearest neighbors. To alleviate this dilemma, shared nearest neighbor (SNN)-based clustering methods were developed. For example, a very early attempt was made by Jarvis and Patrick [44], who introduced a nonparametric clustering algorithm using the concept of similarity based on the sharing of near neighbors. Liu et al. [45] proposed the SNN-based fast search clustering algorithm for finding density peaks, which is based on three newly defined indicators: SNN similarity, local density ρ, and distance from the nearest larger density point δ, to take the information from the nearest neighbors and the shared neighbors into account so as to self-adapt to local surroundings. Specifically, it introduced a two-step allocation method that can accurately recognize and allocate points by counting the number of shared neighbors between the two points and also assign remaining points by finding the clusters to which more neighbors belong. Lv et al. [46] proposed a fast-searching density peak clustering algorithm based on shared nearest neighbors and an adaptive clustering center (DPC-SNNACC) algorithm to automatically determine the number of knee points in the decision graph according to the characteristics of different datasets and the number of clustering centers without human intervention.

2.2. DPC-BS

Applications of DPC to BS in hyperspectral imaging have been recently explored. Several recent works related to using DPC to develop BS techniques have emerged in the literature.
First of all, in order for DPC to be applied to BS, we need to interpret each data sample point xi as a band vector, bi. In this case, the data sample set x i i = 1 N is replaced with the full band set, b l l = 1 L , where b l = b l 1 , b l 2 , , b l L T is the lth band image and can be considered as a band vector with L spectral band components. So, the sample distance between two data points, xi and xj, d i j = d ( x i , x j ) now becomes the band distance (BD) between two bands, bi and bj, d ( b i , b j ) , which is measured by the Euclidean distance.
b i j = d ( b i , b j ) = | | b i b j | | 2 = l = 1 L | | b i l b j l | | 2
Using (7), four DPC-based BS methods were recently developed and are described in detail, as follows.

2.2.1. Exemplar Component Analysis (ECA)

An early DPC-BS method was developed by Sun et al. [47], who considered BS as exemplar component analysis (ECA) of hyperspectral bands with high local density and large distances from bands with higher densities. An early attempt to extend DPC to hyperspectral BS is exemplar component analysis (ECA) proposed by Sun et al. [47], which defines local density in (6) as:
ρ ES ( b i ) ρ i ES = j = 1 , j i L exp b i j / 2 σ 2
where α in (6) is set to 1 and γi in (5), as follows:
γ ES ( b i ) γ i ES = ρ i ES × δ i ES
where ES is the exemplar score, and δ ES ( b i ) δ i ES is defined by (3). The γ i ES in (9) is then used as a BP criterion to prioritize all bands in descending order and then select the first nBS largest γ i ES scores as the desired band subset.

2.2.2. Enhanced Fast Density-Peak Clustering (E-FDPC)

To improve ECA’s ability to find bands as cluster centers in small regions, Jia et al. [48] developed an enhanced fast density-peak clustering (E-FDPC) method by adopting weights to use DPC scores for BS. A second DPC-BS method is the enhanced fast density-peak clustering (E-FDPC) proposed by Jia et al. [48], which is defined as:
s i j = | | b j b i | | 2 = l = 1 L | | b j l b i l | | 2
b i j E - FDPC = s i j / L
ρ E - FDPC ( b i ) ρ i E - FDPC = j = 1 , j i L exp ( b i j E - FDPC / b c E - FDPC ) 2
and
b c E - FPDC = b initial / exp ( n BS / L )
where binitial is an initial value of the cut-off threshold, which can be empirically determined as b initial = 2 % × L × ( L 1 ) th smallest value of s i j i , j . Then, an E-FDPC score is calculated by multiplying the two indicators, ρ i E - FPDC in (12) and δ E - FPDC ( b i ) δ i E - FPDC in (13), together as:
γ E - FPDC ( b i ) γ i E - FPDC = ρ i E - FPDC × δ i E - FPDC 2
where ρ i E - FPDC and δ i E - FPDC are determined by b i j E - FDPC and b c E - FPDC , respectively.

2.2.3. Information-Assisted Density Peak Index (IaDPI)

ECA and E-FDPC rank bands by their calculated DPC scores to locate cluster center bands with larger ranking scores. Recently, Luo et al. [49] developed the information-assisted density peak index (IaDPI) method by including band entropy as an additional indicator to further improve E-FDPC. Recently, Luo et al. [49] developed IaDPI to include band entropy as an additional indicator to further improve E-FDPC, where the entropy of a band image bl is defined as:
H i = H ( b i ) = l = 1 L p i l log p i l
with p i = p i 1 , p l 2 , , p i L T defined as the probability vector of the band image bl and p i l = b i l l = 1 L b i l . The local density for IaDPI is then defined as:
ρ IaDPA ( b i ) ρ i IaDPA = b i j < b c exp b i j / b c α 1 + H i / H c α 2 + ( | H i H j | / H c α 2
where α1 is an adjusting factor set to 2, and α2 is set to 1. Furthermore, it also defines:
T b i j = b i j α 1 H j α 2
to derive:
δ IaDPA ( b i ) δ i IaDPA = T b i j α 1 H j α 2 ( α 1 + α 2 )
where j is determined by:
j = arg min j : ρ j > ρ i T b i j
Finally, the IaDPA score is calculated by:
γ IaDPA ( b i ) γ i IaDPA = ρ i IaDPA × δ i IaDPA

2.2.4. Shared Nearest Neighbors Network (SNNC)

Because k N N ( b i ) and k N N ( b j ) may share the same nearest neighbors, Li et al. [50] proposed an efficient SNN clustering (SNNC) method for hyperspectral optimal BS, which can obtain the local density of each band using SNN to more accurately reflect the local distribution characteristics. In other words, the local density of each band obtained by SNN can more accurately reflect the similarities in the local distribution characteristics, measured by:
S N N ( b i , b j ) = k N N ( b i ) k N N ( b j )
From (21), an SNN-based local density can now be defined as:
ρ k SNNC ( b i ) ρ i k - SNNC = b j k N N ( b i ) exp d ( b i , b j ) S N N ( b i , b j ) + 1
where d ( b i , b j ) = l = 1 L ( b i l b j l ) 2 .
δ k SNNC ( b i ) δ i k - SNNC = min j d ( b i , b j )   if   ρ i k - SNNC < ρ j k - SNNC max j d ( b i , b j )   otherwise
In addition, to acquire a band subset containing a large amount of information, the information entropy was proposed to be used as one of the weight factors. More specifically, an entropy-based BP criterion for SNNC is derived by combining (22) and (23) with the information entropy of band bi defined in (15), referred to as H k - SNNC ( b i ) H i k - SNNC , as follows:
γ k SNNC ( b i ) γ i k - SNN = ρ i k - SNNC × δ i k - SNNC × H i k - SNNC
which corresponds to γ i in (5).
Finally, SNNC also developed a method for automatically selecting the optimal band subset, designed based on the slope change.

2.3. New Variations of DPC-BS

While DPC has shown potential and promise in BS for hyperspectral BS, there are several issues that remain unsolved. The first issue is the use of the Euclidean distance, which may not effectively capture spectral correlation compared to the use of a spectral discrimination measure. The second issue is the lack of a mechanism to address inter-band correlation. It should be noted that although IaDPI and SNNC use band information entropy as an additional indicator, it only measures the uncertainty of each band without considering inter-band spectral correlation. The third issue is determining the value of the cut-off BD, bc, or the number of nearest neighboring band images, k. This is because DPC-BS performance is heavily determined by its BLD, ρi and BD, δi, which are actually calculated using either bc or k. Unfortunately, these two parameters, bc and k, are indeed inter-correlated and cannot be determined independently. In other words, one parameter determines the other. To address this issue, we further derive two versions of DPC-BS, namely bc-DPC-BS and k-DPC-BSD PC-BS, by manipulating bc or k.

2.3.1. bc-DPC-BS

The bc-DPC-BS to be developed in this section is similar to ECA, E-FDPC, and IaDPA in the sense that the cut-off band distance bc is determined and fixed at a certain value a priori. It can be considered as a variant of DPC-BS wherein the Euclidean distance used to calculate the BLD, ρi and BD, δi is replaced with a spectral discrimination measure, m. In this case, dc is replaced by bc and ρi in (1) becomes:
ρ i BLD = j i χ b c ( b c b i j )
where bc is the predetermined band distance (BD) threshold, bij is the spectral distance between bi and bj, defined as:
b i j = m ( b i , b j )
where m can be any type of spectral measure, such as SAM, SID, or SIDAM. In this case, Equations (2) and (3) are modified and re-defined as:
χ b c ( b c b i j ) = 1 ;   if   b c b i j < 0 0 ;   otherwise
δ b c - BD ( b i ) δ i b c - BD = min ρ j BLD > ρ i BLD b i j
For the band with the highest band local density, δ i max - BD in (5) is re-defined as:
δ max - BD ( b i ) δ i max - BD = max j b i j
Analogous to (6), (25) can also be modified as kernel-based BLD. ρ i b c - BLD is defined as:
ρ b c - BLD ( b i ) ρ i b c - BLD = j = 1 , i L exp b i j b c α
where α is set to 2 for our experiments. Combining (30) with (28), we can define the bc-DPC score as:
γ b c - DPC ( b i ) γ i b c - DPC = ρ i b c - BLD × δ i b c - BD
which corresponds to γ i in (5).

2.3.2. k-DPC-BS

The k-DPC-BS presented in this subsection is similar to DPC based on k-nearest neighboring bands. But, it introduces a new aspect by interchanging the roles played by BLD and BD in bc-DPC-BS compared to k-DPC-BS. Specifically, it assumes that the value of k is yielded a priori to calculate the BD of bi, δ i k - BD , which embraces k bands surrounding bi. Then, ρ i b c - BLD can be further calculated by δ i k - BD . It is a sequential process of k δ i k - BD ρ i b c - BLD . In this case, we consider a cluster centered at a band vector bi and expand its range until the cluster contains k bands surrounding bi according to their spectral similarities.
For each bi, let the set contain the k nearest neighboring bands surrounding bi, denoted by B i k . Then, we can define a criterion similar to ρ i in (1) as:
ρ k - BLD ( b i ) ρ i k - BLD = max b j B i k m ( b i , b j )
The minimum distance of bi from all other bands bj with ρ j k - BLD greater than ρ i k - BLD , denoted by δ i k - BD is defined as:
δ k - BD ( b i ) δ i k - BD = min ρ j k - BLD > ρ i k - BLD m ( b i , b j )
Finally, a k-DPC score, denoted by γ k - DPC ( b i ) , can be defined by γ i k - DPC in correspondence to γ i in (5) as:
γ k - DPC ( b i ) γ i k - DPC = ρ i k - BLD × δ i k - BD

2.4. BDPC-BS

According to (31) and (34), the bc-DPC-BS and k-DPC-BS scores do not take into account the inter-band correlation, and neither do ECA, E-FDPC, IaDPA, or k-SNNC. So, to address this issue, a new concept derived from prominent band peaks proposed in [5], called band prominence value (BPV), BPV ( b i ) for each band vector bi is introduced into DPC as a third indicator. By incorporating BPV into BLD, ρi, and BD, δi, we can further extend the two-indicator (BLD,BD)-DPC-BS and bc-DPC-BS and k-DPC-BS in Section 2.3 to three indicator (BLD,BD,BPV)-DPC-BS, bc-BDPC-BS, and k-BDPC-BS. This is similar to SNNC, which includes a third indicator, information entropy H i k - SNNC in (24).
The concept of BPV can be applied to any DPC method. It plots DPC scores obtained by (9) for ECA, (14) for E-FDPC, (20) for IaDPA, (24) for k-SNNC, (31) for bc-DPC-BS, and (34) for k-DPC-BS, etc., as a curve where each point on the curve corresponds to the band prominence value (BPV) of one band. For example, for band C, BPV(C) is calculated and illustrated in Figure 1, which shows that the band prominence value of the peak C can be calculated.
To better understand how BPV is calculated according to Figure 1, a detailed step-by-step implementation for calculating BPV is summarized in Algorithm 1. As a matter of fact, how to calculate the prominence of a peak is available in Matlab and can be performed by findpeaks on https://www.mathworks.com/help/signal/ug/prominence.html (accessed on 5 March 2024).
Algorithm 1: Algorithm for Calculating Prominence Scores of Bands
1.
Input: Let the local minimum (LM) points set be Ω LM = b j l l = 1 n LM and Ω ˜ = Ω Ω LM , where Ω is the full bands set.
2.
For b i Ω ˜ , calculate the prominence by the following procedure:
 a.
Extend a horizontal line from point bi to the left and the right until the line either
   •
crosses the curve because there is a higher point
  or
   •
reaches the left or right end of the curve.
  Then, the left interval and right interval of bi are defined.
 b.
Let γ i be either γ i b c - BDPC or γ i k - BDPC . We can find the BPV of bi,  BPV ( b i ) , as follows:
  i
Find the minimum point bl in the left interval. Let ς l = γ l if γ l < γ i . Otherwise, let ς l = 0 .
    ii.
Find the minimum point bj in the right interval. Let ς j = γ j if γ j < γ i . Otherwise, let ς j = 0 .
    iii.
Output  BPV ( b i ) = γ i max ( ς l , ς j )
Once BPV ( b i ) i = 1 L for all band vectors b i i = 1 L are calculated by the above algorithm, we can multiply BPV ( b i ) with γ i b c - BDPC in (31) or γ i k - BDPC in (34) to yield new BPV-based BP criteria for bc-BDPC-BS, defined as:
η i b c - BDPC = γ i b c - BDPC × BPV ( b i )
and for k-BDPC BS, defined as:
η i k - BDPC = γ i k - BDPC × BPV ( b i )
Assuming nBS is the number of bands to be selected, bc-BDPC and k-BDPC can use (35) and (36) to select the bands with the first nBS largest values of η i b c - BDPC and η i k - BDPC as desired bands for BS, respectively.
A final remark is noteworthy. The key difference between the idea used in BDPC-BS and SMI-BS in [5] is that BDPC-BS calculates BPV for each band as an indicator jointly used with BLD and BD for BS, compared to SMI-BS, which find prominent peaks for BS.

2.4.1. Determination of bc and k

Now, we come to the last issue, which is how to determine the values of the two parameters, bc and k, which have significant impacts on the performance of DPC-BS and BDPC-BS methods. Unfortunately, these two parameters were generally determined a priori or empirically in the past. This section discusses several approaches to determine appropriate values of bc and k for BDPC automatically.

2.4.2. Determination of bc

As noted above, when DPC is applied to BS, the total number of data samples, N, is now replaced by L, the total number of bands. In [35], bc is determined by:
b c = b L d × p % / 100
where the b L d × p % / 100 th distance is selected according to an ascending order among all possible distances between two band vectors. In (37), x is the upper ceiling of x, defined by the smallest integer equal to or greater than x, and L d = L ( L 1 ) / 2 is the total number of distances, and p% is a pre-specified percentage.
As an alternative to (37), IaDPI [49] selects:
b c = V d 2 % × L × ( L 1 ) / exp ( k / L )
where V d ( x ) is the xth smallest distance among all possible distances between two data samples, with L being the total number of bands.
In [41], it assumed that the potential energy in the data domain was similar to the local density of the points in the dataset. As a result, the potential energy of each point can be used as an indicator of the overall distribution of the dataset to estimate the potential energy of the whole dataset. In this case, we can define the potential energy of each band image bi as:
δ i ( σ ) = j = 1 , j i L exp | | b j b i | | / σ 2
where σ plays a similar role as dc does in (1). So, calculating the potential energy of band sets is equivalent to calculating the local density of band sets. In doing so, it introduces the Gini coefficient, denoted by:
G ( σ ) = 1 i = 1 L δ i ( σ ) / Z
where:
Z = i = 1 L δ i ( σ )
is the total potential energy of all band images. We then define:
σ opt = min σ G ( σ )
which determines the desired bc as:
b c = σ opt
Interestingly, (37), (38) and (43), which are used to determine bc, have nothing to do with the number of clusters, nclusters, which is supposed to be the key parameter of BDPC. To address this issue, we developed an automatic algorithm to determine nclusters-based bc, which is very easy and simple to implement. Its idea is to first use the well-known K-means clustering method, also known as ISODATA [51], to group a dataset into nclusters, C j j = 1 n clusters . Then, for each cluster C j , we calculated its cluster center, μ j j = 1 n clusters , by its mean. Finally, we found bc  b c = min j , r C j m ( r , μ j ) .
Despite the fact that ISODATA and K-means methods are essentially the same algorithm, ISODATA is used to avoid confusion between K-means and kNNs because the uppercase “K” used by the K-means method is the number of clusters, while the lowercase italic “k” used by kNNs is the number of nearest neighbors.
Unlike DPC, which determines dc empirically, bc-BDPC automatically determines bc by the number of clusters, nclusters. So, if we assume that each class is specified by one cluster, then bc is in turn determined by the number of classes, nclasses of interest. Once bc is determined, ρ i b c - BLD can be determined accordingly, i.e., b c ρ i b c - BLD .

2.4.3. Determination of k

As an alternative to using bc to determine the value of k, we can directly use the value of k to calculate local density and BD without appealing to bc. For example, k can be selected as:
k = p % 100 × L
with p% predetermined empirically. Such selection is a trial-and-error and not practical.
To resolve this issue, virtual dimensionality (VD) introduced in [27,28] can be used for this purpose. It is first used to estimate the number of bands to be selected, and then k can be determined by:
k BS = 2 × L n BS
where “2” is included to take the adjacent bands on both sides.

2.4.4. Determination of nBS

It should be noted that in addition to VD in (45) to determine the value of k, [50] also developed an automatic rule for selecting an optimal band subset using the slope change, defined as follows:
p = i | k i | | k i + 1 | γ   for   i = 1 , 2 , , L 2
where:
q = i = 1 L 1 | k i | | k i + 1 |   and   γ = q / ( L 2 )
n BS = max { p }

3. Experiments

3.1. Images to Be Used for Experiments

Three popular hyperspectral images that have been studied extensively for HSIC were used for experiments.

3.1.1. Purdue Indiana Indian Pines

The first image is the Purdue Indiana Indian Pines test site with an aerial view, shown in Figure 2a, along with its ground truth of 17 class maps in Figure 2b and their class labels in Figure 2c. Table 1 tabulates the number of data samples in each class, where there are four small classes with less than 10 samples: classes 7, 9, 1, 16, and three classes with more than 1000 samples: classes 14, 2, and 11. So, this scene clearly has an imbalanced class issue in classification. It is an airborne visible/infrared imaging spectrometer (AVIRIS) image scene and has a size of 145 × 145 × 220 pixel vectors, with water absorption bands (bands 104–108 and 150–163, 220). So, a total of 220 bands were used for experiments. It should be noted that in many reports, 200 bands were used by excluding water absorption bands. However, for BS, it is believed that a full set of 220 bands should be used for band integrity.

3.1.2. Salinas

A second image dataset is Salinas in Figure 3a, which is also an AVIRIS scene. It was collected over Salinas Valley, California, with a spatial resolution of 3.7 m per pixel with a spectral resolution of 10 nm. It has a size of 512 × 217 × 224 , with 20 water absorption bands, 108–112, 154–167, and 224. So, a total of 224 bands were used for experiments. Figure 3b,c shows the color composite of the Salinas image along with the corresponding ground truth class labels. Unlike the Purdue data, the Salinas scene does not have the issue of imbalanced classes, as tabulated in Table 2, where the smallest class is class 13 with 916 data samples.

3.1.3. University of Pavia

A third hyperspectral image dataset used for experiments was the University of Pavia image shown in Figure 4, which is an urban area surrounding the University of Pavia, Italy. It was recorded by the ROSIS-03 satellite sensor over an urban area surrounding the University of Pavia, Italy. It has a size of 610 × 340 × 115 with a spatial resolution of 1.3 m per pixel and a spectral coverage ranging from 0.43 to 0.86 μm with a spectral resolution of 4 nm (the 12 most noisy channels were removed before experiments). Figure 4b provides its ground-truth map of nine classes along with color class labels in Figure 4c. Table 3 also tabulates the number of data samples in parentheses collected for each class. Like the Salinas scene, this scene also has very large classes, with only one small class having less than 1000 data samples: class 9 with 947 samples. However, this scene has a more complicated BKG than the other two studied scenes, as already shown in [15,16], where the precision rate of this scene was much lower than the other two scenes.

3.2. Experimental Results and Discussions

In order to conduct credible experiments for a comparative analysis, three crucial elements are considered and investigated:
(a)
Number of bands needed to be selected, nBS.
Since BS is unsupervised, nBS must be determined without prior knowledge. In this case, VD was used for this purpose.
(b)
Classifiers used for HSIC.
There are many classifiers available and reported in the literature. Selecting a good candidate to be used for HSIC will be a challenge. In this paper, its selection is based on three aspects. One is that it should be effective and perform reasonably well compared to many existing state-of-the-art classifiers. Another is that it should be simple enough to be implemented so that those who are interested in our work can repeat experiments. A third one is that it does not require too many parameters to be tuned, unlike many deep learning-based HSIC methods. It turns out that spectral-spatial classifiers meet these requirements and are preferred to deep learning-based classifiers when it comes to experiments. This is because bc-BDPC and k-BDPC methods are steady with the parameters bc and k being determined automatically. In this case, spectral-spatial classifiers are more appropriate to be used as classifiers than deep learning-based classifiers, which require a number of model parameters to be tuned empirically. Specifically, the Iterative EPF (IEPF) method in [52] was selected as the classifier for experiments because it has been shown to significantly improve EPF-based methods [53]. There were four IEPF methods proposed in [52] that performed similarly. The IEPF-G-g was chosen as its representative for our comparison because IEFP-G-g used guided filters, as opposed to IEPF-B-g, which used bilateral filters [5]. The IEPF-G-g was implemented with the selection of training samples for each class, which followed the same procedure that was used in [52]. Table 4 tabulates the number of training samples selected for each class across the three datasets, and the remaining data samples were used as test data samples.
(c)
BS methods selected for comparison
There are two classes of BS methods of interest. One comprises DPC-based BS methods: ECA, E-FDPC, IaDPI, DPC-KNN, G-DPC-KNN, and Shared Nearest Neighbor Network (SNNC) [50]. The other class is made up of the most recently developed BS methods, such as sequential CCBSS (SQ-CCBSS) [14], successive CCBSS (SC-CCBSS) [14], ref. [8], MDPP [11], DSEBS [12], Linearly Constrained Minimum Variance Sequential Feedforward Band Selection (LCMV-SFBS) [15], Linearly Constrained Minimum Variance Sequential Backward Band Selection (LCMV-SBBS) [15], Feedforward Class Signature-Constrained Band Prioritization (FCSCBS-BP) [16], and Backward Class Signature-Constrained Band Prioritization (BCSCBS-BP) [16]. Since Self-Mutual Information (SMI)-BS was shown in [5] to perform better than the above-mentioned BS methods, it is sufficient to choose SMI-BS for comparison.
With all things considered, this section compares six DPC-based methods, bc-BDPC and k-BDPC, against ECA, E-FDPC, IaDPI, DPC-KNN, G-DPC-KNN, SNNC, and BP-based SMI-BS using three spectral discrimination measures: SAM, SID, and SIDAM. Table 5 lists the parameter setting used for all compared BS methods along with the computer environment: Intel® Core(TM) i7-9750H CPU 2.6 GHz with RAM: 16 G.

3.2.1. Purdue Indian Pines

According to [15,16], VD was estimated to be 18. Figure 5 and Figure 6 plot η i b c - BDPC in (39) and η i k - BDPC in (40), which were produced by bc-BDPC and k-BDPC using (a) SID, (b) SAM, and (c) SIDAM for the Purdue Indian Pines scene, respectively, where the red dots are 18 selected bands according to the peaks of BPV, as tabulated in Table 6. In addition, Table 6 also includes 18 bands selected by SMI-BS, ECA, E-FDPC, IaDPI, DPC-kNN, G-DPC-kNN, SNNC, and Uniform BS, where parameters used by various methods are specified in Table 5.
In order to demonstrate the effectuveness of BDPC, HSIC was used to illustrate its application, where IEPF-G-g was implemented as the desired classifier. Figure 7a–r shows the classification maps produced by (a) full bands, (b) UBS, (c) bc-BDPC-SID, (d) bc-BDPC-SAM, (e) bc-BDPC-SIDAM, (f) k-BDPC-SID, (g) k-BDPC-SAM, (h) k-BDPC-SIDAM, (i) SMI-BS, (j) ECA, (k) E-FDPC, (l) IaDPI, (m) DPC-KNN ( kBS), (n) DPC-KNN (k = 2%L), (o) G-DPC-KNN (kBS), (p) G-DPC-KNN (k = 10), (q) SNNC (kBS), and (r) SNNC (k = 3), respectively. Upon visual inspection, all the methods seemed to perform reasonably well on classification.
However, there are some subtle and appreciable differences in the classification of class 2 (highlighted in blue), class 10 (highlighted in yellow), and class 11 (highlighted in purple), where different methods misclassified different data samples. So, assessing which one performs better on these classes is a challenge. This is because the classification maps in Figure 7 only provide qualitative analyses, and it is still very difficult to evaluate the overall effectiveness of different methods based only on their classification maps.
To address this issue, Table 7 tabulates the classification results of bc-BDPC and k-BDPC using SID, SAM, and SIDAM, along with the results produced by full bands and uniform BS for quantitative comparison. The commonly used overall accuracy (OA), average accuracy (AA), and Kappa coefficient were used as evaluation criteria, with the best results boldfaced in red and the second best boldfaced in black. As we can see, the best result was produced by bc-BDPC using SID, which also outperformed using full bands and uniform BS. Most impressively, all the various versions of bc-BDPC and k-BDPC performed better than full bands and uniform BS, albeit at the expense of additional computer running time (seconds), documented at the bottom row of Table 7 for reference.
To further conduct a comparative analysis, Table 8 also tabulates the (OA, AA, Kappa) classification results obtained by the existing DPC-based BS methods: ECA, E-FDPC, IaDPI, DPC-KNN, G-DPC-KNN, and SNNC, where the best and second-best results are boldfaced in red and black, respectively, with running time (seconds) included at the bottom row for reference. Since SMI-BS has been shown to perform better than many existing BS methods in [5], SMI-BS results are also included in Table 8 for comparison. As shown in Table 8, E-FDPC and DPC-KNN using k = 2%L were the best. Nevertheless, SMI-BS, IaDPI, and SNNC also performed very well, with nearly the same performance as E-FDPC and DPC-KNN. Now, if we compare Table 7 to Table 8, we can see that bc-BDPC using SID performed better than the best results of E-FDPC and DPC-KNN in Table 7.

3.2.2. Salinas

The VD estimated for Salinas was 21 [15,16]. Table 9 also includes 21 bands selected by SMI-BS, ECA, E-FDPC, IaDPI, DPC-kNN, G-DPC-kNN, SNNC, and uniform BS for comparison, where parameters used by various methods are specified in Table 5. Figure 8 and Figure 9 plot η i b c - BDPC and η i k - BDPC produced by bc-BDPC and k-BDPC using (a) SID, (b) SAM, and (c) SIDAM for Salinas, respectively, where the red dots are the selected bands and are tabulated in Table 9.
Figure 10a–r shows the classification maps produced by (a) full bands, (b) UBS, (c) bc-BDPC-SID, (d) bc-BDPC-SAM, (e) bc-BDPC-SIDAM, (f) k-BDPC-SID, (g) k-BDPC-SAM, (h) k-BDPC-SIDAM, (i) SMI-BS, (j) ECA, (k) E-FDPC, (l) IaDPI, (m) DPC-KNN, (kBS) (n) DPC-KNN (k = 2%L), (o) G-DPC-KNN (kBS), (p) G-DPC-KNN (k = 10), (q) SNNC (kBS), and (r) SNNC (k = 3), respectively. Like the Purdue data, all the methods seemed to perform comparably on classification except for the two largest classes: class 8 (highlighted in dark blue) and class 15 (highlighted in brown), where different methods showed visible differences. In particular, (c) bc-BDPC-SID, (p) G-DPC-KNN (k = 10), and (r) SNNC (k = 3) were among the best performers for the classification of classes 8 and 15. However, this was not true when the overall performance was considered.
Table 10 tabulates the (OA, AA, Kappa) classification results of bc-BDPC and k-BDPC using, SID, SAM, and SIDAM, along with the results produced by full bands and uniform BS for quantitative comparison, where the best and second-best results are boldfaced in red and black, respectively. As shown in Table 10, the best results were produced by bc-BDPC using SAM and full bands. Interestingly, unlike the Purdue data, all the various versions of bc-BDPC and k-BDPC did not perform as well as full bands and uniform BS. The quantitative results in Table 10 demonstrated that the qualitative visual inspection in Figure 10 was not reliable.
To further conduct a comparative analysis, Table 11 also tabulates the (OA, AA, Kappa) classification results obtained by ECA, E-FDPC, IaDPI, DPC-KNN, G-DPC-KNN, and SNNC, where the best and second-best results are boldfaced in red and black, respectively, with running time (seconds) included at the bottom row for reference. Since SMI-BS has been shown to perform better than many existing BS methods in [5], SMI-BS results are also included in Table 11 for comparison.
As shown in Table 11, SNNC (k = 3) was the best, followed by DPC-KNN using kBS as the second-best. Interestingly, SNNC (k = 3) not only outperformed all the DPC-based methods in Table 11 but also performed better than using full bands. According to Table 10 and Table 11, it is worth noting that all the methods had difficulty with classifying classes 8 and 15, which are adjacent each other. This is particularly true for k-BDPC. Because of the poor classification of these two largest classes, k-BDPC did not perform well compared to other methods.

3.2.3. U. of Pavia

Once again, the VD estimated for U. of Pavia was 14 [15,16]. Figure 11 and Figure 12 plot η i b c - BDPC and η i k - BDPC produced by bc-BDPC and k-BDPC, using (a) SID, (b) SAM, and (c) SIDAM for the U. of Pavia scene, respectively, where the red dots are the selected bands and are tabulated in Table 12. Table 12 also includes 14 bands selected by SMI-BS, ECA, E-FDPC, IaDPI, DPC-kNN, G-DPC-kNN, SNNC, and uniform BS.
Figure 13a–r shows the classification maps produced by (a) full bands, (b) UBS, (c) bc-BDPC-SID, (d) bc-BDPC-SAM, (e) bc-BDPC-SIDAM, (f) k-BDPC-SID, (g) k-BDPC-SAM, (h) k-BDPC-SIDAM, (i) SMI-BS, (j) ECA, (k) E-FDPC, (l) IaDPI, (m) DPC-KNN (kBS), (n) DPC-KNN (k = 2%L), (o) G-DPC-KNN (kBS), (p) G-DPC-KNN (k = 10), (q) SNNC (kBS), and (r) SNNC (k = 3), respectively. Overall, all the methods performed reasonably well upon visual inspection, except for the classification of the second largest class: class 6 (highlighted in brown), which had subtle differences.
However, to evaluate overall performance, it is extremely difficult to name which method is the best. In this case, Table 13 tabulates the (OA, AA, Kappa) classification results of bc-BDPC and k-BDPC using SID, SAM, and SIDAM, along with the results produced by full bands and uniform BS for quantitative comparison with the best results boldfaced in red and the second-best results boldfaced in black. As we can see, the best results were produced using full bands, followed very closely by bc-BDPC using SIDAM with only differences within 0.2%. Table 14 also tabulates the (OA, AA, Kappa) classification results obtained by ECA, E-FDPC, IaDPI, DPC-KNN, G-DPC-KNN, and SNNC, where the best and second-best results are boldfaced in red and black, respectively, with computer running time in seconds included at the bottom row for reference. As shown in Table 14, E-FDPC was the best, followed by the second best DPC-KNN using kBS. If we compare Table 13 to Table 14, bc-BDPC using SAM and SIDAM performed slightly better in OA than E-FDPC but slightly worse in AA than E-FDPC. As a matter of fact, bc-BDPC generally outperformed all other BS methods in Table 14.

3.2.4. Discussions

The experiments in this section yield several interesting observations and findings.
  • Experimental results showed that the three datasets exhibited different characteristics, which resulted in different performances of various BS methods and classifiers. For example, the Purdue data had four small classes with less than 100 data samples and three large classes with more than 1000 data samples, compared to Salinas and U. of Pavia, both of which had classes with more than 900 data samples. So, the Purdue data ran into an imbalanced class issue. As a result, full bands are generally good for balanced classes and performed better than all the test BS methods for Salinas and U. of Pavia but performed worse than SMI-BS and most DPC-based BS methods for the Purdue data.
  • Also, interestingly, both bc-BDPC and k-BDPC performed very well and better than full bands for the Purdue data. This indicates that bc-BDPC and k-BDPC may perform better for datasets with imbalanced classes. However, bc-BDPC and k-BDPC did not perform as well as full bands for Salinas and U. of Pavia. Nevertheless, bc-BDPC still managed to perform comparably to full bands. Surprisingly, k-BDPC did not perform well as expected, specifically for Salinas, as its performance significantly degraded due to the classification of classes 8 and 15.
  • Since BDPC combines the advantages of the two indicators of DPC, cluster density ρ and cluster distance δ, and the band prominent peaks of SMI-BS, BPV, it performs better than DPC-based BS and SMI-BS, as expected.
  • According to experiments, bc-BDPC generally performed better than k-BDPC for all the three datasets. This means that the cut-off distance, bc, is more crucial than the k nearest neighbors in designing DPC-BS. More specifically, b c ρ i b c - BLD δ i k - BD is more effective than k δ i k - BD ρ i b c - BLD .
  • DPC-KNN using kBS performed better than DPC-KNN using k = 2%L for Salinas and U. of Pavia but worse for the Purdue data. This demonstrates that DPC-KNN using kBS works well for large datasets compared to DPC-KNN using k = 2%L, which performs better for small datasets. In addition, G-DPC-KNN using kBS performed better than DPC-KNN using k = 2%L for U. of Pavia but worse for the Purdue and Salinas data scenes.
  • It is notewothy that experiments showed that DPC-KNN and G-DPC-KNN using kBS did not perform as well as DPC-KNN and G-DPC-KNN did for the Purdue data, but did perform better than DPC-KNN and G-DPC-KNN for U. of Pavia. This indicates that DPC-KNN and G-DPC-KNN may work well for datasets mixed with small and large classes, while DPC-KNN and G-DPC-KNN using kBS may work well for datasets with large classes. As for the Salinas data, they all performed comparably.
  • Analogously to DPC-KNN and G-DPC-KNN with/without using kBS discussed above, SNNC using kBS performed better than SNNC using (48) for U. of Pavia but worse for the Purdue and Salinas data scenes. Nevertheless, both methods performed very closely.
  • According to [54], the evaluation metrics, OA, AA, and Kappa coefficient used to evaluate classification performance are referred to as a priori measures or producer’s accuracy, based on ground truth. There is another type of evaluation metric, called precision rate (PR), referred to as a posteriori measure or user’s accuracy, based on classified data samples and addresses classification with background [55,56]. Since PR is generally used to evaluate object detection coupled with recall rate in computer vision, its usage and discussions are beyond the scope of this paper. So, no discussions of PR are included.
  • As shown by experiments, our proposed k-BDPC may not be the best. However, overall, bc-BDPC is generally the best, even if some improvements in accuracy may not be significant. There are two reasons for this. One is that most compared methods reported in the literature have chosen or tuned their parameters based on empirical performance. As a result, there would be little room to improve these methods significantly. Another reason is that our proposed methods, bc-BDPC or k-BDPC, determine their parameters, bc and k, automatically, without the need for manual tuning. But, the major advantage is that there is no need for tuning these parameters; thus, no robustness issues need to be addressed.
  • It should be noted that our developed bc-BDPC and k-BDPC are indeed steady algorithms that do not require parameters to be tuned. More specifically, the parameter bc used in bc-BDPC and the parameter k used in k-BDPC are both determined automatically in Section 2.4.1 without the need for empirical tuning. In this case, there are no robustness issues.
  • While BDPC-BS methods have shown promise in BS, there are also some weaknesses. One is that the parameters, bc for bc-BDPC and k for k-BDPC, determined by the proposed automatic rules may not be optimal, as shown in the conducted experiments. Another weakness is that, according to experimental results, k-BDPC generally does not perform as well as bc-BDPC. This may be closely related to the value of k. In this case, an alternative automatic rule needs be developed for k-BDPC.

4. Conclusions

This paper presents two BDPC-based BS methods, bc-BDPC and k-BDPC, both of which extend current existing DPC-based BS methods. The innovation and originality of this paper can be derived from the following novelties:
(a)
The first and foremost novelty is the introduction of the new concept of band prominence value (BPV) as a third indicator for DPC, which has never been explored in the literature. It can be combined with BLD and BD to yield a new BDPC score, calculated by η = BLD × BD × BPV , which can better rank bands for BS. Experimental results also demonstrate that BPV indeed enhances and further improves BS performance compared to methods that only use BLD and BD for band prioritization.
(b)
A BDPC-BS method, bc-BDPC, is developed. Specifically, an automatic rule is derived for bc-BDPC so that the cut-off band distance bc can be determined without prior knowledge and empirical determination. In particular, bc is determined by the number of clusters corresponding to the nclasses used for HSIC.
(c)
Analogous with bc-BDPC, another BDPC-BS method, k-BDPC, is developed, where an automatic rule is also derived for k-BDPC so that the number of nearest neighboring bands, k, can be determined by the number of selected bands, nBS, which is in turn determined by VD.
(d)
In order for BDPC-BS to better capture spectral characteristics, the commonly used Euclidean distance for calculating distance between data points is replaced by three types of spectral discrimination measures, namely SAM, SID, and SIDAM, to calculate spectral correlation among bands. This task cannot be accomplished by DPC-BS methods.
As a final remark, it is interesting to note that the three performance indicators, ρiin (1) and δi in (3) used by DPC, coincidentally correspond to Fisher’s ratio, within-class distance, and inter-class distance. According to recent work [57], ρi in (1), δi, and γi in (5) can be reinterpreted as three class features: intra-class feature (Intra-CF), which measures class variability within a class; inter-class feature (Inter-CF), which measures class separability between classes; and γi in (5) total class feature (TCF), which multiples Intra-CF and Inter-CF, respectively, for hyperspectral image classification (HSIC). As a consequence, a new concept of Fisher’s ratio-based DPC can be derived for HSIC. Efforts along this direction are currently underway.

Author Contributions

Conceptualization, C.-I.C.; methodology, C.-I.C. and Y.-M.K.; software, Y.-M.K. and K.Y.M.; validation, Y.-M.K. and K.Y.M.; formal analysis C.-I.C. and Y.-M.K.; investigation, C.-I.C. and Y.-M.K.; writing—C.-I.C.; writing—review and editing, C.-I.C. All authors have read and agreed to the published version of the manuscript.

Funding

The work of Chein-I Chang is supported by YuShan Fellow Program, sponsored by the Ministry of Education in Taiwan and also partly supported by the National Science and Technology Council (NSTC) under Grant 111-2634-F-006-012.

Data Availability Statement

Data available in a publicly accessible repository that does not issue DOIs. Publicly available datasets were analyzed in this study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Chang, C.-I. (Ed.) Advances in Hyperspectral Image Processing Techniques; Wiley: Hoboken, NJ, USA, 2023. [Google Scholar]
  2. Chang, C.-I. Hyperspectral Data Processing: Signal Processing Algorithm Design and Analysis; Wiley: Hoboken, NJ, USA, 2013. [Google Scholar]
  3. Chang, C.I.; Du, Q.; Sun, T.S.; Althouse, M.L.G. A Joint Band Prioritization and Band Decorrelation Approach to Band Selection for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 1999, 37, 2631–2641. [Google Scholar] [CrossRef]
  4. Feng, J.; Jiao, L.C.; Zhang, X.; Sun, T. Hyperspectral Band Selection Based on Trivariate Mutual Information and Clonal Selection. IEEE Trans. Geosci. Remote Sens. 2014, 52, 4092–4105. [Google Scholar] [CrossRef]
  5. Chang, C.I.; Kuo, Y.M.; Chen, S.; Liang, C.-C.; Ma, K.; Hu, P. Self-Mutual Information-Based Band Selection for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 5979–5997. [Google Scholar] [CrossRef]
  6. Guo, B.; Gunn, S.R.; Damper, R.I.; Nelson, J.D.B. Band Selection for Hyperspectral Image Classification Using Mutual Information. IEEE Geosci. Remote Sens. Lett. 2006, 3, 522–526. [Google Scholar] [CrossRef]
  7. Chang, C.I.; Wang, S. Constrained Band Selection for Hyperspectral Imagery. IEEE Trans. Geosci. Remote Sens. 2006, 44, 1575–1585. [Google Scholar] [CrossRef]
  8. Yang, H.; Du, Q.; Su, H.; Sheng, Y. An Efficient Method for Supervised Hyperspectral Band Selection. IEEE Geosci. Remote Sens. Lett. 2011, 8, 138–142. [Google Scholar] [CrossRef]
  9. Pudil, P.; Novovicova, J.; Kittler, J. Floating Search Methods in Feature Selection. Pattern Recogn. Lett. 1994, 15, 1119–1125. [Google Scholar] [CrossRef]
  10. Yuan, Y.; Zhu, G.; Wang, Q. Hyperspectral Band Selection by Multitask Sparisty Pursuit. IEEE Trans. Geosci. Remote Sens. 2015, 53, 631–644. [Google Scholar] [CrossRef]
  11. Zhu, G.; Huang, Y.; Lei, J.; Bi, Z.; Xu, F. Unsupervised Hyperspectral Band Selection by Dominant Set Extraction. IEEE Trans. Geosci. Remote Sens. 2016, 54, 227–239. [Google Scholar] [CrossRef]
  12. Wang, L.; Chang, C.I.; Lee, L.-C.; Wang, Y.; Xue, B.; Song, M.; Yu, C.; Li, S. Band Subset Selection for Anomaly Detection in Hyperspectral Imagery. IEEE Trans. Geosci. Remote Sens. 2017, 55, 4887–4898. [Google Scholar] [CrossRef]
  13. Chang, C.I.; Lee, L.-C.; Xue, B.; Song, M.; Chen, J. Channel Capacity Approach to Hyperspectral Band Subset Selection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 4630–4644. [Google Scholar] [CrossRef]
  14. Wang, L.; Li, H.C.; Xue, B.; Chang, C.-I. Constrained Band Subset Selection for Hyperspectral Imagery. IEEE Geosci. Remote Sens. Lett. 2017, 14, 2032–2036. [Google Scholar] [CrossRef]
  15. Yu, C.; Song, M.; Chang, C.-I. Band Subset Selection for Hyperspectral Image Classification. Remote Sens. 2018, 10, 113. [Google Scholar] [CrossRef]
  16. Yu, C.; Song, M.; Wang, Y.; Chang, C.-I. Class Signature-Constrained Background-Suppressed Approach to Band Selection for Classification of Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 14–31. [Google Scholar] [CrossRef]
  17. Gong, M.; Zhang, M.; Yuan, Y. Unsupervised Band Selection Based on Evolutionary Multiobjective Optimization for Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 544–557. [Google Scholar] [CrossRef]
  18. Zhang, M.; Ma, J.; Gong, M. Unsupervised Hyperspectral Band Selection by Fuzzy Clustering with Particle Swarm Optimization. IEEE Geosci. Remote Sens. Lett. 2017, 14, 773–777. [Google Scholar] [CrossRef]
  19. Xu, Y.; Du, Q.; Younan, N.H. Particle Swarm Optimization-Based Band Selection for Hyperspectral Target Detection. IEEE Geosci. Remote Sens. Lett. 2017, 14, 554–558. [Google Scholar] [CrossRef]
  20. Su, H.; Yong, B.; Du, Q. Hyperspectral Band Selection Using Improved Firefly Algorithm. IEEE Geosci. Remote Sens. Lett. 2016, 13, 68–72. [Google Scholar] [CrossRef]
  21. Gao, J.; Du, Q.; Gao, L.; Sun, X.; Wu, Y.; Zhang, B. Ant Colony Optimization for Supervised and Unsupervised Hyperspectral Band Selection. In Proceedings of the 2013 5th Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Gainesville, FL, USA, 26–28 June 2013; pp. 1–4. [Google Scholar]
  22. Wang, Q.; Li, Q.; Li, X. Hyperspectral Band Selection via Adaptive Subspace Partition Strategy. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 4940–4950. [Google Scholar] [CrossRef]
  23. Wang, Q.; Li, Q.; Li, X. A Fast Neighborhood Grouping Method for Hyperspectral Band Selection. IEEE Trans. Geosci. Remote Sens. 2020, 58, 608–619. [Google Scholar] [CrossRef]
  24. Wang, Q.; Zhang, F.; Li, X. Hyperspectral Band Selection via Optimal Neighborhood Reconstruction. IEEE Trans. Geosci. Remote Sens. 2020, 58, 8465–8476. [Google Scholar] [CrossRef]
  25. Baisantry, M.; ASao, A.K.; Shukla, D. Band Selection Using Combined Divergence-Correlation Index and Sparse Loadings Representation for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 4630–4644. [Google Scholar] [CrossRef]
  26. Rodriguez, A.; Laio, A. Clustering by Fast Search and Find of Density Peaks. Science 2014, 344, 1492–1496. [Google Scholar] [CrossRef] [PubMed]
  27. Chang, C.-I. Hyperspectral Imaging: Spectral Techniques for Detection and Classification; Kluwer: Norwell, MA, USA, 2003. [Google Scholar]
  28. Chang, C.-I.; Du, Q. Estimation of Number of Spectrally Distinct Signal Sources in Hyperspectral Imagery. IEEE Trans. Geosci. Remote Sens. 2004, 42, 608–619. [Google Scholar] [CrossRef]
  29. Chang, C.-I. An Information-Theoretic Approach to Spectral Variability, Similarity, and Discrimination for Hyperspectral Image Analysis. IEEE Trans. Inf. Theory 2000, 46, 1927–1932. [Google Scholar] [CrossRef]
  30. Du, Y.; Chang, C.-I.; Ren, H.; Chang, C.-C.; Jensen, J.O. New Hyperspectral Discrimination Measure for Spectral Characterisation. Opt. Eng. 2004, 43, 1777–1786. [Google Scholar]
  31. Wei, X.; Peng, M.; Huang, H.; Zhou, Y. An Overview on Density Peaks Clustering. Neurocomputing 2023, 554, 126633. [Google Scholar] [CrossRef]
  32. Tobin, J.; Zhang, M. A Theoretical Analysis of Density Peaks Clustering and the Component-Wise Peak-Finding Algorithm. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 1109–1120. [Google Scholar] [CrossRef]
  33. Xu, X.; Liao, H.; Yan, X. An Automatic Density Peaks Clustering Based on a Density-Distance Clustering Index. AIMS Math. 2023, 8, 28926–28950. [Google Scholar] [CrossRef]
  34. Wang, Y.; Wang, D.; Zhou, Y.; Zhang, X.; Quek, C. VDPC: Variational Density Peak Clustering Algorithm. Inf. Sci. 2023, 621, 627–651. [Google Scholar] [CrossRef]
  35. Du, M.; Ding, S.; Jia, H. Study on Density Peaks Clustering Based on K-Nearest Neighbors and Principal Components. Knowl.-Based Syst. 2016, 99, 135–145. [Google Scholar] [CrossRef]
  36. Liu, L.; Yu, D. Density Peaks Clustering Algorithm Based on Weighted K-Nearest Neighbors and Geodesic Distance. IEEE Access 2020, 8, 168282–168296. [Google Scholar] [CrossRef]
  37. Sun, L.; Qin, X.; Ding, W.; Xu, J.; Zhang, S. Density Peaks Clustering Based on K-Nearest Neighbors and Self-Recommendation. Int. J. Mach. Learn. Cybern. 2021, 12, 1913–1938. [Google Scholar] [CrossRef]
  38. Yuan, X.; Yu, H.; Liang, J.; Xu, B. A Novel Density Peaks Clustering Algorithm Based on K Nearest Neighbors with Adaptive Merging Strategy. Int. J. Mach. Learn. Cybern. 2021, 12, 2825–2841. [Google Scholar] [CrossRef]
  39. Wang, L.; Zhou, W.; Wang, H.; Parmar, M.; Han, H. A Novel Density Peaks Clustering Halo Node Assignment Method Based on K-Nearest Neighbor Theory. IEEE Access 2019, 7, 174380–174390. [Google Scholar] [CrossRef]
  40. Yin, L.; Wang, Y.; Chen, H.; Deng, W. An Improved Density Peak Clustering Algorithm for Multi-Density Data. Sensors 2022, 22, 8814. [Google Scholar] [CrossRef] [PubMed]
  41. Jiang, D.; Zang, W.; Sun, R.; Wang, Z.; Liu, X. Adaptive Density Peaks Clustering Based on K-Nearest Neighbor and Gini Coefficient. IEEE Access 2020, 8, 113900–113917. [Google Scholar] [CrossRef]
  42. Anandarao, S.; Chellasamy, S.H. A Comprehensive Study on Density Peak Clustering and its Variants. Int. J. Intell. Syst. Appl. 2023, 11, 216–224. [Google Scholar]
  43. Wang, Z.; Wang, H.; Du, H.; Chen, S.; Shi, X. A Novel Density Peaks Clustering Algorithm for Automatic Selection of Clustering Centers Based on K-Nearest Neighbors. Math. Biosci. Eng. 2023, 20, 11875–11894. [Google Scholar] [CrossRef]
  44. Jarvis, R.A.; Patrick, E.A. Clustering Using a Similarity Measure Based on Shared Near Neighbors. IEEE Trans. Comput. 1973, 22, 1025–1034. [Google Scholar] [CrossRef]
  45. Liu, R.; Wang, H.; Yu, X. Shared-Nearest-Neighbor-Based Clustering by Fast Search and Find of Density Peaks. Inf. Sci. 2018, 450, 200–226. [Google Scholar] [CrossRef]
  46. Lv, Y.; Liu, M.; Xiang, Y. Fast Searching Density Peak Clustering Algorithm Based on Shared Nearest Neighbor and Adaptive Clustering Center. Symmetry 2020, 12, 2014. [Google Scholar] [CrossRef]
  47. Sun, K.; Geng, X.; Ji, L. Exemplar Component Analysis: A Fast Band Selection Method for Hyperspectral Imagery. IEEE Geosci. Remote Sens. Lett. 2015, 12, 998–1002. [Google Scholar]
  48. Jia, S.; Tang, G.; Zhu, J.; Li, Q. A Novel Ranking-Based Clustering Approach for Hyperspectral Band Selection. IEEE Trans. Geosci. Remote Sens. 2016, 54, 88–102. [Google Scholar] [CrossRef]
  49. Luo, X.; Xue, R.; Yin, J. Information-Assisted Density Peak Index for Hyperspectral Band Selection. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1870–1874. [Google Scholar] [CrossRef]
  50. Li, Q.; Wang, Q.; Li, X. An Efficient Clustering Method for Hyperspectral Optimal Band Selection via Shared Nearest Neighbor. Remote Sens. 2019, 11, 350. [Google Scholar] [CrossRef]
  51. Duda, R.O.; Hart, P.E. Pattern Classification and Scene Analysis; John Wiley & Sons: Hoboken, NJ, USA, 1973. [Google Scholar]
  52. Zhong, S.; Chang, C.-I.; Zhang, Y. Iterative edge preserving filtering approach to hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2019, 16, 90–94. [Google Scholar] [CrossRef]
  53. Kang, X.; Li, S.; Benediktsson, J.A. Spectral–Spatial Hyperspectral Image Classification with Edge-Preserving Filtering. IEEE Trans. Geosci. Remote Sens. 2014, 52, 2666–2677. [Google Scholar] [CrossRef]
  54. Chang, C.-I. A Statistical DetectionTheory Approach to Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 2057–2074. [Google Scholar] [CrossRef]
  55. Chang, C.-I.; Liang, C.C.; Hu, P. Iterative Convolutional Neural Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5513526, 1–26. [Google Scholar]
  56. Chang, C.-I.; Liang, C.C.; Hu, P. Iterative Gaussian Laplacian Pyramid Network for Hypersspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–22. [Google Scholar]
  57. Zhong, S.; Chang, C.-I.; Li, J.J.; Shang, X.; Chen, S.; Song, M.; Zhang, Y. Class feature weighted hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 4728–4745. [Google Scholar] [CrossRef]
Figure 1. The vertical line represents the prominence of peak C, which is the difference between the height of C and the height of v1.
Figure 1. The vertical line represents the prominence of peak C, which is the difference between the height of C and the height of v1.
Remotesensing 16 00942 g001
Figure 2. AVIRIS image scene: Purdue Indiana Indian Pines test site. (a) Band 186 (2162.56 nm). (b) Ground truth map. (c) Classes by colors.
Figure 2. AVIRIS image scene: Purdue Indiana Indian Pines test site. (a) Band 186 (2162.56 nm). (b) Ground truth map. (c) Classes by colors.
Remotesensing 16 00942 g002
Figure 3. Ground-truth of Salinas scene with 16 classes. (a) Salinas scene. (b) Ground-truth image. (c) Classes by colors.
Figure 3. Ground-truth of Salinas scene with 16 classes. (a) Salinas scene. (b) Ground-truth image. (c) Classes by colors.
Remotesensing 16 00942 g003
Figure 4. Ground-truth of University of Pavia scene with nine classes. (a) University of Pavia scene. (b) Ground-truth map. (c) Classes by colors.
Figure 4. Ground-truth of University of Pavia scene with nine classes. (a) University of Pavia scene. (b) Ground-truth map. (c) Classes by colors.
Remotesensing 16 00942 g004
Figure 5. η i b c - BDPC values by bc-BDPC for the Purdue data using (a) SID, (b) SAM, (c) SIDAM.
Figure 5. η i b c - BDPC values by bc-BDPC for the Purdue data using (a) SID, (b) SAM, (c) SIDAM.
Remotesensing 16 00942 g005
Figure 6. η i k - BDPC values by k-BDPC for the Purdue data using (a) SID, (b) SAM, (c) SIDAM.
Figure 6. η i k - BDPC values by k-BDPC for the Purdue data using (a) SID, (b) SAM, (c) SIDAM.
Remotesensing 16 00942 g006
Figure 7. IEPF-G-g classification maps produced by different BS methods for Purdue data. (a) Full bands (OA = 95.71%), (b) UBS (OA = 96.81%), (c) bc-BDPC-SID (OA = 98.08%),(d) bc-BDPC-SAM (OA = 98.12%), (e) bc-BDPC-SIDAM (OA =97.70%), (f) k-BDPC-SID (OA = 97.53%), (g) k -BDPC-SAM (OA = 97.91%), (h) k-BDPC-SIDA (OA = 97.78%), (i) SMI-BS (OA = 97.57%), (j) ECA (OA = 96.87%), (k) E-FDPC (OA = 97.80%), (l) IaDPI (OA = 97.59%), (m) DPC-KNN (kBS) (OA = 91.77%), (n) DPC-KNN (k = 2%L) (OA = 98.05%), (o) G-DPC-KNN (kBS) (OA = 89.33%), (p) G-DPC-KNN (k = 10) (OA = 97.65%), (q) SNNC (kBS) (OA = 97.14%), (r) SNNC (k = 3) (OA = 97.66%).
Figure 7. IEPF-G-g classification maps produced by different BS methods for Purdue data. (a) Full bands (OA = 95.71%), (b) UBS (OA = 96.81%), (c) bc-BDPC-SID (OA = 98.08%),(d) bc-BDPC-SAM (OA = 98.12%), (e) bc-BDPC-SIDAM (OA =97.70%), (f) k-BDPC-SID (OA = 97.53%), (g) k -BDPC-SAM (OA = 97.91%), (h) k-BDPC-SIDA (OA = 97.78%), (i) SMI-BS (OA = 97.57%), (j) ECA (OA = 96.87%), (k) E-FDPC (OA = 97.80%), (l) IaDPI (OA = 97.59%), (m) DPC-KNN (kBS) (OA = 91.77%), (n) DPC-KNN (k = 2%L) (OA = 98.05%), (o) G-DPC-KNN (kBS) (OA = 89.33%), (p) G-DPC-KNN (k = 10) (OA = 97.65%), (q) SNNC (kBS) (OA = 97.14%), (r) SNNC (k = 3) (OA = 97.66%).
Remotesensing 16 00942 g007
Figure 8. η i b c - BDPC values by bc-BDPC for Salinas using (a) SID, (b) SAM, (c) SIDAM.
Figure 8. η i b c - BDPC values by bc-BDPC for Salinas using (a) SID, (b) SAM, (c) SIDAM.
Remotesensing 16 00942 g008
Figure 9. η i k - BDPC values by k-BDPC for Salinas using (a) SID, (b) SAM, (c) SIDAM.
Figure 9. η i k - BDPC values by k-BDPC for Salinas using (a) SID, (b) SAM, (c) SIDAM.
Remotesensing 16 00942 g009
Figure 10. Classification maps produced by IEPF-G-g using different BS criterion for Salinas data. (a) Full bands (OA = 96.24%), (b) UBS (OA = 95.69%), (c) bc-BDPC-SID (OA = 95.28%), (d) bc-BDPC-SAM (OA = 96.32%), (e) bc-BDPC-SIDAM (OA = 96.09%), (f) k-BDPC-SID (OA = 91.89%), (g) k -BDPC-SAM (OA = 90.09%), (h) k-BDPC-SIDAM (OA = 90.50%), (i) SMI-BS (OA = 95.32%), (j) ECA (OA = 96.32%), (k) E-FDPC (OA =96.47%), (l) IaDPI (OA = 95.86%), (m) DPC-KNN (kBS) (OA =96.86%), (n) DPC-KNN (k = 2%L) (OA = 95.16%), (o) G-DPC-KNN (kBS) (OA = 95.05%), (p) G-DPC-KNN (k = 10) (OA = 95.81%), (q) SNNC (kBS) (OA = 95.29%), (r) SNNC (k = 3) (OA = 96.91%).
Figure 10. Classification maps produced by IEPF-G-g using different BS criterion for Salinas data. (a) Full bands (OA = 96.24%), (b) UBS (OA = 95.69%), (c) bc-BDPC-SID (OA = 95.28%), (d) bc-BDPC-SAM (OA = 96.32%), (e) bc-BDPC-SIDAM (OA = 96.09%), (f) k-BDPC-SID (OA = 91.89%), (g) k -BDPC-SAM (OA = 90.09%), (h) k-BDPC-SIDAM (OA = 90.50%), (i) SMI-BS (OA = 95.32%), (j) ECA (OA = 96.32%), (k) E-FDPC (OA =96.47%), (l) IaDPI (OA = 95.86%), (m) DPC-KNN (kBS) (OA =96.86%), (n) DPC-KNN (k = 2%L) (OA = 95.16%), (o) G-DPC-KNN (kBS) (OA = 95.05%), (p) G-DPC-KNN (k = 10) (OA = 95.81%), (q) SNNC (kBS) (OA = 95.29%), (r) SNNC (k = 3) (OA = 96.91%).
Remotesensing 16 00942 g010
Figure 11. η i b c - BDPC values by bc-BDPC for U. of Pavia using (a) SID, (b) SAM, (c) SIDAM.
Figure 11. η i b c - BDPC values by bc-BDPC for U. of Pavia using (a) SID, (b) SAM, (c) SIDAM.
Remotesensing 16 00942 g011
Figure 12. η i k - BDPC values by k-BDPC for U. of Pavia using (a) SID, (b) SAM, (c) SIDAM.
Figure 12. η i k - BDPC values by k-BDPC for U. of Pavia using (a) SID, (b) SAM, (c) SIDAM.
Remotesensing 16 00942 g012
Figure 13. Classification maps produced by IEPF-G-g using different BS criterion for University of Pavia data. (a) Full bands (OA = 98.23%), (b) UBS (OA = 96.87%), (c) bc-BDPC-SID (OA = 97.82%), (d) bc-BDPC-SAM (OA = 97.83%), (e) bc-BDPC-SIDAM (OA = 98.13%), (f) k-BDPC-SID (OA = 89.12%), (g) k -BDPC-SAM (OA = 9707%), (h) k-BDPC-SIDA (OA = 96.63%), (i) SMI-BS (OA = 97.46%), (j) ECA (OA = 95.76%), (k) E-FDPC (OA =97.95%), (l) IaDPI (OA = 97.00%), (m) DPC-KNN (kBS) (OA = 97.79%), (n) DPC-KNN (k = 2%L) (OA = 96.07%), (o) G-DPC-KNN (kBS) (OA = 96.93%), (p) G-DPC-KNN (k = 10) (OA = 96.79%), (q) SNNC (kBS) (OA = 97.39%), (r) SNNC (k = 3) (OA = 96.30%).
Figure 13. Classification maps produced by IEPF-G-g using different BS criterion for University of Pavia data. (a) Full bands (OA = 98.23%), (b) UBS (OA = 96.87%), (c) bc-BDPC-SID (OA = 97.82%), (d) bc-BDPC-SAM (OA = 97.83%), (e) bc-BDPC-SIDAM (OA = 98.13%), (f) k-BDPC-SID (OA = 89.12%), (g) k -BDPC-SAM (OA = 9707%), (h) k-BDPC-SIDA (OA = 96.63%), (i) SMI-BS (OA = 97.46%), (j) ECA (OA = 95.76%), (k) E-FDPC (OA =97.95%), (l) IaDPI (OA = 97.00%), (m) DPC-KNN (kBS) (OA = 97.79%), (n) DPC-KNN (k = 2%L) (OA = 96.07%), (o) G-DPC-KNN (kBS) (OA = 96.93%), (p) G-DPC-KNN (k = 10) (OA = 96.79%), (q) SNNC (kBS) (OA = 97.39%), (r) SNNC (k = 3) (OA = 96.30%).
Remotesensing 16 00942 g013
Table 1. Class labels of Purdue Indiana Indian Pines with number of data samples in each class.
Table 1. Class labels of Purdue Indiana Indian Pines with number of data samples in each class.
class 1 (46)Alfalfaclass 7 (28)grass/pasture-mowedclass 13 (205)wheat
class 2 (1428)corn-notillclass 8 (478)hay-windrowedclass 14 (1265)woods
class 3 (830)corn-minclass 9 (20)oatsclass 15 (386)bldg-grass green-drives
class 4 (237)cornclass 10 (972)soybeans-notillclass 16 (93)stone-steel towers
class 5 (483)grass/pastureclass 11 (2455)soybeans-minclass 17 (10,249)BKG
class 6 (730)grass/treesclass 12 (593)soybeans-clean
Table 2. Class labels of Salinas with number of data samples in each class.
Table 2. Class labels of Salinas with number of data samples in each class.
class 1 (2009)Brocoli_green_weeds_1class 10 (3278)Corn_senesced_green_weeds
class 2 (3726)Brocoli_green_weeds_2class 11 (1068)Lettuce_romaine_4wk
class 3 (1976)Fallowclass 12 (1927)Lettuce_romaine_5wk
class 4 (1394)Fallow_rough_plowclass 13 (916)Lettuce_romaine_6wk
class 5 (2678)Fallow_smoothclass 14 (1070)Lettuce_romaine_7wk
class 6 (3959)StubbleClass 15 (7268)Vinyard_untrained
class 7 (3579)Celeryclass 16 (1807)Vinyard_vertical_trellis
class 8 (11,271)Grapes_untrainedclass 17 (56,975)BKG
class 9 (6203)Soil_vinyard_develop
Table 3. Class labels of University of Pavia with number of data samples in each class.
Table 3. Class labels of University of Pavia with number of data samples in each class.
class 1 (6631)Asphaltclass 5 (1345)Painted metal sheetsClass 9 (947)Shadows
class 2 (18,649)Meadowsclass 6 (5029)Bare SoilClass 10 (164,624)BKG
class 3 (2099)Gravelclass 7 (1330)Bitumen
class 4 (3064)Treesclass 8 (3682)Self-Blocking Bricks
Table 4. Number of training samples of each class for three datasets.
Table 4. Number of training samples of each class for three datasets.
Class12345678910111213141516Total
Purdue2591806568731470118111170668469471025
Salinas676767676867696967696767676868691083
U. of Pavia100100100100100100100100100 900
Table 5. Parameter setting for compared DPC-BS methods.
Table 5. Parameter setting for compared DPC-BS methods.
DPC Methodskbc
bc-BDPC nclusters
IaDPI 2% × L × (L − 1)/exp(nBS/L)
k-BDPC k BS = 2 × L / n BS
DPC-kNN k N N = p % × N
G-DPC-kNNemperical selectionGini coefficient
SNNCk = 3
Table 6. 18 bands selected for the Purdue data.
Table 6. 18 bands selected for the Purdue data.
Selected Bands (nVD = 18)
bc-BDPC-SID219671918828134482111624135123343678378
bc-BDPC-SAM13668178291222112896952618952232135181934
bc-BDPC-SIDAM1896828881361591550212132212212035342336
k-BDPC-SID2935421361223777992327534132331274
k -BDPC-SAM29423513612037779826313233216643134
k-BDPC-SIDAM29354213612137779934232633131322725
SMI-BS591171253840180173201683134121141921601089592
ECA1641296783193521419728110147103112111149208165107
E-FDPC1943221009861136359960311017662135859
IaDPI1711158208875814246771353117697190159254160
DPC-KNN (kBS)150136187146147119145116103110114101112115125111205208
DPC-KNN (k = 2%L)373627546157614247100323473562183
G-DPC-KNN (kBS)204206203205148202112207209167201656970126129132133
G-DPC-KNN (k = 10)1101811003711761145467660365875783479
SNNC (kBS)1043137775838261119605610082764036
SNNC (k = 3)1804984651321068177142212188822075776
uniform BS114274053667992105118131144157170183196209220
Table 7. Quantitative comparative analysis of classification for the Purdue data among bc-BDPC and k-BDPC using, SID, SAM, and SIDAM along with full bands and uniform BS, with the best and second-best results highlighted and boldfaced in red and black, respectively.
Table 7. Quantitative comparative analysis of classification for the Purdue data among bc-BDPC and k-BDPC using, SID, SAM, and SIDAM along with full bands and uniform BS, with the best and second-best results highlighted and boldfaced in red and black, respectively.
IEPF-G-g
Classbc-BDPC-SIDbc-BDPC-SAMbc-BDPC-SIDAMk-BDPC-SIDk-BDPC-SAMk-BDPC-SIDAMUniform BSFull Bands
11.00000.99571.00001.00001.00001.00000.99780.9957
20.96460.96600.96440.95900.96370.95680.92930.9302
30.98860.98330.98650.98570.98770.98830.98390.9619
40.99750.99870.99790.99830.99920.99871.00000.9899
50.97830.97990.97830.97620.97850.97560.97640.9627
60.99960.99930.99990.99960.99970.99930.99860.9947
70.99290.98930.99640.99290.99290.99290.98930.9893
81.00001.00001.00001.00001.00001.00001.00001.0000
91.00001.00001.00001.00001.00001.00001.00001.0000
100.96970.97860.96430.97700.97730.97310.95510.9399
110.96560.96540.95270.94460.95590.95300.94110.9224
120.99210.98740.98920.99140.99170.99210.99160.9855
130.99760.99610.99710.99760.99710.99760.99660.9966
140.99620.99750.99720.99630.99720.99630.99620.9837
150.99690.99400.99820.99510.99530.99530.99690.9886
160.99680.99780.99780.99890.99780.99890.99890.9957
POA0.98080.98120.97700.97530.97910.97680.96810.9571
PAA0.98980.98930.98870.98830.98960.98860.98450.9773
Kappa0.97790.97860.97400.97210.97630.97360.96380.9514
time (s)2738542769
Table 8. Quantitative comparative analysis of classification for the Purdue data among SMI-BS, ECA, E-FDPC, IaDPI, DPC-KNN, G-DPC-KNN, and SNNC, with the best and second-best results highlighted and boldfaced in red and black, respectively.
Table 8. Quantitative comparative analysis of classification for the Purdue data among SMI-BS, ECA, E-FDPC, IaDPI, DPC-KNN, G-DPC-KNN, and SNNC, with the best and second-best results highlighted and boldfaced in red and black, respectively.
IEPF-G-g
ClassSMI-BSECAE-FDPCIaDPIDPC-KNN (kBS)DPC-KNN
(k = 2%L)
G-DPC-KNN (kBS)G-DPC-KNN
(k = 10)
SNNC (kBS)SNNC
(k = 3)
11.00001.00000.99780.99781.00000.99570.96090.99780.99570.9957
20.95750.94400.96290.95380.89650.97230.87630.95670.95020.9491
30.98510.98000.98410.98780.91670.98980.88480.98370.98550.9887
40.99870.99870.99750.99750.97640.99750.94350.99920.99700.9970
50.97970.97830.97620.98160.94820.97230.88740.97160.97120.9789
60.99930.99810.99840.99890.92960.99700.87880.99640.99680.9989
70.99290.98930.99640.99291.00000.98570.97500.98570.98570.9964
81.00001.00001.00001.00000.98620.99980.94271.00001.00001.0000
91.00001.00001.00001.00000.96501.00000.93501.00001.00001.0000
100.96790.95310.97790.97230.90320.97670.86060.96770.96290.9660
110.95050.93740.95450.95090.87270.96040.91240.95760.94310.9598
120.99220.99210.99210.99340.92900.99310.86630.99440.99330.9907
130.99510.99710.99660.99610.98150.99760.94440.99800.99800.9980
140.99630.99470.99680.99470.94350.99570.87360.99600.99420.9953
150.99530.99740.99170.99590.97330.98760.92540.98990.98780.9959
160.99891.00000.99680.99780.96450.99250.95700.99780.99570.9978
POA0.97570.96870.97800.97590.91770.98050.89330.97650.97140.9766
PAA0.98810.98500.98870.98820.94910.98830.91400.98700.98480.9880
Kappa0.97230.96440.97510.97240.90710.97790.87980.97340.96770.9734
time (s)360.20.20.30.20.20.30.30.40.7
Table 9. 21 bands selected for Salinas.
Table 9. 21 bands selected for Salinas.
Selected Bands (nVD = 21)
bc-BDPC-SID9114115715615570159108192164160165311111661101631542241852
bc-BDPC -SAM9213515670162187161109165110111166154311125218108159107153
bc-BDPC -SIDAM921561551571641591631601611651081351111541661927011011222431
k-BDPC-SID15716015916415510816316511016211115416622411122221071671132
k -BDPC -SAM1561641081601551541091651591101661111571631122241531621671071
k-BDPC -SIDAM1561641591601551571081631651621101111541662241122221071671113
SMI-BS15323440631171501571842062232241647194118151176185202
ECA109184886813621132194218552159615393118204224160161166180
E-FDPC1936110454103786481102939105236833838010684
IaDPI154381071948651401989167101112159136963385133175170162
DPC-KNN (kBS)11124897319712910451710564263210813371536106103
DPC-KNN (k = 6%L)15915615715816316216410816511011115515411222221213722092161217
G-DPC-KNN (kBS)204205202150209206207119203118117170362223242721171525
G-DPC-KNN (k = 7)38645749613413363108109510445437888738645749
SNNC (kBS)39121422311712341104638543641265839121422
SNNC (k = 3)1835130141817324484366968219310254183513014
Uniform BS11223344556677889100111122133144155166177188199210224
Table 10. Quantitative comparative analysis of classification for Salinas among bc-BDPC and k-BDPC using SID, SAM, and SIDAM, along with full bands and uniform BS, with the best and second-best results highlighted and boldfaced in red and black, respectively.
Table 10. Quantitative comparative analysis of classification for Salinas among bc-BDPC and k-BDPC using SID, SAM, and SIDAM, along with full bands and uniform BS, with the best and second-best results highlighted and boldfaced in red and black, respectively.
IEPF-G-g
Classbc-BDPC-SIDbc-BDPC-SAMbc-BDPC-SIDAMk-BDPC-SIDk-BDPC-SAMk-BDPC-SIDAMUBSFull Bands
10.99900.99790.99830.99240.98550.97950.99981.0000
20.99090.99140.99370.98370.97490.96120.99640.9967
30.99890.99780.99350.99450.99070.98440.99970.9986
40.99450.99440.99630.99190.98960.98510.99860.9973
50.98720.98860.98510.97570.98010.97860.99160.9899
60.99610.99620.99460.99830.99970.99370.99880.9978
70.99430.99260.99200.99400.99440.98060.99630.9958
80.88760.90580.90480.76490.71550.75590.88860.8944
90.99460.99290.98480.99340.99030.98930.99570.9964
100.98680.98830.97390.97380.97020.96120.98890.9820
110.99340.99250.99440.99490.99660.98290.99820.9960
120.99920.99770.99820.98660.99550.98531.00001.0000
130.98930.99020.98800.99880.99790.98480.99270.9936
140.98740.99040.98800.99310.98570.98590.99320.9894
150.89680.90840.90980.81270.76210.76460.87280.9098
160.98780.98260.97760.98720.99250.98770.99030.9901
POA0.95820.96320.96090.91890.90090.90500.95690.9624
PAA0.98020.98170.97960.96470.95760.95380.98140.9830
Kappa0.95370.95290.95670.91060.89110.89550.95230.9583
time (s)139827522848278
Table 11. Quantitative comparative analysis of classification for Salinas among SMI-BS, ECA, E-FDPC, IaDPI, DPC-KNN, G-DPC-KNN, and SNNC, with the best and second-best results highlighted and boldfaced in red and black, respectively.
Table 11. Quantitative comparative analysis of classification for Salinas among SMI-BS, ECA, E-FDPC, IaDPI, DPC-KNN, G-DPC-KNN, and SNNC, with the best and second-best results highlighted and boldfaced in red and black, respectively.
IEPF-G-g
ClassSMI-BSECAE-FDPCIaDPIDPC-KNN (kBS)DPC-KNN (k = 2%L)G-DPC-KNN (kBS)G-DPC-KNN (k = 7)SNNC (kBS)SNN
(k = 3)
10.99980.99991.00001.00000.99990.99590.99400.99991.00000.9999
20.99470.99480.99790.99890.99770.98570.99770.99730.99790.9976
30.99960.99910.99920.99920.99990.99200.98790.99900.99970.9997
40.99790.99810.99660.99850.99580.99460.99510.99810.99760.9961
50.99090.98740.98960.98790.99290.98250.97710.98840.98820.9926
60.99820.99780.99940.99870.99900.99300.99220.99900.99930.9987
70.99550.99600.99680.99630.99630.99070.99280.99560.99610.9963
80.86110.90040.89970.87730.91780.87840.90370.87100.86520.9160
90.99450.99500.99980.99630.99600.98670.98800.99770.99930.9968
100.98220.98240.98290.98840.98540.96780.98030.98550.98640.9882
110.99610.99470.99630.99700.99780.98980.98900.99650.99710.9970
121.00000.99961.00000.99991.00000.98840.99341.00001.00000.9999
130.99360.99400.99630.99760.99310.99220.99560.99440.99510.9942
140.99150.99200.99510.99330.99180.98410.98730.99280.99170.9931
150.89460.90910.91210.90260.91540.89460.83930.90950.87630.9201
160.99120.98830.99180.98880.99070.97360.98690.99390.99510.9900
POA0.95320.96320.96470.95860.96860.95160.95150.95810.95290.9691
PAA0.98010.98300.98460.98260.98560.97440.97500.98240.98030.9860
Kappa0.94830.95910.96090.95410.96520.94640.94620.95360.94780.9657
time (s)1491.81.82.11.91.81.71.71.91.9
Table 12. 14 bands selected for U. of Pavia.
Table 12. 14 bands selected for U. of Pavia.
Selected Bands (nVD = 14)
bc-BDPC-SID936212371470325726967
bc-BDP-SAM9155311984128258507259381
bc-BDPC-SIDAM9161123717043272569196
k-BDPC-SID172234530615717077369
k -BDPC-SAM1722342757173707466916
k-BDPC-SIDAM172234530166717073697
SMI-BS224191918213740485766829294
ECA6188534633625663496092475754
E-FDPC5093248482243685676916
IaDPI743259884455141947675737772
DPC-KNN (kBS)58931635123482698477552
DPC-KNN (k = 2%L)6063615753476451494589329155
G-DPC-KNN (kBS)2827292630313635333432797877
G-DPC-KNN (k = 7)7571726867767469736058565053
SNNC (kBS)82524913472996931186
SNNC (k = 5)92591723417510067589
Uniform BS191725334149576573818997103
Table 13. Quantitative comparative analysis of classification for U. of Pavia among bc-BDPC and k-BDPC using SID, SAM, and SIDAM, along with full bands and uniform BS, with the best and second-best results highlighted and boldfaced in red and black, respectively.
Table 13. Quantitative comparative analysis of classification for U. of Pavia among bc-BDPC and k-BDPC using SID, SAM, and SIDAM, along with full bands and uniform BS, with the best and second-best results highlighted and boldfaced in red and black, respectively.
IEPF-G-g
Classbc-BDPC-SIDbc-BDPC-SAMbc-BDPC-SIDAMk-BDPC-SIDk-BDPC-SAMk-BDPC-SIDAMUniform BSFull Bands
10.97650.98110.98450.90640.98610.98770.98270.9816
20.97080.96580.97510.85740.95580.94520.94340.9738
30.97840.98810.98420.90410.98180.98170.98740.9915
40.98870.98740.98650.90790.98030.98050.99290.9870
50.99920.99950.99960.93200.99890.99880.99770.9984
60.99190.99400.98980.89310.97710.97270.98320.9967
70.99980.99960.99990.92651.00000.99990.99690.9997
80.97060.98120.97120.97470.96850.97340.98810.9816
90.99820.99950.99920.92350.99550.99620.99600.9998
POA0.97820.97830.98130.89120.97070.96630.96830.9823
PAA0.98600.98850.98780.91400.98260.98180.98540.9900
Kappa0.97130.97150.97540.86020.96160.95580.95840.9767
time (s)851411816538189
Table 14. Quantitative comparative analysis of classification for U. of Pavia among SMI-BS, ECA, E-FDPC, IaDPI, DPC-KNN, G-DPC-KNN, and SNNC, with the best and second-best results highlighted and boldfaced in red and black, respectively.
Table 14. Quantitative comparative analysis of classification for U. of Pavia among SMI-BS, ECA, E-FDPC, IaDPI, DPC-KNN, G-DPC-KNN, and SNNC, with the best and second-best results highlighted and boldfaced in red and black, respectively.
IEPF-G-g
ClassSMI-BSECAE-FDPCIaDPIDPC-KNN (kBS)DPC-KNN
(k = 2%L)
G-DPC-KNN (kBS)G-DPC-KNN
(k = 7)
SNNC (kBS)SNNC
(k = 5)
10.97820.97010.98330.97770.98440.97660.97230.97030.98230.9783
20.96130.93600.96840.95080.96580.93530.95590.95590.95840.9403
30.98790.97910.98840.98340.98410.98600.98700.97310.98710.9809
40.99330.99010.98980.98950.98980.98910.98760.98830.99190.9875
50.99670.99910.99910.99840.99800.99830.99840.99870.99810.9949
60.98730.95680.99220.98510.99230.96870.97050.97920.98470.9692
70.99720.99370.99920.99790.99870.99580.99680.99810.99860.9950
80.97330.96880.97840.98070.97510.97650.97690.96000.97580.9802
90.99570.99390.99870.99880.99610.99310.99740.99460.99710.9937
POA0.97460.95760.97950.97000.97790.96070.96930.96790.97390.9630
PAA0.98570.97640.98860.98470.98710.97990.98250.97980.98600.9800
Kappa0.96660.94460.97300.96060.97090.94860.95980.95790.96560.9516
time(s)921.11.11.81.01.11.01.11.41.4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chang, C.-I.; Kuo, Y.-M.; Ma, K.Y. Band Selection via Band Density Prominence Clustering for Hyperspectral Image Classification. Remote Sens. 2024, 16, 942. https://doi.org/10.3390/rs16060942

AMA Style

Chang C-I, Kuo Y-M, Ma KY. Band Selection via Band Density Prominence Clustering for Hyperspectral Image Classification. Remote Sensing. 2024; 16(6):942. https://doi.org/10.3390/rs16060942

Chicago/Turabian Style

Chang, Chein-I, Yi-Mei Kuo, and Kenneth Yeonkong Ma. 2024. "Band Selection via Band Density Prominence Clustering for Hyperspectral Image Classification" Remote Sensing 16, no. 6: 942. https://doi.org/10.3390/rs16060942

APA Style

Chang, C. -I., Kuo, Y. -M., & Ma, K. Y. (2024). Band Selection via Band Density Prominence Clustering for Hyperspectral Image Classification. Remote Sensing, 16(6), 942. https://doi.org/10.3390/rs16060942

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop