Next Article in Journal
Time-Series Cross-Radiometric Calibration and Validation of GF-6/WFV Using Multi-Site
Next Article in Special Issue
Advanced Machine Learning and Deep Learning Approaches for Remote Sensing II
Previous Article in Journal
Spatiotemporal Analysis of Land Surface Temperature in Response to Land Use and Land Cover Changes: A Remote Sensing Approach
Previous Article in Special Issue
A Parameter-Free Pixel Correlation-Based Attention Module for Remote Sensing Object Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Generative Adversarial Network and Mutual-Point Learning Algorithm for Few-Shot Open-Set Classification of Hyperspectral Images

1
The VIPSL Laboratory, School of Electronic Engineering, Xidian University, Xian 710071, China
2
The MCI Laboratory, School of Aerospace Science and Technology, Xidian University, Xian 710071, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(7), 1285; https://doi.org/10.3390/rs16071285
Submission received: 28 January 2024 / Revised: 29 March 2024 / Accepted: 2 April 2024 / Published: 5 April 2024

Abstract

:
Existing approaches addressing the few-shot open-set recognition (FSOSR) challenge in hyperspectral images (HSIs) often encounter limitations stemming from sparse labels, restricted category numbers, and low openness. These limitations compromise stability and adaptability. In response, an open-set HSI classification algorithm based on data wandering (DW) is introduced in this research. Firstly, a K-class classifier suitable for a closed set is trained, and its internal encoder is leveraged to extract features and estimate the distribution of known categories. Subsequently, the classifier is fine-tuned based on feature distribution. To address the scarcity of samples, a sample density constraint based on the generative adversarial network (GAN) is employed to generate synthetic samples near the decision boundary. Simultaneously, a mutual-point learning method is incorporated to widen the class distance between known and unknown categories. In addition, a dynamic threshold method based on DW is devised to enhance the open-set performance. By categorizing drifting synthetic samples into known and unknown classes and retraining them together with the known samples, the closed-set classifier is optimized, and a (K + 1)-class open-set classifier is trained. The experimental results in this research demonstrate the superior FSOSR performance of the proposed method across three benchmark HSI datasets.

1. Introduction

In recent years, the central role of hyperspectral imaging in remote sensing technology has sparked widespread research interest [1]. The accuracy of object recognition is directly contingent on the processing and analysis of HSI data. Within this realm, hyperspectral image (HSI) classification, recognized as a crucial step, necessitates pixel-wise labeling of object categories [2,3,4,5,6,7]. Due to the high-dimensional and nonlinear characteristics of HSI data [8,9,10,11], the primary focus in HSI classification becomes feature extraction [12]. Deep learning, renowned for itss outstanding adaptive feature extraction ability, enables data-driven HSI classification methods to extract inherent features crucial for automatic recognition [13]. With the continuous optimization of network architectures and learning patterns, there has been a notable enhancement in the classification accuracy of models when applied to standard HSI datasets.
However, HSI classification requires a substantial amount of labeled data, posing a significant challenge. To address this issue, classification methods based on few-shot learning (FSL) have been proposed and applied to HSI classification based on machine learning and deep learning. Yet, the closed-set training models presume that the training data encompass all categories that may arise in the future. In reality, deep learning models may encounter new categories that were not encountered during training. Consequently, to address these challenges, few-shot open-set recognition (FSOSR) [14,15,16,17,18] methods for HSI classification are garnering increasing attention.
When comparing traditional hyperspectral image (HSI) open-set recognition (OSR) [19,20,21,22,23] with few-shot open-set recognition (FSOSR) for HSI [24,25,26,27], a significant challenge arises in addressing the coexistence of spectrally fine-grained known categories and anomalies. Current FSOSR methods, which employ reconstruction-based strategies for open-set HSI classification, exhibit limitations in practical applications. Specifically, generative models underperform in capturing information about unknown categories and fail to effectively utilize known category information to model the space of unknown categories, leading to instability in addressing open-set challenges. Additionally, the process of threshold determination is complex and lacks flexibility, making it difficult to set an efficient and learnable threshold, further limiting the adaptability of these methods. On the other hand, existing generative adversarial networks (GANs) primarily adopt a fully supervised approach for generating within-distribution data. However, in few-shot scenarios, estimating data density from limited samples presents significant challenges, impeding the optimal training of these methods. More critically, these methods continue to use closed-set models, rendering them ineffective in generating data for open spaces. These limitations undoubtedly weaken the performance of HSI in open-set classification.
To tackle the aforementioned challenges, this research proposes an open-set HSI classification algorithm based on reciprocal point learning (RPL) and data wandering (DW) (RPLDW). Firstly, a K-class classifier suitable for closed-set problems is trained, with its internal encoder used for extracting image classification features to estimate the feature distribution of known classes. Subsequently, these feature distributions are utilized to optimize the classifier and fine-tune its training. Next, synthetic samples near the decision boundary are generated using a GAN with sample density constraints to solve the few-shot problem, as illustrated in Figure 1.
Concurrently, a framework of RPL is introduced to classify the open-set HSI. The risk of open space is reduced by simulating the out-of-category space to increase the distance between known and unknown categories. In addition, a dynamic threshold method based on DW is devised to build a separate space for each category, thereby enhancing the model’s open-set performance. Employing a DW method, the drifting synthetic samples are categorized into known and unknown categories and inputted together with known samples into the classifier for a new round of training. At this stage, both the closed-set classifier and also a (K+1) class classifier suitable for open-set scenarios are trained in this research, sharing parameters of the early network layers. This approach enhances the data reliability and richness by including generated synthetic samples, ultimately improving the performance of the classification algorithm. The primary contributions of this research can be summarized as follows:
(1)
A method that generates boundaries is proposed based on sample density constraints to calculate sample density. Leveraging information entropy, this approach confines generated samples outside the edge regions of known classes, ensuring proximity to the known classes. Consequently, the classifier can more accurately demarcate the range of known classes.
(2)
The RPL method is adopted to address the open-set HSI classification challenge. It distinguishes the space of unknown categories by constructing inverse prototypes and enlarging the distance between known and unknown categories, thereby reducing the risk of open space. Meanwhile, it reinforces the boundaries of known categories and utilizes dynamic thresholds to remove unknown categories, avoiding manual determination of thresholds and enhancing the robustness of the algorithm.
(3)
A novel sample distribution analysis method is built to screen out edge samples based on DW, enhancing the sample boundaries and improving the classification accuracy.
(4)
Extensive experiments on three benchmark HSIs demonstrate the outstanding performance of the proposed RPLDW in FSOSR tasks, effectively fulfilling the open-set few-shot recognition. Therefore, in the real-world scenario of open-set HSI, RPLDW proves to be more robust and effective in minimizing open risks.

2. Related Works

2.1. Few-Shot Open-Set Recognition

The FSOSR issue is explored in [18] by creating Gaussian clusters that utilize a limited support set. Despite effectively learning the closed-set distribution, the prototypes in PEELER fail to remove outliers, particularly in rejecting spectrally fine-grained outliers. Another approach [17] implements a thresholding technique that relies on the disparity between original prototypes and queried transformed prototypes to identify outliers. Similarly, current FSOSR methods applied to HSI [28,29] also depend on a comparable threshold to exclude outliers. Nonetheless, determining an optimal threshold that can accurately reject outliers presenting a marginal spectral difference to the known land cover samples remains a significant challenge. Furthermore, utilizing the reciprocal points classification loss [30] results in a less compact known class distribution in FSL, contributing to reduced closed-set accuracy.

2.2. Generative Open-Set Recognition

A sophisticated open-set classifier should understand the nuanced closed-set distribution while navigating the open space effectively. Presently, generative OSR methods often address three aspects independently: (1) outlier rejection based on reconstruction error [31,32,33], (2) pseudo-open-set generation [23,34], and (3) enrichment of the closed-set distribution [35,36]. Predominantly, generative OSR methods employ GAN [37] or autoencoder [38] algorithms, leveraging large-scale training datasets. However, the class conditional auto-encoder (C2AE) [38] is susceptible to selection bias of a known sample from the training set, resulting in low reconstruction error on known samples. Research [37] introduces synthetic features and penalizes the discriminator to encourage outlier rejection. The adversarial samples are not proposed to be always considered outliers but can serve as representatives of known classes. A critical challenge in generative FSOSR is the non-convergence of GAN when working with limited training data [39]. The training strategy employed by [38] appears to offer a potential solution to this challenge.

2.3. Data Generated

Research [40] exclusively generates pseudo-known samples by rapidly adapting to a new domain using optimization-based meta-learning techniques, specifically [41] and Reptile [42]. However, it lacks the capability to produce out-of-distribution data. In contrast, [43] focuses solely on generating synthetic data for few-shot classification, while [44] creates synthetic images by combining two noise vectors with a feature vector through unsupervised meta-learning. Research [45] introduces an anti-collapse regularizer to maintain discriminability and diversity in a few-shot setting. The proposed regularizer minimizes the logarithmic similarity among generated adversarial samples and the logarithmic similarity of unit variance Gaussian noise. Furthermore, the approach of fusing input images with random interpolation coefficients in [46] facilitates the generation of arbitrary landcover classes. Additionally, [47] amplifies singular values associated with age, pose, and skin tone derived from singular value decomposition performed on [48]. Nevertheless, the range of end members in HSI datasets is infinite, leading to intractable singular values.

3. Methodology

3.1. Subsection

Herein, a dataset D t r = x ( i ) , y ( i ) i = 1 K containing K-known classes is given (where x ( i ) is a set of known classes, with N i samples in each, and y ( i ) refers to a set with the same one-hot vectors, representing labels corresponding to the set x ( i ) ). Under this given condition, it should correctly classify the K-known classes while recognizing and classifying new classes as the (K + 1) class.
Firstly, a K-class classifier suitable for a closed set is trained in this research, with its encoder being employed to extract image classification features and estimate the feature distribution of all known classes. Meanwhile, the training of the classifier is adjusted based on it. Subsequently, a GAN constrained by sample density is employed to generate synthetic samples near the decision boundaries. Leveraging RPL, this research models the unknown class space to augment the feature distance between the known data after boundary expansion and the unknown data in the open-set space. Then, employing the data walk method, the unclassified synthetic samples are categorized into known and unknown classes and sent to the classifier with the known samples for a new round of training. At this stage, the closed-set classifier and, concurrently, a (K + 1) class classifier suitable for the open set are trained. The training of these two classifiers shares the parameters of the previous layers of the network. The overall network architecture is illustrated in Figure 2.
The multitask network employed in this research requires an input of dimensions 9 × 9 × Channel. The first part of the network is an encoder/feature extractor equipped with two residual units and a global average pooling layer. After extracting latent features, a fully connected layer with a softmax function is set, serving as a classifier to output probabilities to the known classes. The reconstruction task of this research involves gradually increasing the spatial dimension of the latent features through a deconvolution layer. The output of the reconstruction task is a 9 × 9 × Channel instance, w intended to closely resemble the input data by minimizing the loss. Extreme value theory (EVT) is adopted to model the reconstruction loss, thus separating known and unknown classes.
The overall algorithm is outlined in Algorithm 1 below. W f refers to the synthetics generated by a GAN constrained by sample density, which is composed of W f C and W f O based on DW: W f C is categorized as known classes as edge samples to reinforce the sample boundaries, thus improving the performance of the closed-set classification while W f O simulates the unknown classes and serves as the K + 1 class to train the open-set classifier f O S .
Algorithm 1 Open-set recognition algorithm for generating decision boundaries based on sample density
Input:
D t r = x i , y i i = 1 K : Dataset of known classes
θ : Partition Threshold
α : Density Representative Parameter
Output:
f O S : Open-Set Classifier
1:Calculate the sample distribution to establish an information entropy field
2:GAN generates decision boundary samples based on sample information entropy constraints.
3:translates to “Train closed set classifier f C S
4:RPL
5:FOR  x i D  DO
6: Calculate the score of the newly generated fake samples.
7: Divide W f c into W f c and W f o
8:END FOR
9:On the basis of f C S , train the open-set classifier f O S with newly generated samples.

3.2. Generation of Known Class Sample Boundaries Based on Sample Density

This first step is to train a closed-set classifier through metric learning. Subsequently, the values in the pre-logit layer of the classifier are taken as sample points, and all sample points are collected as the sample space to calculate the sample density at any position in the sample feature space based on sample features. Next, the distribution of known classes is analyzed, and the objective function of GAN is constrained to facilitate the generation of samples within the boundary area of known classes and their periphery. In this way, it strengthens the boundary of known samples while increasing the reality and richness of simulated unknown class samples. This section details how to constrain GAN to generate sample boundaries, which is realized by calculating the sample density and training the classifier and GAN.

3.2.1. Training of the Closed-Set Classifier

In this research, a closed-set classifier is trained firstly through metric learning, with the following loss function:
L 1 = E D t r ( x , y ) log y m y n exp ( w ( m ) w ( n ) 2 2 / T ) y m = y n exp ( w ( m ) w ( n ) 2 2 / T ) + E x , y ~ p t r i ( x , y ) [ log p C ( y | x ) ]
where
T = 1 N t o t a l m , n = 1 N t o t a l w ( m ) w ( n ) 2 2
This facilitates the classifier in accurate classification while gathering samples with the same label as much as possible, resulting in distinct classes in the sample space as separate clusters. Following this, the parameters of the feature extractor in the classifier are frozen, the distance between each sample point and the center sample is computed, and the classifier is trained.
L 1 = E D t r ( x , y ) log y m y n exp ( w ( m ) w ( n ) 2 2 / T j ) y m = y n exp ( w ( m ) w ( n ) 2 2 / T j ) + λ E x , y ~ p t r i ( x , y ) [ log p C ( y | x ) ] + ( 1 λ ) E x , y ~ p t r i ( x , y ) [ r ( y | x ) ] + β E P t r i ( x ) [ ϕ ( x ) ]
where λ , β > 0 is a balancing parameter. Sample points that are close to the center sample have a high-density score. The classifier’s score and the sample density score are bundled together. When the sample is at the sample boundary or floating outside the cluster, the sample density is low, the classifier score is low, and the corresponding predicted label information entropy value is large, increasing the difficulty of being classified into a known class. With such a classifier, training a GAN to generate boundary samples becomes feasible, further refining the constraints on the classifier’s ability to recognize the boundaries of known class samples.

3.2.2. Calculation of Sample Density

There are two challenges when calculating the sample density. Firstly, due to the different aggregation degrees in each cluster, some clusters exhibit a relatively dense distribution of central feature points, while others have spare edges. Consequently, the density estimation of different regions needs to consider the relative positions of distinct clusters. Secondly, a small cluster of sample points is found to be gathered, forming a pseudo-sample center, as described in Figure 3.
The density peak clustering algorithm (DPC) is employed to calculate the maximum diameter d max of the entire sample space. α points are selected on average contained in each neighborhood, and d max   ×   α i = 1 K N i serves as the neighborhood radius of a sample point x j . Meanwhile, the number of sample points neighboring the point is counted as the local density ρ j of the point. Finally, the entropy of the region is calculated using the following equation:
H j = l = 1 ρ j h ρ j l lnh ρ j l ρ j
where
h ρ j l = x l j = 1 ρ j x j
Each data point ρ j can be sorted in ascending order, and the minimum distance δ j between this sample point and other points with higher density is calculated. The density peak score for this point is calculated as:
γ j = ρ j × δ j H j
Afterward, the top K points with the highest scores are selected as the center samples of the known class. The distance R t r = r ( i ) i = 1 K between each sample point and the center sample of its class is calculated. Meanwhile, the maximum radius r max ( i ) in the i-th class is selected and r max ( i ) × α N i refers to the neighborhood radius of each sample point. After that, the number of sample points N j contained in the neighborhood of the j-th sample is calculated. The aggregation score of the neighborhood where each point is located is calculated as follows:
θ j = N j × N i r max ( i ) × α
Subsequently, the distribution of the known class is estimated through Gaussian kernel density. All sample points are assigned a Gaussian function φ j ( x ) with a mean of itself and a variance of θ j . The sum of the Gaussian functions of all sample points forms the energy field, which can be expressed as follows:
ϕ ( x ) = j = 1 N t a t a l φ j ( x )
In the above equation, N total signifies the number of all sample points in the sample space, based on which the density representation of each position within the sample space can be acquired.

3.2.3. Training of GAN

GAN, typically used to generate pixel images that are as similar as possible to the original images, primarily encompasses discriminators (D) and generators (G). The generator maps the latent variables z to the output space through the prior distribution P pre ( z ) and outputs G ( z ) . The discriminator D ( x ) [ 0 , 1 ] represents the probability that the sample comes from the real target distribution. Through the showdown between the generator and the discriminator, the distribution P G ( x ) of the samples generated by the generator gradually approaches the target distribution P t r i ( x ) . Its objective function is described as follows:
min G   max D E P t r i ( x ) [ log D ( x ) ] + E P p r e ( z ) [ log ( 1 D ( x ) ) ]
This research aims to produce samples on the periphery of the known classes for the generator. For this purpose, the original GAN objective function is modified accordingly, as given in Equation (10) below.
min G   max D E P t r i ( x ) log D ( x ) + E P G ( x ) [ log ( 1 D ( x ) ] ( a ) + β E P G ( x ) [ ϕ ( x ) ] ( b )
In the above expression, β > 0 is the balance parameter, which means that the trained classifier becomes a class one at β = 0 . This objective function can be delineated into two components. The first term (a) corresponds to the loss of the original GAN, making the samples generated by the generator as similar as possible to the real samples. The second term (b) calculates the sample density. Throughout the training, the generator and the discriminator undergo multiple training iterations under the constraint of sample density. This iterative process eventually guides the generator in producing samples on the low-density boundary and discrete samples.
Equation (10) emphasizes generating authentic images to augment the closed-set training data. To ensure that the generated and the real images are ultimately projected into the same feature space, the conditional generative network should be trained on the basis of the well-trained autoencoder. However, this research deviates from relying on autoencoder training and instead opts to generate synthetic samples by referencing the latent feature layer of the known class. The W chosen refers to the feature value corresponding to the pre-logit layer of the closed-set classifier. The final equation is as follows:
min G   max D E P t r i ( w ) log D ( w ) + E P G ( w ) [ log ( 1 D ( w ) ] ( a ) + β E P G ( w ) [ ϕ ( w ) ] ( b )
This approach enables the generator to produce synthetic samples in areas with low sample density, effectively balancing the distribution of internal samples within the known classes and consequently enhancing the accuracy of the classifier. Additionally, generating samples in the low-density regions outside the known classes can train the classifier to recognize unknown class samples by categorizing them as unknown classes. This process provides negative feedback, allowing the model to perceive the boundary area of the known classes.

3.3. Reciprocal Point Learning

Reciprocal points [30] represent the inverse prototypes of each known category, and their utilization enables the constraints of the space outside the category while maintaining the accuracy of the known category. During the training, RPL [49], based on reciprocal points, pushes all known categories towards the edge of the space. As a result, this strategy ensures that the unknown category and reciprocal points collectively define the space outside the category, minimizing the overlap between known and unknown distributions. Consequently, this enhances the distinguishability of the unknown category. In this research, RPL is adopted to model the space of unknown categories to classify the open-set HSI.
Specifically, given there are n labeled samples and N known categories in the training dataset D S = q 1 , q 2 , , q n , y i 1 , 2 , , N is the corresponding label of q i . For trail data D T = t 1 , t 2 , , t m , the labels of t i belong to 1 , 2 , , N for closed-set HSI classification and to 1 , 2 , , N N + 1 , N + 1 , , N + U for open-set HSI classification, where U denotes the number of unknown categories in a realistic HSI scenario. Thus, the potential unknown data can be denoted as D U . For each known category k ( 1 k N ) , S k is the embedding space in the space R D , and the corresponding open space O k can be denoted as O k = R S k .
Therefore, as an inverse prototype of the known category k , the reciprocal point P k can be regarded as a potential representation of the sub-dataset D L k D U [14]. As demonstrated in Figure 4a, the reciprocal point P k should be closer to the empty space O k and far away from S k , as follows.
max ( ζ ( D L k D U ) , P k ) d d ζ ( D S k , P k )
where ζ denotes a set of distances such that the known category k is maximally distant from P .
Therefore, a pivotal objective of RPL is to expand the distance between known categories and their corresponding reciprocal points. To achieve this, the cosine distance d is employed to compute the spatial position and angular orientation between instance q i and the corresponding reciprocal point P k .
d ( φ ( q i ) , P k ) = 1 φ ( q i ) P k φ ( q i ) P k
where φ ( ) denotes a feature representation network with a learnable parameter set Θ , and φ ( q i ) refers to a network feature of sample q i .
After describing the distance metric, it is noteworthy that as the distance d ( φ ( x i ) , P k ) increases, the probability of labeling q i is calculated as the category k . The category probability can be calculated using the softmax function, considering the nature of the reciprocal points and the instance-related weights in the expression below.
p ( y = k q i , φ , P ) = exp ( d ( φ ( q i ) , P k ) / ξ i ) j = 1 N exp ( d ( φ ( q i ) , P k ) / ξ i )
Therefore, the supervised loss function can be defined as the negative logistic loss of the labels:
L C ( q , P ) = log 2 ( 1 + exp ( q P ) )
The learnable parameter set of the feature representation network φ ( ) is updated by minimizing L C . Therefore, φ ( ) can classify the known categories through supervised learning with multi-class samples and their corresponding true labels. More importantly, as illustrated in Figure 4b, all known categories are pushed to the periphery of the space by their respective reciprocal points throughout the training. In addition, reciprocal points in the space outside the category are also pushed into the known space.

3.4. Sample Distribution Using Data Wandering

Since the potential unknown categories are in the space outside the categories, the risk of encountering open space persists. In response, a learnable dynamic threshold strategy is further proposed in this research. By constructing a separation space, the open-set HSI classification approach is optimized to lower the risk of the open space and facilitate the recognition of unknown categories. Different from the past approaches that relied on multi-class classifier prediction results for in-class splitting, DW is introduced for sample distribution analysis and further constrains the open space, as explicated in Figure 5.
By constructing an information energy field and setting a threshold θ , the sample space is divided into within-class and outside-class after incorporating the synthetic samples. The basic concept is depicted in Figure 6.
In the figure above, the horizontal axis d represents the distance from the sample to the relative sample center, while the vertical axis s shows the magnitude of the energy value. The interval 0 , L refers to the central sample x ( i ) of the known class, the interval L , L + M contains the known class edge samples and some generated samples ( w ( i ) + w f C ), and the interval L + M , is composed of generated samples w f O only. Existing methods can be employed to directly classify the interval L , as an unknown class, so samples from the known class that are close to the edge are classified as unknown. In this research, some generated samples from the known class are selected to enhance the accuracy of recognition.
Although RPL can classify synthetic samples in the open-set space into unknown class W f O and known class W f C . Due to the existence of fuzzy sample boundaries, the boundary samples are categorized as comprehensively as possible, and a data walking method is adopted to iterate. Each sample point is moved in the direction of high energy at a specific step length. Following this, a new data distribution is formed, and the energy field is recalculated. These steps are iteratively repeated several times. After multiple iterations, based on the data redistribution in the open-set space, synthetic samples within the known range are identified as boundary samples through RPL, while the remaining samples are considered new classes.
The data wandering algorithm, as an optimization technique based on energy fields, demonstrates its unique value and efficiency. According to the hypothesis proposed by the decision algorithm, clustering clusters should be centered around high-density data, with low-density data surrounding the high-density ones. In the information energy field, a cluster’s region should exhibit a distribution state where the central energy field strength is high, and the edge energy field strength is low. This paper utilizes d ( φ ( q i ) , P k ) to construct the information energy field. After forming the sample information energy field values, it is assumed that each sample point can move in the numerical space. By comparing the energy field strength of the sample point with its surrounding eight neighborhoods, the sample point is made to wander toward the direction of higher field strength. After several wanderings, all sample points will move towards regions of higher data energy at a certain step length, forming the final data distribution shape, and the final clustering results are obtained by dividing the regions. Calculating the average energy of samples and setting a threshold can be defined: if the average energy of a clustering cluster is higher than 1.5 times the total average entropy, it is defined as a high-energy cluster; otherwise, it is a low-energy cluster. After the data wandering, due to the energy distribution deviation, i.e., the average energy of the high-energy cluster is far higher than that of the low-energy cluster, new samples in the process of data wandering will tend to move towards high-energy fields. This will result in the following: (1) Data wandering can effectively solve the clustering affiliation of discrete points, making it easier to form clustering clusters after data wandering. (2) In the process of data wandering, high-energy clustering clusters will influence low-energy clustering clusters, causing the clustering results of low-energy clusters to become more dispersed.
Therefore, the RPL constraints for the separation space based on DW can be represented as follows:
max ( θ ( D L k D U D L U k ) , P k ) R , R θ ( D S k , P k ) θ = ϕ ( W f O ) ϕ ( W f C ) log ( ϕ ( W f O ) ϕ ( W f C ) )
The process of the data walking method is illustrated in Figure 7 and Figure 8.

4. Experimental Results and Corresponding Analysis

4.1. Dataset

To carry out the open-set HSI multiclassification experiments proposed and to thoroughly validate our methodology, three HSI datasets, Pavia University (PU), Salinas (SA), and Indian Pines (IP) are selected in this research.
The PU dataset, captured by the reflective optical system imaging spectrometer (ROSIS) near the university, exhibits spatial dimensions of 610 × 340 with a spatial resolution of 1.3 m. It consists of 115 channels covering a wavelength range of 0.43~0.86 μm. Before classification experiments, 12 bands affected by noise and water absorption are eliminated first. Buildings without previous labels are annotated as unknown due to their distinct spectral profiles from known land cover.
The SA dataset is acquired by the Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) over the Salinas Valley in California, USA. The image possesses spatial dimensions of 512 × 217 with a spatial resolution of 3.7 m, and it includes 224 bands spanning a wavelength range of 0.4~2.5 μm. Similarly, 20 bands are removed prior to experiments due to water absorption. Man-made materials without previous annotations are labeled as unknown samples.
The IP dataset, obtained via AVIRIS at the Indian Pines test site in Tippecanoe County, Northwest Indiana, USA, has spatial dimensions of 145 × 145 with a spatial resolution of 20 m. It comprises 220 bands spanning a wavelength range of 0.43~0.86 μm. Before classification experiments, 20 bands are eliminated due to water absorption. Classes with fewer than 10 instances were discarded as tail classes and considered unknown, resulting in 8 known classes. Roads that are not previously labeled are identified as unknown.
Table 1 and Table 2 provides the class distribution for each dataset, while Figure 9a, Figure 10a and Figure 11a are their false color images. Figure 9b, Figure 10b and Figure 11b display the standard true values of the closed-set constant index classification, along with the respective class representations denoted by the true values.

4.2. Evaluation Metrics

To evaluate the performance of RPLDW, the standard OSR indicators are adopted herein, namely closed-set overall accuracy (ClosedOA), open-set overall accuracy (OpenOA), and Kappa. ClosedOA and OpenOA represent the percentage of known category samples and that of open-set samples that are correctly classified, respectively.
O p e n O A = k = 1 K + 1 T P k + T N k k = 1 K + 1 T P k + T N k + F P k + F N k
where T P k , T N k , F P k , and F N k signify the true positive, true negative, false positive, and false negative of the k th known class, respectively. All the outliers are combinedly considered as the ( K + 1 )th class in OpenOA.

4.3. Experiments

The ADAM (Adaptive Moment Estimation) optimizer is utilized for back-propagation, with a maximum training round of 100 times. The learning rate is set at 0.0001 for 1-shot and 0.003 for 5-shot. During the training, all experiments are composed in a random manner from the training samples and repeated 10 times to reduce experimental errors. Subsequently, the average results are calculated for analysis and comparison.
To simulate the conditions for open-set HSI classification, the last category of features is designated as unknown, based on generally accepted standard true values of features, to ensure experiment fairness. In addition, setting a single unknown category aids in maintaining the overall effectiveness of the open-set classification image and facilitates in-depth analysis.
To fully assess the performance of our method in open-set HSI classification, several important comparison methods are employed in this work, involving two categories:
(1)
Comparison with traditional closed-set HSI classification methods: the three-dimensional octave convolutional space-spectral attention networks (3DOC-SSAN) [50] and the spectral–spatial transformer network (SSTN) based on transformer backbone [51] are adopted to emphasize the superiority of open-set classification methods in dealing with unknown categories. In this research, a fully connected (FC) layer is introduced at the end of the network for score prediction, where the maximum score serves as the final label. According to the general solution for open-set issues [52], an extra softmax function is added to these networks, and the maximum probability value in the probability vector functions as the prediction label. Referring to [52], the threshold of the softmax function is set to 0.5 to recognize outputs considered as unknown categories (i.e., when the maximum probability is less than 0.5);
(2)
Comparison with open-set HSI classification methods: OpenMax [22], SSLR, and MDL4OW [29] are originally developed for large-scale OSR by fitting the Weibull distribution in the prototype network, making them suitable for FSOSR. RDOSR [25] performs OSR on HSI datasets in the latent space using a limited number of supervised samples. PEELER [26], SnaTCHer [25], and OCN [53] are developed in the context of FSOSR and can be easily evaluated on HSI datasets.
The above comparisons are conducted by adhering to the optimal settings outlined in each original piece of literature. Consistency in the combinations of randomly selected training samples for each method is maintained across 10 experiments to ensure fairness. Notably, each method’s closed-set and open-set performance is validated by recording the closed-set and open-set metrics, respectively. The experimental results are summarized in Table 3, Table 4, Table 5, Table 6, Table 7 and Table 8.

4.4. Experimental Results

The RPLDW is compared with the most advanced 1-shot and 5-shot FSOSR methods in Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, respectively. The HSI datasets demonstrate superior performance compared to other OSR methods. The following valuable conclusions can be drawn from the observed results.
(1)
Limitations of Closed-Set Methods and the Impact of Sample Size on Accuracy: While the use of the softmax function at the end of networks enhances the ability of other closed-set methods to recognize unknown categories, these methods exhibit significant fluctuations when confronted with potential unknown categories. This is especially pronounced in the FSOSR context with limited samples. Methods like 3DOC-SSAN and SSTN perform well with ample samples but become unstable with fewer samples, leading to decreased performance in open-set environments. This highlights the inherent limitations of closed-set classifiers in FSOSR contexts.
(2)
Performance Comparison of Open-Set Methods and Traditional Classification Methods: Both traditional classification methods and those designed for open sets generally do not perform as well in open-set environments as they do in closed-set ones. This phenomenon is more evident in the context of FSOSR. Dedicated open-set methods, such as RPLDW, demonstrate greater resilience and stability, especially under limited sample conditions, by considering the capability to process unknown categories during the design phase.
(3)
Correlation Between Open-Set and Closed-Set Performance: Both conventional classification methods and specialized open-set classification methods reveal a close relationship between open-set and closed-set performance. In FSOSR contexts, if a method excels in open-set performance, it often also performs well in closed-set scenarios, but the reverse is not necessarily true.
(4)
Superiority of Open-Set Methods: Open-set methods exhibit stable accuracy in handling both open and closed-set classifications, which is particularly crucial in FSOSR environments. Compared to closed-set methods, specialized open-set methods effectively exclude unknown categories while maintaining high accuracy in recognizing known categories.
(5)
Advantages of the RPLDW Method: In the FSOSR environment, the RPLDW method outperforms all other studied methods. It scores highest in closed and open-set metrics and achieves the highest accuracy in identifying unknown categories. RPLDW enhances the boundaries of known categories, eliminating the need to manually set specific thresholds. It constructs an independent space in limited sample situations and employs RPL to calculate reciprocal point distances. This approach significantly improves accuracy by determining labels within the known range in open-set scenarios.
Therefore, it can be concluded that RPLDW performs more robustly and effectively when dealing with open risks in the real open world.
To compare the effects of different methods more intuitively, the classification maps depicting the performance of the better-performing open-set classification methods in the experiment are generated, as explicated in Figure 12. This visual approach provides a better comprehension and comparison of the performance differences among various methods.
Figure 12 reveals that high closure metrics usually align with high-quality closed-set classification maps, whereas low closure metrics may result in subpar outcomes. Despite SSTN demonstrating superior open performance compared to other closed-set methods, it still struggles to effectively exclude unknown categories.
Improved closed-set methods can identify potential unknown categories, as evidenced by the black areas in the open-set classification maps. However, the overall open-set performance of closed-set methods falls short of ideal. The primary issue lies in misclassifying a considerable number of known categories as unknown, causing inconsistencies between the black areas of the open-set classification maps and the actual situation.
The negative impact of the OSE slightly influences the performance of the open-set methods. The classification maps of open-set methods achieve superior visual effects in both the ideal closed-set assumption and the realistic open HSI world. They usually correspond with the actual distribution of ground objects and exhibit less noise.
RPLDW demonstrates excellent visual effects on both closed-set and open-set classification maps, providing further evidence of its robustness and superior performance in the realistic open HSI world. Notably, MDL4OW and SSLR, with fixed thresholds, tend to overlook certain potential areas of unknown categories. In contrast, our proposed RPLDW incorporates a dynamic, learnable threshold, enabling it to discover more potential areas of unknown categories. This adaptive approach proves effective in navigating the complex open world with new ground object categories.

4.5. Open-Set Multiclassification and Visualization Analysis

Herein, categories 11, 12, 13, and 14 of the SA datasets are identified as unknown for open-set multiclassification experimental analysis. The experiments are compared with the latest open-set recognition algorithms, including FCPN [15], DLRSPs-DAEs [54], POSM [15], and SSMLP-RPL [14]. The other detailed settings are consistent with the above-mentioned experiments, and the results are outlined in Table 9.
To further intuitively demonstrate the discriminative feature space constructed by RPLDW, the t-SNE dimensionality reduction visualization method is utilized to further highlight its advantages in identifying known and unknown categories. This is implemented by taking SA data as an example, and the results are visualized in Figure 13.
Examining the original SA data reveals suboptimal discriminability in its feature space, primarily presenting as overlap and too much closeness among different categories. However, using the RPL training and DW strategy, the data samples are successfully projected into a more distinctive high-dimensional feature space, enabling the open-set HSI classification. More specifically, the distance between known categories and their corresponding antipodes is broadened, pushing known categories towards the fringe of the space. In addition, the distance between unknown categories and antipodes is confined within a limited boundary. Therefore, the feature space created by RPLDW enhances the distance between known categories and unknown categories, facilitating the identification of unknown categories.
This optimized feature space ensures the correct classification of known categories while effectively excluding unknown categories. More importantly, the label assignment for the instance to be classified is determined by its distance to the antipode. Setting a reasonable threshold can effectively reject unknown categories, thereby mitigating the risk of open space. However, a single and fixed threshold could potentially lead to overlooking potential unknown categories due to variations in the distance between known categories and their antipodes. A learnable dynamic threshold strategy is proposed in this research to address this issue. It can reject unknown categories to the maximum extent without affecting the recognition of known categories.

4.6. Small Sample Analysis

For experiments on HSI data, the samples in each category are randomly disrupted, and the number of samples from each category is set as 20, 50, 150, and 200 labeled with the real class label, forming the training dataset. Importantly, the training dataset does not include samples labeled as “Unknown”. To create the test dataset, the remaining labeled data are combined with the samples labeled as “Unknown”. Furthermore, five repeated tests are conducted following the aforementioned steps to mitigate the potential impact of random factors and ensure the robustness of the experimental outcomes. The algorithm’s performance is assessed using the average classification results. To assess the efficacy of the algorithm proposed in this research, its experimental results are compared with those of several algorithms mentioned in the previous section, employing three experimental datasets. The specific comparisons are detailed in Table 10, Table 11 and Table 12.
The experimental results further reinforce the efficacy of the proposed RPLDW algorithm, showing its pronounced advantage over other methods at a consistent training rate. Notably, as the training dataset increases from 150 to 200, the RPLDW algorithm maintains stable classification results, underscoring its robust performance even with relatively small training datasets. This stability is a significant strength of our algorithm, opening up possibilities for practical applications in scenarios with limited data and quick decision-making requirements or situations characterized by continuous data growth and continuous learning needs.
The success of the proposed method in achieving high accuracy can be attributed to the effectiveness of the reciprocal points and reinforced boundary algorithms proposed. The reciprocal point algorithm can effectively project data instances into a high-dimensional feature space, enlarging the separation between known and unknown categories. This clear distance difference provides a clearer foundation for classification decisions, thereby enhancing overall accuracy. Simultaneously, the reinforced boundary algorithm contributes to the stability of classification by optimizing decision boundaries and maintaining high classification accuracy even with limited training samples. This approach fortifies more robust classification decisions in environments with scarce data, solidifying the strong performance of our method under these circumstances.
In summary, the harmonious integration of reciprocal points and reinforced boundary algorithms empowers the proposed method to demonstrate exceptional performance, particularly in scenarios with limited data. This success is a testament to algorithmic design and our ability to deeply understand the dynamics of data and the needs of classification.

4.7. RPLDW Model Ablation Analysis

The RPLDW method introduced herein incorporates GAN and RPL, which significantly enhances its performance in HSI data classification in the OSE. Ablation experiments were conducted on these two modules by systematically excluding them from the RPLDW framework. Subsequently, open-set classification on the “Unknown” labeled data across three datasets was conducted, using the accuracy of open-set classification as the evaluation metric. The experimental results are summarized in Table 13, revealing the substantial contributions of both GAN and RPL to the classification capability of RPLDW in the OSE. The removal of either the GAN or the RPL module from RPLDW leads to a marked decrease in the model’s accuracy for classification in the OSE.
After the removal of the GAN module, the model exhibited a notable decrease in the average accuracy for classification in the OSE across various training data sizes. Specifically, for training data sizes of 20, 50, 100, 150, and 200, the declines in accuracy were 22.32%, 23.78%, 22.97%, 22.57%, and 20.7%, respectively. Similarly, excluding the RPL module resulted in accuracy reductions of 27.11%, 26.89%, 25.62%, 24.23%, and 22.95% for the same training data sizes, respectively. RPL differentiates the unknown category space by constructing inverse prototypes and widening the distance between known and unknown categories, thereby mitigating risks in open spaces. Conversely, GAN assesses the similarity of test data from a metric domain perspective, facilitating the detection and classification of unknown label data.
In summary, the RPLDW model proposed herein leveraged RPL to construct a separation space and incorporated GAN for outstanding performance in HSI open-set classification. RPLDW demonstrates superior performance in unknown class identification, model stability, classification efficacy, and convergence speed compared to other algorithms. Despite its effectiveness in identifying known and unknown categories in HSI data with enhanced boundaries, some stray data points in the sample space remain inadequately classified. Additionally, applying a threshold method for open-set recognition in RPLDW can impact the accuracy. These limitations provide opportunities for future enhancement and refinement.

5. Conclusions

In the field of HSI classification, despite the significant strides made by deep learning methods, they still heavily rely on the ideal CSE and demand a substantial volume of labeled data. In response to these challenges, this research proposes the RPLDW method. Firstly, a K-class classifier tailored for the closed set is trained, leveraging its internal encoder to extract features. These features, indicative of known class distributions, are then employed to fine-tune the training of the classifier. Secondly, synthetic samples close to the decision boundary are generated using the GAN based on sample density constraints. These synthetic samples are processed to augment the training data. This addresses the scarcity of samples and enhances the training data. Concurrently, an RPL framework is introduced to lower the risk of open space by simulating category external space, thereby enlarging the distance between known and unknown categories. Finally, a dynamic threshold method is designed based on data roaming by dividing the synthetic samples into known and unknown categories and inputting them together with known samples into the classifier for a new round of training.
The experimental results reveal the strong competitiveness of RPLDW in both the CSE and OSE, even surpassing the current state-of-the-art HSI classification methods. This proves the effectiveness and robustness of the proposed RPLDW in lowering risks in the real open world and improving classification accuracy. However, it should be noted that RPLDW may erroneously label distinct land objects as unknown categories at their boundaries, highlighting an area that warrants further research and improvement.

Author Contributions

Conceptualization, T.X.; methodology, T.X.; software, T.X.; investigation, T.X.; resources, J.L. and Y.W.; data curation, T.X.; writing—original draft preparation, T.X.; writing—review and editing, Y.D.; supervision, J.L. and Y.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (No. U21A20514) and the Key Chain Innovation Projects of Shaanxi (Program No. 2022ZDLGY01-14).

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available. Given that the funding-supported project is currently in progress, data will be provided upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Hong, D.; Chanussot, J.; Zhu, X.X. An overview of multimodal remote sensing data fusion: From image to feature, from shallow to deep. In Proceedings of the 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS 2021, Brussels, Belgium, 11–16 July 2021; pp. 1245–1248. [Google Scholar]
  2. Zhong, Y.; Wang, X.; Xu, Y.; Wang, S.; Jia, T.; Hu, X.; Zhao, J.; Wei, L.; Zhang, L. Mini-UAV-borne hyperspectral remote sensing: From observation and processing to applications. IEEE Geosci. Remote Sens. Mag. 2018, 6, 46–62. [Google Scholar] [CrossRef]
  3. Hong, D.; He, W.; Yokoya, N.; Yao, J.; Gao, L.; Zhang, L.; Chanussot, J.; Zhu, X. Interpretable hyperspectral artificial intelligence: When nonconvex modeling meets hyperspectral remote sensing. IEEE Geosci. Remote Sens. Mag. 2021, 9, 52–87. [Google Scholar] [CrossRef]
  4. Zhang, L.; Zhang, L.; Tao, D.; Huang, X. On combining multiple features for hyperspectral remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 2012, 50, 879–893. [Google Scholar] [CrossRef]
  5. Zhang, Y.; Li, W.; Zhang, M.; Wang, S.; Tao, R.; Du, Q. Graph information aggregation cross-domain few-shot learning for hyperspectral image classification. IEEE Trans. Neural Netw. Learn. Syst. 2022, 35, 1912–1925. [Google Scholar] [CrossRef]
  6. Zhang, S.; Deng, Q.; Ding, Z. Multilayer graph spectral analysis for hyperspectral images. EURASIP J. Adv. Signal Process. 2022, 2022, 92. [Google Scholar] [CrossRef]
  7. Ma, S.; Tong, L.; Zhou, J.; Yu, J.; Xiao, C. Self-Supervised Spectral-Spatial Graph Prototypical Network for Few-Shot Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5518915. [Google Scholar] [CrossRef]
  8. Wang, M.; Wang, Q.; Chanussot, J.; Hong, D. L0-l1 hybrid total variation regularization and its applications on hyperspectral image mixed noise removal and compressed sensing. IEEE Trans. Geosci. Remote Sens. 2021, 59, 7695–7710. [Google Scholar] [CrossRef]
  9. Shang, X.; Song, M.; Wang, Y.; Yu, C.; Yu, H.; Li, F.; Chang, C.-I. Target-constrained interference-minimized band selection for hyperspectral target detection. IEEE Trans. Geosci. Remote Sens. 2021, 59, 6044–6064. [Google Scholar] [CrossRef]
  10. Hong, D.; Gao, L.; Yao, J.; Yokoya, N.; Chanussot, J.; Heiden, U.; Zhang, B. Endmember-guided unmixing network (EGU-Net): A general deep learning framework for self-supervised hyperspectral unmixing. IEEE Trans. Neural Netw. Learn. Syst. 2022, 33, 6518–6531. [Google Scholar] [CrossRef] [PubMed]
  11. He, W.; Chen, Y.; Yokoya, N.; Li, C.; Zhao, Q. Hyperspectral super-resolution via coupled tensor ring factorization. Pattern Recognit. 2021, 122, 108280. [Google Scholar] [CrossRef]
  12. Liu, B.; Yu, X.; Zhang, P.; Yu, A.; Fu, Q.; Wei, X. Supervised deep feature extraction for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 1909–1921. [Google Scholar] [CrossRef]
  13. Zhang, L.; Zhang, L.; Du, B. Deep learning for remote sensing data: A technical tutorial on the state of the art. IEEE Geosci. Remote Sens. Mag. 2016, 4, 22–40. [Google Scholar] [CrossRef]
  14. Sun, Y.; Liu, B.; Wang, R.; Zhang, P.; Dai, M. Spectral-Spatial MLP-like Network with Reciprocal Points Learning for Open-Set Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5513218. [Google Scholar] [CrossRef]
  15. Xie, Z.; Duan, P.; Liu, W.; Kang, X.; Wei, X.; Li, S. Feature consistency-based prototype network for open-set hyperspectral image classification. IEEE Trans. Neural Netw. Learn. Syst. 2023, 1–11. [Google Scholar] [CrossRef] [PubMed]
  16. Krishnendu, C.S.; Sowmya, V.; Soman, K.P. Deep Learning-Based Open Set Domain Hyperspectral Image Classification Using Dimension-Reduced Spectral Features. In Smart Computer Vision; Springer International Publishing: Cham, Switzerland, 2023; pp. 273–293. [Google Scholar]
  17. Pal, D.; Bose, S.; Banerjee, B.; Jeppu, Y. Extreme Value Meta-Learning for Few-Shot Open-Set Recognition of Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5512516. [Google Scholar] [CrossRef]
  18. Liu, Y.; Tang, Y.; Zhang, L.; Liu, L.; Song, M.; Gong, K.; Peng, Y.; Hou, J.; Jiang, T. Hyperspectral open set classification with unknown classes rejection towards deep networks. Int. J. Remote Sens. 2020, 41, 6355–6383. [Google Scholar] [CrossRef]
  19. Yue, J.; Fang, L.; He, M. Spectral-spatial latent reconstruction for open-set hyperspectral image classification. IEEE Trans. Image Process. 2022, 31, 5227–5241. [Google Scholar] [CrossRef]
  20. Fang, L.; Zhu, D.; Yue, J.; Zhang, B.; He, M. Geometric-spectral reconstruction learning for multi-source open-set classification with hyperspectral and LiDAR data. IEEE/CAA J. Autom. Sin. 2022, 9, 1892–1895. [Google Scholar] [CrossRef]
  21. Ji, F.; Zhao, W.; Wang, Q.; Emery, W.J.; Peng, R.; Man, Y.; Wang, G.; Jia, K. Spectral-Spatial Evidential Learning Network for Open-Set Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2024, 62. [Google Scholar] [CrossRef]
  22. Bendale, A.; Boult, T.E. Towards open set deep networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1563–1572. [Google Scholar]
  23. Ge, Z.Y.; Demyanov, S.; Chen, Z.; Garnavi, R. Generative openmax for multi-class open set classification. arXiv 2017, arXiv:1707.07418. [Google Scholar]
  24. Hassen, M.; Chan, P.K. Learning a neuralnetwork-based representation for open set recognition. In Proceedings of the 2020 SIAM International Conference on Data Mining, SIAM, Cincinnati, OH, USA, 7–9 May 2020; pp. 154–162. [Google Scholar]
  25. Jeong, M.; Choi, S.; Kim, C. Few-shot open-set recognition by transformation consistency. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 12566–12575. [Google Scholar]
  26. Liu, B.; Kang, H.; Li, H.; Hua, G.; Vasconcelos, N. Few-shot open-set recognition using metalearning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8798–8807. [Google Scholar]
  27. Zhu, D.; Du, B.; Dong, Y.; Zhang, L. Spatial–spectral joint reconstruction with interband correlation for hyperspectral anomaly detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5529513. [Google Scholar] [CrossRef]
  28. Kaviani Baghbaderani, R.; Qu, Y.; Qi, H.; Stutts, C. Representative-discriminative learning for open-set land cover classification of satellite imagery. In European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 1–17. [Google Scholar]
  29. Liu, S.; Shi, Q.; Zhang, L. Few-shot hyperspectral image classification with unknown classes using multitask deep learning. IEEE Trans. Geosci. Remote Sens. 2020, 59, 5085–5102. [Google Scholar] [CrossRef]
  30. Chen, G.; Qiao, L.; Shi, Y.; Peng, P.; Li, J.; Huang, T.; Pu, S.; Tian, Y. Learning open set network with discriminative reciprocal points. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 507–522. [Google Scholar]
  31. Deecke, L.; Vandermeulen, R.; Ruff, L.; Mandt, S.; Kloft, M. Image anomaly detection with generative adversarial networks. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, Dublin, Ireland, 10–14 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–17. [Google Scholar]
  32. Sabokrou, M.; Khalooei, M.; Fathy, M.; Adeli, E. Adversarially learned one-class classifier for novelty detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3379–3388. [Google Scholar]
  33. Perera, P.; Morariu, V.I.; Jain, R.; Manjunatha, V.; Wigington, C.; Ordonez, V.; Patel, V.M. Generative-discriminative feature representations for open-set recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11814–11823. [Google Scholar]
  34. Neal, L.; Olson, M.; Fern, X.; Wong, W.K.; Li, F. Open set learning with counterfactual images. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 613–628. [Google Scholar]
  35. Pidhorskyi, S.; Almohsen, R.; Adjeroh, D.A.; Doretto, G. Generative probabilistic novelty detection with adversarial autoencoders. arXiv 2018, arXiv:1807.02588. [Google Scholar]
  36. Zenati, H.; Romain, M.; Foo, C.S.; Lecouat, B.; Chandrasekhar, V. Adversarially learned anomaly detection. In Proceedings of the 2018 IEEE International conference on data mining (ICDM), Singapore, 17–20 November 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 727–736. [Google Scholar]
  37. Kong, S.; Ramanan, D. Opengan: Open-set recognition via open data generation. arXiv 2021, arXiv:2104.02939. [Google Scholar]
  38. Oza, P.; Patel, V.M. C2ae: Class conditioned auto-encoder for open-set recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2307–2316. [Google Scholar]
  39. Ding, C.; Li, Y.; Wen, Y.; Zheng, M.; Zhang, L.; Wei, W.; Zhang, Y. Boosting few-shot hyperspectral image classification using pseudo-label learning. Remote Sens. 2021, 13, 3539. [Google Scholar] [CrossRef]
  40. Liang, W.; Liu, Z.; Liu, C. Dawson: A domain adaptive few shot generation framework. arXiv 2020, arXiv:2001.00576. [Google Scholar]
  41. Finn, C.; Abbeel, P.; Levine, S. Modelagnostic meta-learning for fast adaptation of deep networks. In Proceedings of the International Conference on Machine Learning, PMLR, Sydney, Australia, 6–11 August 2017; pp. 1126–1135. [Google Scholar]
  42. Nichol, A.; Achiam, J.; Schulman, J. On first-order meta-learning algorithms. arXiv 2018, arXiv:1803.02999. [Google Scholar]
  43. Zhang, R.; Che, T.; Ghahramani, Z.; Bengio, Y.; Song, Y. Metagan: An adversarial approach to few-shot learning. Adv. Neural Inf. Process. Syst. 2018, 31. [Google Scholar]
  44. Phaphuangwittayakul, A.; Guo, Y.; Ying, F. Fast adaptive meta-learning for few-shot image generation. IEEE Trans. Multimedia 2021, 24, 2205–2217. [Google Scholar] [CrossRef]
  45. Li, K.; Zhang, Y.; Li, K.; Fu, Y. D2GAN: A few-shot learning approach with diverse and discriminative feature synthesis. In Proceedings of the International Conference on Learning Representations ICLR 2020, Addis Ababa, Ethiopia, 30 April 2020. [Google Scholar]
  46. Hong, Y.; Niu, L.; Zhang, J.; Zhao, W.; Fu, C.; Zhang, L. F2gan: Fusing-and-filling gan for few-shot image generation. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 2535–2543. [Google Scholar]
  47. Robb, E.; Chu, W.S.; Kumar, A.; Huang, J.B. Few-shot adaptation of generative adversarial networks. arXiv 2020, arXiv:2010.11943. [Google Scholar]
  48. Karras, T.; Laine, S.; Aittala, M.; Hellsten, J.; Lehtinen, J.; Aila, T. Analyzing and improving the image quality of stylegan. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8110–8119. [Google Scholar]
  49. Chen, G.; Peng, P.; Wang, X.; Tian, Y. Adversarial reciprocal points learning for open set recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 8065–8081. [Google Scholar] [CrossRef] [PubMed]
  50. Tang, X.; Meng, F.; Zhang, X.; Cheung, Y.-M.; Ma, J.; Liu, F.; Jiao, L. Hyperspectral image classification based on 3-D octave convolution with spatial–spectral attention network. IEEE Trans. Geosci. Remote Sens. 2021, 59, 2430–2447. [Google Scholar] [CrossRef]
  51. Zhong, Z.; Li, Y.; Ma, L.; Li, J.; Zheng, W.-S. Spectral–spatial transformer network for hyperspectral image classification: A factorized architecture search framework. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5514715. [Google Scholar] [CrossRef]
  52. Vaze, S.; Han, K.; Vedaldi, A.; Zisserman, A. Open-set recognition: A good closed-set classifier is all you need. arXiv 2021, arXiv:2110.06207. [Google Scholar]
  53. Pal, D.; Bundele, V.; Sharma, R.; Banerjee, B.; Jeppu, Y. Few-shot open-set recognition of hyperspectral images with outlier calibration network. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2022; pp. 3801–3810. [Google Scholar]
  54. Lin, S.; Zhang, M.; Cheng, X.; Shi, L.; Gamba, P.; Wang, H. Dynamic Low-Rank and Sparse Priors Constrained Deep Autoencoders for Hyperspectral Anomaly Detection. IEEE Trans. Instrum. Meas. 2023, 73, 2500518. [Google Scholar] [CrossRef]
Figure 1. Generating samples near the decision boundary.
Figure 1. Generating samples near the decision boundary.
Remotesensing 16 01285 g001
Figure 2. Overall network architecture.
Figure 2. Overall network architecture.
Remotesensing 16 01285 g002
Figure 3. Pseudo-sample center.
Figure 3. Pseudo-sample center.
Remotesensing 16 01285 g003
Figure 4. Graphical representation of RPL. (a) Reciprocal points of a known category. (b) Mutual exclusion between all known categories and reciprocal points. Where ★ signifies the open-set space of unknown samples, ▲ symbolizes the reciprocal points, and ● marks the known samples.
Figure 4. Graphical representation of RPL. (a) Reciprocal points of a known category. (b) Mutual exclusion between all known categories and reciprocal points. Where ★ signifies the open-set space of unknown samples, ▲ symbolizes the reciprocal points, and ● marks the known samples.
Remotesensing 16 01285 g004
Figure 5. A learnable dynamic thresholding strategy for open space, (a) schematic of RPL, (b) reinforced boundaries via GAN, and (c) high-dimensional feature space zoomed in via DW, where the blue region marks the constructed separation space.
Figure 5. A learnable dynamic thresholding strategy for open space, (a) schematic of RPL, (b) reinforced boundaries via GAN, and (c) high-dimensional feature space zoomed in via DW, where the blue region marks the constructed separation space.
Remotesensing 16 01285 g005
Figure 6. Analysis combined with data walk sample distribution.
Figure 6. Analysis combined with data walk sample distribution.
Remotesensing 16 01285 g006
Figure 7. Sample distribution and its information energy field before data walking.
Figure 7. Sample distribution and its information energy field before data walking.
Remotesensing 16 01285 g007
Figure 8. Sample distribution and its information energy field after six iterations.
Figure 8. Sample distribution and its information energy field after six iterations.
Remotesensing 16 01285 g008
Figure 9. PU dataset. (a) False-color image. (b) Closed-set ground truth.
Figure 9. PU dataset. (a) False-color image. (b) Closed-set ground truth.
Remotesensing 16 01285 g009
Figure 10. SA dataset. (a) False-color image. (b) Closed-set ground truth.
Figure 10. SA dataset. (a) False-color image. (b) Closed-set ground truth.
Remotesensing 16 01285 g010
Figure 11. IP dataset. (a) False-color image. (b) Closed-set ground truth.
Figure 11. IP dataset. (a) False-color image. (b) Closed-set ground truth.
Remotesensing 16 01285 g011
Figure 12. Thematic classification maps of different methods. (a) Ground truth map. (b) SSTN. (c) SnaTCHer. (d) PEELER. (e) RDOSR. (f) OpenMax. (g) MDL4OW. (h) SSLR. (i) RPLDW.
Figure 12. Thematic classification maps of different methods. (a) Ground truth map. (b) SSTN. (c) SnaTCHer. (d) PEELER. (e) RDOSR. (f) OpenMax. (g) MDL4OW. (h) SSLR. (i) RPLDW.
Remotesensing 16 01285 g012
Figure 13. T-SNE visualization results of feature space on SA data. (a) Original feature space. (b) Feature space constructed by RPLDW (the 16th category is unknown). (c) Feature space constructed by RPLDW (the 11th, 12th, 13th, and 14th categories are defined as the unknown ones).
Figure 13. T-SNE visualization results of feature space on SA data. (a) Original feature space. (b) Feature space constructed by RPLDW (the 16th category is unknown). (c) Feature space constructed by RPLDW (the 11th, 12th, 13th, and 14th categories are defined as the unknown ones).
Remotesensing 16 01285 g013
Table 1. Statistics of the three HSI databases.
Table 1. Statistics of the three HSI databases.
PUSAIP
IndexClass NameSamplesIndexClass NameSamplesIndexClass NameSamples
1Asphalt66311Brocoli-green-weeds-120091Alfalfa46
2Meadows18,6492Brocoli-green-weeds-237262Corn-notill1428
3Gravel20993Fallow19763Corn-mintill830
4Trees30644Fallow-rough-plow13944Corn237
5Painted-metal-sheet13255Fallow-smooth26785Grass-pasture483
6Bare-soil50296Stubble39596Grass-trees730
7Bitumen13307Celery35797Grass-pasture-mowed28
8Self-blocking-brick36828Grapes-untrained11,2718Hay-windrowed478
9Shadows9479Soil-vineyard-develop62039Oats20
10Corn-senesced-green-weeds327810Soybean-notill972
11Lettuce-romaine-4wk106811Soybean-mintill2455
12Lettuce-romaine-5wk192712Soybean-clean593
13Lettuce-romaine-6wk91613Wheat205
14Lettuce-romaine-7wk107014Woods1265
15Vineyard-untrained726815Buildings-Grass-Trees-Drives386
16Vineyard-vertical-trellis180716Stone-Steel-Towers93
Table 2. Physical parameters of the three HSI databases.
Table 2. Physical parameters of the three HSI databases.
PUSA IP
Number of channels115224102
Spectral ranges (μm)0.43~0.860.4~2.50.4~2.5
Spatial resolution610 × 340512 × 2171096 × 1096
Table 3. The 1-shot accuracy (%) of different methods in IP scenarios. (The 16th category is defined as unknown).
Table 3. The 1-shot accuracy (%) of different methods in IP scenarios. (The 16th category is defined as unknown).
Category3DOC-SSANSSTNOCNSnaTCHerPEELERRDOSROpenMaxMDL4OW SSLR RPLDW
1 65.7090. 2189.2481.3180.6990.2375.9975.1575.8389.10
2 60.8058.0872.6078.5278.0661.7166.7466.3566.3578.57
3 68.8561.0876.3579.4681.7564.8967.5469.4869.4878.18
4 54.4270.0587.5684.4781.5874.4371.8069.3469.3485.22
5 58.1764.2480.3079.4578.5968.2567.5366.8066.8082.50
6 59.0269.4786.8387.0388.7373.8173.9875.4275.4285.10
7 61.3771.2889.1079.8779.5675.7467.8967.6367.6386.98
8 63.3571.2089.0088.9887.4375.6575.6474.3174.3187.79
9 68.8870.2187.7789.1086.8874.6075.7473.8573.8584.92
10 52.1462.9678.7172.1463.1566.9061.3253.6853.6877.56
11 45.3252.9166.1361.6054.6356.2152.3659.4959.4974.52
12 45.7868.0185.0179.0162.2072.2667.1652.8752.8783.07
13 70.1071.0388.7988.8888.8875.4775.5575.5575.5586.55
14 54.7868.5985.7485.0981.3572.8872.3269.1569.1583.12
15 67.9069.2586.5681.3578.9473.5869.1567.1067.1084.14
Unknown66.2969.8183.2186.3285.4880.1473.3772.6672.6688.62
Closed OA64.4765.2981.6180.8276.3469.3768.7064.8964.8982.30
Closed AA71.5966.9083.6283.1077.8171.0870.6366.1366.1384.09
Closed Kappa61.3963.6679.5778.8974.2167.6367.0563.0863.0879.82
Open OA53.1359.6074.5075.9574.9863.3364.5663.7363.7377.03
Open AA46.6860.7575.9380.0075.8764.5468.0064.4964.4980.57
Open Kappa50.4459.1873.9775.6072.4462.8864.2661.5761.5776.36
Table 4. The 1-shot accuracy (%) of different methods in SA scenarios. (The 16th category is defined as unknown).
Table 4. The 1-shot accuracy (%) of different methods in SA scenarios. (The 16th category is defined as unknown).
Category3DOC-SSANSSTNOCNSnaTCHerPEELERRDOSROpenMaxMDL4OW SSLR RPLDW
1 75.8278.2591.8289.1289.1290. 8676.0179.3282.6189.60
2 65.5478.3772.6278.5378.0861.7366.7582.3384.3788.44
3 61.3678.3776.3779.4881.7764.9167.5685.2784.7689.73
4 61.9777.5787.5884.4981.6074.4471.8284.5384.1088.78
5 62.1175.8480.3279.4778.6168.2767.5582.5684.2988.92
6 78.0478.1186.8587.0588.7573.8273.9983.3683.1185.25
7 77.0978.1489.1279.8979.5875.7577.0082.3981.6786.79
8 57.1863.5789.0289.0087.4575.6775.6567.2765.6878.92
9 69.8078.2987.7989.1286.9074.6275.7585.2184.7088.60
10 66.7374.8378.7272.1663.1766.9261.3482.5482.0285.56
11 58.8177.6666.1561.6167.6456.2352.3784.0580.5188.73
12 56.1177.9485.0379.0362.2172.2867.1885.4585.5289.32
13 74.2278.1888.8188.9188.9175.4975.5785.1285.1888.10
14 57.3377.5985.7685.1181.3772.9072.3484.8984.8486.60
15 41.7568.7386.5881.3778.9673.6069.1672.0671.2580.43
Unknown41.3743.8153.0362.6053.4349.6355.2354.0949.4056.21
Closed OA59.8574.0281.6380.8476.3669.3968.7179.3783.1286. 94
Closed AA63.4875.6983.6483.1277.8271.0970.6582.9482.4688.22
Closed Kappa57.7573.4979.5978.9074.2367.6575.0777.5576.3585.49
Open OA53.2369.9181.0082.5881.5268.8570.1975.6076.0582.96
Open AA53.7270.8182.5686.9882.4970.1773.9377.2677.6683.18
Open Kappa50.7769.1480.4282.0278.7668.3669.8674.4974.9882.14
Table 5. The 1-shot accuracy (%) of different methods in PU scenarios. (The 9th category is defined as unknown).
Table 5. The 1-shot accuracy (%) of different methods in PU scenarios. (The 9th category is defined as unknown).
Category3DOC-SSANSSTNOCNSnaTCHerPEELERRDOSROpenMaxMDL4OW SSLR RPLDW
1 47.2078.6972.8156.9374.7678.1274.2082.2378.1177.17
2 58.6680.0574.0859.8976.0480.8578.3585.1082.4882.44
3 48.1780.0972.9450.1476.0975.7175.9979.7079.9879.65
4 77.1881.8279.8276.7677.7380.3579.9484.5884.1578.18
5 81.0388.1083.6882.0083.7067.8079.2471.3783.4187.41
6 65.6784.4678.4852.3980.2479.7378.2183.9382.3384.06
7 73.0687.3481.5561.3682.9780.5579.3084.7983.4784.69
8 49.1177.5570.2250.6573.6874.3067.6478.2171.2072.09
Unknown45.0349.2859.7853.1829.9135.9843.6037.8845.9062.30
Closed OA57.7583.1383. 1765. 6978.9877.4878.4281.5682.5585.35
Closed AA60.2384.4680.0461.9380.2478.2778.9282.3983.0784.79
Closed Kappa48.8181.5277.0253.3877.4577.0277.2481.0781.3183.63
Open OA48.9359.8773.4657.9375.8873.6774.6777.5578.6080.28
Open AA49.6857.6068.8854.6373.7273.3473.5077.2077.3778.67
Open Kappa50.6457.4370.5051.3473.5672.6372.9676.4576.8077.93
Table 6. The 5-shot accuracy (%) of different methods in IP scenarios. (The 16th category is defined as unknown).
Table 6. The 5-shot accuracy (%) of different methods in IP scenarios. (The 16th category is defined as unknown).
Category3DOC-SSANSSTNOCNSnaTCHerPEELERRDOSROpenMaxMDL4OW SSLR RPLDW
1 73.0099. 5699. 0199.0099.0099.3284.4383.5084.2699.00
2 67.5664.5480.6787.2486.7368.5774.1573.7273.7287.30
3 76.5067.8684.8388.2990.8372.1175.0577.2177.2186.87
4 60.4677.8397.2993.8690.6482.7079.7877.0477.0494.69
5 64.6371.3889.2288.2887.3275.8475.0474.2274.2291.67
6 65.5877.1896.4896.7098.5982.0182.2083.8083.8094.55
7 68.1879.2099.0088.7488.4084.1575.4375.1475.1496.64
8 70.3879.1198.8998.8797.1484.0684.0482.5782.5797.54
9 76.5478.0297.5299.0096.5382.8984.1582.0582.0594.36
10 57.9469.9687.4580.1670.1774.3368.1459.6459.6486.18
11 50.3558.7873.4868.4460.7062.4658.1766.1066.1082.80
12 50.8775.5794.4687.7969.1180.2974.6258.7458.7492.30
13 77.8978.9298.6598.7698.7683.8583.9583.9583.9596.17
14 60.8776.2295.2794.5490.3980.9880.3676.8376.8392.36
15 75.4576.9496.1890.3987.7181.7576.8374.5574.5593.49
Unknown73.6677.5792.4695.9194.9889.0481.5280.7380.7398.47
Closed OA71.6372.5490.6889.8084.8277.0876.3372.1072.1091.44
Closed AA71.5974.3392.9192.3386.4578.9778.4873.4873.4893.43
Closed Kappa68.2270.7388.4187.6582.4675.1574.5070.0970.0988.69
Open OA59.0366.2282.7884.3983.3170.3671.7370.8170.8185.59
Open AA51.8667.5084.3788.8984.3071.7175.5671.6671.6689.52
Open Kappa56.0465.7582.1984.0080.4969.8671.4068.4268.4284.84
Table 7. The 5-shot accuracy (%) of different methods in SA scenarios. (The 16th category is defined as unknown).
Table 7. The 5-shot accuracy (%) of different methods in SA scenarios. (The 16th category is defined as unknown).
Category3DOC-SSANSSTNOCNSnaTCHerPEELERRDOSROpenMaxMDL4OW SSLR RPLDW
1 83.3285.9999.0197.9497.9499. 3283.5287.1690.7898.46
2 72.0286.1279.8086.3085.8067.8373.3690.4892.7197.19
3 67.4286.1283.9287.3489.8571.3374.2493.7193.1598.6
4 68.1085.2496.2492.8589.6781.8178.9292.8992.4297.56
5 68.2583.3588.2687.3386.3875.0274.2391.6692.6397.71
6 85.7685.8395.4495.6697.5381.1381.3191.6091.3393.68
7 84.7185.8797.9487.7987.4583.2584.6290.5489.7595.37
8 62.8369.8697.8397.8196.1083.1583.1473.9272.1886.72
9 76.7086.0396.4797.9495.4982.0083.2593.6493.0797.36
10 73.3382.2386.5179.3069.4273.5367.4090.7190.1394.02
11 64.6285.3472.6967.7030.3761.7957.5592.3688.4797.5
12 61.6685.6593.4486.8568.3779.4373.8293.9093.9898.15
13 81.5685.9197.5997.7097.7082.9583.0493.5493.6096.81
14 63.0085.2694.2593.5289.4280.1179.4993.2993.2395.17
15 45.8875.5295.1589.4286.7780.8776.0179.1978.3088.39
Unknown45.4648.1458.2868.7958.7154.5460.6959.4454.2961.77
Closed OA65.7781.3489.7088.8383.9176.2575.5187.2291. 9196. 39
Closed AA69.7683.1891.9191.3485.5278.1277.6491.1490.6196.94
Closed Kappa63.4780.7687.4686.7181.5774.3473.7085.2283.9093.94
Open OA58.5076.8289.0190.7489.5875.6677.1383.0883.5791.16
Open AA59.0477.8190.7295.5890.6577.1181.2484.9085.3491.41
Open Kappa55.7975.9888.3890.1386.5575.1276.7781.8582.3990.26
Table 8. The 5-shot accuracy (%) of different methods in PU scenarios. (The 9th category is defined as unknown).
Table 8. The 5-shot accuracy (%) of different methods in PU scenarios. (The 9th category is defined as unknown).
Category3DOC-SSANSSTNOCNSnaTCHerPEELERRDOSROpenMaxMDL4OW SSLR RPLDW
1 53.0388.4281.8063.9684.0087.7783.3792.3987.7686.71
2 65.9189.9483.2467.2985.4490.8488.0495.6292.6792.63
3 54.1289.9981.9656.3485.4985.0785.3889.5589.8789.49
4 86.7291.9389.6986.2587.3390.2889.8295.0394.5587.84
5 91.0598.9994.0292.1394.0476.1889.0380.1993.7298.21
6 73.7994.9088.1858.8690.1689.5987.8894.3092.5094.45
7 82.0998.1391.6368.9493.2290.5189.1095.2793.7995.16
8 55.1887.1478.9056.9182.7883.4976.0087.8880.0081.00
Unknown50.5955.3767.1659.7533.6040.4348.9942.5651.5770.00
Closed OA64.8993.4193.9471. 5188.7487.0688.1191.6492.7595.90
Closed AA67.6794.9089.9469.5990.1687.9488.6792.5793.3495.27
Closed Kappa54.8491.6086.5459.9787.0286.5486.7991.0991.3693.97
Open OA54.9869.7482.5465.0985.2582.7783.8987.1388.3190.20
Open AA55.8267.1977.4061.3882.8382.4082.5886.7486.9388.39
Open Kappa56.9067.0079.2157.6882.6581.6181.9885.9086.2987.56
Table 9. Multiclassification accuracy of different methods in SA scenarios (%).
Table 9. Multiclassification accuracy of different methods in SA scenarios (%).
CategoryFCPNDLRSPs-DAEsPOSMSSMLP-RPLRPLDW
1.0099.00 99.00 98.46 97.01 98.45
2.0087.30 87.02 97.19 97.39 98.19
3.0086.87 88.85 98.60 98.33 98.11
4.0094.69 92.67 97.56 97.42 97.51
5.0091.67 89.50 97.71 97.61 97.43
6.0094.55 96.57 93.68 94.91 95.94
7.0096.64 92.52 95.37 94.92 97.86
8.0097.54 97.34 86.72 81.35 89.74
9.0094.36 95.45 97.36 97.67 98.52
10.0086.18 78.18 94.02 94.45 96.39
15.0082.80 86.75 97.50 95.32 91.89
16.0092.30 80.71 98.15 98.54 98.00
Unknown78.47 76.73 61.77 74.20 79.32
Closed OA91.44 88.13 91. 3992. 6392.9
Closed AA93.43 89.94 96.94 96.16 95.98
Closed Kappa88.69 85.58 93.94 91.13 92.14
Open OA85.59 84.45 91.16 89.57 93.65
Open AA89.52 86.91 91.41 90.62 91.57
Open Kappa84.84 82.67 90.26 88.50 90.34
Table 10. Overall classification accuracy on IP dataset (%).
Table 10. Overall classification accuracy on IP dataset (%).
Sample Size MDL4OW SSLR FCPNDLRSPs-DAEsPOSMSSMLP-RPLRPLDW
20Unknown93.05 92.15 96.51 93.7094.9594.0397.48
Open OA81.87 80.83 84.41 81.9683.5482.4884.73
Open AA86.24 81.79 86.03 83.5288.0083.4688.62
Open Kappa81.50 78.09 83.81 81.3783.1679.6883.99
50Unknown95.40 94.48 98.95 96.0697.3596.4199.21
Open OA83.94 82.87 86.55 84.0385.6584.5688.15
Open AA88.42 83.85 88.20 85.6390.2385.5692.20
Open Kappa83.56 80.06 85.93 83.4285.2681.6987.39
100Unknown97.49 96.81 91.38 98.4399.4898.7899.52
Open OA86.01 84.91 88.68 86.1087.7686.6490.72
Open AA90.60 85.92 90.37 87.7492.4587.6794.89
Open Kappa85.61 82.03 88.04 85.4887.3683.7189.93
150Unknown97.55 97.17 93.82 90.8099.5499.1699.82
Open OA88.08 86.95 90.81 88.1689.8788.7293.72
Open AA92.78 87.98 92.55 89.8594.6789.7898.02
Open Kappa87.67 84.00 90.16 87.5389.4685.7292.90
200Unknown97.59 97.54 96.26 93.1699.5899.5399.94
Open OA90.14 88.99 92.94 90.2391.9890.8194.14
Open AA94.95 90.05 94.72 91.9696.8991.8998.47
Open Kappa89.73 85.98 92.27 89.5991.5687.7393.33
Table 11. Overall classification accuracy on SA dataset (%).
Table 11. Overall classification accuracy on SA dataset (%).
Sample Size MDL4OW SSLR FCPNDLRSPs-DAEsPOSMSSMLP-RPLRPLDW
20Unknown60.06 55.23 61.26 60.6460.6856.1561.15
Open OA84.84 85.35 90.04 87.4286.5787.0990.25
Open AA86.70 87.16 91.20 88.5488.4788.9390.49
Open Kappa83.59 84.15 89.05 86.4585.3085.8789.36
50Unknown60.97 56.60 61.90 61.2761.6157.5563.62
Open OA87.41 87.94 92.77 90.0689.2089.7393.89
Open AA89.33 89.80 93.96 91.2391.1591.6394.15
Open Kappa86.13 86.70 91.75 89.0787.8888.4792.97
100Unknown61.58 57.51 62.33 61.6862.2358.4865.47
Open OA89.13 89.66 94.58 91.8390.9491.4996.63
Open AA91.08 91.56 95.80 93.0192.9493.4396.89
Open Kappa87.81 88.40 93.55 90.8289.6190.2095.67
150Unknown62.19 58.43 62.76 62.1062.8559.4267.02
Open OA90.84 91.38 96.40 93.6092.6993.2598.91
Open AA92.83 93.32 97.65 94.8094.7395.2299.18
Open Kappa89.50 90.10 95.34 92.5791.3391.9497.93
200Unknown62.95 59.57 63.30 62.6263.6260.5867.33
Open OA92.98 93.54 98.68 95.8094.8895.4599.36
Open AA95.02 95.52 99.95 97.0496.9697.4799.63
Open Kappa91.61 92.22 97.59 94.7593.4894.1198.38
Table 12. Overall classification accuracy on PU dataset (%).
Table 12. Overall classification accuracy on PU dataset (%).
Sample Size MDL4OWSSLRFCPNDLRSPs-DAEsPOSMSSMLP-RPLRPLDW
20Unknown61.2960.0366.0765.0262.1361.0569.30
Open OA84.5385.6891.5188.8586.2687.4289.30
Open AA84.1684.3488.9186.3285.8886.0687.50
Open Kappa83.3483.7288.7286.1385.0485.4386.68
50Unknown72.5471.5567.1666.0873.4172.6072.10
Open OA87.1088.2794.2891.5488.8790.0792.90
Open AA86.7186.9091.6088.9388.4888.6791.04
Open Kappa85.8786.2591.4088.7487.6288.0190.18
100Unknown73.3872.5667.8966.7974.2673.6374.20
Open OA88.8090.0096.1393.3390.6291.8495.61
Open AA88.4188.6093.4090.6890.2190.4193.69
Open Kappa87.5587.9593.2090.4889.3489.7492.81
150Unknown74.2173.5768.6267.5075.1174.6675.95
Open OA90.5191.7397.9895.1392.3693.6197.87
Open AA90.1190.3095.1992.4291.9592.1595.90
Open Kappa89.2489.6494.9992.2291.0691.4795.00
200Unknown75.2574.8369.5368.3876.1875.9576.30
Open OA92.6593.90100.2997.3794.5495.8198.32
Open AA92.2392.4397.4494.6094.1294.3296.34
Open Kappa91.3491.7597.2394.4093.2093.6295.44
Table 13. RPLDW ablation results (%).
Table 13. RPLDW ablation results (%).
Sample SizeDATASETRPLDWRPLDW without RPLRPLDW without GAN
Accuracy
Rate
Accuracy
Variation
Mean
Variation
Accuracy
Rate
Accuracy
Variation
Mean
Variation
20Pavia University83.0460.72−22.32−23.8657.46−25.58−27.11
Salinas86.3361.44−24.8958.54−27.79
Indian Pines75.2750.89−24.3847.32−27.95
50Pavia University82.4060.95−21.45−23.7857.48−24.92−26.89
Salinas85.9761.73−24.2458.73−27.24
Indian Pines75.5549.91−25.6447.05−28.5
100Pavia University81.3457.85−23.49−22.9755.39−25.95−25.62
Salinas85.3762.75−22.6259.43−25.94
Indian Pines76.0053.19−22.8151.03−24.97
150Pavia University80.2757.39−22.88−22.5755.60−24.67−24.23
Salinas84.7763.43−21.3461.65−23.12
Indian Pines76.4653.04−23.4251.57−24.89
200Pavia University79.2158.94−20.27−20.7055.99−23.22−22.95
Salinas84.1763.42−20.7562.01−22.16
Indian Pines76.9255.79−21.1353.44−23.48
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, T.; Wang, Y.; Li, J.; Du, Y. Generative Adversarial Network and Mutual-Point Learning Algorithm for Few-Shot Open-Set Classification of Hyperspectral Images. Remote Sens. 2024, 16, 1285. https://doi.org/10.3390/rs16071285

AMA Style

Xu T, Wang Y, Li J, Du Y. Generative Adversarial Network and Mutual-Point Learning Algorithm for Few-Shot Open-Set Classification of Hyperspectral Images. Remote Sensing. 2024; 16(7):1285. https://doi.org/10.3390/rs16071285

Chicago/Turabian Style

Xu, Tuo, Ying Wang, Jie Li, and Yuefan Du. 2024. "Generative Adversarial Network and Mutual-Point Learning Algorithm for Few-Shot Open-Set Classification of Hyperspectral Images" Remote Sensing 16, no. 7: 1285. https://doi.org/10.3390/rs16071285

APA Style

Xu, T., Wang, Y., Li, J., & Du, Y. (2024). Generative Adversarial Network and Mutual-Point Learning Algorithm for Few-Shot Open-Set Classification of Hyperspectral Images. Remote Sensing, 16(7), 1285. https://doi.org/10.3390/rs16071285

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop