Next Article in Journal
A Vehicle Recognition Model Based on Improved YOLOv5
Previous Article in Journal
Demand Response Management via Real-Time Pricing for Microgrid with Electric Vehicles under Cyber-Attack
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Kernel Reverse Neighborhood Discriminant Analysis

School of Computer Science and Cyber Engineering, Guangzhou University, Guangzhou 510006, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(6), 1322; https://doi.org/10.3390/electronics12061322
Submission received: 31 December 2022 / Revised: 13 February 2023 / Accepted: 28 February 2023 / Published: 10 March 2023
(This article belongs to the Section Artificial Intelligence)

Abstract

:
Currently, neighborhood linear discriminant analysis (nLDA) exploits reverse nearest neighbors (RNN) to avoid the assumption of linear discriminant analysis (LDA) that all samples from the same class should be independently and identically distributed (i.i.d.). nLDA performs well when a dataset contains multimodal classes. However, in complex pattern recognition tasks, such as visual classification, the complex appearance variations caused by deformation, illumination and visual angle often generate non-linearity. Furthermore, it is not easy to separate the multimodal classes in lower-dimensional feature space. One solution to these problems is to map the feature to a higher-dimensional feature space for discriminant learning. Hence, in this paper, we employ kernel functions to map the original data to a higher-dimensional feature space, where the nonlinear multimodal classes can be better classified. We give the details of the deduction of the proposed kernel reverse neighborhood discriminant analysis (KRNDA) with the kernel tricks. The proposed KRNDA outperforms the original nLDA on most datasets of the UCI benchmark database. In high-dimensional visual recognition tasks of handwritten digit recognition, object categorization and face recognition, our KRNDA achieves the best recognition results compared to several sophisticated LDA-based discriminators.

1. Introduction

Linear discriminant analysis (LDA) [1] is a supervised classical linear learning method stemming from the Fisher criterion [2], which can be used not only for supervised classification problems but also for feature dimensional reduction. Its idea is not difficult to understand. For a set of given training samples, LDA attempts to find the optimal projection matrix that can maximize the between-class scatter matrix and minimize the within-class scatter matrix, simultaneously. As a supervised feature extraction method, LDA has demonstrated its feasibility and efficiency in pattern classification [3], and it has been widely applied in hyperspectral image classification [4], EEG signals analysis [5], re-identification [6], etc. LDA is also a critical evaluation approach to visual classification, such as handwriting recognition, face recognition and object categorization [7].
LDA assumes that the samples of all training data follow a multivariate Gaussian distribution that has the same covariance but different mean values for different classes [8]; once this statement is not satisfied, for example, when several different clusters are encountered from samples in the same class, LDA gives undesirable results [9]. To solve this problem, a two-step subclass division method was proposed [10]; the first step applies k-means for clustering and the second step uses the EM-alike framework to optimize the subclasses. Recently, the reverse nearest neighbors approach has been exploited with LDA to form neighborhood linear discriminant analysis (nLDA) [7]. As an unsupervised outlier detection method, reverse nearest neighbor (RNN) can eliminate the “isolated points” in the training set. nLDA expects that a sample and its RNN should be as close as possible, while the RNN from two samples that belong to different classes should be as far as possible in the projected space. As the scatter matrix is defined based on RNN, which does not directly consider the whole class, the problem of datasets containing multimodal classes can be overcome by nLDA. nLDA is a method where the scatter matrix is directly defined on the neighborhood, which is a new localized discriminator. The smallest subclass can be thought of as a neighborhood; thus, it can perform discriminative learning without obeying the independently and identically distributed (i.i.d.) assumption of sample data.
However, since nLDA is extended from LDA, which is a linear parametric method, thus it may fail to handle nonlinear data. It is well known that the complex appearance variations caused by deformation, illumination and visual angle often generate non-linearity [11,12]. If we ignore the non-linearity of the data, the linear discriminators often give a lower performance. One practical solution is to map the points of the original input data to a higher-dimensional feature space and then learn the linear classifier on that space for discriminant learning. However, addressing the non-linearity in the higher-dimensional feature space is non-trivial, especially when the dimension of the mapped space is infinite. Many methods have been proposed to use the kernel trick to handle this problem [11,12,13], which is also exploited in our proposed approach.
To handle the nonlinear data, circumvent the assumption of i.i.d. of sample data and solve the multimodal classes problem simultaneously, in this paper, we extend nLDA to its kernel version, called kernel reverse neighborhood discriminant analysis (KRNDA). The Gaussian kernel is employed to map the input data into a higher-dimensional feature space where non-linearity can be alleviated. The deduction of KRNDA with kernel tricks is presented in detail. We give extensive evaluations on the UCI dataset, and the visual classification tasks of handwriting recognition, face recognition and object categorization to verify the effectiveness of the proposed approach.
The rest of this paper is organized as follows. Section 2 reviews the related work. Section 3 briefly introduces the basic principles of LDA and nLDA methods. Subsequently, we propose our KRNDA method; the deduction with the kernel trick is presented in Section 4. Experiments are conducted in Section 5. Section 6 concludes this paper.

2. Related Work

Research works on LDA have been studied for decades [14,15,16] and plentiful approaches have been proposed to address different problems of LDA. For pattern recognition, the small sample size (SSS) problem is an essential issue for LDA. The SSS problem generally occurs when the feature dimension of each sample is very high, while there are insufficient training samples of each class. This result often leads to the singularity or ill-condition of the within-class scatter matrix, and the LDA can not be solved by the eigen-decomposition. To address this problem, Fisherfaces [1] were proposed to conduct principal component analysis (PCA) for reducing the feature dimension of each sample and then implement the LDA algorithm. The singularity or ill-condition of the within-class scatter matrix caused by the SSS problem can also be solved by adding a small value to the diagonal of the within-class scatter matrix, which is known as regularized LDA [17]. Pang et al. [18] combined the clustering and regularization terms to alleviate the deviation of the scatter matrix caused by the SSS problem. There are many other approaches proposed to circumvent the singularity of the within-class scatter matrix and the instability of its reverse caused by the SSS. For example, direct LDA (DLDA) [19] and null space LDA (NLDA) [20] select different subspaces of the scatter matrix to avoid the singularity. Eigenfeature regularization and extraction (ERE) [21] retains and reweights the entire eigenspace of the within-class scatter matrix for non-singularity to persist. This approach achieved excellent performance on face recognition.
The multimodal classes problem of LDA arouses a lot of attention. Marginal fisher analysis (MFA) [22] employs graph embedding to characterize the nearest neighbors to avoid the i.i.d. of the sample data. Exploiting manifolds to represent local structures is also one of the solutions. For instance, local fisher discriminant analysis (LFDA) [23] effectively combines the ideas of locality preserving projections (LPPs) [24] and LDA; it attains inter-class separation and intra-class local structure preservation by defining a local intra-class and a local inter-class scatter matrix to handle the multimodal classes problem. Locality sensitive discriminant analysis (LSDA) [25] projects the dataset to a lower-dimensional subspace to preserve local manifold structure and discriminant information. Neighbors with the same label should be as close as possible; that is the aim of LFDA and LSDA. Nonparametric discriminant analysis (NDA) [9] solves the multimodal problem of LDA by using k-nearest neighbors, which constructs a kind of nonparametric inter-class divergence in a local area. Nonparametric discriminant analysis for face recognition [26] extends NDA to address multi-class situations. Recently, nLDA [7] directly defined the scatter matrix on the neighborhood to solve the multimodal classes problem and achieved remarkable results.
To handle the non-linearity problem, many LDA-based approaches are extended to their kernel versions to learn the nonlinear structures by the powerful tool of the kernel trick [27]; for example, generalized discriminant analysis (GDA) [11] detailed the deduction of kernel LDA. Kernel Fisherface (KLDA) [28], kernel direct-LDA (KDDA) [13], null space kernel LDA (NKDA) [12] and complete discriminant evaluation and feature extraction (CDEFE) [29] are the kernel versions of PCA+LDA, DLDA, NDA and ERE, respectively. In the previous studies, researchers worked out an alternative formulation of kernel LPP (KLPP) used to develop a framework for KPCA+LPP algorithms. Ref. [30] achieved good results in recognition tasks, such as face recognition and radar target recognition. In solving the nonlinear problems encountered in image classification, the KAHISD/KCHISD of the Cevikalp and Triggs [31] article is an extension of the affine/convex hull-based image classification task to its kernel version. Zhu et al. [32] proposed the KCH-ISCRC with kernel tricks, which well addresses collaborative image set-based representation and classification (ISCRC). It is easy to see that Gaussian kernels can successfully solve nonlinear problems in various applications [28,29,31,32].
LDA has new developments with recent technologies. Dorfer et al. [33] proposed a deep linear discriminant analysis with the deep neural network. Alarcón and Destercke [34] proposed a Gaussian discriminator to fuse the near-ignorance priors and robust Bayesian analysis. Belous et al. [35] proposed a framework called dual spatial discriminative projection learning and successfully applied it to image classification tasks. Hu et al. [36] proposed a cross-modal discriminative network for cross-modal learning.

3. LDA and nLDA

In this section, we briefly introduce the algorithm of LDA [1] and nLDA [7].

3.1. LDA

The idea and algorithm of LDA are simple and intuitive. Let X = x 1 , x 2 , x 3 , , x N R D × N denotes the D-dimensional training dataset, where N is the number of total training samples and D is the dimension of each sample. The class label of the training sample x i R D is denoted by y i with y ( x i ) ( 1 , 2 , , C ) , where C represents the total class number of samples. We use N i to describe the number of training samples in class i. The execution of LDA is to use the most discriminative projection matrix ϱ = ϱ 1 , ϱ 2 , ϱ 3 ϱ d R D × d ( d D ) to project the training set X R D × N from high-dimensional space to a lower-dimensional feature space by:
ϱ T X R d × N .
The optimal projected matrix ϱ of LDA can be obtained by maximizing the equation:
J L D A ( ϱ ) = ϱ T S b ϱ ϱ T S w ϱ .
where the between-class scatter matrix S b and within-class scatter matrix S w are defined, respectively, by:
S b = 1 C i = 1 C u i u u i u T .
S w = j = 1 C y ( x i ) = j x i u j x i u j T .
where u i is the mean vector of the i-th class:
u i = 1 N i y ( x j ) = i x j .
u is the mean of all training samples:
u = 1 N i = 1 N x i .
In LDA, Equation (2) can be reduced to the following eigen-decomposition problem:
S b ϱ = λ S w ϱ .
By retaining d eigenvectors corresponding to the largest eigenvalues, the most discriminative projection matrix ϱ is formed.

3.2. nLDA

LDA constructs the within-class scatter matrix with the information provided by the whole class, while nLDA defines the local neighborhood scatter by using the reverse k-nearest neighbor ( R N N k ) set. The nearest neighbors (NN) represents the samples with high similarity. For N samples in dataset X , the k-NN of point x p X is defined by N N k x p , X , which ranks the similarities of x p to the other points in X from high to low and selects the top k nearest samples to represent the nearest neighbors of point x p . The Euclidean distance is the common approach for measuring the similarity of k-NN. The definition of the R N N k [37] is based on the k-NN. Given a dataset X and the point x p X , the definition of the R N N k set of x p is presented by:
R N N k x p , X = x q x q X \ x p , x p N N k x q , X .
The number of the reverse nearest neighbors may not be equal to k. It depends on the distribution of the sample’s data. In some exceptional cases, the number of reverse nearest neighbors is empty. Read [7] for more detailed information.
For the training set X = x 1 , x 2 , x 3 , , x N , the associated label of x i is y i { 1 , , C } . In the projected space, nLDA has the same optimization objective as LDA, that is, to find the optimal projection matrix ϱ which maximizes the equation:
J n L D A ( ϱ ) = ϱ T S b n L D A ϱ ϱ T S w n L D A ϱ .
where S b n L D A and S w n L D A represent the between-neighborhood scatter matrix and within-neighborhood scatter matrix, respectively. In [7], to avoid the consuming calculation, the approximate between-neighborhood scatter matrix is preferred. They are defined by Equations (10) and (11).
S b n L D A = i = 1 R N N k x i , X y i t N x j N N x i , X X y i | R N N k x j , X y j t y i y j u i u j u i u j T .
S w n L D A = i = 1 R N N k x i , X y i t N x j R N N k x i , X y i x j u i x j u i T .
where R N N k x j , X y j is the number of local samples obtained from the RNN and X y j is the set of image samples with the same label as sample x j . In R N N k x j , X y j t , t is the threshold of the number of RNN samples, i.e., when the number of RNN in R N N k ( x i , X y i ) reaches t, the local neighborhood is constructed and the samples in the local neighborhood are used to calculate the mean vector u i which is formulated by:
u i = 1 R N N k x i , X y i x j R N N k x i , X y i x j .
x j N N x i , X X y i indicates that x j is the nearest neighbor from other classes of x i and u j is the mean vector of the RNN of x j in set X y j . u j can be formulated by:
u j = 1 R N N k ( x j , X y j ) x i R N N k ( x j , X y j ) x i .
Then, maximizing Equation (9) can be reduced to the following generalized eigen-decomposition problem:
S b n L D A ϱ = λ S w n L D A ϱ .
By retaining d orthogonal eigenvectors according to the largest eigenvalues, the projection matrix ϱ can be used to extract discriminative features. Finally, the Euclidean distance between a test sample x * and u i is represented as d i s t ϱ T u i , ϱ T x * , the label of x * is determined by the nearest distance between x * and the mean of each RNN set in the gallery data, which can be expressed as:
min R N N k x i , X y i t dist ϱ T u i , ϱ T x * . .

4. Kernel Reverse Neighborhood Discriminant Analysis

In this section, we detail the proposed kernel reverse neighborhood discriminant analysis (KRNDA). We use the kernel trick to deal with the non-linearity underlying the input data of nLDA. First, the kernel method based on nLDA requires the definition of a mapping for each observed sample x being mapped from the original feature space to a higher-dimensional feature space F :
φ : R D F , x φ ( x ) .
Subsequently, the mapped training dataset of X can be represented as φ ( X ) = φ ( x 1 ) , φ ( x 2 ) , , φ ( x N ) . Usually, the mapping function is not explicitly specified [38], especially when the mapped feature space is infinite-dimensional. Hence, many kernel functions are defined for implicitly calculating the inner product of two image vectors in the mapped space F . For two images x i and x j , the inner product of them in the mapped space can be expressed as:
k x i , x j = φ x i , φ x j = φ x i T φ x j .
where · is the inner product. How to select a suitable kernel function to implicitly calculate the inner product of two vectors in higher-dimensional feature space is a critical factor for kernel methods [39]. The typical and effective Gaussian kernel function is used in this paper:
k g x i , x j = exp x i x j 2 2 2 σ 2 .
One of the advantages of the Gaussian kernel is that it has only one free parameter σ , which is easily tuned.
The proposed KRNDA aims to find an optimal projection matrix ω = ω 1 , ω 2 , , ω h , where h is the reserved dimension of the final extracted feature, such that the mapped data in space F can be projected onto a more discriminative and lower-dimensional feature space by ω T φ X . The optimal ω can be obtained by maximizing the following equation:
J K R N D A ω = ω T S b K R N D A ω ω T S w K R N D A ω .
where S b K R N D A is the between-neighborhood scatter matrix and S w K R N D A is the within-neighborhood scatter matrix in space F .
According to [11], the projection matrix ω can be linearly represented by all the training data in the mapped space F . Therefore we can obtain:
ω i = j = 1 N α i j φ x j .
where α i j denotes the coefficient of the linear combination. For a sample x m in higher-dimensional space φ x m , the projected vector can be expressed by:
g m = ω T φ x m = ω 1 , ω 2 , , ω h T φ x m = ω 1 T φ x m , ω 2 T φ x m , , ω h T φ x m T .
Obviously, the final extracted feature g m does not seem to be explicitly expressed, since each ω i and the φ x m contain the implicit mapping φ . . However, the kernel trick can circumvent this problem. According to Equations (17) and (20), each ω i T φ x m in Equation (21) can be calculated as:
ω i T φ x m = j = 1 N φ ( x j ) T α i j T φ ( x m ) = j = 1 N α i j φ ( x j ) T φ m = α i 1 φ ( x 1 ) T φ ( x m ) + α i 2 φ ( x 2 ) T φ ( x m ) + + α i N φ ( x N ) T φ ( x m ) = α i 1 , α i 2 , , α i N φ ( x 1 ) T φ ( x m ) φ ( x 2 ) T φ ( x m ) φ ( x N ) T φ ( x m ) = α i T K m .
where α i = α i 1 , α i 2 , , α i N T . According to Equation (17), we have k x i , x m = φ x i T φ x m ; K m denotes the inner product between all training samples and the m-th training sample, in the mapped space F ; then the extracted feature g m can be explicitly expressed as:
g m = ω T φ x m = α 1 T K m , α 2 T K m , , α h T K m T = α T K m .
where α = α 1 , α 2 , , α h T . We define the between-neighborhood scatter matrix of our KRNDA in Equation (19) and formulate the ω T S b K R N D A ω as:
ω T S b K R N D A ω = i = 1 R N N k φ x i , φ X y i t N φ ( x j ) N N k φ x i , φ ( X ) φ X y i R N N k φ x j , φ X y j t y i y j ω T ( φ ( u i ) φ ( u j ) ω T ( φ ( u i ) φ ( u j ) T = α T B α .
where
B = i = 1 R N N k φ x i , φ X y i t N φ ( x j ) N N k φ x i , φ ( X ) φ X y i R N N k φ x j , φ X y j t y i y j P i P j P i P j T
In Equation (24), φ ( u i ) denotes the mean vector of the local neighborhood computed by reverse k-nearest neighbors ( R N N k ) of φ ( x i ) in dataset φ ( X y i ) with the same label y i . Certainly, the number of the R N N k should be equal to or larger than threshold t. Likewise, φ ( u j ) denotes the mean vector of the local neighborhood computed by R N N k of φ ( x j ) in dataset φ ( X y j ) with the same label y j . It should be noted that the labels of these two R N N k neighborhoods are different ( y i y j ), which exhibits the between-class information. The calculation of P i and P j in Equation (25) can be formulated by calculating ω T φ ( u i ) and ω T φ ( u j ) . Using the relationship of Equation (23), we give the deduction of ω T φ ( u i ) , for instance:
ω T φ ( u i ) = ω T φ ( x m ) R N N k ( φ ( x i ) , φ ( X y i ) ) φ ( x m ) R N N k φ ( x i ) , φ ( X y i ) = α T φ ( x m ) R N N k ( φ ( x i ) , φ ( X y i ) ) K m R N N k φ ( x i ) , φ ( X y i )
Hence, P i = φ ( x m ) R N N k ( φ ( x i ) , φ ( X y i ) ) K m R N N k φ ( x i ) , φ ( X y i ) . Similarly, P j = φ ( x m ) R N N k ( φ ( x j ) , φ ( X y j ) ) K m R N N k φ ( x j ) , φ ( X y j ) .
We employ the definition of the approximate between-neighborhood scatter in [7] to construct S b K R N D A ; hence, for each φ ( u i ) , the number of the between-class reverse nearest neighborhoods is restricted to k by using N N k φ x i , φ ( X ) φ X y i . This approach reduces the amount of computation when the dataset is very large [7]. The calculation of R N N k ( φ ( x i ) , φ ( X y i ) ) follows the definition of the R N N k set in Equation (8), where the distance of nearest neighbors is calculated in the mapped higher-dimensional feature space with the kernel trick.
Likewise, the within-neighborhood scatter can also be explicitly expressed by the kernel trick; the equation ω T S w K R N D A ω in Equation (19) can be formulated as:
ω T S w K R N D A ω = i = 1 R N N k φ x i , φ X y i t N φ ( x j ) R N N k φ x i , φ X y i ω T φ ( x j ) φ ( u i ) ω T φ ( x j ) φ ( u i ) T = α T W α .
where:
W = i = 1 R N N k φ x i , φ X y i t N φ ( x j ) R N N k φ x i , φ X y i K j Q i K j Q i T
where Q i = φ ( x m ) R N N ( φ ( x i ) , φ ( X y i ) K m R N N k φ ( x i ) , φ ( X y i ) | is the mean of higher-dimensional features in R N N k of φ ( x i ) with the same label y i . According to Equations (24), (25), (27) and (28), the object function of Equation (19) can be rewritten as:
J K R N D A ( ω ) = α T B α α T W α .
Hence, the optimization of maximizing Equation (29) can be reduced to the generalized eigen-decomposition problem:
B α = λ W α .
The optimal projection matrix can be obtained by applying the eigen-decomposition to W 1 B . We reserve h eigenvectors according to the largest eigenvalues to form the final coefficient matrix α = α 1 , α 2 , , α h .
At the testing stage, for a test sample x * , we first map it to the higher-dimensional space F by φ ( x * ) , then the final extracted feature can be expressed as ω T φ ( x * ) ; according to the kernel trick, it can be rewritten as α T K * . K * denotes the kernel vector of inner production between all gallery (or training) samples and test sample x * in the mapped space F . We compare the Euclidean distance between α T K * and the projected mean vector ω T φ ( u i ) , where ω T φ ( u i ) can be explicitly rewritten by α T K u i . φ ( u i ) is the mean vector of the i-th gallery RNN with the same label, and K u i denotes the kernel vector of inner production between all gallery samples and the mean vector u i in the mapped space F . Therefore the label of test sample φ ( x * ) can be determined by the nearest Euclidean distance of:
min R N N k φ ( x i ) , φ ( X y i ) t dist α T K u i , α T K * .

5. Experiments

Experiments are conducted on two handwritten digit datasets and the University of California at Irvin (UCI) benchmark database. In addition, the COIL-20 (Columbia Object Image Library) [40] object recognition dataset and the ORL face recognition dataset [41] are also adopted for evaluation.
To verify the effectiveness of our method, we compare the performances of our KRNDA with the original nLDA. Several LDA-based discriminators are added for comparison, including LDA, LFDA, ccLDA and L 2 , 1 RLDA . The LFDA is a traditional local discriminator which has been successfully applied to pedestrian re-identification [42]. We implemented the code of ccLDA [18], which uses clustering to solve the SSS problem. L 2 , 1 RLDA is a discriminator with excellent discriminant effect and the performance of L 2 , 1 norm is better than that of l 2 -norm [43]. The codes of LFDA, L 2 , 1 RLDA and nLDA are provided by the authors.
The parameters in the above discriminators are set as follows: for LDA, the final extracted features are preserved according to 95% energy of the eigenvalues. For LFDA, we set the parameter K = 2 to achieve the optimal effect in our experiments (K is a parameter used in the local scaling method). To avoid the singularity caused by the SSS problem, regularization is applied to the within-class scatter matrix: S w = S w + τ I , where I is the identity matrix and the τ is a small value, i.e, 10 3 × t r a c e ( S w ) . For ccLDA, we follow the original experimental settings: K < C ( here, K is the cluster number and C is the class number) and 98% energy of the eigenvalues is reserved for experiments. In L 2 , 1 RLDA , we use the original parameter settings provided by the authors.
There are four main parameters related to the proposed KRNDA. The first one is the parameter t, which controls the isolated points in Equations (24) and (27); we set t = 1 according to the article [7], which has proved that it does not affect the performance much. We also applied this setting to our KRNDA in all experiments. The second parameter is the value of k in R N N k ; [7] has proved that the value of k has better performance in the range { 10 , 20 , 30 , 40 } ; hence, we adopted similar setting with nLDA in our KRNDA. The third parameter is the σ in Gaussian kernel, see Equation (18); we evaluate this parameter in Section 5.4 and apply the best value for our experiments. The last parameter is the dimension of the final extracted feature, similar to nLDA, and we preserve 95% energy of the eigenvalues to construct the final features. All of the experiments were implemented on an Intel(R) Core(TM) i7-11700K (3.60GHz) PC, using Matlab R2020b.

5.1. Experiments on Handwritten Digit Recognition

Two handwritten digit datasets, Mnist and USPS, are used in our evaluation. The purpose of KRNDA is to solve the nonlinear problem; however, it is the extension of the nLDA which performs well on the multimodal class problem. Therefore, we construct the classification problem of handwritten digit recognition as a binary classification task to verify the performance of KRNDA in solving the multimodal classes problem. For binary digit recognition, it is to determine whether the numbers are odd or even. The odd class is 1, 3, 5, 7 and 9, and the even class is 2, 4, 6, 8 and 10. Some digital images of the two classes are shown in Figure 1. As the five numbers in each of the odd class and even class are quite different, it can be considered that there are five subclasses (clusters) in each class.
The Mnist dataset has a total of 60,000 samples in the training set; the details of the training and testing samples are shown in Table 1. Adopting the whole set directly for training is time-consuming and may lead to being out of memory. Hence, we constructed subsets for our experiments. We randomly selected different numbers of samples for experiments and created four subsets: Mnist600, Mnist1000, Mnist6000 and Mnist10000. This means that 600, 1000, 6000 and 10,000 samples are selected from the total training set for training, respectively. For the corresponding testing sets, we randomly selected 100, 166, 1000 and 1666 samples from the original 10,000 testing samples to form four testing subsets for evaluation. The USPS dataset has a total of 9298 handwritten digit images. We conducted our experiments according to the protocol in [7], where the training set contains 7291 images and the testing set contains 2007 images, and the detailed training and testing numbers of each class are shown in Table 2.
Firstly, we compare the proposed method (KRNDA) with the Gaussian kernel function, sigmoid kernel function (SRNDA) and polynomial kernel function (PRNDA) on the USPS and Mnist datasets. We chose the parameters corresponding to the highest recognition rates during our experiments. For the Mnist datasets, we performed 10-fold cross-validations to obtain the average recognition rates and standard deviations.
From Table 3, we can find that the Gaussian kernel function outperforms the polynomial and sigmoid kernel functions. The poor performance of the sigmoid kernel function may be caused by the multimodal classes data. The Gaussian kernel demonstrates a better ability to correct the non-linearity and it has only one adjustable parameter, which is easy to use for evaluations. Consequently, in the following experiments of this paper, we employ the Gaussian kernel function with KRNDA for all evaluations.
Here, we present the recognition rates on the USPS handwritten digit dataset, and the average recognition rates and standard deviations of 10-fold cross-validations on Mnist handwritten digit datasets. As shown in Table 4, compared with the original nLDA, our KRNDA consistently achieves better recognition rates, especially on the Mnist600 and Mnist1000 subsets, because the small training subset may cause more non-linearity. These results demonstrate the effectiveness of the proposed KRNDA on handling the nonlinear data problem. We can see that the proposed KRNDA also outperforms other LDA-based methods, such as LDA, LFDA, ccLDA and L 2 , 1 RLDA . This result illustrates that our KRNDA inherits the advantage in solving the multimodal classes problem.

5.2. Experiments on COIL-20 and ORL Datasets

In this subsection, we use the COIL-20 object dataset and the ORL face dataset to evaluate more complicated image classification tasks. The COIL-20 consists of 20 different objects, such as car, cat, cup and eraser, etc. For each object, there are 72 images of different viewpoints [44]. The images are normalized and resized to 20 × 20 pixels. We randomly selected 50% of the images per class for training and the remaining images for testing. Some sample images are shown in Figure 2.
The ORL face dataset contains 400 images of 40 individuals. Images were captured at different times, and other variations of expression (open or closed eyes, smiling or non-smiling) and facial details (glasses or no glasses) [45] are involved. The face images of the ORL database are normalized and resized to 20 × 20 pixels. We randomly selected 50% of the images per class for training and the remaining images for testing. Some sample images are shown in Figure 3.
For the COIL-20 object dataset and the ORL face dataset, we also performed 10-fold cross-validations to obtain the average recognition rates and standard deviations. As shown in Table 5, the proposed KRNDA outperforms the original nLDA, and the recognition rates reach the highest, 98.62% and 99.90% on ORL and COIL-20, respectively. We can see that, in Figure 2 and Figure 3, the COIL-20 dataset with multi-view samples causes much more distortion than the ORL dataset. Hence the accuracy increment of KRNDA on the COIL-20 dataset is larger than the ORL dataset compared to nLDA. Our KRNDA also performs well on face recognition which contains non-linearity caused by lighting and pose variation. These results have shown the effectiveness of the proposed kernel efficient solver. Due to the benefits of RNN and the non-linearity solver, the proposed KRNDA can handle different applications of pattern recognition.

5.3. Experiments on UCI Benchmark Datasets

We selected 40 benchmark datasets from the University of California at Irvine (UCI) machine learning repository [46] for our experiments. The details of the class number, sample number and feature dimension of each dataset are listed in Table 6. The number of classes is from 2 to 10. The number of features is from 2 to 256, which is less than the visual classification tasks. For each dataset, we randomly selected approximately 50% of the samples for training and the remaining samples for testing. We also employed the LDA-based discriminators, which were introduced previously, for comparison.
The classification results of the UCI datasets are reported in Table 7. In each row, the best classification accuracies are highlighted. As can be seen, the proposed KRNDA outperforms other discriminators in most cases, and the average classification accuracy of all datasets achieves 81.63%, which is better than all other discriminators. The degradation of KRNDA, such as in the datasets of Bupa, Sonar and breast-tissue, etc., may be caused by the lower non-linearity (caused by lower-dimensional features of UCI datasets) or the non-optimal value setting of parameter σ .

5.4. Parameter Evaluation

The parameter σ of the Gaussian kernel plays an important role in exhibiting the performance of the proposed KRNDA. It can be seen as the controller of handling the non-linearity. Therefore, we conducted experiments to tune the optimal value of σ . Experiments were conducted on a subset of the Mnist database; we randomly selected 600 samples for training and 100 samples for testing, and 10 cross-validations were conducted to obtain the mean accuracy.
Experimental results are shown in Figure 4. The Y-axis is the recognition accuracy and the X-axis is the value of σ . As can be seen, the best value σ = 6 reaches the highest accuracy with 95.1%; hence, in most of our experiments, the value of σ is set as σ = 6 . Certainly, this value is non-optimal for different applications and different datasets; however, it achieves considerable performances in most cases of our experiments.

6. Conclusions

In this paper, we extended nLDA to its kernel version for handling the non-linearity of the input data, which is called kernel reverse neighborhood discriminant analysis (KRNDA). We mapped the original samples to a higher-dimensional feature space for discriminative analysis, however, since the mapping function generalized by the specific kernel functions can not be explicitly expressed. We gave a comprehensive deduction of the kernel extension with the kernel trick. We evaluated the performances of our KRNDA on a multimodal classes problem of handwritten digit recognition with three different kernel functions; the Gaussian kernel function has better performances than the polynomial and sigmoid kernel functions; this result demonstrates the better capacity to correct the non-linearity by the Gaussian kernel. As a result, we evaluated the KRNDA with the Gaussian kernel function on visual recognition tasks of USPS and Mnist handwritten digits datasets, a COIL-20 object categorization dataset and an ORL face recognition dataset. The performances of the proposed KRNDA outperform the original nLDA and other previous LDA-based discriminators, such as LDA, LFDA, ccLDA and L 2 , 1 RLDA . In addition, we evaluated our KRNDA on non-visual and lower-dimensional datasets of the UCI benchmark database. Due to lower non-linearity or the non-optimal value setting of parameter σ , the proposed KRNDA is inferior to other discriminators on some datasets of the UCI database; however, the average performance of KRNDA is still considerable.

Author Contributions

Conceptualization, W.L., H.T.; Methodology, W.L., H.T.; Software, W.L., M.X. and J.F.; Validation, J.D. and S.Y.; Data curation, W.L., M.X. and J.F.; Writing—original draft, W.L.; Writing—review & editing, W.L., H.T., J.D., S.Y. and G.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Natural Science Foundation of Guangdong Province under Grant 2021A1515011859 and 2020A1515010423, in part by the National Natural Science Foundation of China under Grant 61701126, 62176071, in part by the research projects in Guangzhou University, China, under Grant RP2020123, and in part by the Guangzhou Basic Research Program Jointly Funded by City and University under Grant 202102010395.

Data Availability Statement

http://yann.lecun.com/exdb/mnist/ (accessed on 30 December 2022); https://github.com/hanzheteng/Pattern-Recognition/tree/master/usps (accessed on 30 December 2022); https://paperswithcode.com/dataset/orl (accessed on 30 December 2022); https://www.cs.columbia.edu/CAVE/software/softlib/coil-20.php (accessed on 30 December 2022); http://archive.ics.uci.edu/ml/index.php (accessed on 30 December 2022).

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Belhumeur, P.N.; Hespanha, J.P.; Kriegman, D.J. Eigenfaces vs. fisherfaces: Recognition using class specific linear projection. IEEE Trans. Pattern Anal. Mach. Intell. 1997, 19, 711–720. [Google Scholar] [CrossRef] [Green Version]
  2. Fukunaga, K. Statistical pattern recognition. In Handbook of Pattern Recognition and Computer Vision; World Scientific: Singapore, 1993; pp. 33–60. [Google Scholar]
  3. Wang, J.; Liu, Z.; Zhang, K.; Wu, Q.; Zhang, M. Robust sparse manifold discriminant analysis. Multimed. Tools Appl. 2022, 81, 20781–20796. [Google Scholar] [CrossRef]
  4. Huang, K.K.; Ren, C.X.; Liu, H.; Lai, Z.R.; Yu, Y.F.; Dai, D.Q. Hyperspectral image classification via discriminative convolutional neural network with an improved triplet loss. Pattern Recognit. 2021, 112, 107744. [Google Scholar] [CrossRef]
  5. Nkengfack, L.C.D.; Tchiotsop, D.; Atangana, R.; Louis-Door, V.; Wolf, D. EEG signals analysis for epileptic seizures detection using polynomial transforms, linear discriminant analysis and support vector machines. Biomed. Signal Process. Control. 2020, 62, 102141. [Google Scholar] [CrossRef]
  6. Li, W.H.; Zhong, Z.; Zheng, W.S. One-pass person re-identification by sketch online discriminant analysis. Pattern Recognit. 2019, 93, 237–250. [Google Scholar] [CrossRef] [Green Version]
  7. Zhu, F.; Gao, J.; Yang, J.; Ye, N. Neighborhood linear discriminant analysis. Pattern Recognit. 2022, 123, 108422. [Google Scholar] [CrossRef]
  8. Zollanvari, A.; Dougherty, E.R. Generalized consistent error estimator of linear discriminant analysis. IEEE Trans. Signal Process. 2015, 63, 2804–2814. [Google Scholar] [CrossRef]
  9. Fukunaga, K.; Mantock, J. Nonparametric discriminant analysis. IEEE Trans. Pattern Anal. Mach. Intell. 1983, 6, 671–678. [Google Scholar] [CrossRef]
  10. Tao, Y.; Yang, J.; Chang, H. Enhanced iterative projection for subclass discriminant analysis under EM-alike framework. Pattern Recognit. 2014, 47, 1113–1125. [Google Scholar] [CrossRef]
  11. Baudat, G.; Anouar, F. Generalized discriminant analysis using a kernel approach. Neural Comput. 2000, 12, 2385–2404. [Google Scholar] [CrossRef]
  12. Liu, W.; Wang, Y.; Li, S.Z.; Tan, T. Null space-based kernel fisher discriminant analysis for face recognition. In Proceedings of the Sixth IEEE International Conference on Automatic Face and Gesture Recognition, Seoul, Republic of Korea, 17–19 May 2004; pp. 369–374. [Google Scholar]
  13. Lu, J.; Plataniotis, K.N.; Venetsanopoulos, A.N. Face recognition using kernel direct discriminant analysis algorithms. IEEE Trans. Neural Netw. 2003, 14, 117–126. [Google Scholar] [PubMed] [Green Version]
  14. Li, Y.; Tian, X.; Liu, T.; Tao, D. On better exploring and exploiting task relationships in multitask learning: Joint model and feature learning. IEEE Trans. Neural Networks Learn. Syst. 2017, 29, 1975–1985. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Liu, Z.; Shi, K.; Zhang, K.; Ou, W.; Wang, L. Discriminative sparse embedding based on adaptive graph for dimension reduction. Eng. Appl. Artif. Intell. 2020, 94, 103758. [Google Scholar] [CrossRef]
  16. Zhang, L.; Liu, Z.; Pu, J.; Song, B. Adaptive graph regularized nonnegative matrix factorization for data representation. Appl. Intell. 2020, 50, 438–447. [Google Scholar] [CrossRef]
  17. Friedman, J.H. Regularized discriminant analysis. J. Am. Stat. Assoc. 1989, 84, 165–175. [Google Scholar] [CrossRef]
  18. Pang, Y.; Wang, S.; Yuan, Y. Learning regularized LDA by clustering. IEEE Trans. Neural Netw. Learn. Syst. 2014, 25, 2191–2201. [Google Scholar] [CrossRef]
  19. Yu, H.; Yang, J. A direct LDA algorithm for high-dimensional data—with application to face recognition. Pattern Recognit. 2001, 34, 2067–2070. [Google Scholar] [CrossRef] [Green Version]
  20. Chen, L.F.; Liao, H.Y.M.; Ko, M.T.; Lin, J.C.; Yu, G.J. A new LDA-based face recognition system which can solve the small sample size problem. Pattern Recognit. 2000, 33, 1713–1726. [Google Scholar] [CrossRef]
  21. Jiang, X.; Mandal, B.; Kot, A. Eigenfeature regularization and extraction in face recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 30, 383–394. [Google Scholar] [CrossRef]
  22. Yan, S.; Xu, D.; Zhang, B.; Zhang, H.J.; Yang, Q.; Lin, S. Graph embedding and extensions: A general framework for dimensionality reduction. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 29, 40–51. [Google Scholar] [CrossRef] [Green Version]
  23. Sugiyama, M. Dimensionality reduction of multimodal labeled data by local fisher discriminant analysis. J. Mach. Learn. Res. 2007, 8, 1027–1061. [Google Scholar]
  24. He, X.; Niyogi, P. Locality preserving projections. Adv. Neural Inf. Process. Syst. 2003, 16, 153–160. [Google Scholar]
  25. Cai, D.; He, X.; Zhou, K.; Han, J.; Bao, H. Locality sensitive discriminant analysis. In Proceedings of the IJCAI, Hyderabad, India, 6–12 January 2007; Volume 2007, pp. 1713–1726. [Google Scholar]
  26. Li, Z.; Lin, D.; Tang, X. Nonparametric discriminant analysis for face recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 755–761. [Google Scholar] [PubMed]
  27. Breneman, J. Kernel Methods for Pattern Analysis; Cambridge University Press: Cambridge, UK, 2005. [Google Scholar]
  28. Yang, M.H. Kernal Eigenfaces vs. Kernal Fisherfaces: Face Recognition Using Kernal Methods, Automatrix Face and Gesture Recognition, 202. In Proceedings of the Fourth International Conference on Computation in Electromagnetics—CEM 2002, Bournemouth, UK, 8–11 April 2002; pp. 208–213. [Google Scholar]
  29. Jiang, X.; Mandal, B.; Kot, A. Complete discriminant evaluation and feature extraction in kernel space for face recognition. Mach. Vis. Appl. 2009, 20, 35–46. [Google Scholar] [CrossRef]
  30. Feng, G.; Hu, D.; Zhang, D.; Zhou, Z. An alternative formulation of kernel LPP with application to image recognition. Neurocomputing 2006, 69, 1733–1738. [Google Scholar] [CrossRef]
  31. Cevikalp, H.; Triggs, B. Face recognition based on image sets. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 2567–2573. [Google Scholar]
  32. Zhu, P.; Zuo, W.; Zhang, L.; Shiu, S.C.K.; Zhang, D. Image set-based collaborative representation for face recognition. IEEE Trans. Inf. Forensics Secur. 2014, 9, 1120–1132. [Google Scholar] [CrossRef] [Green Version]
  33. Dorfer, M.; Kelz, R.; Widmer, G. Deep linear discriminant analysis. arXiv 2015, arXiv:1511.04707. [Google Scholar]
  34. Alarcón, Y.C.C.; Destercke, S. Imprecise gaussian discriminant classification. In Proceedings of the International Symposium on Imprecise Probabilities: Theories and Applications, PMLR, Ghent, Belgium, 3–6 July 2019; pp. 59–67. [Google Scholar]
  35. Belous, G.; Busch, A.; Gao, Y. Dual subspace discriminative projection learning. Pattern Recognit. 2021, 111, 107581. [Google Scholar] [CrossRef]
  36. Hu, P.; Peng, X.; Zhu, H.; Lin, J.; Zhen, L.; Wang, W.; Peng, D. Cross-modal discriminant adversarial network. Pattern Recognit. 2021, 112, 107734. [Google Scholar] [CrossRef]
  37. Korn, F.; Muthukrishnan, S. Influence sets based on reverse nearest neighbor queries. ACM Sigmod Rec. 2000, 29, 201–212. [Google Scholar] [CrossRef]
  38. Yang, M.; Zhu, P.; Van Gool, L.; Zhang, L. Face recognition based on regularized nearest points between image sets. In Proceedings of the 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Shanghai, China, 22–26 April 2013; pp. 1–7. [Google Scholar]
  39. Tan, H.; Gao, Y. Kernelized Fast Algorithm for Regularized Hull-Based Face Recognition With Image Sets. IEEE Access 2018, 6, 36395–36407. [Google Scholar] [CrossRef]
  40. Nene, S.; Nayar, S. Columbia Object Image Library (Coil-20). Technical Report CUCS-005-96. 1996. Available online: https://www1.cs.columbia.edu/CAVE/publications/pdfs/Nene_TR96.pdf (accessed on 30 December 2022).
  41. Samaria, F.S.; Harter, A.C. Parameterisation of a stochastic model for human face identification. In Proceedings of the 1994 IEEE Workshop on Applications of Computer Vision, Sarasota, FL, USA, 5–7 December 1994; pp. 138–142. [Google Scholar]
  42. Pedagadi, S.; Orwell, J.; Velastin, S.; Boghossian, B. Local fisher discriminant analysis for pedestrian re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 3318–3325. [Google Scholar]
  43. Nie, F.; Wang, Z.; Wang, R.; Wang, Z.; Li, X. Towards Robust Discriminative Projections Learning via Non-Greedy 2,1-Norm MinMax. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 2086–2100. [Google Scholar] [CrossRef] [PubMed]
  44. Liu, X.; Ma, Z.; Niu, G. Mixed region covariance discriminative learning for image classification on riemannian manifolds. Math. Probl. Eng. 2019, 2019, 1261398. [Google Scholar] [CrossRef] [Green Version]
  45. Hu, H.; Zhang, P.; Ma, Z. Direct kernel neighborhood discriminant analysis for face recognition. Pattern Recognit. Lett. 2009, 30, 902–907. [Google Scholar] [CrossRef]
  46. Dua, D.; Graff, C. UCI Machine Learning Repository; University of California: Irvine, CA, USA, 2017; Available online: http://archive.ics.uci.edu/ml (accessed on 30 December 2022).
Figure 1. The digit samples in USPS and Mnist; the left is odd digits and right is even digits.
Figure 1. The digit samples in USPS and Mnist; the left is odd digits and right is even digits.
Electronics 12 01322 g001
Figure 2. Examples of three objects from the COIL-20 database.
Figure 2. Examples of three objects from the COIL-20 database.
Electronics 12 01322 g002
Figure 3. Examples of three subjects from the ORL face database.
Figure 3. Examples of three subjects from the ORL face database.
Electronics 12 01322 g003
Figure 4. The evaluation of parameter σ .
Figure 4. The evaluation of parameter σ .
Electronics 12 01322 g004
Table 1. Training and test sample numbers of the Mnist dataset.
Table 1. Training and test sample numbers of the Mnist dataset.
ClassNumber of Training SetNumber of Test Set
05923980
167421135
259581032
361311010
45842982
55421892
65918958
762651028
85851974
959491009
Table 2. Training and test sample numbers of the USPS dataset.
Table 2. Training and test sample numbers of the USPS dataset.
ClassNumber of Training SetNumber of Test Set
0644177
11194359
21005264
3731198
4658166
5652200
6556160
7664170
8645147
9542166
Table 3. The accuracy (%) comparison of different kernel method recognition problems.
Table 3. The accuracy (%) comparison of different kernel method recognition problems.
DataUSPSMnist600Mnist1000Mnist6000Mnist10000
SRNDA81.9657.50 ± 4.4355.24 ± 4.3858.71 ± 3.6058.62 ± 2.63
PRNDA96.8191.30 ± 3.1093.01 ± 2.8895.55 ± 0.7396.21 ± 0.29
KRNDA97.6595.70 ± 2.0596.20 ± 1.5598.15 ± 0.3498.06 ± 0.32
Table 4. The accuracy (%) comparison of USPS and Mnist digit recognition problems.
Table 4. The accuracy (%) comparison of USPS and Mnist digit recognition problems.
DataUSPSMnist600Mnist1000Mnist6000Mnist10000
LDA89.8373.80 ± 4.5377.53 ± 2.6982.11 ± 1.2382.39 ± 0.79
LFDA94.4777.50 ± 2.5883.01 ± 2.9183.92 ± 1.0185,59 ± 1.20
ccLDA89.9370.60 ± 7.2679.22 ± 2.5282.52 ± 0.8182.70 ± 1.29
L 2 , 1 RLDA 96.7190.86 ± 2.0392.97 ± 1.8096.63 ± 0.4897.17 ± 0.29
nLDA96.6681.40 ± 4.1389.10 ± 1.7195.47 ± 0.9496.16 ± 0.43
KRNDA97.6595.70 ± 2.0596.20 ± 1.5598.15 ± 0.3498.06 ± 0.32
Table 5. The accuracy (%) comparison of ORL and COIL-20 datasets recognition problems.
Table 5. The accuracy (%) comparison of ORL and COIL-20 datasets recognition problems.
DataORLCOIL-20
LDA96.87 ± 0.9795.52 ± 0.85
LFDA98.58 ± 0.7896.38 ± 1.00
ccLDA94.30 ± 1.7993.82 ± 0.60
L 2 , 1 RLDA 97.23 ± 1.6399.38 ± 0.21
nLDA96.78 ± 1.4696.39 ± 0.67
KRNDA98.62 ± 0.7399.90 ± 0.13
Table 6. The details of UCI benchmark datasets.
Table 6. The details of UCI benchmark datasets.
DatasetClassesSamplesFeature
Aggregation745,170,102,273,34,130,342
Bupa2145,2006
Indian2167,41610
Robotnavigation4826,2097,2205,32824
Sonar297,11160
balance-scale349,288,2884
balloons29,74
bank24000,52116
breast-cancer2201,859
breast-tissue621,15,18,16,14,229
acute261,596
splice3767,768,165560
vertebral-column-2clases2210,1006
Indian2167,41610
acute_nephritis270,506
car465,1210,384,696
cardiotocography-3clases31655,295,17621
chess-krvkp21527,166936
climate-simulation2494,4618
blood2570,1785
arrhythmia23664,373620
waveform-noise31692,1653,165540
mammographic2516,4455
cylinder-bands2200,31235
fertility288,129
iris350,50,504
knowledge342,55,755
ionosphere2126, 22634
ozone22463,7372
monks-32266,2886
nursery44320,328,4266,40448
parkinsons248,14723
promoters253,5357
semeion10161,162,159,159,161,159,161,158,155,158256
statlog-australian-credit2222,46814
tic-tac-toe2332,6269
statlog-german-credit2700,30024
pittsburg-bridges-TYPE616,11,44,13,11,107
mushroom24208,391622
Table 7. The classification accuracy (%) of different discriminators on UCI benchmark datasets.
Table 7. The classification accuracy (%) of different discriminators on UCI benchmark datasets.
DatasetLDALFDAccLDA L 2 , 1 RLDA nLDAKRNDA
Aggregation99.4999.4999.4999.4999.7599.75
Bupa59.8855.2362.2157.5672.0955.24
Indian66.3263.2357.3964.6067.0170.10
Robotnavigation71.7486.4784.4285.4885.3787.02
Sonar76.9279.8183.6577.8876.9280.77
balance-scale87.8289.7479.8178.2185.5889.10
balloons37.5050.0050.0050.0062.5075.00
bank83.9886.6486.5985.4986.6488.32
breast-cancer58.7462.9464.3464.3463.6465.74
breast-tissue49.0666.0460.3856.6050.9415.09
acute100.00100.00100.00100.00100.00100.00
splice74.1171.1663.7062.1375.4983.70
vertebral-column-2clases75.4878.0678.0681.2980.0081.94
Indian66.3263.2357.3964.6067.0170.10
acute_nephritis100.00100.00100.00100.00100.00100.00
car79.9894.5693.6393.5294.6896.41
balance-scale87.8289.7479.8178.2185.5889.10
cardiotocography-3clases87.4987.3988.8188.3388.4382.88
chess-krvkp91.2495.6289.9987.1194.8795.49
climate-simulation91.8590.0086.6787.0491.4890.37
blood71.6672.9971.9373.5375.4066.84
arrhythmia69.7873.4672.2772.1187.8683.00
waveform-noise80.5677.9261.6875.6079.9284.52
mammographic71.2577.0875.0077.7174.1774.79
cylinder-bands64.8466.8068.7568.7568.3658.59
fertility84.0074.0080.0082.0080.0086.00
iris97.3394.6794.6794.6785.1496.00
knowledge93.0289.5388.3774.4296.0082.56
ionosphere85.1483.4383.4382.2961.8689.71
ozone95.3595.4395.9894.9596.3797.79
monks-383.7597.8382.3188.4597.8398.56
nursery86.0591.5688.8582.8993.7895.38
parkinsons83.5176.2984.5479.3882.4785.54
promoters62.2662.2652.8377.3662.2675.47
semeion83.9285.6865.5889.5788.4493.72
statlog-australian-credit53.0461.7455.0758.8460.2964.35
tic-tac-toe97.9198.5498.54100.0098.5498.96
statlog-german-credit64.2069.4068.6066.8068.8072.20
pittsburg-bridges-TYPE51.9246.1548.0842.3138.4646.15
mushroom89.78100.00100.00100.00100.00100.00
AVERAGE77.8880.1077.5778.5980.6081.63
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, W.; Tan, H.; Feng, J.; Xie, M.; Du, J.; Yang, S.; Yan, G. Kernel Reverse Neighborhood Discriminant Analysis. Electronics 2023, 12, 1322. https://doi.org/10.3390/electronics12061322

AMA Style

Li W, Tan H, Feng J, Xie M, Du J, Yang S, Yan G. Kernel Reverse Neighborhood Discriminant Analysis. Electronics. 2023; 12(6):1322. https://doi.org/10.3390/electronics12061322

Chicago/Turabian Style

Li, Wangwang, Hengliang Tan, Jianwei Feng, Ming Xie, Jiao Du, Shuo Yang, and Guofeng Yan. 2023. "Kernel Reverse Neighborhood Discriminant Analysis" Electronics 12, no. 6: 1322. https://doi.org/10.3390/electronics12061322

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop