Next Article in Journal
Information Rates for Channels with Fading, Side Information and Adaptive Codewords
Previous Article in Journal
Uniform Error Estimates of the Finite Element Method for the Navier–Stokes Equations in R2 with L2 Initial Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Classification with Linearity-Enhanced Logits to Softmax Function

1
School of Mathematics and Statistics, Yunnan Unverisity, Kunming 650504, China
2
School of Information Science and Engineering, Yunnan Unverisity, Kunming 650504, China
3
The Key Lab of Intelligent Systems and Computing of Yunnan Province, Yunnan University, Kunming 650504, China
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(5), 727; https://doi.org/10.3390/e25050727
Submission received: 11 February 2023 / Revised: 20 March 2023 / Accepted: 27 March 2023 / Published: 27 April 2023

Abstract

:
Recently, there has been a rapid increase in deep classification tasks, such as image recognition and target detection. As one of the most crucial components in Convolutional Neural Network (CNN) architectures, softmax arguably encourages CNN to achieve better performance in image recognition. Under this scheme, we present a conceptually intuitive learning objection function: Orthogonal-Softmax. The primary property of the loss function is to use a linear approximation model that is designed by Gram–Schmidt orthogonalization. Firstly, compared with the traditional softmax and Taylor-Softmax, Orthogonal-Softmax has a stronger relationship through orthogonal polynomials expansion. Secondly, a new loss function is advanced to acquire highly discriminative features for classification tasks. At last, we present a linear softmax loss to further promote the intra-class compactness and inter-class discrepancy simultaneously. The results of the widespread experimental discussion on four benchmark datasets manifest the validity of the presented method. Besides, we want to explore the non-ground truth samples in the future.

1. Introduction

In the past few years of artificial intelligence research, Convolutional Neural Networks (CNNs) have played a crucial role in deep learning classification tasks. Benefiting from advanced network architecture [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16] and discriminative capacity [17], CNNs have dramatically upgraded the performance across various visual classification tasks, such as object recognition [18,19], face verification [20,21], molecular biology [22,23], and hand-written digit recognition [24]. A recent current towards learning is to strengthen CNN with more discriminative power and more applied scenarios. In the research of machine learning, Long [25] proposed a new self-training semi-supervised deep learning (SSDL) method to further explore the fault diagnosis models. Xu [26] presented the Global Contextual Multiscale Fusion Network (GCMFN) to better accommodate noisy and unbalanced scenarios. In addition, several studies have employed CNN in the medical field. Fan [27] attached SVM to the fully connected layer to better identify the cancer datasets. Sekhar [28] proposed a novel transfer learning method to detect brain tumors. As arguably one of the most crucially used components in CNN architecture, softmax is widely used in image classification tasks.
Intuitively, the softmax loss is a popular choice to learn discriminative features in the pioneering work [29], but the original softmax loss only discriminates between partial features and does not separate inter-class features enough. Several variants have been offered to enhance the discriminative capacity of the softmax loss. The center loss [30] was proposed to compact the intra-class distance by calculating the L2 distance between the feature vector and its class center. By the cooperative penalization of the softmax, center loss achieves stronger discriminability and obtains a smaller intra-class distance. However, updating the factual center is impractical as the training number grows.
Some research has also adapted distance constraints to a pair or a triplet of samples to improve the discriminative power so that similar samples are as compact as possible and dissimilar samples are as spread apart as possible. For example, contrastive loss [31] further distinguishes similar samples from different samples by feature extraction. On the other hand, regarding triplet loss, Refs. [32,33,34] presented the triplet training samples for the first time. This method guarantees an anchor sample is far from a negative sample and is close to a positive sample in a triplet of samples. Furthermore, batches of all triplet loss [35] and hard triplet loss [36] were proposed to administer more constraints and achieved a stronger generalization capability of features. Based on triplet loss, N-pair loss [37] employs a positive sample and multiple negative samples for an anchor sample to train the network. Specifically, N-pair loss applies N-1 negative samples in each training phase, which selects more information and increases the convergence speed. However, both contrastive loss and triplet loss cannot oblige on each sample of features, which will lead to unstable convergence as the size of the samples grows dramatically. Although these methods can achieve higher discriminative power and diminish this problem, they also complicate the network and make training more difficult.
On the other hand, various studies have attempted to reformulate the softmax by implementing the margin-based loss function. Unlike previous loss methods, these studies aim to improve the discriminability of softmax loss by presenting the angular penalty, which was set between feature vectors and corresponding weight vectors of the last fully connected layer. By the angular margin m, the margin-based loss functions enhance inter-class distance and try to achieve stronger discriminability. For instance, SphereFace [38,39] and L-Softmax [40] first came up with the concept of angular margins by a multiplicative angular penalty, which further separated various classes and compacted the same classes. The novel loss functions enhance the discriminability of features by the change of the decision boundary, but these can lead to an unstable training process due to the difficulty of optimization. CosFace [41], AM-Softmax [42], and Soft-Softmax [43] suggested enhancing angular discriminative power by use of an additive cosine margin. Benefitting from the cosine margin, this can thereby further develop the discriminative power and provide an intuitive explanation. Building on the previous method, ArcFace [44,45] presented an additive angular margin that effectively unites the multiplicative angular margin, cosine margin, and angular margin. Profiting from the advantages of a unified framework, ArcFace plays a crucial role in deep classification and achieved sample features with stronger discriminability. For the ArcFace, the feature margin between different classes was set to the same and fixed, and this may not adapt to the real situation of various classes. In addition, several lines of research have been improved in various directions based on ArcFace. For example, Dyn-Arcface [46] replaced the flexible margin penalty based on the distance between each class center and the other class centers. ElasticFace loss [47] relaxes the fixed single margin by deploying a random margin drawn from a normal distribution. To reflect the more real properties of class separability, Groupface [48] suggested enriching the feature representations with group-aware representations based on the Arcface. AdaptiveFace put forward hard prototype mining (HPM) to adaptively adjust the margins between various classes to solve the problem of imbalanced training data in deep classification. Moreover, Uniformface [49] presented equalized distances between various class centers by adding a new loss function on SphereFace. ASL [50] mitigates the bias induced by data imbalance and increases interclass diversity. It is obvious that the flexible models present better performance compared to the fixed margin. Generally, the margin-based loss functions enhance the inter-class discrepancy by proposing an angular penalty, which is between feature vectors and the corresponding weight vectors of the last fully connected layer. However, these methods only penalize the partial samples in the angular space, which leads to unfair considerations for every class.
Based on this, we present a linear Orthogonal-Softmax loss to achieve stronger discriminability. Inspired by the Taylor-Softmax [51], the proposed Orthogonal-Softmax presents various orthogonal polynomials approximation for the e z of softmax, which is designed by executing Gram–Schmidt orthogonalization. By employing an approximated linear logit, the proposed Orthogonal-Softmax has a sturdier linear relationship than the softmax loss and the Taylor-Softmax. On the other hand, benefitting from the thinking of CosFace and AM-Softmax, we added margin m to the new loss and achieved Orthogonal-M. Compared to the Orthogonal-Softmax, Orthogonal-M increases the inter-class separation and achieves stronger discriminative power. The principal contributions can be outlined as follows:
(1) The proposed Orthogonal-Softmax applies Gram–Schmidt orthogonalization in the softmax loss, which presents the approximated orthogonal polynomials for the exponential function of softmax. Additionally, in order to verify the fitting effect of the new loss functions, we compare various series of orthogonal polynomials to the Taylor series.
(2) In order to achieve the stronger discriminative power, we employ the idea of inter-class margin m to the Orthogonal-Softmax and obtain Orthogonal-M. The proposed Orthogonal has a better geometric attribute, which enhances inter-class discrepancy and intra-class compactness.
(3) Extensive experiments are conducted on four benchmark datasets (MNIST, Fashion-MNIST, CIFAR10, and CIFAR100). The results demonstrate the effectiveness of the Orthogonal-Softmax and Orthogonal-M, which have better performances over the Taylor-Softmax and softmax loss.

2. Related Work

In recent years, the softmax loss has been widely used as a key method to learn discriminative features for multiclass classification. Several margin-based methods have been presented to enhance the discriminative power of the softmax loss. These studies have added a margin penalty into various classes to create inter-class feature separability. SphereFace [38,39], CosFace [41], AM-Softmax [42], and ArcFace [44,45] all introduce an additive angular margin between the features and their corresponding weights under various manners. On the other hand, Taylor-Softmax [51], LinCos-Softmax [52] and LinArc [53] have proposed an approximated linear model, which creates a stricter relationship by Taylor expansion.
In addition, margin-based softmax loss functions enforce better intra-class compactness and inter-class diversity, but these studies have not effectively emphasized every sample according to its practical importance. To a certain extent, Taylor approximated softmax loss enhances the linear relationship with the angle, but they may not have enough discriminative power. Based on margin-based softmax loss and approximated linear softmax loss, we introduce a novel loss function through Gram–Schmidt orthogonalization. By combining the strengths of both, Orthogonal-Softmax has a better approximate effect and enhances the discriminative power through experiments on four datasets.

3. Overview of the Proposed Method

In this section, we will introduce the relative definition and derivation of orthogonal polynomials first. Based on the previous definition, then we will present the proposed Orthogonal-Softmax and the whole process of the Gram–Schmidt orthogonalization algorithm.

3.1. Introduction of Orthogonal Polynomials

Orthogonal polynomials are generally calculated by Gram–Schmidt orthogonalization, and we mainly introduce the idea of the nearest distance between orthogonal polynomials and the target function as follows:
Following the definition of Axler [54]: U is a subspace of the inner product V, for any vector v V , and we have:
v P U v     v u
and
P U v = v j , e 1 e 1 + + v j , e j 1 e j 1
where P U v is the orthogonal projection on U and e i is the orthonormal basis of v.
Based on the definition, it can be inferred that P U v is the shortest distance from v to V. In this way, we assume that the vector v is the exponential function in the softmax, and the orthogonal polynomial is the P U v . Finding the shortest distance from subspace V to vector v means finding the best approximated orthogonal polynomial for the exponential function e in the inner product space.

3.2. Orthogonal-Softmax

In more detail, the softmax loss is defined by the formula:
L S o f t m a x = e z i j = 1 K e z j f o r i = 1 , , K
where z = z 1 , z K R K and z i denotes the deep feature of the ith input vector z, and y i is the corresponding class. K is the sum of the classes.
In order to enhance the discriminability of features and produce an excellent result, we designed a linear approximated logit by applying the orthogonal polynomials. The Orthogonal-Softmax loss can be defined as:
L O r t h o g o n a l - S o f t m a x = f n ( z ) j = 1 K f n ( z )
where
f n ( z ) = i = 0 n v , e i e i
and
e i = β i β i
and
β i = α i α i , β 1 β 1 , β 1 β 1 α i , β i 1 β i 1 , β i 1 β i 1 ,
where α i is a linearly independent list of vectors in inner-product space V, and we can apply the Gram–Schmidt procedure to the α i to obtain the orthogonal basis β i and orthonormal basis e i . Furthermore, U is the subspace of V where v V and P U v = v , e 1 e 1 + + v , e i e i is the orthographic projection on U to V. Generally, the orthographic projection is the shortest distance from V to U, and we can denote P U v as the best approximation, so it is the smallest distance to the exponential function; hence, we can use the orthogonal polynomial to approximate the softmax.
We can present the whole process of the second-order approximated logit, thus a group of basis was given by:
α 0 = 1 , α 1 = x , α 2 = x 2
i = 0 :
α 0 = 1 , β 0 = α 0 = 1 , e 0 = β 0 β 0 = 1
i = 1 :
α 1 = x , β 1 = α 1 α 1 , β 0 β 0 , β 0 β 0 = x 1 2 , e 1 = β 1 β 1 = 2 3 x 1 2
i = 2 :
α 2 = x 2 , β 2 = α 2 α 2 , β 0 β 0 , β 0 β 0 α 2 , β 1 β 1 , β 1 β 1 = x 2 x + 1 6 e 2 = β 2 β 2 = 6 5 x 2 x + 1 6 ,
where after a simple and intuitive calculation, an approximated linear polynomial can be presented as:
f n ( z ) = v , e 0 e 0 + v , e 1 e 1 + v , e 2 e 2 = 0.839 x 2 + 0.851 x + 1.012993
The second-order orthogonal polynomials approximation for softmax has been proved by this, and the whole process of Gram–Schmidt Orthogonalization can be calculated in this way. The various series of the orthogonal basis β i ( x ) and orthonormal basis e i ( x ) are presented in Table 1. Various series of orthogonal polynomials f i ( x ) will be directly presented in the classification experiments of Section 3.

3.3. Comparing with Taylor-Softmax

By using Taylor series approximation, the Taylor-Softmax [39] can be defined as:
L T a y l o r - S o f t m a x = f n ( z ) j = 1 K f n ( z )
where
f n ( z ) = i = 0 n z i i ! f o r i = 1 , 2 , , K
The linear approximated effects of orthogonal polynomials and Taylor polynomials on various series are shown in Figure 1. Black lines denote the exponential functions of softmax, red lines denote the orthogonal polynomials of Orthogonal-Softmax, and blue lines denote the Taylor polynomials of Taylor-Softmax. Compared to Taylor polynomials, the orthogonal polynomials present the better approximation for the exponential functions on all series. In addition, as the series increases, both polynomials achieve stronger approximated effects, which we will demonstrate further in the following experiments.
Back to Equation (3), when dealing with the binary-classes scenarios with the class 1 and class 2, the original Softmax loss presents a decision boundary by z 1 = z 2 . To make the intra-class more compact, CosFace [41] and AM-Softmax [42] introduce an additive margin m and the decision boundary given by:
c l a s s 1 : z 1 > z 1 m > z 2 c l a s s 2 : z 2 > z 2 m > z 1
where m is a fixed parameter presented to control the margin of inter-class. By introducing z 1 m to relace z 1 , the logits can be defined as:
L A M - S o f t m a x = e z i m e z i m + j = 1 K e z j f o r i = 1 , 2 , , K .
Based on the thinking of AM-Softmax, we also introduced margin m into Orthogonal Softmax and Taylor-Softmax as Orthogonal-M and Taylor-M, respectively. As illustrated in Figure 2, by employing an additive margin, the proposed Orthogonal-M enhances inter-class discrepancy and intra-class compactness.

4. Experiments

4.1. Implementation Details

In the following experiment, we employ MNIST, FashionMNIST, CIFAR10, and CIFAR100 as our training datasets in order to create a fair comparison with Taylor-Softmax [39]. It is important to note that f n ( z ) is definite for n = 2 k in Taylor-Softmax, so we will look into the orthogonal polynomials series expansion of 2, 4, and 6 on the experiment design method for the convenience of comparison. Our aim was not to achieve the best accuracy for these datasets but to explore new ways to enhance the discriminative power of Softmax.

4.2. Experimental Setting

As shown in Table 2, we present the network structure corresponding to each dataset. During the experiment process, we evaluate the generalized softmax loss in visual classification. For the CNN construction, we adopt the VGG-net, as well as the Taylor-Softmax for convenient comparison. During all experiments, we assume the PRelu as the activation and the batch size is 256. Convolution neural network training is finished with SGD with momentum 0.9 and weight decay 0.0005.
The margin parameter m plays a crucial role in the proposed the Orthogonal-M. Following the setup of the previous work, we varied margin m from 0 to 0.6 with a deep size of 0.1 , which was set in both Taylor-M and Orthogonal-M. As shown in Figure 3, for the sixth-order orthogonal series of the proposed Orthogonal-M, MNIST and CIFAR10 achieved the highest accuracy levels at m = 0.4 and m = 0.5 , respectively. This indicates that different datasets have various intrinsic properties, and it may not be effective to treat all datasets with the same fixed parameter m. Thus, we present the best accuracy levels of each dataset in various m values. In all these topologies, we replace the final Softmax function with each of its alternatives in our experiments.

4.3. Evaluation Results

The experimental results shown in Table 3 are the verification accuracies on four datasets under various methods.
As shown in Table 3, we display the performance of our proposed Orthogonal-Softmax with series variables 2 to 6 to correspond to Taylor-Softmax. The bold numbers in each column are the highest verification result for various models on each dataset. From the quantitative comparison among all the methods on four datasets, Orthogonal-Softmax presents better performance than Softmax loss and Taylor-Softmax. Benefitting from the introduction of margin-based loss functions, the proposed Orthogonal-M further enhances the discriminative power based on the Orthogonal-Softmax. For the MNIST and Fashion-MNIST, it is well known that these two datasets are simple and typical in deep classification tasks, so the results in the proposed methods are not significantly improved. On CIFAR10, the proposed Orthogonal-M achieves an improvement of 90.45%, which is 3.58% over the Softmax loss and 3.28% over the Taylor-Softmax. Additionally, on CIFAR100, the proposed Orthogonal-M is 8.45% higher than the Softmax loss and 7.08% higher than the Taylor-Softmax. It is worth noting that the proposed Orthogonal-M partially achieves lower performance on series 2 and 4 when compared with Orthogonal-Softmax. The reason for this is that series 6 has a better approximation effect and is more steady, and the added margin presents a better reflection of margin-based Softmax. As observed, all the verification experiments show that the proposed linear logits can appropriately distinguish features and outperform the other existing methods.

5. Conclusions

In this paper, we proposed the Orthogonal-Softmax loss function, which employs an approximated linear logit to effectively replace the Softmax loss. By taking Gram–Schmidt orthogonalization, the proposed Orthogonal-Softmax was able to achieve a better linear relationship and stronger discriminative power. We also supplied a mathematical explanation to dissect the whole process of Gram–Schmidt orthogonalization and its advantages. Experimental results and analyses on four well-known benchmark datasets (MNIST, Fashion-MNIST, CIFAR10, and CIFAR100) demonstrate the superiority of the proposed Orthogonal-Softmax and Orthogonal-M when compared to other loss functions. Therefore, how the conundrum of non-ground-truth classes can be resolved, as well as how bigger datasets can be explored will be our future work.

Author Contributions

The contributions of the respective authors are as follows: conceptualization, H.S.; methodology, H.S.; software, H.S.; validation, H.S.; formal analysis, H.S.; investigation, H.S.; writing—original draft preparation, H.S.; writing—review and editing, H.S. and S.W.; supervision, H.S.; funding acquisition, H.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been supported by grants from the National Natural Science Foundation of China (62062067).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

We would like to express thanks to the anonymous reviewers for their invaluable comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, M.; Deng, W. Deep face recognition: A survey. Neurocomputing 2021, 429, 215–244. [Google Scholar] [CrossRef]
  2. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the 2012 Advances in Neural Information Processing Systems (NeurIPS), Nevada, NV, USA, 3–6 December 2012; Volume 25, pp. 1097–1105. [Google Scholar]
  3. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  4. Nair, V.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 2013 International Conference on Machine Learning (ICML), Haifa, GA, Israel, 14–16 June 2013. [Google Scholar]
  5. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  6. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  7. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 June 2015; pp. 1440–1448. [Google Scholar]
  8. Zhong, Y.; Chen, J.; Huang, B. Towards end-to-end face recognition through aligement learning. IEEE Signal Process. Lett. 2017, 24, 1213–1217. [Google Scholar] [CrossRef]
  9. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. (IJCV) 2015, 25, 211–252. [Google Scholar] [CrossRef]
  10. Hinton, G.E.; Srivastava, N.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R.R. Improving neural networks by preventing co-adaptation of feature detectors. arXiv 2013, arXiv:1207.0580. [Google Scholar]
  11. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  12. Zeiler, M.D.; Fergus, R. Stochastic pooling for regularization of deep convolutional neural networks. arXiv 2013, arXiv:1301.3557. [Google Scholar]
  13. Wang, S.; Liu, P. A New Feature Extraction Method Based on the Information Fusion of Entropy Matrix and Covariance Matrix and Its Application in Face Recognition. Entropy 2015, 17, 4664–4683. [Google Scholar]
  14. Ma, X.; Yu, C.; Chen, X.; Zhou, L. Large-Scale Person Re-Identification Based on Deep Hash Learning. Entropy 2019, 21, 449–463. [Google Scholar]
  15. Peralta, B.; Saavedra, A.; Caro, L.; Soto, A. Mixture of Experts with Entropic Regularization for Data Classification. Entropy 2019, 21, 190. [Google Scholar]
  16. Cao, R.; He, C.; Wei, P.; Su, Y.; Xia, J. Prediction of circRNA-Disease Associations Based on the Combination of Multi-Head Graph Attention Network and Graph Convolutional Network. Biomolecules 2022, 12, 932. [Google Scholar] [CrossRef]
  17. Sun, Y.; Chen, Y.; Wang, X.; Tang, X. Deep learning face representation by joint identification-verfication. In Proceedings of the 2014 Advances in Neural Information Processing Systems (NeurIPS), Montreal, QC, Canada, 8–13 December 2014; Volume 27. [Google Scholar]
  18. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers:Surpassing human -level performance on imagenet recognition. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1026–1034. [Google Scholar]
  19. Sermanet, P.; Eigen, D.; Zhang, X.; Mathieu, M.; Fergus, R.; LeCun, Y. Overfeat: Integrated recognition, localization and detection using convolutional networks. arXiv 2013, arXiv:1312.6229. [Google Scholar]
  20. Taigman, Y.; Yang, M.; Ranzato, M.A.; Wolf, L. Deepface: Closing the gap to human-level performance in face verification. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014; pp. 1701–1708. [Google Scholar]
  21. Sun, Y.; Wang, X.; Tang, X. Deeply learned face representations are sparse, selective, and robust. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 507–516. [Google Scholar]
  22. Wang, S.; Cao, Z.; Li, M.; Yue, Y. G-DipC: An Improved Feature Representation Method for Short Sequences to Predict the Type of Cargo in Cell-Penetrating Peptides. IEEE/ACM Trans. Comput. Biol. Bioinform. 2020, 17, 739–747. [Google Scholar] [CrossRef]
  23. He, Y.; Wang, S. SE-BLTCNN: A channel attention adapted deep learning model based on PSSM for membrane protein classification. Comput. Biol. Chem. 2022, 98, 107680. [Google Scholar] [CrossRef]
  24. Wan, L.; Matthew, Z.; Zhang, S.; Yann, L.; Rob, F. Regularization of neural networks using dropconnect. In Proceedings of the 2013 International Conference on Machine Learning (ICML), Atlanta, GA, USA, 7–19 June 2013; pp. 1058–1066. [Google Scholar]
  25. Long, J.; Chen, Y.; Yang, Z.; Huang, Y.; Li, C. A novel self-training semi-supervised deep learning approach for machinery fault diagnosis. Int. J. Prod. Res. 2022. [Google Scholar] [CrossRef]
  26. Xu, Y.; Yan, X.; Feng, K.; Zhang, Y.; Zhao, X.; Sun, B.; Liu, Z. Global contextual multiscale fusion networks for machine health state identification under noisy and imbalanced conditions. Reliab. Eng. Syst. Safe 2023, 231, 108972. [Google Scholar] [CrossRef]
  27. Fan, J.; Lee, J.H.; Lee, Y.K.; Zhao, X.; Sun, B.; Liu, Z.A. Transfer Learning Architecture Based on a Support Vector Machine for Histopathology Image Classification. Appl. Sci. 2021, 11, 6380. [Google Scholar] [CrossRef]
  28. Sekhar, A.; Biswas, S.; Hazra, R.; Sunaniya, A.K.; Mukherjee, A.; Yang, L. Brain Tumor Classification Using Fine-Tuned GoogLeNet Features and Machine Learning Algorithms: IoMT Enabled CAD System. IEEE J. Biomed. Health 2022, 26, 983–991. [Google Scholar] [CrossRef]
  29. Sun, Y.; Chen, Y.; Wang, X.; Tang, X. Deep learning face representation from predicting 10,000 classes. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  30. Wen, Y.; Zhang, K.; Li, Z.; Qiao, Y. A discriminative feature learning approach for deep face recognition. In Proceedings of the 2016 European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
  31. Hadsell, R.; Chopra, S.; LeCun, Y. Dimensionality reduction by learning an invariant mapping. In Proceedings of the 2006 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New York, NY, USA, 17–22 June 2006; Volume 2, pp. 1735–1742. [Google Scholar]
  32. Schroff, F.; Kalenichenko, D.; Philbin, J. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 507–516. [Google Scholar]
  33. Hoffer, E.; Ailon, N. Deep metric learning using triplet network. In Proceedings of the 2015 International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015; pp. 84–92. [Google Scholar]
  34. Wang, J.; Song, Y.; Leung, T.; Rosenberg, C.; Wang, J.; Philbin, J.; Chen, B.; Wu, Y. Learning fine-grained image similarity with deep ranking. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  35. Hermans, A.; Bayler., L.; Leibe., B. In Defense of the Triplet Loss for Person Re-Identification. arXiv 2017, arXiv:1703.07737. [Google Scholar]
  36. Song, H.O.; Xiang, Y.; Jegelka, S.; Savarese, S. Deep Metric Learning via Lifted Structured Feature Embedding. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 4004–4012. [Google Scholar]
  37. Sohn, K. Improved deep metric learning with multi-class n-pair loss objective. In Proceedings of the 30th International Conference on Neural Information Processing Systems(NIPS), Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  38. Liu, W.; Wen, Y.; Yu, Z.; Li, M.; Raj, B.; Song, L. SphereFace: Deep hypersphere embedding for face recognition. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 212–220. [Google Scholar]
  39. Liu, W.; Wen, Y.; Raj, B.; Singh, R.; Weller, A. SphereFace Revived: Unifying Hyperspherical Face Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 2458–2474. [Google Scholar] [CrossRef]
  40. Liu, W.; Wen, Y.; Yu, Z.; Yang, M. Large-Margin softmax Loss for Deep Classification. In Proceedings of the 2013 International Conference on Machine Learning (ICML), Atlanta, GA, USA, 19–24 June 2016; pp. 507–516. [Google Scholar]
  41. Wang, H.; Wang, Y.; Zhou, Z.; Ji, X.; Gong, D.; Zhou, J.; Li, Z.; Liu, W. CosFace: Large margin cosine loss for deep face recognition. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 5265–5274. [Google Scholar]
  42. Wang, F.; Cheng, J.; Liu, W.; Liu, H. Additive margin softmax for face verification. IEEE Signal Processing Lett. 2018, 25, 926–930. [Google Scholar] [CrossRef]
  43. Liang, X.; Wang, X.; Lei, Z.; Liao, S.; Li, S. Soft-Margin softmax for Deep Classification. In Proceedings of the 2017 Advances in Neural Information Processing Systems (NeurIPS), Long Beach, CA, USA, 4–9 December 2017; pp. 413–421. [Google Scholar]
  44. Deng, J.; Guo, J.; Xue, N.; Zafeiriou, S. ArcFace: Additive angular margin loss for deep face recognition. In Proceedings of the 2019 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 4690–4699. [Google Scholar]
  45. Deng, J.; Guo, J.; Yang, J.; Xue, N.; Kotsia, I.; Zafeiriou, S. ArcFace: Additive Angular Margin Loss for Deep Face Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 5962–5979. [Google Scholar] [PubMed]
  46. Jiao, J.; Liu, W.; Mo, Y.; Jiao, J.; Deng, Z.; Chen, X. Dyn-arcFace: Dynamic additive angular margin loss for deep face recognition. Multim. Tools Appl. 2021, 80, 25741–25756. [Google Scholar] [CrossRef]
  47. Boutros, F.; DAamer, N.; Kirchbuchner, F.; Kuijper, A. ElasticFace: Elastic Margin Loss for Deep Face Recognition. In Proceedings of the 2022 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–20 June 2022; pp. 1578–1587. [Google Scholar]
  48. Kim, Y.; Park, W.; Roh, M.; Shin, J. GroupFace: Learning latent groups and constructing group-based representations for face recognition. In Proceedings of the 2020 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 13–19. [Google Scholar]
  49. Duan, Y.; Lu, J.; Zhou, J. UniformFace: Learning Deep Equidistributed Representation for Face Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 3415–3424. [Google Scholar]
  50. Jiang, P.; Liu, G.; Wang, Q.; Wu, J. Accurate and Reliable Facial Expression Recognition Using Advanced Softmax Loss With Fixed Weights. IEEE Signal Process. Lett. 2020, 27, 725–729. [Google Scholar] [CrossRef]
  51. Banerjee, K.; Gupta, R.R.; Karthik, V.; Mishra, B. Exploring alternatives to softmax function. arXiv 2020, arXiv:2011.11538. [Google Scholar]
  52. Ou, W.; Po, L.; Zhou, C.; Zhang, Y.; Feng, L.; Rehman, Y.A.U.; Zhao, Y.Z. LinCos-Softmax: Learning Angle-Discriminative Face Representations with Linearity-enhanced Cosine Logits. IEEE Access. 2020, 8, 109758–109769. [Google Scholar] [CrossRef]
  53. Chopra, R.; Dhar, J.; Patel, V. LinArc—Deep Face Recognition Using LinCos And ArcFace. In Proceedings of the Advanced Communication Technologies and Signal Processing (ACTS), Rourkela, India, 15–17 December 2021. [Google Scholar]
  54. Axler, S. Linear Algebra Done Right, 2nd ed.; Springer: New York, NY, USA, 1997. [Google Scholar]
Figure 1. Curves for comparing the approximate effects of Orthogonal-Softmax and Taylor-Softmax on various series. The Orthogonal Softmax presents a preferable approximation on all six series.
Figure 1. Curves for comparing the approximate effects of Orthogonal-Softmax and Taylor-Softmax on various series. The Orthogonal Softmax presents a preferable approximation on all six series.
Entropy 25 00727 g001
Figure 2. The comparison of decision boundary on the binary class scenarios. Class 1 and class 2 are represented by green and blue zones, respectively, and gray areas are design margins. As shown in (b), m is an additive margin to further increase inter-class distance. (a) Orthogonal-Softmax; (b) Orthogonal-M.
Figure 2. The comparison of decision boundary on the binary class scenarios. Class 1 and class 2 are represented by green and blue zones, respectively, and gray areas are design margins. As shown in (b), m is an additive margin to further increase inter-class distance. (a) Orthogonal-Softmax; (b) Orthogonal-M.
Entropy 25 00727 g002
Figure 3. Accuracy (%) of Orthogonal-M with various margin parameters m on MNIST and CIFAR10 under sixth-order orthogonal series approximation.
Figure 3. Accuracy (%) of Orthogonal-M with various margin parameters m on MNIST and CIFAR10 under sixth-order orthogonal series approximation.
Entropy 25 00727 g003
Table 1. The presented Orthogonal Basis β i ( x ) and Orthonormal Basis e i ( x ) for various series i.
Table 1. The presented Orthogonal Basis β i ( x ) and Orthonormal Basis e i ( x ) for various series i.
Function β i ( x ) e i ( x )
Series
i = 0 11
i = 1 x 1 2 2 3 x 1 2
i = 2 x 2 x + 1 6 6 5 x 2 x + 1 6
i = 3 x 3 2 3 x 2 + 3 5 x 1 20 20 7 x 3 2 3 x 2 + 3 5 x 1 20
i = 4 x 4 2 x 3 + 9 7 2 7 x + 1 70 210 x 4 2 x 3 + 9 7 2 7 x + 1 70
i = 5 x 5 5 2 + 20 9 x 3 5 6 x 2 + 5 42 x 1 252 253 11 x 5 5 2 + 20 9 x 3 5 6 x 2 + 5 42 x 1 252
i = 6 x 6 3 x 5 + 75 22 x 4 20 11 x 3 + 5 11 x 2 1 22 x + 1 924 924 13 x 6 3 x 5 + 75 22 x 4 20 11 x 3 + 5 11 x 2 1 22 x + 1 924
Table 2. Topologies of four various datasets.
Table 2. Topologies of four various datasets.
DatasetsMNISTFashion-MNISTCIFAR10CIFAR100
Layer
Conv0.X[3×3,64] ×1[3 × 3,64] ×1[3 × 3,64] ×1[3 × 3,96] ×1
Conv1.X[3 × 3,64] ×3[3 × 3,64] ×3[3 × 3,64] ×4[3 × 3,96] ×4
Pool12 × 2 Max,Stride 22 × 2 Max,Stride 22 × 2 Max,Stride 22 × 2 Max,Stride 2
Conv2.X[3 × 3,64] ×3[3 × 3,64] ×3[3 × 3,96] ×4[3 × 3,128] ×4
Pool22 × 2 Max,Stride 22 × 2 Max,Stride 22 × 2 Max,Stride 22 × 2 Max,Stride 2
Conv3.X[3 × 3,64] ×3[3 × 3,96] ×3[3 × 3,128] ×4[3 × 3,384] ×4
Pool32 × 2 Max,Stride 22 × 2 Max,Stride 22 × 2 Max,Stride 22 × 2 Max,Stride 1
Fully Connected256256256512
Fully Connected101010100
Note: Conv1.X, Conv2.X, and Conv3.X indicate convolution units that may contain multiple convolution layers. The [3,3,96] ×3 indicates three cascaded convolution layers with 96 filters of size 3 × 3.
Table 3. Verification results (%) of different loss functions.
Table 3. Verification results (%) of different loss functions.
DatasetVariantsAccuracy246
MNISTsoftmax99.41%
Taylor99.59%99.54%99.59%99.50%
Taylor-M99.67%99.67%99.59%99.53%
Orthogonal99.68%99.68%99.64%99.68%
Orthogonal-M99.71%99.69%99.64%99.71%
Fashion-MNISTsoftmax93.55%
Taylor93.92%93.84%93.71%93.92%
Taylor-M93.98%93.52%93.23%93.98%
Orthogonal94.40%94.40%94.30%94.33%
Orthogonal-M94.45%93.81%94.32%94.45%
CIFAR10softmax86.87%
Taylor87.17%86.86%87.06%87.17%
Taylor-M87.47%87.47%86.86%87.08%
Orthogonal89.75%89.75%89.72%89.55%
Orthogonal-M90.45%89.69%89.71%90.45%
CIFAR100softmax48.57%
Taylor49.94%44.70%49.24%49.94%
Taylor-M49.95%44.77%49.95%49.56%
Orthogonal56.67%54.53%56.50%56.67%
Orthogonal-M57.02%54.47%56.81%57.02%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shao, H.; Wang, S. Deep Classification with Linearity-Enhanced Logits to Softmax Function. Entropy 2023, 25, 727. https://doi.org/10.3390/e25050727

AMA Style

Shao H, Wang S. Deep Classification with Linearity-Enhanced Logits to Softmax Function. Entropy. 2023; 25(5):727. https://doi.org/10.3390/e25050727

Chicago/Turabian Style

Shao, Hao, and Shunfang Wang. 2023. "Deep Classification with Linearity-Enhanced Logits to Softmax Function" Entropy 25, no. 5: 727. https://doi.org/10.3390/e25050727

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop