Next Article in Journal
A Three-Dimensional Cartesian Mesh Generation Algorithm Based on the GPU Parallel Ray Casting Method
Next Article in Special Issue
Improved Single Sample Per Person Face Recognition via Enriching Intra-Variation and Invariant Features
Previous Article in Journal
Full-Scale Train Derailment Testing and Analysis of Post-Derailment Behavior of Casting Bogie
Previous Article in Special Issue
Deep Learning for Facial Recognition on Single Sample per Person Scenarios with Varied Capturing Conditions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Double Additive Margin Softmax Loss for Face Recognition

College of Information Engineering, Yangzhou University, Yangzhou 225127, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(1), 60; https://doi.org/10.3390/app10010060
Submission received: 30 October 2019 / Revised: 14 December 2019 / Accepted: 17 December 2019 / Published: 19 December 2019
(This article belongs to the Special Issue Advanced Biometrics with Deep Learning)

Abstract

:
Learning large-margin face features whose intra-class variance is small and inter-class diversity is one of important challenges in feature learning applying Deep Convolutional Neural Networks (DCNNs) for face recognition. Recently, an appealing line of research is to incorporate an angular margin in the original softmax loss functions for obtaining discriminative deep features during the training of DCNNs. In this paper we propose a novel loss function, termed as double additive margin Softmax loss (DAM-Softmax). The presented loss has a clearer geometrical explanation and can obtain highly discriminative features for face recognition. Extensive experimental evaluation of several recent state-of-the-art softmax loss functions are conducted on the relevant face recognition benchmarks, CASIA-Webface, LFW, CALFW, CPLFW, and CFP-FP. We show that the proposed loss function consistently outperforms the state-of-the-art.

1. Introduction

Face recogniton problems are ubiquitous in the computer vision domain. In the past few years, Deep Convolutional Neural Networks (DCNNs) have set the community of face recognition (FR) on fire [1]. Due to effectively layered end-to-end learning network frameworks and carefully deep feature extracting techniques from local to global, which are the most important ingredients for their success, DCNNs has immensely improved the state of the art in real-world face recognition scenarios. Numerous layered network architectures for face recognition tasks such as AlexNet [2], VGG [3], InceptionNet [4], ResNet [5], and DenseNet [6], have been proposed. Among them, the most representative one is AlexNet, originally proposed by Krizhevsky et al. AlexNet has become a pioneered architecture developed for image classification and was the winner of the ImageNet Large Scale Challenge in 2012.
It is well known that the effective feature representation for face images plays an important role in FR. Over the recent years, a hot research trend towards DCNNs has been devoted to learning with more discriminative deep features. Intuitively, the learnt deep features for FR are desired if the maximal within-class variance is less than the minimal between-class variance. However, learning deep features satisfying this condition is generally not easy owing to the inherently large intra-class variation and high inter-class similarity [7] in many FR applications. Despite the softmax function with the cross-entropy loss (called softmax loss) is popularly used in the training of a DCNN, recent studies [8,9] made it clear that the current softmax loss is insufficient to encourage the desired deep features meeting the above condition. To boost the discriminative ability of DCNNs and inspired by the previous idea, the center loss [10], pairwise loss [11] and triplet loss [12] were proposed. They unanimously proposed the enhancement of discrimination power of deep features by minimizing winth-class variance and maximizing between-class variance in the Euclidean space of features. While these methods is superior to the traditional softmax loss over classification performance, they usually suffer from some drawbacks. The center loss only explicitly enhanced the intra-class compactness while disregarding the inter-class separability. For the pairwise loss and triplet loss, They require the careful mining of pairs or triplets of samples, which is highly time-consuming.
Due to the fact that few existing softmax losses can effectively achieve the discriminative condition that the maximal within-class variance is less than the minimal between-class variance under the conventional Euclidean metric space, more recently, approaches have been proposed to address this problem by transforming the original Euclidean space of features to an corresponding angular space [10,13,14,15]. Specifically, both Large-Margin Softmax Loss [13] and A-Softmax Loss [14] are an angular softmax loss that enables DCNNs to learn angular deep features by imposing an angular margin constraint for larger inter-class variance. Compared to the Euclidean margin suggested by [2,16], the learnt angular features is more discriminative with the angular margin because the angular metric with cosine similarity is intrinsically more suitable to the softmax loss. During training of A-Softmax Loss, the orginal Softmax loss must be combined to ensure the convergence. To overcome the optimisation problem of A-Softmax Loss, the Additive Margin Softmax loass (AM-Softmax) [15] is proposed. The loss integrates a angular margin to the softmax loss in an additve manner. Its implementation and optimisation are much easier than A-Softmax Loss since A-Softmax Loss integrates the angular margin in a multiplicative way. AM-Softmax is easily reproducible and achieves state-of-the-art performance.
Motivated by AM-Softmax loss, This paper propose a new additive angular margin Loss, namely double additive margin softmax loss (DAM-Softmax). The idea behind the proposed loss is to impose an additive margin m to both the intra-class angular variation and inter-class angular variation simultaneously to enhance the intra-class compactness and inter-class discrepancy of the learned features. Compared to AM-Softmax loss, our loss has a stronger geometrical significance and will lead to obtain more discriminative features. Experimental results on some relevant face recognition benchmarks show that the proposed loss achieves better classification performance than the current state-of-the-art losses.
The rest of this paper is organized as follows. In Section 2, We will briefly introduce the related works such as the original softmax Loss, L-Softmax Loss, A-Softmax Loss and AM-Softmax Loss. Then we discuss the proposed loss, Double Additive Margin Softmax Loss in detail in Section 3. Finally, extensive experiments are presented in Section 4.

2. Preliminaries

In order to clearly understand the proposed DAM-Softmax loss, we will briefly review the classical softmax loss and AM-softmax loss. The classical softmax loss is formulated by
L 1 = 1 N i = 1 N l o g e w y i f i c = 1 C e w c f i
where w c ( c = 1 , , C , C is the number of classes) denotes the weight vector of the last fully connected classifier layer, f i is the learned deep feature input vector of the last fully connected classifier layer corresponding to the original input x i with the label y i , and N is the number of training samples in a minibatch. The inner product, w c f i , between w c and f i can be also factorized into w c f i c o s ( θ c ) where θ c is the angle between w c and f i , the loss can thus be rewritten as
L 2 = 1 N i = 1 N l o g e w y i f i c o s ( θ y i ) c = 1 C e w c f i c o s ( θ c )
The A-Softmax is a new loss function derived from the classical Softmax loss which proposed to impose a constraint to make w c = 1 and generalize the modified softmax loss to angular softmax (A-Softmax) loss by replacing f i c o s ( θ y i ) with f i ψ ( θ y i ) ,
L 3 = 1 N i = 1 N l o g e f i ψ ( θ y i ) e f i ψ ( θ y i ) + c = 1 , c y i C e f c c o s ( θ c )
where the authors proposed to define ψ ( θ ) as ψ ( θ ) = ( 1 ) k c o s ( m θ ) 2 k , θ [ k π m , ( k + 1 ) π m ] and k [ 0 , m 1 ] for removing the restriction which θ must be in the range of [ 0 , π m ] .
In the AM-Softmax loss, the authors suggested to introduce an additive margin to its decision boundary by defining ψ ( θ ) as c o s ( θ ) m . In addition, both the deep feature vector f i and weight vectors w c are normalized during the implementation. Thus, The AM-Softmax loss is given by
L 4 = 1 N i = 1 N l o g e s · ( c o s ( θ y i ) m ) e s · ( c o s ( θ y i ) m ) + c = 1 , c y i C e s · c o s ( θ c ) = 1 N i = 1 N l o g e s · ( w y i f i m ) e s · ( w y i f i m ) + c = 1 , c y i C e s · w c f i
where s is a hyper-parameter for scaling the cosine values.
In order to simultaneously enlarge the between-class angular margin and compress the within-class angular variation, It is clear to learn that both the A-Softmax loss and AM-Softmax loss share a common idea to generalize the original softmax loss to angular softmax loss by introducing an integer m to quantitatively control the decision boundary. specifically, In binary class case, a learned feature f from class 1 is given and θ i is the angle between f and w i . A-softmax loss requires c o s ( m θ 1 ) > c o s ( θ 2 ) to correctly classify f . AM-softmax loss instead proposes c o s ( θ 1 ) m > c o s ( θ 2 ) to correctly classify f . Both of them explicitly enforce the intra-class compactness to achieve more discriminative deep features by imposing an intra-class angular margin in the multiplicative manner and in the additive manner, respectively. Compared with the A-Softmax loss, AM-Softmax loss is simpler which is simpler and reaches better performance. In addition, It is much easier to implement because the computation of the gradient for back-propagation is no longer required.

3. Double Additive Margin Softmax Loss

One can obviously learn that AM-Softmax loss can obtain better performance by incorporating a single additive margin to its intra-class angular variation. Inspired by that, we propose to impose an additive margin to both the intra-class angular variation and inter-class angular distribution simultaneously to enhance the intra-class compactness and inter-class discrepancy. To give a formal formulation for the idea, we first define a function g ( θ ) = c o s ( θ ) . The Equation (4) can be rewritten as
L 5 = 1 N i = 1 N l o g e s · ψ ( θ y i ) e s · ψ ( θ y i ) + c = 1 , c y i C e s · g ( θ c )
where ψ ( θ y i ) = c o s ( θ y i ) and g ( θ c ) = c o s ( θ c ) .
As analyzed above, we impose an additive margin m to both the intra-class and inter-class angular variation angular distribution simultaneously. Then we have the formulations:
ψ ( θ y i ) = c o s ( θ y i ) m g ( θ c ) = c o s ( θ c ) + m
Compared to AM-Softmax loss, our formulation is also simple while explicitly encourages intra-class compactness and inter-class separability simultaneously, we thus term the loss as Double Additive Margin Softmax loss (DAM). Finally, the proposed loss function can be formulated by
L 6 = 1 N i = 1 N l o g e s · ( c o s ( θ y i ) m ) e s · ( c o s ( θ y i ) m ) + c = 1 , c y i C e s · ( c o s ( θ c ) + m ) = 1 N i = 1 N l o g e s · ( w y i T f i m ) e s · ( w y i f i m ) + c = 1 , c y i C e s · ( w c f c + m )

3.1. Geometric Interpretation

Our double additive margin has a more explicit geometric interpretation on the hypersphere manifold. To simplify the geometric interpretation, we project the features onto two dimensional space and discuss the binary classification case on the hypersphere manifold where there are only w 1 and w 2 , and w 1 = w 2 = 1 . Thus, the classification performance depends totally on the angles θ 1 between f and w 1 , and θ 2 between f and w 2 .
Classification Stringency. In order to correctly classify the learned feature vector f as class 1, the traditional softmax loss requires θ 1 < θ 2 which easily has θ 1 θ 2 < 0 , the AM-Softmax loss requires c o s ( θ 1 ) m < c o s ( θ 2 ) which can be rewriten as c o s ( θ 1 ) c o s ( θ 2 ) < m , while our DAM-Softmax loss needs c o s ( θ 1 ) m < c o s ( θ 2 ) + m which can be reformulated into c o s ( θ 1 ) c o s ( θ 2 ) < 2 m . It is obvious to see that the DAM-Softmax loss is more stringent than the orginal softmax loss and AM-Softmax loss for satisfying the corresponding classification criteria.
Classification Boundary. In Figure 1, we draw a schematic diagram to show the classification boundary of the classical softmax loss, AM-Softmax loss and the proposed DAM-Softmax loss. The classification boundary of the traditional softmax loss is denoted as the vector p 0 . In this case, we have w 1 p 0 = w 2 p 0 at the decision boundary p 0 ( w 1 c l a s s 1 , w 2 c l a s s 2 ). For the AM-Softmax, the boundary becomes a marginal region instead of a single vector. At the new boundary p 1 for class 1, one has w 1 p 1 m = w 2 p 1 , which gives m = ( w 1 w 2 ) p 1 = c o s ( θ w 1 , p 1 ) c o s ( θ w 2 , p 1 ) . If we further assume that all the classes have the same intra-class variance and the boundary for class 2 is at p 2 , we can get c o s ( θ w 2 , p 1 ) = c o s ( θ w 1 , p 2 ) . Thus, m = c o s ( θ w 1 , p 1 ) c o s ( θ w 1 , p 2 ) , which is the difference of the cosine scores for class 1 between the two sides of the margin region. For our DAM-Softmax loss, the boundary becomes a wider marginal region than the one of AM-Softmax loss. At the new boundary p 3 for class 1, one has w 3 p 3 m = w 4 p 3 + m ( w 3 c l a s s 1 , w 4 c l a s s 2 ), which gives 2 m = ( w 3 w 4 ) p 3 = c o s ( θ w 3 , p 3 ) c o s ( θ w 4 , p 3 ) . If we further assume that all the classes have the same intra-class variance and the boundary for class 2 is at p 4 , we can get c o s ( θ w 4 , p 3 ) = c o s ( θ w 3 , p 4 ) . Thus, 2 m = c o s ( θ w 3 , p 3 ) c o s ( θ w 3 , p 4 ) , which is the difference of the cosine scores for class 1 between the two sides of the margin region. Obviously, the DAM-Softmax loss leads to a larger classification margin between class 1 and class 2.

3.2. Feature Distribution Visualization on MNIST Dataset

In order to better study and verify the effectiveness of the proposed DAM-Softmax loss function, we conducted an experiment on the MNIST dataset [17] to visualize the learned feature distributions. We chose the 7-layer CNN models with the original Softmax loss, AM-Softmax loss and DAM-Softmax loss for training and required to output two-dimensional deep features for visualization. After the 2-dimensional features were obtained, we then made the normalization to them and ploted them on a circle in the two dimensional space.
The visualization from Figure 2 can well demonstrate that our DAM-Softmax outperforms AM-Softmax [15] when the heperparameters s and m is 30 and 0.4, respectively. Compared to AM-Softmax [15], the DAM-Softmax loss can lead to the larger inter-class margin and smaller intra-class variance property to the features without tuning too many hyper-parameters.

3.3. Algorithm

The proposed DAM-Softmax loss is extremely easy to implement in the popular deep learning frameworks, e.g., Pytorch [18] and Tensorflow [19]. The algorithm for DAM-Softmax loss is given as follow.
Algorithm 1: The steps of the DAM-Softmax Loss
Input : Feature Scale s, Margin Parameter m in Equation (7), Randomly initialized weights w ,
               Input images f , Batch size N
      1. Normalize the input image f ( f ^ = f f ), and make the new f = f ^
      2. Normalize the weight w ( w ^ = w w ), and make the new w = w ^
      3. According to the Equation (7), introducing the variable substitutions ( the new f i = f ^ i
      and the new w y i = w ^ y i ) and get the c o s ( θ y i ) = w y i f i
      4. According to the Equation (7), introducing the variable substitutions ( the new f c = f ^ c
      and the new w c = w ^ c ) and get c o s ( θ c ) = w c f c
      5. Calculate “ c o s ( θ y i ) m ”, and get “ s · ( c o s ( θ y i ) m ) ” in the Equation (6) and the Equation (7)
      6. Calculate “ c o s ( θ c ) + m ”, and get “ s · ( c o s ( θ c ) + m ) ” in the Equation (6) and the Equation (7)
      7. Construct loss functioin: L = 1 N i = 1 N l o g e s · ( w y i f i m ) e s · ( w y i f i m ) + j y i e s · ( w c f c + m )
Output : Loss function L

4. Experiment

In this section, we first introduce the experiment settings. Then, we will discuss the effect of the hyperparameters. Finally, we will evaluate the performance of our loss function with several existing state-of-the-art loss functions on the benchmark datasets.

4.1. Implementation Settings

Datasets

Training Datasets. The CASIA-WebFace [7] dataset used for training consists of 49,4414 color face images from 10,575 classes.
Test Datasets. LFW dataset [17] contains 13,233 web-collected images from 5749 different identities, with large variations in pose, expression and illuminations. CFP dataset [18] consists of 500 subjects, each with 10 frontal and 4 profile images. The evaluation protocol includes frontal-frontal (FF) and frontal-profile (FP) face verification, each having 10 folders with 350 same-person pairs and 350 different-person pairs. In our experiments, we employ to test the performance the most challenging subsets CFP-FP which contains images of celebrities in frontal and profile views, CPLFW [20] and CALFW [21] which have higher pose and age variations with same identities from LFW dataset. Specific details of the three datasets above are shown in Table 1 and some example images from CFP-FP, CPLFW and CALFW are given in Figure 3, Figure 4 and Figure 5, respectively.
Data Prepossessing. We adopt the data preprocessing method used in [14,22] to detect faces and facial landmarks in images and align them. Then, we crop the aligned face images and resize them to 112 × 112 , and proceed to perform the normalization for the cropped face images by subtracting 128 and dividing 128.
Dataset Overlap Removal. To develop open-set evaluations, we use the overlap checking code provided by F. Wang [22] to get ride of the overlapped subjects between the training dataset of CASIA-WebFace and the testing dataset of LFW.

4.2. Network Architecture and Parameter Settings

For the fair comparison, the CNN architecture used in all experiments of this paper is the ResNet-face18 model specially designed for the training of face recognition, which is a modified ResNet [5]. The model has an improved residual block of BN-Conv-BN-PReLu-Conv-BN structure in which the kernel size and stride in the first convolutional layer is 3 × 3 and 1 instead of the original 7 × 7 and 2, and the stride in the second convolutional layer is set to be 2 instead of 1. In addition, PReLu [23] is used to replace the original ReLu. All implementations in the paper are conducted by Pytorch [18]. We set the batch size to be 256 and the weight decay parameter 5 × 10 4 . The initial learning rate is set as 10 1 . We set the learning decay rate to be 0.05 which means that the learning rate will be reduced by 5% when the loss value increases. In addition, the total epoch is set as 110. The SGD [9] is used in the optimization process of ResNet-face18.

4.3. Effect of Hyperparameter m

According the discussion in Section 3, our proposed DAM-Softmax loss has two hyperparameters which are the scale s and the margin m. More importantly, the two hyperparameters plays an key role for the performance of the proposed loss. Several recent works [15,24] have already discussed the scale s, we thus follow [15,24] to directly set it to 30 and will no longer discuss it in this paper. In this case, we can focus on the other hyperparameter, margin m. We train the ResNet-face18 model with the DAM-Softmax loss on CASIA-Webface dataset to conduct experiments to seek the best angular margin. For comparison, we train the same network model with the AM-Softmax loss on the same dataset. In Table 2 and Table 3, we list the performance of AM-Softmax loss and the proposed DAM-Softmax loss. under the variation of m from 0.1 to 0.25. As illustrated in Table 2 and Table 3, we can see that the recognition rate increases gradually from m = 0.25 to 0.3 and arrive at the saturated at m = 0.4, the performance then begins to drop from 0.4 to 0.5. For DAM-Softmax loss, the classification accuracy improves significantly from m = 0.1 to 0.17 and reaches the best at m = 0.18, from 0.18 to 0.25, the performance turns to decrease. Therefore, we fix the margins of AM-Softmax loss and DAM-Softmax loss as 0.4 and 0.18, respectively. The experiments for both AM-Softmax loss and DAM-Softmax loss can result in excellent performance without observing any difficulty in convergence. The proposed loss get to the best verification accuracy on CASIA-Webface training dataset.

4.4. Comparison with State of the Art Loss Functions on LFW Dataset

In this part, we evaluate the performance of the proposed DAM-Softmax loss and the state-of-the-art loss functions. Following the previous experimental setting, we train a ResNet-Face18 model under the guidance of the original softmax, L-Softmax, A-Softmax, AM-Softmax and DAM-Softmax on the training dataset of CAISAWebFace. The experimental results on the test dataset of LFW are shown in Table 4.
From Figure 6, It can be seen that the verification accuracy of DAM-Softmax loss is over 80% after one epoch while AM-Softmax loss requires 20 epoches to achieve the similar accuracy. In the 40th epoch, DAM-Softmax loss reaches the best performance which is still superior to the one of AM-Softmax loss. Figure 7 reports the training loss under the variation of epoch. When epoch = 75, the training loss of the original Softmax loss approaches stabilization with the value of about 13 by using softmax, the AM-softmax’s training loss reach stabilization at around 10 when epoch is 55, while DAM-softmax get to stabilization at the 40th epoch and has a lower training loss. Therefore, This can demonstrate that the proposed loss has a faster convergence speed than AM-Softmax loss. As can be seen in Table 4, our proposed DAM-Softmax loss consistently arrives at competitive results compared to the other losses, which demonstrates the effectiveness of DAM-Softmax loss.

4.5. Comparison with State of the Art Loss Functions on CFP-FP, CPLFW and CALFW Datasets

In order to further verify the effectiveness and robustness of DAM-Softmax, we compare the performance of the proposed losse with related baseline methods, e.g., the original softmax, L-Softmax, A-Softmax, AM-Softmax and DAM-Softmax on three datasets which have large-pose, large-age and different-angle. The experimental results are listed in the Table 5. The details of CFP-FP, CPLFW and CALFW datasets are listed in the Table 1.
As seen in Table 5, The proposed DAM-Softmax loss obtains the best performance. From Table 5, we can see that DAM-Softmax works much better on three datasets than AM-Softmax loss. Thus, we further demonstrate that our DAM-Softmax loss has stronger robustness.

5. Conclusions and Future Work

In this paper, we present a novel Double Additive Angular Margin Loss function for face recognition. specifically, we propose to simultaneously impose a angular margin to the intra-class and inter-class variation on the hypersphere manifold, which can effectively enhance the discriminative power of learned deep features. Competitive performance on several popular face benchmarks verify the superiority and robustness of our approach.

Author Contributions

Conceiving the idea, S.Z., and C.C.; Writing original draft, S.Z. and C.C.; Writing final original manuscript, S.Z., and C.C.; Supervision, C.C.; Data curation, S.Z. and G.H.; All authors discussed and revised the results, and have read and approved the published version of the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (60875004).

Conflicts of Interest

The authors declare there is no conflicts of interest regarding the publication of this paper.

References

  1. Schroff, F.; Kalenichenko, D.; Philbin, J. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 815–823. [Google Scholar]
  2. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Nice, France, 2012; pp. 1097–1105. [Google Scholar]
  3. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  4. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on cOmputer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  5. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  6. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  7. Yi, D.; Lei, Z.; Liao, S.; Li, S.Z. Learning face representation from scratch. arXiv 2014, arXiv:1411.7923. [Google Scholar]
  8. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  9. Bottou, L. Large-scale machine learning with stochastic gradient descent. In Proceedings of the COMPSTAT’2010, Paris, France, 22–27 August 2010; pp. 177–186. [Google Scholar]
  10. Wen, Y.; Zhang, K.; Li, Z.; Qiao, Y. A discriminative feature learning approach for deep face recognition. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 499–515. [Google Scholar]
  11. Hadsell, R.; Chopra, S.; LeCun, Y. Dimensionality reduction by learning an invariant mapping. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06). IEEE, Santa Barbara, CA, USA, 26–27 October 2006; Volume 2, pp. 1735–1742. [Google Scholar]
  12. Cheng, D.; Gong, Y.; Zhou, S.; Wang, J.; Zheng, N. Person re-identification by multi-channel parts-based cnn with improved triplet loss function. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1335–1344. [Google Scholar]
  13. Liu, W.; Wen, Y.; Yu, Z.; Yang, M. Large-margin softmax loss for convolutional neural networks. arXiv 2016, arXiv:1612.02295. [Google Scholar]
  14. Liu, W.; Wen, Y.; Yu, Z.; Li, M.; Raj, B.; Song, L. Sphereface: Deep hypersphere embedding for face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 212–220. [Google Scholar]
  15. Wang, F.; Cheng, J.; Liu, W.; Liu, H. Additive margin softmax for face verification. IEEE Signal Process. Lett. 2018, 25, 926–930. [Google Scholar] [CrossRef] [Green Version]
  16. Sun, Y.; Chen, Y.; Wang, X.; Tang, X. Deep learning face representation by joint identification-verification. arXiv 2014, arXiv:1406.4773. [Google Scholar]
  17. Huang, G.B.; Learned-Miller, E. Labeled Faces in the Wild: Updates and New Reporting Procedures; Dept. Comput. Sci. Univ. Massachusetts Amherst: Amherst, MA, USA, 2014; pp. 14–003. [Google Scholar]
  18. Sengupta, S.; Chen, J.C.; Castillo, C.; Patel, V.M.; Chellappa, R.; Jacobs, D.W. Frontal to profile face verification in the wild. In Proceedings of the 2016 IEEE Winter Conference on Applications of Computer Vision (WACV). IEEE, Lake Placid, NY, USA, 7–9 March 2016; pp. 1–9. [Google Scholar]
  19. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, D.; Greg, S.; Davis, A.; Dean, J.; et al. Tensorflow: Large-scale machine learning on heterogeneous distributed systems. arXiv 2016, arXiv:1603.04467. [Google Scholar]
  20. Kafai, M.; Eshghi, K.; Le, A.; Bhanu, B. A Reference-Based Framework for Pose Invariant Face Recognition. In Proceedings of the IEEE International Conference Workshops on Automatic Face and Gesture Recognition, Ljubljana, Slovenia, 4–8 May 2015. [Google Scholar]
  21. Zheng, T.; Deng, W.; Hu, J. Cross-Age LFW: A Database for Studying Cross-Age Face Recognition in Unconstrained Environments. arXiv 2017, arXiv:1708.08197. [Google Scholar]
  22. Deng, J.; Guo, J.; Xue, N.; Zafeiriou, S. Arcface: Additive angular margin loss for deep face recognition. arXiv 2018, arXiv:1801.07698. [Google Scholar]
  23. Xu, B.; Wang, N.; Chen, T.; Li, M. Empirical evaluation of rectified activations in convolutional network. arXiv 2015, arXiv:1505.00853. [Google Scholar]
  24. Wang, F.; Xiang, X.; Cheng, J.; Yuille, A.L. Normface: l 2 hypersphere embedding for face verification. In Proceedings of the 25th ACM International Conference on Multimedia, Mountain View, CA, USA, 23–27 October 2017; pp. 1041–1049. [Google Scholar]
Figure 1. Geometric difference.
Figure 1. Geometric difference.
Applsci 10 00060 g001
Figure 2. Features visualization in MNIST dataset: Softmax (left), AM-Softmax (middle) and DAM-Softmax (right). Specifically, we set the feature (output of the network) dimension to 2 and then draw the classification result as a scatter plot.
Figure 2. Features visualization in MNIST dataset: Softmax (left), AM-Softmax (middle) and DAM-Softmax (right). Specifically, we set the feature (output of the network) dimension to 2 and then draw the classification result as a scatter plot.
Applsci 10 00060 g002
Figure 3. Samples of the same people in the CFP-FP.
Figure 3. Samples of the same people in the CFP-FP.
Applsci 10 00060 g003
Figure 4. Samples of different people in the CALFW.
Figure 4. Samples of different people in the CALFW.
Applsci 10 00060 g004
Figure 5. Samples of different people in the CPLFW.
Figure 5. Samples of different people in the CPLFW.
Applsci 10 00060 g005
Figure 6. Accuracy rate of the test phase.
Figure 6. Accuracy rate of the test phase.
Applsci 10 00060 g006
Figure 7. Training loss v . s Epoch.
Figure 7. Training loss v . s Epoch.
Applsci 10 00060 g007
Table 1. Face datasets for training and testing.
Table 1. Face datasets for training and testing.
DatasetsIdentityImage
CPLFW [20]574911625
CALFW [21]574912174
CFP-FP [18]5007000
Table 2. Experimental results of different values of m with AM-Softmax.
Table 2. Experimental results of different values of m with AM-Softmax.
Parameter m0.250.30.350.40.450.5
Accuracy Rate97.21%97.34%97.45%97.68%97.57%97.59%
Table 3. Experimental results of different values of m (110 epoch) with DAM-Softmax.
Table 3. Experimental results of different values of m (110 epoch) with DAM-Softmax.
Parameter m0.10.130.150.170.180.20.220.25
Accuracy Rate97.68%97.74%97.82%97.94%97.97%97.89%97.81%97.65%
Table 4. Some results of comparative testing experiment.
Table 4. Some results of comparative testing experiment.
ModelAccuracy Rate
Softmax ( resnet-face18, 110 epoch )97.08%
L-Softmax ( resnet-face18, 110 epoch ) [10]97.33%
A-Softmax ( resnet-face18, 110 epoch ) [14]97.52%
AM-Softmax ( resnet-face18, 110 epoch ) [15]97.68%
DAM-Softmax ( resnet-face18, 47 epoch )97.83%
DAM-Softmax ( resnet-face18, 110 epoch )97.97%
Table 5. Verification results of different datasets based on different algorithms.
Table 5. Verification results of different datasets based on different algorithms.
MethodCALFWCPLFWCFP-FP
Softmax88.21%77.54%89.54%
AM-Softmax89.72%80.21%92.12%
DAM-softmax90.17%82.08%93.26%

Share and Cite

MDPI and ACS Style

Zhou, S.; Chen, C.; Han, G.; Hou, X. Double Additive Margin Softmax Loss for Face Recognition. Appl. Sci. 2020, 10, 60. https://doi.org/10.3390/app10010060

AMA Style

Zhou S, Chen C, Han G, Hou X. Double Additive Margin Softmax Loss for Face Recognition. Applied Sciences. 2020; 10(1):60. https://doi.org/10.3390/app10010060

Chicago/Turabian Style

Zhou, Shengwei, Caikou Chen, Guojiang Han, and Xielian Hou. 2020. "Double Additive Margin Softmax Loss for Face Recognition" Applied Sciences 10, no. 1: 60. https://doi.org/10.3390/app10010060

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop