Next Article in Journal
Performance Evaluation of Deep Learning Based Network Intrusion Detection System across Multiple Balanced and Imbalanced Datasets
Previous Article in Journal
Research on Multi-Scene Electronic Component Detection Algorithm with Anchor Assignment Based on K-Means
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DeepKnuckle: Deep Learning for Finger Knuckle Print Recognition

1
Faculty of Informatics, Eotvos Lorand University, 1117 Budapest, Hungary
2
Faculty of Information Technology, Mutah University, Mutah, Karak 6171, Jordan
3
Department of Computer Science, Applied College, University of Tabuk, Tabuk 71491, Saudi Arabia
4
Department of Accounting, Mutah University, Mutah, Karak 6171, Jordan
5
The National Centre for Genomics and Bioinformatics, King Abdulaziz City for Science and Technology, Riyadh 11442, Saudi Arabia
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(4), 513; https://doi.org/10.3390/electronics11040513
Submission received: 15 January 2022 / Revised: 29 January 2022 / Accepted: 2 February 2022 / Published: 9 February 2022

Abstract

:
Biometric technology has received a lot of attention in recent years. One of the most prevalent biometric traits is the finger-knuckle print (FKP). Because the dorsal region of the finger is not exposed to surfaces, FKP would be a dependable and trustworthy biometric. We provide an FKP framework that uses the VGG-19 deep learning model to extract deep features from FKP images in this paper. The deep features are collected from the VGG-19 model’s fully connected layer 6 (F6) and fully connected layer 7 (F7). After applying multiple preprocessing steps, such as combining features from different layers and performing dimensionality reduction using principal component analysis (PCA), the extracted deep features are put to the test. The proposed system’s performance is assessed using experiments on the Delhi Finger Knuckle Dataset employing a variety of common classifiers. The best identification result was obtained when the Artificial neural network (ANN) classifier was applied to the principal components of the averaged feature vector of F6 and F7 deep features, with 95% of the data variance preserved. The findings also demonstrate the feasibility of employing these deep features in an FKP recognition system.

1. Introduction

Automated identification solutions have become critical for security and privacy in today’s digitally linked world [1]. Biometrics is a type of person recognition method; it is a security solution that differs from typical authentication and identification techniques such as passwords, ID cards, and PIN codes [2].
Biometrics are techniques that use automated ways to objectively validate a system utilizing biological features. It uses physiological or behavioral aspects of humans as a means of authenticating personal identity [3]. Physiological characteristics include those retrieved from the human body, such as iris [4], faces [5,6], retinas, veins [7], fingerprints [8,9,10,11], palm prints, finger knuckle print, and DNA, ECG [12,13,14,15], while behavioral characteristics include voice, stride, signature, and keystroke [16,17,18,19,20]. Biometrics traits are commonly employed in systems such as Security in IoT [21,22,23,24,25], e-banking [26,27,28,29], cloud security [30,31,32,33], access control [34,35,36,37], network security systems [38,39,40,41,42], and ID cards [43,44,45,46,47], and other applications related to IoT [48,49,50,51,52].
Hand-based authentication systems are models that recognize fingerprints [53,54,55,56,57], palm prints, hand geometry, hand form, and hand veins [58,59]. For a long time, hand-based systems have been in the limelight [60]. The success of the systems is due to the trustworthy qualities that give stability, acceptance, simplicity, and robustness [61]. Many corporations, industries, and government agencies now rely on hand-based technology for a variety of purposes, including security [62].
Recently, it was discovered that the finger-knuckle (FKP) [63] is also a biometric element that may be employed in a safe authentication system [64]. It is a hand-based characteristic in which images include information about an individual’s finger knuckle lines and textures, which may become a unique anatomical feature and be used to identify a person. These characteristics were proposed and examined in order to overcome some of the limits and shortcomings of earlier hand-based technologies [65], as well as the inconsistencies of the low-cost [2] and small-size imaging equipment used to capture the structure [66]. Local feature extraction is performed on the FKP system, followed by a mixture of local and global feature extractions, geometric feature extraction, and ultimately an enhanced acquisition device that includes both major and minor intrinsic features [11,67,68,69,70].
Fingerprint, palm print, hand geometry, and hand vein are examples of biometrics that have been thoroughly investigated [71]. FKP can be regarded as a different biometric identifying technology because of its uniqueness [72]. The following are the FKP’s distinct benefits over other biometrics: the surface of FKP is difficult to abrade since individuals normally grasp objects with the inner side of their hands. Because of the non-contact nature of the FKP collection [73], users are more likely to accept it [74]. As such, FKP is regarded as one of the most feasible and effective personal identification technologies in the future [75,76,77,78,79].
The finger knuckle print is a global, one-of-a-kind, and persistent biometric pattern that is utilized for extremely exact personal identification. Recent FKP research has focused on robust feature extraction, contactless/unconstrained acquisition, and fusion techniques [80]. However, the literature includes limited work [81,82,83,84,85].
In order to support corporate, industrial, economic, and social change for competitive advantage of firms and nations, and to improve overall human progress, technological innovation plays an important role in society for satisfying needs, achieving goals, and solving problems of adopters directed to supporting corporate, industrial, economic, and social change [86,87,88,89,90]. Awareness technological and social change requires a thorough understanding of scientific developments and new technological trajectories [91,92,93,94,95].
Biometrics technology as being part of technological innovation [96,97], have received a lot of attention in recent years, especially to current techniques like machine learning and artificial intelligence. As a result of the continual search, contemporary methods known as deep learning were developed [98,99,100]. These approaches received a lot of interest since they can be used to classify image textures [101,102,103,104,105]. Deep learning is a branch of machine learning research that uses learning algorithms to scan many layers of representation to model complicated relationships in data [106,107]. As a result, high-level traits and ideas have been identified based on the lowest of them. Deep structure refers to the hierarchical structure of features, and most of these models are built on the approach of supervised or unsupervised representation learning. Convolutional architectures are one of the most important tools for deep learning success in image classification. Principal component analysis has recently become one of the most popular deep learning methodologies (PCA) [108]. Deep network training is computationally intensive from the outset and necessitates a large volume of tagged data [109,110,111,112,113].
Feature extraction uses deep learning models (convolution neural networks (CNNs)) strategy, which is employed when there is a lack of training data or resources [54]. It can be done by using a pre-trained model, such as VGG, Inception, SqueezeNet, and ResNet, which have been trained on a large dataset project [114] like ImagNet. VGG CNN is made up of various primary structures, each of which is made up of several linked convolutional layers and full-connected layers. The convolutional kernel has a size of 3 × 3, while the input has a size of 224 × 224 × 3. In general, the number of layers is centered at 16~19 [115]. VGGs have lately demonstrated outstanding performance in many computer vision applications. Many firms, including Adobe, Apple, Facebook, Baidu, Google, IBM, Microsoft, NEC, Netflix, and NVIDIA, have lately employed VGG as one of their deep learning methodologies [80]. As a pre-processing model, (VGG-19 CNN) is utilized. The network depth has been increased as compared to typical convolutional neural networks. It employs an alternating structure of numerous convolutional layers and non-linear activation layers, which outperforms a single convolutional layer. The layer structure may extract image features more effectively, apply Max pooling for down sampling, and alter the linear unit (ReLU) as the activation function, that is, choose the greatest value in the image area as the pooled value of the area. The down sampling layer is primarily used to increase the network’s anti-distortion capabilities to the image while keeping the sample’s key characteristics and minimizing the number of parameters [116].
This study aims at providing an investigation structure for the use of deep learning in supporting FKP recognition using VGG-19 (f6 and f7); thereof, our study intends to achieve the following objectives:
  • Determine the extent to which deep learning can support FKP recognition using the deep VGG-19 method.
  • Examine the impact of dimensionality reduction on the discriminative power of deep features.
  • Identify the best performing classifier on FKP deep features.
  • Determine the authentication performance when using FKP deep features.
The remainder of the paper is broken into the following sections: The literature review on FKP recognition is presented in Section 2. The methodology used to implement the proposed approach, the dataset used, the performance evaluation measures, and the structure of our experimental model are all shown in Section 3. The experimental model’s results are presented and discussed in Section 4. Finally, part 5 depicts the study’s findings as well as its future endeavors.

2. Related Work

There has been a lot of study done on FKP recognition for both identification and authentication security, whether for IoT or more general security [116,117,118,119,120,121]. These are only a few instances.
Lalithamani et al. presented new work in the field of biometric authentication system based on the master finger joint pattern—compared to current systems with simple classifiers such as SVM, PCA, and LDA. Given its accuracy, the deep learning method is best suited for this authentication. The Convolution Neural Network (CNN) extracts features and compares them optimally with trained images. CNN is trained by backpropagation algorithm with random gradient descent and minibatch learning with the help of neural network [7].
Hammouche et al. suggested a novel technique for FKP authentication based on phase congruency with a Gabor Filter bank [122]. Furthermore, Zhang et al. proposed a novel computing framework with the goal of implementing a new efficient feature extraction approach for FKP recognition. The authors conducted a thorough examination of three often utilized local features: local orientation, local phase, and phase congruency. In addition, they developed a method for effectively calculating all characteristics utilizing phase congruency [123]. For an FKP identification system, Muthukumar and Kavipriya used the Gabor feature with an SVM classifier [124].
Heidari and Chalechale presented a unique FKP biometric system in which the feature extraction is a mix of the entropy-based pattern histogram (EPH) and a set of statistical texture characteristics (SSTF). The genetic algorithm (GA) was used to find the best characteristics among the retrieved features. This has been tested on PolyU dataset [73]. While, Singh and Kant developed a multimodal biometric system for person authentication based on FKP and iris characteristics, in which the PCA approach was utilized for feature extraction and the Neuro fuzzy neural network (NFNN) classifier was employed in the identification stage [125].
Chlaoua et al. have developed a simple deep learning method known as principal component analysis. In their proposed approach, PCA has been involved to learn two-stages of filter banks then a simple binary hashing and block histograms for clustering of feature vectors are used. The resultant features are finally fed to the classification step, i.e., linear multiclass Support Vector Machine (SVM). The authors also studied a multimodal biometric system based on matching score level fusion [114].
In Reference [126], a simple CNN end-to-end model is introduced for FKP recognition. The dataset Poly-U FKP is used to evaluate the proposed model and the results gives 99.83% (0.76 and 99.18%, respectively) as the best (standard deviation, mean, respectively) accuracy of recognition. Despite this model being based on a straightforward method for data augmentation with a reduced number of trainable parameters, the latter is composed of a few sets of layers: two connected layers and three convolutional ones. Chalabi et al. also used the PCANet-SVM approach to create a system based on score level fusion of minor and major finger knuckles [62].
Hamidi et al. used two types of pre-trained models, VGG–16 and VGG19, with deep convolutional neural networks to extract features from Finger-Knuckle-Print images in order to construct an efficient multimodal identification system. The results presented in this work reveal that unimodal and multimodal identification systems based on matching score level fusion function extremely well [80].
Fei et al. introduced a new direction convolution difference vector to effectively depict the direction details of finger knuckle images. They then presented a feature learning approach for encoded discriminative direction features for finger knuckle image recognition. The final experimental findings indicate that the proposed method outperforms other Finger knuckle image identification algorithms, demonstrating the usefulness of hash learning-based methods [127].
Many deep learning models have been used for different computer vision tasks. For example, ResNet [128] is a robust model that used an interesting idea of adding the identity block to the CNN architecture in order to make the CNN deeper and reduce the problem of learning degradation in deep CNNs.
Likewise, MobileNet [129] is another interesting CNN which used inverted residual in which the residual connections are created between the bottleneck layers. The small size of this model makes it suitable to be used on mobile devices.
Moreover, ShuffleNet [130] is another CNN model which is suitable for mobile devices. The low computational power of this architecture makes it the right option for many applications. The main reason of this low computational cost is due to two operations used in this CNN, which are channel shuffle and pointwise group convolution.
EfficientNet [131] is a state-of-the-art CNN which introduced exciting ideas to improve the performance of CNNs. EfficientNet suggests that the performance of a CNN can be improved by increasing the depth, number of layers, and the width, number of filters, as well as the resolution of the input image. This architecture has many versions, varying from b0 to b7, each of which has its own depth, width, and resolution scales.
FKP is generally fused with other biometrics to reinforce the intended security system in order to increase the performance of a biometric security system; these approaches include, but are not limited to, the work of [73,132,133,134,135].
Despite the large number of FKP recognition methods in the literature [136,137], none are perfect, and each has its own limitations, which are primarily caused by issues inherited from computer vision and machine learning. As a result, there is still room for improvement, especially when using deep learning techniques on FKP images.

3. The Proposed Method

The proposed study aims at investigating the ability of deep features extracted from VGG-19 to improving the performance of FKP biometric recognition.

3.1. Dataset Description

To achieve the objectives of the study, especially as it employs deep learning, we adopt a standard FKP image dataset called IIT Delhi Finger Knuckle Dataset version1.0.20 [138]. The size of this dataset and its diverse FKP images alignment with VGG-19 layered network for deep architecture design. This helps to overcome the limitations associated with acquiescing real-time as this dataset is publicly available for research community. Therefore, our experiment was carried out on IIT Delhi Finger Knuckle Dataset version1.0.20. IIT Delhi Finger Knuckle Dataset contains 790 images for 158 subjects, 5 images each. The size of each image is 80 × 100 pixels. Figure 1 shows sample images of the used dataset.

3.2. The Proposed FKP Recognition

This section presents the proposed framework and procedures for applying the current investigation to obtain the results of using the VGG-19 deep learning-based method for FKP identification or authentication. Figure 2 shows the proposed framework.
In order to extract features, we used the Pre-trained VGG-19 [139] which was pre-trained on the ImageNet dataset [140]. For feature extraction and training, we used MATLAB 2017a and WEKA tool. We used the Weka tool version (3.6.9) to train and test using the extracted features, and to classify the obtained deep features using some of its supported classification methods, such as ANN, RF, NB, and KNN. All the WEKA methods are used with their default parameters.
MATLAB relies on data format in matrices; so, we use it to run VGG-19 (layers F6 and F7) for feature extraction, and the dimensionality reduction using PCA, in addition to the computation of average, max and min of the deep features merging (discussed later). The output of this phase is two sets of features vectors:
  • F6, which is collected from layer 6.
  • F7, which is collected from layer 7.
Each set has 790 rows, which are the number of images and examples in the FKP dataset, and 4096 columns, which are the number of features obtained by the pre-trained VGG-19 for each layer. The resultant feature vectors are used for training and testing purposes. F8, on the other hand is a 1000-node classification layer that represents the likelihood of one class in the ImageNet dataset. It is not a common practice to use it here because it does not contain deep features.
After getting the feature vectors, we converted them to an arff file to be compatible with the WEKA tool [48].
The vectors obtained using the pre-trained VGG-19 are very large, 4096 dimensions, therefore, we used WEKA to reduce the number of features. In PCA, the more principal components (PCs) we keep the more dimensionality we get, and vice versa. Therefore, we opt for including all PCs in a manner that retains 95%, 97%, or 99% variance of the data (the 4096 features). Accordingly, we extracted new features sets: PCA99, PCA97, and PCA95, which are reduced versions of the original features. Following this, we used the following classifiers on the new PCA99, PCA97, and PCA95 datasets:
  • ANN (Multilayer Perceptron: Weka default parameters: number of hidden layers = (features + classes)/2, learning rate = 0.3).
  • RF (Weka default parameters: number of trees in the random forest = 100).
  • NB (Weka default parameters: batch size = 100).
  • KNN (Weka default parameters: K = 1, linear search with Euclidean distance).
Table 1 shows the number of features after applying the three variance percentages (99%, 97%, and 95%) of PCA to the feature vector of the FKPS datasets (F6 and F7).

3.3. Merging F6 and F7

Due to the varying performance of the F6 and F7 features, as will be presented soon, we opt for creating new feature sets by merging both F6 and F7, wishing to obtain deep feature vectors with an improved performance, the merging is done here using any of three methods:
  • Minimum:
The minimum merging work by taking the minimum value from each pair of values from F6 and F7. Algorithm 1 explains the minimum merging procedure.
Algorithm 1 Merging of F6 and F7 using minimum method
Input: Feature vector F6 and Feature vector F7
Output: Merged feature vector
  1: Merged feature vector = zeros(length(F6))
  2: for i = 0 to length(F6))
  3: Merged feature vectori = min(F6i,F7i)
  4: end for
  5: return Merged feature vector
  • Maximum
The maximum merging works by taking the maximum value from each pair of values from F6 and F7. Algorithm 2 explains the maximum merging procedure.
Algorithm 2 Merging of F6 and F7 using maximum method
Input: Feature vector F6 and Feature vector F7
Output: Merged feature vector
  1: Merged feature vector = zeros(length(F6))
  2: for i = 0 to length(F6))
  3: Merged feature vectori = max(F6i,F7i)
  4: end for
  5: return Merged feature vector
  • Average of both F6 and F7:
The average merging works by taking the average value from each pair of values from F6 and F7. Algorithm 3 explains the average merging procedure.
Algorithm 3 Merging of F6 and F7 using average method
Input: Feature vector F6 and Feature vector F7
Output: Merged feature vector
  1: Merged feature vector = zeros(length(F6))
  2: for i = 0 to length(F6))
  3: Merged feature vectori = mean(F6i,F7i)
  4: end for
  5: return Merged feature vector
The three previously mentioned algorithms loop through two feature vectors, F6 and F7. Then each value in the new merged vector is obtained by taking the minimum, maximum, or average of each pair of values. By applying Algorithms 1–3, we get three new deep feature sets; again, we used the dimensionality reduction on these datasets using both PCA. Table 2 shows the number of features after applying PCA with the three variances of data percentages (99%, 97%, and 95%) on the merged feature vectors (F6 and F7), according to average, minimum, and maximum. We call these new deep feature sets MPCA99%, MPCA97%, and MPCA95%.

3.4. Identification and Authentication

The evaluation process of this type of model is done through many different performance measures such as Accuracy, Precision, Recall, and F-score, the values of these measures ranging from 0 to 1, 0 represents the worst performance and 1 represents the best performance. We used 10-Fold cross-validation to verify the reliability of the entered data and validate each method and its recognition ability. In each fold, 79 out of 790 images were tested and 711 images used for training, we mean by “image”, the resultant deep feature vector that represents each image.
For authentication, five random profiles were chosen, each with five FKP images and a class 1 label. The remaining cases were labeled as class 0. We compared the first profile to all instances in class 0, then profile 2, and so on for all five profiles, using WEKA software. Here we used 5-fold cross validation because we had only 5 images for each subject [141].

4. Results and Discussion

In this section, we present and discuss the experimental results of the proposed framework.

4.1. Experiments Settings

The presented results show the performance of machine learning classifiers on 20 different feature datasets created from the obtained different deep features of the IIT DELHI image dataset. The feature datasets created are listed as follows:
  • The pure F6 feature vector with 4096 dimensions.
  • The pure F7 feature vector with 4096 dimensions.
  • F6 after applying PCA 95%.
  • F6 after applying PCA 97%.
  • F6 after applying PCA 99%.
  • F7 after applying 95%.
  • F7 after applying PCA 97%.
  • F7 after applying PCA 99%.
  • Average feature vector (F6⊕F7).
  • Minimum feature vector (F6⊕F7).
  • Maximum feature vector (F6⊕F7).
  • Average feature vector (F6⊕F7) after applying PCA 95%.
  • Average feature vector (F6⊕F7) after applying PCA 97%.
  • Average feature vector (F6⊕F7) after applying PCA 99%.
  • Minimum feature vector (F6⊕F7) after applying PCA 95%.
  • Minimum feature vector (F6⊕F7) after applying PCA 97%.
  • Minimum feature vector (F6⊕F7) after applying PCA 99%.
  • Maximum feature vector (F6⊕F7) after applying PCA 95%.
  • Maximum feature vector (F6⊕F7) after applying PCA 97%.
  • Maximum feature vector (F6⊕F7) after applying PCA 99%.
The following tables demonstrate the performance metrics for different classifiers while using the above-designed feature vectors.
As shown in Table 3, the identification results of applying VGG-19(F6) on the IIT DELHI dataset show that the KNN achieves the highest accuracy with (81.5%); followed by the RF with (73.4%). While the lowest accuracy was achieved by NB with (68.6%). As for time, KNN spends less time to conduct the training and testing phases with (0.02 s); while RF spends the maximum time to accomplish the task of predicting with (1.8 s), the training time of the KNN classifier is always zero because it has no training model, each test example is compared directly to the other examples in the training set, this is one reason why KNN is slow in testing, particularly when we have a large number of examples in the training set [142,143,144,145].
As shown in Table 4, the identification results of applying VGG-19(F7) on the IIT DELHI dataset show that the KNN achieves the highest accuracy with (72.8%); followed by the RF with (69.2%). While the lowest accuracy was achieved by NB with (60.08%). For both F6 and F7, the KNN with acceptable accuracy outperformed all other classifiers. Figure 3 illustrates the accuracy of identification using VGG-19-F6 and VGG-19-F7 deep features, showing that F6 is better for identification regardless of the classifier employed.

4.2. Results of PCA Features

When PCA is applied to the F6 feature vectors, we get a new set of feature vectors with fewer dimensions. The number of features is determined by the proportion of data variance maintained, the higher the data variance, the more features we use. The new dataset is used then as an input to the classification process.
Table 5 shows the results of the PCA on F6 in terms of identification. We utilize PCA to minimize the feature vector while preserving the highest data variance (95%, 97%, and 99%). The results of different principal components in Table 5 demonstrate that all classifiers performance have significantly improved in terms of time.
Table 5 shows that all classifiers achieve low training time, especially for RF and NB; this is due to the dimensionality reduction achieved by the PCA. The highest accuracy resultant from applying PCA of all percentage of data variance used (95%, 97%, and 99%) is in favor of ANN with (90%, 88.6%, and 78.1%) respectively, which can be attributed to the reduced feature vector by the PCA, which removed redundant features since keeping at least 95% of data variance. However, ANN consumed the highest training time in this experiment, it is worth mentioning that we could not use the ANN classifier with raw features because of its unacceptable training time (days), this is another advantage of dimensionality reduction since it allows for faster training and hence more classifiers to be used.
Similarly, we applied PCA of 95%, 97%, 99% on F7 as well, Table 6 shows the identification results.
Table 6 shows that all classifiers improved in terms of training time, if compared to the F6, which is due to the reduced number of features acquired after applying PCA on layer F7 compared to layer F6 (see Table 1), which may be justified by the nature of the deep features collected at layer F7 of the VGG-19. Moreover, the performance of some classifiers, namely NB and RF have improved in terms of accuracy in most cases. Again, the highest accuracy resultant from applying PCA of (95%, 97%, and 99%) are in favor of ANN with (84.9%, 87.3%, and 89.4%) respectively. However, ANN consumes more training time compared to the other classifiers.
In summary, it can be said that KNN achieves the best results given all performances measures except these of ANN. ANN significantly outperforms all the other classifiers in terms of Accuracy, Precision, Recall, and F-measures.

4.3. Results after Merging F6 and F7

Both deep features (F6, F7) are merged by setting the average, maximum, and minimum as merging rules to evaluate the identification of the FKP using KNN, NB, and RF classifiers. Table 7 shows the identification results.
Table 7 shows the identification results of the merged dataset based on averaging F6 and F7, in which KNN has the highest accuracy (80.6%), followed by RF and then by the NB with the lowest accuracy.
Table 8 shows the identification result based on maximizing F6 and F7, where KNN has achieved the highest accuracy of (80.3%). On the other hand, the NB has obtained a low accuracy of (67.7%), and the highest training running time is recorded by the RF (1.4 s).
In Table 9, we can see the identification results based on minimizing F6 and F7, where KNN obtained the highest accuracy (77.8%). On the other hand, the NB obtained the lowest accuracy with (64.7%), and the RF performance was in between.
Afterwards, we applied the PCA with (95%, 97%, and 99%) on the merged dataset for each resultant feature vectors, Maximum, Minimum, and Average. The results are shown in Table 10, Table 11 and Table 12.
As can be seen in Table 10, the ANN classifier achieved the highest accuracy compared to the other classifiers; however, it takes a longer training time.
Similar to the maximum deep features, the minimum deep features show that the ANN achieved the highest accuracy compared to the other classifiers; however, it takes a longer training time as shown in Table 11.
Again, as stated in Table 12, ANN achieved the highest accuracy level among other classifiers; however, it takes longer training time. However, it is worth noting that the ANN and all the classifiers used are in favor of lower data variance (PCA 95%), particularly when using F6 or merging with F6, perhaps this is due to losing more redundant features. One exception is when using the PCA on F7, where we notice improvement in the classifiers’ performance, less redundant data on F7 is perhaps the reason. Table 13 summarizes all of the identification accuracy results.
Table 14 shows a direct comparison of the proposed method to some of the other methods that used the same database, which include Surrounded Patterns Code (SPC) [146], Enhanced Local Line Binary Pattern (ELLBP) [147], Local Binary Patterns (LBP) [148], Centralized Binary Patterns (CBP) [149], Center-Symmetric Local Binary Pattern (CSLBP) [150], and Local Binary Patterns for Finger Outer Knuckle (LBP-FOK) [151]. Moreover, we compared our method to some of the-state-of-the-art deep learning-based feature extraction methods such as ResNet18 [128], MobileNetV2 [129], ShuffleNet [111], and EfficientNetb0 [131]. To extract the deep features using these deep learning-based methods, we employed their pre-trained versions. The features are retrieved from the layer immediately preceding the classification layer. The features are then classified using the ANN after PCA 95% is applied.
As presented in Table 14, the proposed method achieves superior results to six hand-crafted methods which are SPC, ELLBP, LBP, CBP, CSLBP, and LBP-FOK. In addition, other deep learning approaches outperformed all hand-crafted methods. Furthermore, the proposed method outperforms MobileNet, ShuffleNet, and Resnet18 when compared to other deep learning methods. EfficientNet, on the other hand, achieved slightly better 1% results than our proposed method.

4.4. Authentication

Knowing that Authentication is different from Identification in biometric systems, in this study, to conduct an authentication experiment, five profiles were randomly chosen, each of which has five FKP images labelled as class 1, belonging to one of the subjects. The rest of the images belonging to the rest of the subjects are labelled as class 0. The resultant image dataset is called profile 1, which is used to authenticate subject 1 using the same approach, for the sake of simplicity, we created five profiles only, each authenticate a specific subject. We then obtained the deep features using the same aforementioned methodology, that is we used F6 and F7 features after applying PCA 95%. For the performance assessment, the 5-fold cross-validation was applied with KNN to obtain the averaged authentication results. Note that the main aim here is to find out the ability of these features to give a unique representation of each subject in the FKP images dataset, and therefore, it can be reliably employed for authentication purposes.
For evaluation purpose, the Equal Error Rate (EER) is considered as a measure for typical biometric systems, it is normally used to predetermine the threshold values based on the false acceptance rate and the false rejection rate in a particular biometric system [152]. Table 15 and Table 16 present the Authentication EER based on the data profiles created for the purpose of authentication in this study.
As can be seen from Table 15 and Table 16, the average authentication for EER is very small (0.24% when using F6-PCA, and 0.23% when using F7-PCA), having such a small error rate means that the accuracy of the authentication system is very high, this is expected since the number of classes in authentication systems is typically two, while in the identification problem it is equal to the number of subjects in the dataset.
Based on the results in Table 17, our proposed approach, which investigates the use of deep learning for FKP authentication, has a low error rate and a high false rejection rate when compared to the other methods such as the work of [153]. However, to the best of our knowledge, very few previous studies tackled the feasibility of deep learning (VGG-19) in the improvement of FKP recognition using ANN, KNN, NB, or RF. However, several studies discussed the feasibility of FKP recognition using machine learning classifiers. However, as a matter of fact, the lower EER, the more accurate the process is. Therefore, compared to [153]; our proposed approach maintains a lower EER rate in both cases of F6 and F7 using PCA. Similarly, lower EER when compared with EER by [154].

4.5. Discussion

For the identification/authentication of FKP, our Investigation, which is based on the evidence in hand (the dataset used, and the method used) concludes the following:
  • The ANN is the best classifier to be used for deep features extracted using VGG-19, if it is provided with the reduced version of the features, otherwise, i.e., if it is applied on the original pure deep features, which obtained from the VGG-19 layer 6 or 7 or any merging of them both, the training time would be unacceptably long.
  • Using the PCA does not only reduce the dimensionality, and therefore, the training time significantly, but also allows for increasing the identification accuracy, particularly when using the ANN classifier.
  • In general, the original pure F6 provides more distinctive features compared to F7, this is evident from the first 8 rows in Table 13.
  • Merging the deep features from layers 6 and 7 allows for more distinctive features, particularly, when using the averaging rule.
  • The dimensionality reduction using PCA on the averaged deep features is the best combination to provide the most distinctive features compared to the other approaches investigated.
  • The extracted features from the deep features using PCA with 95% of data variance are more distinctive and smaller in number if compared to the other two percentages (97% and 99%) used, this is evident from the highest accuracy achieved by the ANN classifier on the deep features obtained by averaging (F6+F7) after applying PCA 95%, see Table 13, row 14.
  • As for the authentication of FKP our investigation concludes that using deep features obtained from either layer 6 or layer 7 leads to a reasonable FKP authentication system, particularly when applying PCA on these deep features.

5. Conclusions

In this paper, deep learning network VGG-19 is investigated to be used for FKP identification or authentication system, using the deep features collected at layers 6 and 7, with and without dimensionality reduction tools such as the PCA, merging both deep features (F6 and F7) is also investigated using different rules such as average, maximum, and minimum.
For the identification system, several machine learning classifiers were applied (such as ANN, KNN, NB, and RF) on deep features obtained from the open-access dataset IIT Delhi Finger Knuckle Dataset. The experimental results show that the best identification result was obtained when applying the ANN classifier on the principal components of the averaged feature vector of F6 and F7 deep features preserving 95% of the data variance. The results also show the potential of using these deep features for an FKP authentication system.
Furthermore, comparisons to state-of-the-art deep learning-based methods demonstrate the superiority of the proposed solution; however, we discovered that another deep learning feature extraction method, EfficientNet, slightly outperforms ours, hence we recommend more research to be conducted on our method and the other deep features method, perhaps incorporating more than one deep features method for FKP recognition.
Aside from the recognition accuracy, the study’s limitations include the use of only one FKP dataset, one Deep learning method (VGG), and one dimensionality reduction method (PCA), which we believe is insufficient to make substantial conclusions. As a result, future research will involve using a more standard FKP dataset with a larger number of images and subjects. Moreover, more deep learning networks will be investigated such as AlexNet [155], GoogleNet [156], etc. Besides, other dimensionality reduction techniques can be used for reducing the dimensionality of the deep features such as DCT, wavelet transform, etc., which need to be examined when applied into the deep neural network itself. Moreover, our future work will focus on merging other biometrics with the FKP such as face, fingerprint, and hand shape altogether, to obtain their deep features in order to provide a more accurate biometric system.

Author Contributions

All authors have contributed equally to this paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Informed Consent Statement

Not applicable, because the FKP database used is a public and open access image dataset utilized by many researchers. Consent statements obtained by the original authors of the dataset [138].

Data Availability Statement

The dataset used in this paper is publicly available at the dataset website: http://www4.comp.polyu.edu.hk/~csajaykr/IITD/iitd_knuckle.htm, accessed on 13 January 2022.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Angelov, P.P.; Gu, X. Deep rule-based classifier with human-level performance and characteristics. Inf. Sci. 2018, 463, 196–213. [Google Scholar] [CrossRef] [Green Version]
  2. Trabelsi, S.; Samai, D.; Meraoumia, A.; Bensid, K.; Benlamoudi, A.; Dornaika, F.; Taleb-Ahmed, A. Finger-Knuckle-Print Recognition Using Deep Convolutional Neural Network. In Proceedings of the 2020 1st International Conference on Communications, Control Systems and Signal Processing (CCSSP), El Oued, Algeria, 15–16 May 2020. [Google Scholar] [CrossRef]
  3. Hassanat, A.B.; Btoush, E.; Abbadi, M.A.; Al-Mahadeen, B.M.; Al-Awadi, M.; Mseidein, K.I.; Almseden, A.M.; Ta-rawneh, A.S.; Alhasanat, M.B.; Prasath, V.S. Victory sign biometrie for terrorists identification: Preliminary results. In Proceedings of the 2017 8th International Conference on Information and Communication Systems (ICICS), Irbid, Jordan, 4–6 April 2017. [Google Scholar]
  4. Fakroon, M.; Gebali, F.; Mamun, M. Multifactor authentication scheme using physically unclonable functions. Internet Things 2020, 13, 100343. [Google Scholar] [CrossRef]
  5. Chakraborty, S.; Singh, S.K.; Kumar, K. Facial Biometric System for Recognition Using Extended LGHP Algorithm on Raspberry Pi. IEEE Sens. J. 2020, 20, 8117–8127. [Google Scholar] [CrossRef]
  6. Elngar, A.A.; Kayed, M. Vehicle Security Systems using Face Recognition based on Internet of Things. Open Comput. Sci. 2020, 10, 17–29. [Google Scholar] [CrossRef]
  7. Sapkale, M.; Rajbhoj, S.M. A biometric authentication system based on finger vein recognition. In Proceedings of the International Conference on Inventive Computation Technologies (ICICT), Coimbatore, India, 26–27 August 2016. [Google Scholar]
  8. Prakash, N.S.; Venkatram, N. Establishing Efficient Security Scheme in Home IOT Devices through Biometric Finger Print Technique. Indian J. Sci. Technol. 2016, 9, 1–8. [Google Scholar] [CrossRef] [Green Version]
  9. Kanchana, S. Fingerprint Based Biometric Authentication in Iot for Resolving Security Challenges. Int. J. Res. Anal. Rev. 2018, 5, 1000–1003. [Google Scholar]
  10. Shi, S.; Cui, J.; Zhang, X.-L.; Liu, Y.; Gao, J.-L.; Wang, Y.-J. Fingerprint Recognition Strategies Based on a Fuzzy Commitment for Cloud-Assisted IoT: A Minutiae-Based Sector Coding Approach. IEEE Access 2019, 7, 44803–44812. [Google Scholar] [CrossRef]
  11. Lalithamani, N.; Balaji, R.; Ramya, M.; Sruthi, S.; Aiswarya, A. Finger Knuckle Biometric Authentication using Convolution Neural Network. Int. J. Pure Appl. Math. 2017, 117, 31–35. [Google Scholar]
  12. Karimian, N. Cardiovascular PPG biometric key generation for IoT in healthcare domain. In Proceedings of the Mobile Multimedia/Image Processing, Security, and Applications, Baltimore, MD, USA, 13 May 2019. [Google Scholar] [CrossRef]
  13. Zhang, Q. Deep Learning of Electrocardiography Dynamics for Biometric Human Identification in era of IoT. In Proceedings of the 2018 9th IEEE Annual Ubiquitous Computing, Electronics & Mobile Communication (UEMCON), New York, NY, USA, 8–10 October 2018. [Google Scholar] [CrossRef]
  14. Jiang, X.; Liu, X.; Fan, J.; Ye, X.; Dai, C.; Clancy, E.A.; Farina, D.; Chen, W. Enhancing IoT Security via Cancelable HD-sEMG-Based Biometric Authentication Password, Encoded by Gesture. IEEE Internet Things J. 2021, 8, 16535–16547. [Google Scholar] [CrossRef]
  15. Barros, A.; Rosario, D.; Resque, P.; Cerqueira, E. Heart of IoT: ECG as biometric sign for authentication and identification. In Proceedings of the 15th International Wireless Communications and Mobile Computing Conference, IWCMC 2019, Tangier, Morocco, 24–28 June 2019. [Google Scholar]
  16. Shah, D.; Haradi, V. IoT Based Biometrics Implementation on Raspberry Pi. Procedia Comput. Sci. 2016, 79, 328–336. [Google Scholar] [CrossRef] [Green Version]
  17. Wazid, M.; Das, A.K.; Bhat, V.K.; Vasilakos, A.V. LAM-CIoT: Lightweight authentication mechanism in cloud-based IoT environment. J. Netw. Comput. Appl. 2020, 150, 102496. [Google Scholar] [CrossRef]
  18. Sujarani, R.; Manivannan, D.; Manikandan, R.; Vidhyacharan, B. Lightweight Bio-Chaos Crypt to Enhance the Security of Biometric Images in Internet of Things Applications. Wirel. Pers. Commun. 2021, 119, 2517–2537. [Google Scholar] [CrossRef]
  19. Subhasini, M.S. Lightweight Biometric System for Internet of Things (IoT) Security. Int. J. Res. Appl. Sci. Eng. Technol. 2019, 7, 163–169. [Google Scholar] [CrossRef]
  20. Alsahlani, A.Y.F.; Popa, A. LMAAS-IoT: Lightweight multi-factor authentication and authorization scheme for real-time data access in IoT cloud-based environment. J. Netw. Comput. Appl. 2021, 192, 103177. [Google Scholar] [CrossRef]
  21. Alsellami, B.M.; Deshmukh, P.D. The Recent Trends in Biometric Traits Authentication Based on Internet of Things (IoT). In Proceedings of the International Conference on Artificial Intelligence and Smart Systems, ICAIS 2021, Coimbatore, India, 25–27 March 2021. [Google Scholar]
  22. Hossain, M.S.; Muhammad, G.; Rahman, S.M.M.; Abdul, W.; Alelaiwi, A.; Alamri, A. Toward end-to-end biomet rics-based security for IoT infrastructure. IEEE Wirel. Commun. 2016, 23, 44–51. [Google Scholar] [CrossRef]
  23. Hassen, O.A.; Abdulhussein, A.A.; Darwish, S.M.; Othman, Z.A.; Tiun, S.; Lotfy, Y.A. Towards a Secure Signature Scheme Based on Multimodal Biometric Technology: Application for IOT Blockchain Network. Symmetry 2020, 12, 1699. [Google Scholar] [CrossRef]
  24. Karimian, N.; Tehranipoor, M.; Woodard, D.; Forte, D. Unlock Your Heart: Next Generation Biometric in Re-source-Constrained Healthcare Systems and IoT. IEEE Access 2019, 7, 49135–49149. [Google Scholar] [CrossRef]
  25. Vorakulpipat, C.; Takahashi, T.; Rattanalerdnusorn, E.; Thaenkaew, P.; Inoue, D. Usable and Secure Cloud-based Biometric Authentication Solution for IoT Devices. In Proceedings of the 2018 IEEE Symposium on Computers and Communications (ISCC), Natal, Brazil, 25–28 June 2018. [Google Scholar] [CrossRef]
  26. Lee, A. A comparative study on biometric based authentication method for IOT service. Int. J. Innov. Technol. Explor. Eng. 2019, 8, 1–11. [Google Scholar]
  27. Chauhan, T.; Singh, A. Investigating scope of enhanced biometric security for IOT based Digital Wallet. In Proceedings of the 12th International Conference on Advances in Computing, Control, and Telecommunication Technologies, ACT 2021, Bangkok, Thailand, 27–28 August 2021. [Google Scholar]
  28. Spehar, J.; Fuks, A.; Vauclair, M.; Meijer, M.; Van Beek, J.; Shao, B. Power Challenges Caused by IOT Edge Nodes: Securing and Sensing Our World. In Proceedings of the 2019 31st International Symposium on Power Semiconductor Devices and ICs, Shanghai, China, 19–23 May 2019. [Google Scholar] [CrossRef]
  29. Ghazali, T.K.; Zakaria, N.H. Security, comfort, healthcare, and energy saving: A review on biometric factors for smart home environment. J. Comput. 2018, 29, 189–208. [Google Scholar]
  30. Yang, W.; Wang, S.; Hu, J.; Ibrahim, A.; Zheng, G.; Macedo, M.J.; Johnstone, M.N.; Valli, C. A cancelable iris-and ste-ganography-based user authentication system for the internet of things. Sensors 2019, 19, 2985. [Google Scholar] [CrossRef] [Green Version]
  31. Srivastava, M. Physiological Biometric an Adequate Security Measure for IoT Applications. In Proceedings of the 2nd International Conference on Advanced Computing and Software Engineering (ICACSE), Sultanpur, Uttar, India, 8–9 February 2019. [Google Scholar]
  32. Nadia, R.; Tama, B.A.; Song, J.S. Seamless Human Impedance-based IoT Authentication with Machine Learning Techniques. In Proceedings of the International Conference on ICT Convergence, Jeju Island, Korea, 21–23 October 2020. [Google Scholar]
  33. Abduljabbar, Z.A.; Jin, H.; Ibrahim, A.; Hussien, Z.A.; Hussain, M.A.; Abbdal, S.H.; Zou, D. Secure biometric image re-trieval in IoT-cloud. In Proceedings of the ICSPCC 2016—IEEE International Conference on Signal Processing, Communications and Computing, Hong Kong, China, 5–8 August 2016. [Google Scholar]
  34. Xue, Q.; Ju, X.; Zhu, H.; Zhu, H.; Li, F.; Zheng, X. A Biometric-Based IoT Device Identity Authentication Scheme. Lecture Notes of the Institute for Computer Sciences, Social-Informatics and Telecommunications Engineering. LNICST 2019, 287, 139–149. [Google Scholar] [CrossRef]
  35. Golec, M.; Gill, S.S.; Bahsoon, R.; Rana, O. BioSec: A Biometric Authentication Framework for Secure and Private Com-munication among Edge Devices in IoT and Industry 4.0. IEEE Consum. Electron. Mag. 2020, 1, 1–6. [Google Scholar] [CrossRef]
  36. Mandal, S.; Bera, B.; Sutrala, A.K.; Das, A.K.; Choo, K.-K.R.; Park, Y. Certificateless-Signcryption-Based Three-Factor User Access Control Scheme for IoT Environment. IEEE Internet Things J. 2020, 7, 3184–3197. [Google Scholar] [CrossRef]
  37. Batool, S.; Saqib, N.A.; Khattack, M.K.; Hassan, A. Identification of Remote IoT Users Using Sensor Data Analytics. Lect. Notes Netw. Syst. 2020, 69, 328–337. [Google Scholar] [CrossRef]
  38. Yang, W.; Wang, S.; Sahri, N.M.; Karie, N.M.; Ahmed, M.; Valli, C. Biometrics for Internet-of-Things Security: A Review. Sensors 2021, 21, 6163. [Google Scholar] [CrossRef]
  39. Wazid, M.; Das, A.K.; Odelu, V.; Kumar, N.; Conti, M.; Jo, M. Design of Secure User Authenticated Key Management Protocol for Generic IoT Networks. IEEE Internet Things J. 2017, 5, 269–282. [Google Scholar] [CrossRef]
  40. Punithavathi, P.; Geetha, S. Partial DCT-based cancelable biometric authentication with security and privacy preservation for IoT applications. Multimed. Tools Appl. 2019, 78, 25487–25514. [Google Scholar] [CrossRef]
  41. Ross, A.; Banerjee, S.; Chowdhury, A. Security in smart cities: A brief review of digital forensic schemes for biometric data. Pattern Recognit. Lett. 2020, 138, 346–354. [Google Scholar] [CrossRef]
  42. Sharma, S.K.; Khuntia, B. Service layer security architecture for iot using biometric authentication and cryptography technique. Adv. Intell. Syst. Comput. 2021, 1171, 291–301. [Google Scholar]
  43. Srikrishnaswetha, K.; Kumar, S.; Ghai, D. Secured Electronic Voting Machine Using Biometric Technique with Unique Identity Number and IOT. Lecture Notes in Networks and Systems. In Innovations in Electronics and Communication Engineering; Springer: Singapore, 2020; Volume 107, pp. 311–326. [Google Scholar] [CrossRef]
  44. Bracciale, L.; Loreti, P.; Pisa, C.; Shahidi, A. Secure Path: Block-Chaining IoT Information for Continuous Authentication in Smart Spaces. IoT 2021, 2, 326–340. [Google Scholar] [CrossRef]
  45. Goyal, M. Enhancing security in Internet of Things using authentication techniques: A review survey. J. Phys. Conf. Ser. 2021, 2007, 012023. [Google Scholar] [CrossRef]
  46. Sowjanya, G.; Nagaraju, S. Design and implementation of door access control and security system based on IOT. In Proceedings of the International Conference on Inventive Computation Technologies, ICICT 2016, Coimbatore, India, 26–27 August 2016; pp. 1–4. [Google Scholar]
  47. Joshy, A.; Jalaja, M.J. Design and implementation of an IoT based secure biometric authentication system. In Proceedings of the 2017 IEEE International Conference on Signal Processing, Informatics, Communication and Energy Systems, SPICES 2017, Kollam, India, 8–10 August 2017; pp. 1–13. [Google Scholar] [CrossRef]
  48. Hassanat, A.B.A. On Identifying Terrorists Using Their Victory Signs. Data Sci. J. 2018, 17, 27. [Google Scholar] [CrossRef] [Green Version]
  49. Tarawneh, A.S.; Chetverikov, D.; Verma, C.; Hassanat, A.B. Stability and reduction of statistical features for image clas-sification and retrieval: Preliminary results. In Proceedings of the 2018 9th International Conference on Information and Communication Systems (ICICS), Irbid, Jordan, 3–5 April 2018. [Google Scholar]
  50. Al-Shamaileh, M.Z.; Hassanat, A.B.; Tarawneh, A.S.; Rahman, M.S.; Celik, C.; Jawthari, M. New Online/Offline text-dependent Arabic Handwriting dataset for Writer Authentication and Identification. In Proceedings of the 2019 10th International Conference on Information and Communication Systems (ICICS), Irbid, Jordan, 11–13 June 2019. [Google Scholar] [CrossRef]
  51. Hassanat, A.; Al-Awadi, M.; Btoush, E.; Btoush, A.A.; Alhasanat, E.; Altarawneh, G. New Mobile Phone and Webcam Hand Images Databases for Personal Authentication and Identification. Procedia Manuf. 2015, 3, 4060–4067. [Google Scholar] [CrossRef] [Green Version]
  52. Al-Btoush, A.I.; Abbadi, M.A.; Hassanat, A.B.; Tarawneh, A.S.; Hasanat, A.; Prasath, V.B.S. New Features for Eye-Tracking Systems: Preliminary Results. In Proceedings of the2019 10th International Conference on Information and Communication Sys-tems (ICICS), Irbid, Jordan, 11–13 June 2019. [Google Scholar]
  53. Lee, S.; Jang, S.-W.; Kim, D.; Hahn, H.; Kim, G.-Y. A novel fingerprint recovery scheme using deep neural network-based learning. Multimed. Tools Appl. 2021, 80, 34121–34135. [Google Scholar] [CrossRef]
  54. Impedovo, D.; Dentamaro, V.; Abbattista, G.; Gattulli, V.; Pirlo, G. A comparative study of shallow learning and deep transfer learning techniques for accurate fingerprints vitality detection. Pattern Recognit. Lett. 2021, 151, 11–18. [Google Scholar] [CrossRef]
  55. Nogay, H.S. Comparative Experimental Investigation of Deep Convolutional Neural Networks for Latent Fingerprint Pattern Classification. Trait. Signal 2021, 38, 1319–1326. [Google Scholar] [CrossRef]
  56. Zhong, C.; Xu, P.; Zhu, L. A deep convolutional generative adversarial network-based fake fingerprint generation method. In Proceedings of the 2021 IEEE International Conference on Computer Science, Electronic Information Engineering and Intelligent Control Technology (CEI), Fuzhou, China, 24–26 September 2021. [Google Scholar] [CrossRef]
  57. Zeng, L.; Al-Rifai, M.; Nolting, M.; Nejd, W. Triplet Loss for Effective Deployment of Deep Learning Based Driver Identification Models. In Proceedings of the 2021 IEEE International Intelligent Transportation Systems Conference (ITSC), Indianapolis, IN, USA, 19–22 June 2021. [Google Scholar]
  58. Tarawneh, A.S.; Chetverikov, D.; Hassanat, A.B. A Pilot Comparative Study of Different Deep Features for Palmprint Identification in Low-Quality Images. In Proceedings of the Ninth Hungarian Conference on Computer Graphics and Geometry, Budapest, Hungary, 10 March 2018. [Google Scholar]
  59. Chawla, B.; Tyagi, S.; Jain, R.; Talegaonkar, A.; Srivastava, S. Finger Vein Recognition Using Deep Learning. In Proceedings of the International Conference on Artificial Intelligence and Applications, Singapore, 20–22 October 2021; Springer: Berlin/Heidelberg, Germany, 2021. [Google Scholar]
  60. Bhaskar, B.; Veluchamy, S. Hand based multibiometric authentication using local feature extraction. In Proceedings of the 2014 Interna-tional Conference on Recent Trends in Information Technology, Chennai, India, 10–12 April 2014. [Google Scholar]
  61. Aftab, A.; Khan, F.A.; Khan, M.K.; Abbas, H. Hand-based multibiometric systems: State-of-the-art and future challenges. PeerJ Comput. Sci. 2021, 7, e707. [Google Scholar] [CrossRef]
  62. Chalabi, N.E.; Attia, A.; Bouziane, A. Multimodal finger dorsal knuckle major and minor print recognition system based on PCANET deep learning. ICTACT J. Image Video Process 2020, 10, 2153–2158. [Google Scholar]
  63. Woodard, D.L.; Flynn, P.J. Finger surface as a biometric identifier. Comput. Vis. Image Underst. 2005, 100, 357–384. [Google Scholar] [CrossRef]
  64. Cheng, K.H.; Kumar, A. Contactless biometric identification using 3D finger knuckle patterns. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 42, 1868–1883. [Google Scholar] [CrossRef]
  65. Duta, N. A survey of biometric technology based on hand shape. Pattern Recognit. 2009, 42, 2797–2806. [Google Scholar] [CrossRef]
  66. Jaswal, G.; Kaul, A.; Nath, R. Knuckle Print Biometrics and Fusion Schemes–Overview, Challenges, and Solutions. ACM Comput. Surv. 2016, 49, 1–46. [Google Scholar] [CrossRef]
  67. Obaidat, M.S.; Rana, S.P.; Maitra, T.; Giri, D.; Dutta, S. Biometric security and internet of things (IOT). In Biometric-Based Physical and Cybersecurity Systems; Springer: Berlin/Heidelberg, Germany, 2018; pp. 477–509. [Google Scholar]
  68. Gagandeep; Arora, J.; Kumar, R. Biometric Fingerprint Attendance System: An Internet of Things Application. Lect. Notes Netw. Syst. 2019, 32, 523–530. [Google Scholar] [CrossRef]
  69. Srianusha, K.; Saddamhussain, S.K.; Kumar, K.P. Biometric car security and monitoring system using IOT. In Proceedings of the International Conference on Vision Towards Emerging Trends in Communication and Networking, ViTECoN 2019, Vellore, India, 30–31 March 2019. [Google Scholar]
  70. Meena, G.; Choudhary, S. Biometric authentication in internet of things: A conceptual view. J. Stat. Manag. Syst. 2019, 22, 643–652. [Google Scholar] [CrossRef]
  71. Hassanat, A.B.; Prasath, V.S.; Al-Mahadeen, B.M.; Alhasanat, S.M.M. Classification and gender recognition from veiled-faces. Int. J. Biom. 2017, 9, 347–364. [Google Scholar] [CrossRef]
  72. Zhang, L.; Zhang, L.; Zhang, D.; Zhu, H. Online finger-knuckle-print verification for personal authentication. Pattern Recognit. 2010, 43, 2560–2571. [Google Scholar] [CrossRef] [Green Version]
  73. Heidari, H.; Chalechale, A. Biometric Authentication Using a Deep Learning Approach Based on Different Level Fusion of Finger Knuckle Print and Fingernail. Expert Syst. Appl. 2021, 191, 116278. [Google Scholar] [CrossRef]
  74. Kumar, A.; Ravikanth, C. Personal authentication using finger knuckle surface. IEEE Trans. Inf. Forensics Secur. 2009, 4, 98–110. [Google Scholar] [CrossRef] [Green Version]
  75. Zhai, Y.; Cao, H.; Cao, L.; Ma, H.; Gan, J.; Zeng, J.; Piuri, V.; Scotti, F.; Deng, W.; Zhi, Y. A novel finger-knuckle-print recognition based on batch-normalized CNN. In Proceedings of the Hinese Conference on Biometric Recognition, Urumqi, China, 11–12 August 2018; Springer: Cham, Switzerland. [Google Scholar]
  76. Liang, Y.; Samtani, S.; Guo, B.; Yu, Z. Behavioral Biometrics for Continuous Authentication in the Internet-of-Things Era: An Artificial Intelligence Perspective. IEEE Internet Things J. 2020, 7, 9128–9143. [Google Scholar] [CrossRef]
  77. Tait, B.L. Aspects of Biometric Security in Internet of Things Devices. Advanced Sciences and Technologies for Security Applications. In Digital Forensic Investigation of Internet of Things (IoT) Devices; Springer: Cham, Switzerland, 2021; pp. 169–186. [Google Scholar]
  78. Shalaby, A.; Gad, R.; Hemdan, E.E.D.; El-Fishawy, N. An efficient multi-factor authentication scheme based CNNs for securing ATMs over cognitive-IoT. Peer J Comput. Sci. 2021, 7, e381. [Google Scholar] [CrossRef]
  79. Menaga, S.; Paruvathavardhini, J.; Pragaspathy, S.; Dhanapal, R.; Immanuel, D.J. An efficient biometric based authenticated geographic opportunistic routing for IoT applications using secure wireless sensor network. Mater. Today Proc. 2021; in press. [Google Scholar] [CrossRef]
  80. Hamidi, A.; Khemgani, S.; Bensid, K. Transfer Learning Using VGG Based on Deep Convolutional Neural Network for Finger-Knuckle-Print Recognition. In Proceedings of the 2nd International Conference on Computer Science’s Complex Systems and their Applications, Oum El Bouaghi, Algeria, 25–26 May 2021. [Google Scholar]
  81. Verma, S.K.; Dutta, M. Finger-Knuckle-Print Based Recognition System using LBP and SURF. Int. J. Comput. Sci. Inf. Technol. 2015, 5, 2863–2867. [Google Scholar]
  82. Prarthana, R.J.; Dhanzil, A.; Mahesh, N.; Raghul, S. An Automated Garage Door and Security Management System (A dual control system with VPN IoT & Biometric Database). In Proceedings of the 2nd International Conference on Electronics, Communication and Aerospace Technology, ICECA 2018, Coimbatore, India, 29–31 March 2018. [Google Scholar] [CrossRef]
  83. Hamidi, H. An approach to develop the smart health using Internet of Things and authentication based on biometric tech-nology. Future Gener. Comput. Syst. 2019, 91, 434–449. [Google Scholar] [CrossRef]
  84. Li, X.; Niu, J.; Kumari, S.; Wu, F.; Sangaiah, A.K.; Choo, K.-K.R. A three-factor anonymous authentication scheme for wireless sensor networks in internet of things environments. J. Netw. Comput. Appl. 2017, 103, 194–204. [Google Scholar] [CrossRef]
  85. Savitha, M.; Senthilkumar, M. A unique secure multimodal biometrics-based user authenticated key exchange protocol for generic HIoT networks. Int. J. Emerg. Trends Eng. Res. 2020, 8, 1610–1619. [Google Scholar]
  86. Coccia, M. Technological Innovation. In The Blackwell Encyclopedia of Sociology; John Wiley & Sons, Ltd.: Hoboken, NJ, USA, 2021; pp. 299–317. [Google Scholar]
  87. Coccia, M. Sources of technological innovation: Radical and incremental innovation problem-driven to support competitive advantage of firms. Technol. Anal. Strat. Manag. 2016, 29, 1048–1061. [Google Scholar] [CrossRef]
  88. Coccia, M.; Finardi, U. New technological trajectories of non-thermal plasma technology in medicine. Int. J. Biomed. Eng. Technol. 2013, 11, 337. [Google Scholar] [CrossRef]
  89. Coccia, M. Multiple working hypotheses for technology analysis. J. Econ. Bibliogr. 2020, 7, 111–126. [Google Scholar]
  90. Coccia, M.; Finardi, U. Emerging nanotechnological research for future pathways of biomedicine. Int. J. Biomed. Nanosci. Nanotechnol. 2012, 2, 299. [Google Scholar] [CrossRef]
  91. Coccia, M. Varieties of capitalism’s theory of innovation and a conceptual integration with leadership-oriented executives: The relation between typologies of executive, technological and socioeconomic performances. Int. J. Public Sect. Perform. Manag. 2017, 3, 148–168. [Google Scholar] [CrossRef]
  92. Coccia, M. Spatial relation between geo-climate zones and technological outputs to explain the evolution of technology. Int. J. Transit. Innov. Syst. 2015, 4, 5–21. [Google Scholar] [CrossRef] [Green Version]
  93. Coccia, M.; Roshani, S.; Mosleh, M. Scientific Developments and New Technological Trajectories in Sensor Research. Sensors 2021, 21, 7803. [Google Scholar] [CrossRef] [PubMed]
  94. Coccia, M.; Wang, L. Path-breaking directions of nanotechnology-based chemotherapy and molecular cancer therapy. Technol. Forecast. Soc. Chang. 2015, 94, 155–169. [Google Scholar] [CrossRef] [Green Version]
  95. Coccia, M. Comparative Hypotheses for Technology Analysis. In Global Encyclopedia of Public Administration, Public Policy, and, Governance; Farazmand, A., Ed.; Springer: Cham, Switzerland, 2020. [Google Scholar]
  96. Coccia, M. Disruptive firms and industrial change. J. Econ. Soc. Thought 2017, 4, 437–450. [Google Scholar]
  97. Coccia, M. A theory of classification and evolution of technologies within a Generalised Darwinism. Technol. Anal. Strat. Manag. 2018, 31, 517–531. [Google Scholar] [CrossRef]
  98. Tarawneh, A.S.; Hassanat, A.B.; Celik, C.; Chetverikov, D.; Rahman, M.S.; Verma, C. Deep face image retrieval: A comparative study with dictionary learning. In Proceedings of the 2019 10th International Conference on Information and Communication Systems (ICICS), Ibrid, Jordan, 11–13 June 2019. [Google Scholar]
  99. Angelov, P.; Gu, X. MICE: Multi-layer multi-model images classifier ensemble. In Proceedings of the 2017 3rd IEEE International Conference on Cybernetics (CYBCONF), Exeter, UK, 21–23 June 2017. [Google Scholar]
  100. Hassanat, A.; Prasath, V.; Mseidein, K.; Al-Awadi, M.; Hammouri, A. A hybrid wavelet-shearlet approach to robust digital image watermarking. Informatica 2017, 41, 3–24. [Google Scholar]
  101. Hassanat, A.; Jassim, S. Color-based lip localization method. In Proceedings of the SPIE—The International Society for Optical Engineering, Orlando, FL, USA, 28 April 2010. [Google Scholar]
  102. Hassanat, A.B.; Alkasassbeh, M.; Al-Awadi, M.; Alhasanat, E.A. Color-based object segmentation method using artificial neural network. Simul. Model. Pract. Theory 2016, 64, 3–17. [Google Scholar] [CrossRef]
  103. Narloch, P.; Hassanat, A.; Altarawneh, A.S.A.; Anysz, H.; Kotowski, J.; Almohammadi, K. Predicting Compressive Strength of Cement-Stabilized Rammed Earth Based on SEM Images Using Computer Vision and Deep Learning. Appl. Sci. 2019, 9, 5131. [Google Scholar] [CrossRef] [Green Version]
  104. Hassanat, A.; Prasath, V.B.S.; Alkasassbeh, M.; Altarawneh, A.S.A.; Al-Shamailh, A.J. Magnetic energy-based feature extraction for low-quality fingerprint images. Signal Image Video Process. 2018, 12, 1471–1478. [Google Scholar] [CrossRef]
  105. Hassanat, A.; Alkasassbeh, M.; Al-Awadi, M.; Alhasanat, E. Colour-based lips segmentation method using artificial neural networks. In Proceedings of the 2015 6th International Conference on Information and Communication Systems, ICICS 2015, Irbid, Jordan, 7–9 April 2015. [Google Scholar]
  106. Ghazi, M.M.; Ekenel, K.H. A comprehensive analysis of deep learning based representation for face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  107. Anand, V.; Kanhangad, V. Cross-Sensor Pore Detection in High-Resolution Fingerprint Images. IEEE Sens. J. 2021, 22, 555–564. [Google Scholar] [CrossRef]
  108. Oquab, M.; Bottou, L.; Laptev, I.; Sivic, J. Learning and transferring mid-level image representations using convolutional neural networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  109. Shin, Y.-H. A Study on the Security Technology of Real-time Biometric Data in IoT Environment. J. Korea Soc. Comput. Inf. 2016, 21, 85–90. [Google Scholar] [CrossRef] [Green Version]
  110. Farid, F.; Elkhodr, M.; Sabrina, F.; Ahamed, F.; Gide, E. A Smart Biometric Identity Management Framework for Personalised IoT and Cloud Computing-Based Healthcare Services. Sensors 2021, 21, 552. [Google Scholar] [CrossRef] [PubMed]
  111. Sengupta, S. A Secured Biometric-Based Authentication Scheme in IoT-Based Patient Monitoring System. Adv.-Telligent Syst. Comput. 2020, 937, 501–518. [Google Scholar]
  112. Yu, S.; Park, K.S.; Park, Y. A Secure Lightweight Three-Factor Authentication Scheme for IoT in Cloud Computing Environment. Sensors 2019, 19, 3598. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  113. Banerjee, S.; Odelu, V.; Das, A.K.; Srinivas, J.; Kumar, N.; Chattopadhyay, S.; Choo, K.-K.R. A Provably Secure and Lightweight Anonymous User Authenticated Session Key Exchange Scheme for Internet of Things Deployment. IEEE Internet Things J. 2019, 6, 8739–8752. [Google Scholar] [CrossRef]
  114. Chlaoua, R.; Meraoumia, A.; Aiadi, K.E.; Korichi, M. Deep learning for finger-knuckle-print identification system based on PCANet and SVM classifier. Evol. Syst. 2018, 10, 261–272. [Google Scholar] [CrossRef]
  115. Luo, H.; Yang, Y.; Tong, B.; Wu, F.; Fan, B. Traffic sign recognition using a multi-task convolutional neural network. IEEE Trans. Intell. Transp. Syst. 2017, 19, 1100–1111. [Google Scholar] [CrossRef]
  116. Xiao, J.; Wang, J.; Cao, S.; Li, B. Application of a Novel and Improved VGG-19 Network in the Detection of Workers Wearing Masks. J. Phys. Conf. Ser. 2020, 1518, 012041. [Google Scholar] [CrossRef]
  117. Yang, W.; Wang, S.; Zheng, G.; Yang, J.; Valli, C. A Privacy-Preserving Lightweight Biometric System for Internet of Things Security. IEEE Commun. Mag. 2019, 57, 84–89. [Google Scholar] [CrossRef]
  118. Rao, M.K.; Santhi, S.G. A Multi-factor Biometric-Based User Authentication Protocol for IoT Networks. Lect. Notes Netw. Syst. 2020, 119, 125–131. [Google Scholar] [CrossRef]
  119. Parei, A.; Hamidi, H. A method for FIDO management through biometric technology in IOT. Iran. J. Inf. Process. Manag. 2018, 33, 803–856. [Google Scholar]
  120. Vijay, M.; Indumathi, G. A highly secure Multi- Factor authentication system using biometrics to enhance privacy in Internet of Things (IOT). Int. Res. J. Multidiscip. Technovation 2019, 1, 26–34. [Google Scholar] [CrossRef]
  121. Taheri, S.; Yuan, J.-S. A Cross-Layer Biometric Recognition System for Mobile IoT Devices. Electronics 2018, 7, 26. [Google Scholar] [CrossRef] [Green Version]
  122. Hammouche, R.; Attia, A.; Akrouf, S. A novel system based on phase congruency and gabor-filter bank for finger knuckle pattern authentication. ICTACT J. Image Video Process 2020, 10, 2125–2131. [Google Scholar]
  123. Zhang, L.; Zhang, L.; Zhang, D.; Guo, Z. Phase congruency induced local features for finger-knuckle-print recognition. Pattern Recognit. 2012, 45, 2522–2531. [Google Scholar] [CrossRef] [Green Version]
  124. Muthukumar, A.; Kavipriya, A. A biometric system based on Gabor feature extraction with SVM classifier for Fin-ger-Knuckle-Print. Pattern Recognit. Lett. 2019, 125, 150–156. [Google Scholar] [CrossRef]
  125. Singh, S.; Kant, C. FKP and Iris based multimodal biometric system using PCA with NFNN. In Proceedings of the International Conference on Sustainable Computing in Science, Technology and Management (SUSCOM), Amity University Rajasthan, Jaipur, India, 26–28 February 2019. [Google Scholar]
  126. Zohrevand, A.; Imani, Z.; Ezoji, M. Deep Convolutional Neural Network for Finger-knuckle-print Recognition. Int. J. Eng. 2021, 34, 1684–1693. [Google Scholar]
  127. Fei, L.; Zhang, B.; Teng, S.; Zeng, A.; Tian, C.; Zhang, W. Learning discriminative finger-knuckle-print descriptor. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–19 May 2019. [Google Scholar]
  128. Kaiming, H.; Xiangyu, Z.; Shaoqing, R.; Jian, S. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  129. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  130. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  131. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  132. Khodadoust, J.; Medina-Pérez, M.A.; Monroy, R.; Khodadoust, A.M.; Mirkamali, S.S. A multibiometric system based on the fusion of fingerprint, finger-vein, and finger-knuckle-print. Expert Syst. Appl. 2021, 176, 114687. [Google Scholar] [CrossRef]
  133. Jaswal, G.; Poonia, R.C. Selection of optimized features for fusion of palm print and finger knuckle-based person au-thentication. Expert Syst. 2021, 38, e12523. [Google Scholar] [CrossRef]
  134. Attia, A.; Akhtar, Z.; Chahir, Y. Feature-level fusion of major and minor dorsal finger knuckle patterns for person au-thentication, Signal. Image Video Process. 2021, 15, 851–859. [Google Scholar] [CrossRef]
  135. Kant, C.; Chaudhary, S. A Multimodal Biometric System Based on Finger Knuckle Print, Fingerprint, and Palmprint Traits. In Innovations in Computational Intelligence and Computer Vision; Springer: Singapore, 2021. [Google Scholar]
  136. Fei, L.; Zhang, B.; Tian, C.; Teng, S.; Wen, J. Jointly learning multi-instance hand-based biometric descriptor. Inf. Sci. 2021, 562, 1–12. [Google Scholar] [CrossRef]
  137. Aizi, K.; Ouslim, M. Score level fusion in multi-biometric identification based on zones of interest. J. King Saud Univ. Comput. Inf. Sci. 2019, 34, 1498–1509. [Google Scholar] [CrossRef]
  138. Kumar, A.; Zhou, Y. IIT Delhi Finger Knuckle Database (Version 1.0). 2009. Available online: http://www4.comp.polyu.edu.hk/~csajaykr/IITD/iitd_knuckle.htm (accessed on 6 January 2022).
  139. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proceedings of the International Conference on Learning Representations ICLR, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  140. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009. [Google Scholar]
  141. Hassanat, A.B.; Tarawneh, A.S.; Abed, S.S.; Altarawneh, G.A.; Alrashidi, M.; Alghamdi, M. RDPVR: Random Data Partitioning with Voting Rule for Machine Learning from Class-Imbalanced Datasets. Electronics 2022, 11, 228. [Google Scholar] [CrossRef]
  142. Hassanat, A. Furthest-Pair-Based Binary Search Tree for Speeding Big Data Classification Using K-Nearest Neighbors. Big Data 2018, 6, 225–235. [Google Scholar] [CrossRef]
  143. Hassanat, A.B.A. Two-point-based binary search trees for accelerating big data classification using KNN. PLoS ONE 2018, 13, e0207772. [Google Scholar] [CrossRef] [Green Version]
  144. Hassanat, A.B.A. Norm-Based Binary Search Trees for Speeding Up KNN Big Data Classification. Computers 2018, 7, 54. [Google Scholar] [CrossRef] [Green Version]
  145. Hassanat, A.B.A. Furthest-Pair-Based Decision Trees: Experimental Results on Big Data Classification. Information 2018, 9, 284. [Google Scholar] [CrossRef] [Green Version]
  146. Al-Nima, R.R.O.; Al-Kaltakchi, M.T.; Al-Sumaidaee, S.; Dlay, S.S.; Woo, W.L.; Han, T.; Chambers, J.A. Personal verification based on multi-spectral finger texture lighting images. IET Signal Process. 2018, 12, 1154–1164. [Google Scholar] [CrossRef] [Green Version]
  147. Al-Nima, R.R.O.; Han, T.; Chen, T.; Dlay, S.S.; Chambers, J.A. Finger Texture Biometric Characteristic: A Survey. arXiv 2020, arXiv:2006.04193. [Google Scholar]
  148. Ojala, T.; Pietikäinen, M.; Harwood, D. A comparative study of texture measures with classification based on featured distributions. Pattern Recognit. 1996, 29, 51–59. [Google Scholar] [CrossRef]
  149. Fu, X.; Wei, W. Centralized Binary Patterns Embedded with Image Euclidean Distance for Facial Expression Recognition. In Proceedings of the 2008 Fourth International Conference on Natural Computation, Jinan, China, 18–20 October 2008. [Google Scholar] [CrossRef]
  150. Heikkilä, M.; Pietikäinen, M.; Schmid, C. Description of Interest Regions with Center-Symmetric Local Binary Patterns. In Computer Vision, Graphics and Image Processing; Springer: Berlin/Heidelberg, Germany, 2006; pp. 58–69. [Google Scholar] [CrossRef] [Green Version]
  151. Al-Nima, R.R.O.; Jarjes, M.K.; Kasim, A.W.; Sheet, S.S.M. Human Identification using Local Binary Patterns for Finger Outer Knuckle. In Proceedings of the 2020 IEEE 8th Conference on Systems, Process and Control (ICSPC), Melaka, Malaysia, 11–12 December 2020. [Google Scholar] [CrossRef]
  152. Tripathi, K. A comparative study of biometric technologies with. Int. J. Comput. Appl. 2011, 14, 10–15. [Google Scholar]
  153. Choras, M.; Kozik, R. Knuckle biometrics based on texture features, In Proceedings of the International Workshop on Emerging Techniques and Challenges for Hand-Based Biometrics, Istanbul, Turkey, 22 August 2010.
  154. Kim, J.; Oh, K.; Teoh, A.B.-J.; Toh, K.-A. Finger-knuckle-print for identity verification based on difference images. In Proceedings of the IEEE 11th Conference on Industrial Electronics and Applications, Hefei, China, 5–7 June 2016. [Google Scholar] [CrossRef]
  155. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  156. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
Figure 1. Sample from the IIT Delhi Finger Knuckle Dataset [138].
Figure 1. Sample from the IIT Delhi Finger Knuckle Dataset [138].
Electronics 11 00513 g001
Figure 2. The Proposed FKP recognition Model.
Figure 2. The Proposed FKP recognition Model.
Electronics 11 00513 g002
Figure 3. Classifier identification accuracy on each of F6 and F7 deep features. Accuracy results obtained from Table 3 and Table 4.
Figure 3. Classifier identification accuracy on each of F6 and F7 deep features. Accuracy results obtained from Table 3 and Table 4.
Electronics 11 00513 g003
Table 1. Number of the features of the training set after the PCA applied.
Table 1. Number of the features of the training set after the PCA applied.
PCAFeature VectorNumber of Features
Raw features vectorVGG19F64096
VGG19F74096
PCA 99%VGG19F6450
VGG19F7184
PCA 97%VGG19F6271
VGG19F795
PCA 95%VGG19F6190
VGG19F768
Table 2. Number of features before and after applying PCA on the merged F6⊕F7.
Table 2. Number of features before and after applying PCA on the merged F6⊕F7.
Merge RuleNumber of Features
Raw featuresMin4096
Max4096
Average4096
MPCA 99%Min557
Max583
Average431
MPCA 97%Min370
Max402
Average250
MPCA 95%Min269
Max299
Average173
Table 3. Identification results of VGG-19-F6 deep features. Bold values signify the best performance.
Table 3. Identification results of VGG-19-F6 deep features. Bold values signify the best performance.
ClassifierAccuracyPrecisionRecallF-MeasureTraining Time
(seconds)
KNN0.8150.8150.8140.8140
RF0.7340.7250.7350.7301.8
NB0.6860.6860.7170.7011.2
Table 4. Identification results of VGG-19-F7 deep features. Bold values signify the best performance.
Table 4. Identification results of VGG-19-F7 deep features. Bold values signify the best performance.
ClassifierAccuracyPrecisionRecallF-MeasureTraining Time
(seconds)
KNN0.7280.7270.7360.7310
RF0.6920.6920.6790.6851.52
NB0.6080.6460.6660.6561.16
Table 5. Identification results of VGG-19-F6 deep features with PCA (95%, 97%, and 99%). Bold values signify the best performance.
Table 5. Identification results of VGG-19-F6 deep features with PCA (95%, 97%, and 99%). Bold values signify the best performance.
ClassifierAccuracyPrecisionRecallF-MeasureTraining Time (seconds)
PCA 95%ANN0.9000.9000.9000.900474.07
KNN0.6860.7410.7660.7530
NB0.2620.3820.4430.4100.09
RF0.4250.4750.4900.4831.45
PCA 97%ANN0.8860.8860.8850.885410.14
KNN0.5300.6710.7080.6890
NB0.2760.4040.4600.4310.05
RF0.6240.6310.6470.6391.72
PCA 99%ANN0.7810.7860.7940.790422.58
KNN0.5120.5290.5870.5570
NB0.0130.0260.0800.0390.03
RF0.4950.4850.4990.4920.99
Table 6. Identification results of VGG-19-F7 deep features with PCA (95%, 97%, and 99%). Bold values signify the best performance.
Table 6. Identification results of VGG-19-F7 deep features with PCA (95%, 97%, and 99%). Bold values signify the best performance.
ClassifierAccuracyPrecisionRecallF MeasuresTraining Time (seconds)
PCA 95%ANN0.8490.8480.8530.850205.33
KNN0.6900.6080.6130.6110
NB0.4420.5520.5750.5640.03
RF0.6330.6290.6380.6340.92
PCA 97%ANN0.8730.8730.8720.872200.39
KNN0.7140.7140.7680.7400
NB0.4100.4990.5360.5170.02
RF0.6290.6190.6280.6230.84
PCA99%ANN0.8940.8940.8920.893193.26
KNN0.6770.6770.7590.7160
NB0.2750.2750.4370.3370.02
RF0.5590.5420.5530.5480.84
Table 7. Identification results of VGG-19-deep features obtained by averaging (F6 and F7). Bold values signify the best performance.
Table 7. Identification results of VGG-19-deep features obtained by averaging (F6 and F7). Bold values signify the best performance.
ClassifierAccuracyPrecisionRecallF-MeasureTraining Time
(seconds)
KNN0.8060.8060.8030.8050
NB0.6900.6900.7200.7050.95
RF0.7290.7290.7170.7231.44
Table 8. Identification results of VGG-19-deep features obtained by maximizing (F6 and F7). Bold values signify the best performance.
Table 8. Identification results of VGG-19-deep features obtained by maximizing (F6 and F7). Bold values signify the best performance.
ClassifierAccuracyPrecisionRecallF-MeasureTraining Time
(seconds)
KNN0.8030.8030.8000.8010.0
NB0.6770.6770.7160.6961.14
RF0.7090.7180.7250.7211.4
Table 9. Identification results of VGG-19-deep features obtained by minimizing (F6 and F7). Bold values signify the best performance.
Table 9. Identification results of VGG-19-deep features obtained by minimizing (F6 and F7). Bold values signify the best performance.
ClassifierAccuracyPrecisionRecallF-MeasureTraining Time
(seconds)
KNN0.7780.7780.7730.7750.0
NB0.6470.6910.7100.7000.94
RF0.6970.6910.6990.6951.4
Table 10. Identification results of VGG-19-deep features obtained by maximizing (F6 and F7) after applying PCA. Bold values signify the best performance.
Table 10. Identification results of VGG-19-deep features obtained by maximizing (F6 and F7) after applying PCA. Bold values signify the best performance.
ClassifierAccuracyPrecisionRecallF MeasureTraining Time
(seconds)
(PCA 95%)ANN0.8940.8940.8920.893781.2
KNN0.6240.7150.7460.7300
NB0.2870.4100.4660.4361.14
RF0.6570.6450.6600.6520.14
(PCA 97%)ANN0.8460.8460.8390.8421598.19
KNN0.4570.6480.6900.6690
NB0.2160.3430.4000.3690.17
RF0.5670.5900.6020.5961.31
(PCA 99%)ANN0.5920.5650.5790.5722249.24
KNN0.1760.5050.5650.5330
NB0.1090.2090.2470.2260.14
RF0.4660.4390.4500.4441.63
Table 11. Identification results of VGG-19-deep features obtained by minimizing (F6 and F7) after applying PCA. Bold values signify the best performance.
Table 11. Identification results of VGG-19-deep features obtained by minimizing (F6 and F7) after applying PCA. Bold values signify the best performance.
ClassifierAccuracyPrecisionRecallF MeasureTraining Time (seconds)
(PCA 95%)ANN0.8890.8890.8850.887695.57
KNN0.6530.7270.7540.7400
NB0.3010.4230.4800.4500.05
RF0.6510.6560.6660.6611.08
(PCA 97%)ANN0.8190.8400.8510.8461054.17
KNN0.4800.6640.7010.6820
NB0.2200.3460.4050.3730.08
RF0.5860.5860.5700.5781.09
(PCA 99%)ANN0.5470.5080.5160.5122118.81
KNN0.1960.5460.6040.5740
NB0.1570.2530.2990.2740.14
RF0.4810.4620.4750.4691.4
Table 12. Identification results of VGG-19-deep features obtained by averaging (F6 and F7) after applying PCA. Bold values signify the best performance.
Table 12. Identification results of VGG-19-deep features obtained by averaging (F6 and F7) after applying PCA. Bold values signify the best performance.
ClassifierAccuracyPrecisionRecallF MeasureTraining Time (seconds)
(PCA 95%)ANN0.9180.9180.9170.917772.1
KNN0.7220.7860.8050.7950
NB0.3840.4850.5270.5050.12
RF0.7010.6830.6950.6891.22
(PCA 97%)ANN0.9080.9080.9070.907622.02
KNN0.5940.6860.7210.7030
NB0.2770.3960.4490.4210.06
RF0.6220.6030.6160.6091.04
(PCA 99%)ANN0.8220.8380.8470.8421368.06
KNN0.2960.5680.6220.5940
NB0.1280.2330.2760.2520.4
RF0.4630.4510.4640.4581.85
Table 13. Identification accuracy results summery. Bold values signify the best performance.
Table 13. Identification accuracy results summery. Bold values signify the best performance.
Feature VectorNuber of FeaturesANNKNNRFNBReduction Percentage
F6 feature vector with 4096 dimensions.4096-0.8150.7340.6860%
F7 with F6 feature vector with 4096 dimensions4096-0.7280.6920.6080%
F6 after applying PCA 95%.1900.9000.6860.4250.26295%
F6 after applying PCA 97%.2710.8860.5300.6240.27693%
F6 after applying PCA 99%.4500.7810.5120.4950.01389%
F7 after applying PCA 95%.680.8490.6900.6330.44298%
F7after applying PCA 97%.950.8730.7140.6190.49998%
F7 after applying PCA 99%.1840.8940.6770.5590.27596%
Average feature vector (F6 + F7).4096-0.8060.7290.6900%
Minimum feature vector (F6 + F7)4096-0.7780.6970.6470%
Maximum feature vector (F6 + F7).4096-0.8030.7090.6770%
Average (F6 + F7) PCA 95%.1730.9180.7220.7010.38496%
Average (F6 + F7) PCA 97%.2500.9080.5940.6220.27794%
Average feature (F6 + F7) PCA 99%.4310.8220.2960.4630.12889%
Minimum (F6 + F7) PCA 95%.2690.8890.7270.6560.42393%
Minimum (F6 + F7) PCA 97%.3700.8190.4800.5860.22091%
Minimum (F6 + F7) PCA 99%.5570.5470.1960.4810.15786%
Maximum (F6 + F7) PCA 95%.2990.8940.6240.6570.28793%
Maximum (F6 + F7) PCA 97%.4020.8460.4570.5670.21690%
Maximum (F6 + F7) PCA 99%.5830.5920.1760.4660.10986%
Table 14. Direct comparison of the proposed method to other methods. Bold values signify the best performance.
Table 14. Direct comparison of the proposed method to other methods. Bold values signify the best performance.
MethodAccuracy (%)
Proposed91.8
SPC54.1
ELLBP70.47
LBP71.63
CBP76.67
CSLBP76.74
LBP-FOK85.97
Resnet1888.48
MobileNetV244.85
ShuffleNet52.28
EfficientNetb092.91
Table 15. Authentication ERR of VGG-19-F6 deep features after applying PCA (95%).
Table 15. Authentication ERR of VGG-19-F6 deep features after applying PCA (95%).
DatasetEER %
Profile 10.5
Profile 20.12
Profile 30.12
Profile 40.12
Profile 50.37
Avg.0.24
Table 16. Authentication ERR of VGG-19-F7 deep features after applying PCA (95%).
Table 16. Authentication ERR of VGG-19-F7 deep features after applying PCA (95%).
ExperimentsEER %
Profile 10.1
Profile 20.12
Profile 30.12
Profile 40.63
Profile 50.2
Avg.0.23
Table 17. Authentication EER of [153] compared to ours. Bold values signify the best performance.
Table 17. Authentication EER of [153] compared to ours. Bold values signify the best performance.
MethodologyEER %
VGG19 F6 + PCA0.24
VGG19 F7 + PCA0.23
[153]1.02
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tarawneh, A.S.; Hassanat, A.B.; Alkafaween, E.; Sarayrah, B.; Mnasri, S.; Altarawneh, G.A.; Alrashidi, M.; Alghamdi, M.; Almuhaimeed, A. DeepKnuckle: Deep Learning for Finger Knuckle Print Recognition. Electronics 2022, 11, 513. https://doi.org/10.3390/electronics11040513

AMA Style

Tarawneh AS, Hassanat AB, Alkafaween E, Sarayrah B, Mnasri S, Altarawneh GA, Alrashidi M, Alghamdi M, Almuhaimeed A. DeepKnuckle: Deep Learning for Finger Knuckle Print Recognition. Electronics. 2022; 11(4):513. https://doi.org/10.3390/electronics11040513

Chicago/Turabian Style

Tarawneh, Ahmad S., Ahmad B. Hassanat, Esra’a Alkafaween, Bayan Sarayrah, Sami Mnasri, Ghada A. Altarawneh, Malek Alrashidi, Mansoor Alghamdi, and Abdullah Almuhaimeed. 2022. "DeepKnuckle: Deep Learning for Finger Knuckle Print Recognition" Electronics 11, no. 4: 513. https://doi.org/10.3390/electronics11040513

APA Style

Tarawneh, A. S., Hassanat, A. B., Alkafaween, E., Sarayrah, B., Mnasri, S., Altarawneh, G. A., Alrashidi, M., Alghamdi, M., & Almuhaimeed, A. (2022). DeepKnuckle: Deep Learning for Finger Knuckle Print Recognition. Electronics, 11(4), 513. https://doi.org/10.3390/electronics11040513

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop