Next Article in Journal
Femtosecond Laser Inscribed Excessively Tilted Fiber Grating for Humidity Sensing
Previous Article in Journal
Design, Optimization, and Experimental Evaluation of Slow Light Generated by π-Phase-Shifted Fiber Bragg Grating for Use in Sensing Applications
Previous Article in Special Issue
One Definition to Join Them All: The N-Spherical Solution for the EEG Lead Field
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid Feature Extractor Using Discrete Wavelet Transform and Histogram of Oriented Gradient on Convolutional-Neural-Network-Based Palm Vein Recognition

1
Department of Electrical Engineering, Universitas Indonesia, Depok 16424, Jawa Barat, Indonesia
2
School of Science, Computing and Engineering Technologies, Swinburne University of Technology, Hawthorn, VIC 3122, Australia
3
Research Center for Biomedical Engineering, Faculty of Engineering, Universitas Indonesia, Depok 16424, Jawa Barat, Indonesia
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(2), 341; https://doi.org/10.3390/s24020341
Submission received: 13 November 2023 / Revised: 29 December 2023 / Accepted: 30 December 2023 / Published: 6 January 2024
(This article belongs to the Special Issue Computational Intelligence Based-Brain-Body Machine Interface)

Abstract

:
Biometric recognition techniques have become more developed recently, especially in security and attendance systems. Biometrics are features attached to the human body that are considered safer and more reliable since they are difficult to imitate or lose. One of the popular biometrics considered in research is palm veins. They are an intrinsic biometric located under the human skin, so they have several advantages when developing verification systems. However, palm vein images obtained based on infrared spectra have several disadvantages, such as nonuniform illumination and low contrast. This study, based on a convolutional neural network (CNN), was conducted on five public datasets from CASIA, Vera, Tongji, PolyU, and PUT, with three parameters: accuracy, AUC, and EER. Our proposed VeinCNN recognition method, called verification scheme with VeinCNN, uses hybrid feature extraction from a discrete wavelet transform (DWT) and histogram of oriented gradient (HOG). It shows promising results in terms of accuracy, AUC, and EER values, especially in the total parameter values. The best result was obtained for the CASIA dataset with 99.85% accuracy, 99.80% AUC, and 0.0083 EER.
Keywords:
CNN; DWT; HOG; palm vein; VeinCNN

1. Introduction

Biometric measurement has been increasingly used for human recognition, such as in system security for authentication systems [1]. Biometrics is a combination of science and technology that measures biofeatures, such as behavioral or physical, to identify an individual [2]. Since they are attached to the human body, biofeatures are considered to be more effective and robust than conventional forms of identity, such as personal identification numbers (PINs), passwords, or cards [3], which are susceptible to being forgotten or stolen [4]. Based on their visibility, biofeatures can be divided into extrinsic and intrinsic biometrics based on their characteristics, each with their advantages and disadvantages. Extrinsic biometrics are visible, while intrinsic biometrics are perceived indirectly. As the most widely used extrinsic biometric, fingerprint recognition is durable and consistent but still faces issues related to image distortion, the clarity of the core of the fingerprint, and insufficient data [5,6]. Among other types of extrinsic biometrics, face recognition relies on the use of a picture and coordinates, and external conditions such as illumination, occlusion, and the person undergoing some physical changes can lead to unreliable recognition [7]. The iris is a thin and secure organ that is least affected by external conditions, but the durability of iris recognition is sensitive to eye clarity, pupil size changes, radiation, illumination, and the alignment of the camera [8]. On the other hand, intrinsic biometrics are more advanced in terms of data acquisition than extrinsic biometrics. Their location inside the body makes these biometrics difficult to forge and imitate, which can avoid fraud [9]. As a result, some advanced methods or tools are needed to acquire and view hidden biometrics, using medical equipment such as electrocardiogram signals, magnetic resonance imaging (MRI), X-ray, and infrared spectroscopy [10,11]. The high cost of acquiring MRI or X-ray images is a consideration of this biometric system [12]. The infrared spectrum can penetrate the human skin, and it is absorbed by hemoglobin and other vascular structures. The different amounts of absorption show the vein pattern in a captured image [13]. Among the different types of intrinsic biometrics, palm veins are one of the most promising. The vein pattern is unique from one human to another. In addition, the advantages of the use of palm veins as a biometric are that it is consistent [14] and the veins only exist in a human who is alive [15,16]; it is also accurate, contactless, cost-effective, convenient, and reliable [17,18]. The veins carry blood containing hemoglobin, which can be captured using an infrared camera with a wavelength spectrum of 750 nm–1 mm [19].
However, the vein pattern is difficult to capture due to some factors causing the infrared light that penetrates the human skin to be imperfectly absorbed by deoxygenated hemoglobin [20], such as scars and hair on the skin [21]. There are several steps to obtain good recognition of the palm veins, including palm vein acquisition, image preprocessing, feature extraction, and classification. Image acquisition can be achieved using the infrared spectrum for palm vein acquisition. The vein image should be filtered during the preprocessing of the entire recognition system. In addition to the appearance of noise, other problems in infrared images are irregular texture and low contrast [22,23]. When using appropriate devices, palm veins can be combined with a palm print to produce images that can be processed [24]. Still, since palm vein images are irregular and noisy, more images must be obtained of the palm veins themselves to lead to better recognition.
The image filter methods commonly used for image preprocessing include the Laplacian filter [25], the Gaussian filter [26], the hybrid Wiener and median filter [27], the Canny detector and Gabor filter [28], and discrete wavelet transform (DWT) [29]. Among others, the DWT method is superior for image processing since it has image segmentation and denoising characteristics [30]. It transforms and decomposes an image into four sub-band images: coefficient approximation, coefficient detail in horizontal, coefficient detail in vertical, and coefficient detail in diagonal. The approximation detail can describe the information in an image.
The process then continues to feature extraction, which can be performed based on two categories of low-level features, namely, structural and textural features. There are several methods for this category of features, such as supervised discriminative sparse principal component analysis neighborhood-preserving embedding (SDSPCA-NPE) [31], local binary pattern (LBP) [32], gray-level co-occurrence matrix (GLCM), and histogram of oriented gradient (HOG) [33]. Based on the results of [33], HOG showed the best result among the texture features that were extracted [33] due to its superiority in detecting the degree of differences among transformations and variants [34], although there are still some reports that low-level features are unrepresentative and unstable [35]. In order to handle this, methods to extract high-level features have been introduced, i.e., deep learning, and can obtain features automatically from a given dataset for each specific application. This method is promising and has been developed using numerous methods, such as deep neural network (DNN), deep belief network (DBN), and convolutional neural network (CNN). Many researchers have expanded biometric classification using the CNN method to obtain more appropriate image recognition. Some biometrics, such as handwriting [36] and finger veins [37], have been researched using CNN with better pattern recognition accuracy [38].
Since palm veins are irregular and undoubtedly unique, using more images will result in better recognition. The use of CNN has been investigated by many researchers for vein recognition, such as finger veins [39,40], hand dorsal veins [41,42], and palm veins [43,44]. Hong et al. [45] developed a CNN for finger vein recognition using three different datasets based on the quality of the images: good, middle, and poor quality. There are three structures of deep learning to be examined: AlexNet, VGG16, and VGG19. VGG16 with fine-tuning obtained the best result among those three methods. Wang et al. developed a system to recognize two contactless palm vein datasets using a Gabor filter as a feature extractor and a lightweight CNN [46]. Wan et al. [39] examined palm dorsal vein images in three datasets. There are various numbers of subjects and images in each dataset. The three datasets have the same image sizes, but they consist of different quality images. The first and third datasets have better quality images than the second dataset. VGG19 obtained a good result, with an accuracy of 99.7%.
Other researchers of vein recognition have proposed various methods using modified CNN structures to overcome the issue of low discriminative ability in palm vein pattern recognition. Wang et al. [47] modified pretrained VGG-16 with spatial pyramid pooling to obtain discriminative features and achieved an EER of 0.068%. Wang et al. in [48] developed VGG16 with minutiae feature kernel to obtain a better result for EER. Another hybrid deep learning method was introduced by Hou and Yan [49]. The developed hybrid deep learning method with a CNN autoencoder was used to recognize finger veins. The input images were processed by the autoencoder to obtain the discriminative features, and the structure of the CNN was simpler. However, very few research reports provide classification analyses on several different datasets. Therefore, this paper proposes a palm vein recognition method based on CNN, DWT, and HOG for hybrid feature extraction and several datasets so that this feature extractor can be used in more general conditions. The datasets in [50,51,52,53,54] are well-known contact and contactless palm vein image datasets, and are often used to show the state of the art in method improvements, as provided in Table 1.
The proposed DWT and HOG hybrid feature extractor is expected to produce an efficient and effective method with a small error rate value in palm vein recognition. The DWT assists in filtering the noise in the original images, and the HOG improves the sensitivity of the gradient magnitude and gradient orientation of the palm vein image information based on image texture. The texture image part, which contains directional information, has a higher value than the nontexture image part, which does not contain directional information. This filtered image is then prepared for the advanced step that contains only the essential information. In this paper, the evaluation of the recognition system is based on several key performance indicators: the accuracy value, area under receiver operating characteristic curve (AUC), and equal error rate (EER).
From the aforementioned research results, we propose a method to recognize palm vein feature biometrics. The key contributions of this paper are as follows:
  • A simple CNN hybrid structure with a feature extraction method to verify the palm vein pattern based on an image. Using the hybrid DWT and HOG as the feature extractor will handle the irregularity and unique properties of the images.
  • The proposed hybrid DWT-HOG VeinCNN is implemented in five datasets of palm vein images in one study to understand the general condition of palm vein images.
  • The proposed CNN structure can maintain satisfactory accuracy while minimizing the equal error rate.
This research paper’s outline is organized as follows: Section 2 shows the methodology, including image acquisition, preprocessing, feature extraction, classification, evaluation, and error metrics. Section 3 shows the results of the paper. Section 4 presents the discussion. The conclusion is shown in Section 5.

2. Materials and Methods

The method consists of five steps for palm vein recognition, as shown in Figure 1. Step 1 is data acquisition, where images are obtained from five datasets [50,51,52,53,54]. In step 2, each image from the dataset is segmented by applying the preprocessing data to obtain only the part of the image that contains the most information, i.e., the ROI.
An image of the center of the palm is obtained from the entire hand image. This partial image is processed through feature extraction in step 3. In step 3, the images are processed by convolutional mathematics to obtain some features based on DWT and HOG. In step 4, the feature recognition process is performed. The output classification is given in step 5. The result shows whether the image could be genuine or an impostor.

2.1. Image Acquisition

This palm vein recognition technique based on images with CNN, DWT, and HOG is proposed because of the great accuracy published in [57,58]. This method is evaluated using five public datasets available online: CASIA, Vera, Tongji, PolyU, and PUT datasets. These public datasets assist researchers in analyzing the proposed algorithm and comparing it with previous algorithms, and the possibility of permission being granted for research or academic purposes is high. An example of each dataset is shown in Figure 2. The details of each dataset are summarized in Table 2.

2.1.1. CASIA

This contactless dataset was obtained and captured by the Chinese Academy of Sciences Institute of Automation (CASIA) [50]. It consists of 100 subjects using their hands, left and right, which are considered as two different individuals. Each palm was captured six times in two sessions with six infrared spectra (460, 630, 700, 850, 940 nm and white light spectra). The images are represented as an 8-bit grayscale image. There are a total of 7200 palm vein images in this dataset in JPEG image format. The resolution of the images is 768 × 576. However, this paper used only 850 and 940 nm infrared spectrum images to obtain a clearer pattern of palm veins and achieve a lower error rate.

2.1.2. Vera

This dataset was collected using a contactless sensor by the University of Applied Sciences Western Switzerland and the Idiap Research Institute [54]. Palm vein images were captured with a wavelength spectrum of 940 nm. There are 110 participants with left and right hands in this study. Each hand was captured five times in two sessions. As a result, the dataset contains 2200 palm vein images in PNG format with an image resolution of 480 × 680.

2.1.3. Tongji

This dataset was collected by Zhang et al. [53] using a contactless sensor for palm veins. The sensor has an infrared camera lens with 940 nm wavelength. Using this spectrum, images of 300 people were obtained. Each palm of an individual was captured 10 times in two sessions. There are a total 12,000 palm vein images in this dataset in BMP image format, and the resolution of the images is 800 × 600.

2.1.4. PolyU

The PolyU dataset was developed by the Hong Kong Polytechnic University [51]. It consists of 250 volunteers with two sessions of imaging. Six images for each palm were captured in each session. In total, the PolyU dataset has 24,000 palm images. It used 880 nm LED infrared. The image size is 352 × 288, and the images are in JPG format (*.jpg).

2.1.5. PUT

The PUT dataset consists of 100 volunteers, with three sessions and 4 images captured in each session [52]. In total, the PUT dataset has 1200 images. It used 880 nm infrared. The image size is 768 × 1024 in BMP format.

2.2. Preprocessing Data

The veins are captured with a contactless sensor in various infrared spectrum wavelengths and different image formats. The preprocessing step consists of a grayscale image with a further segmentation process. The purpose of the segmentation is to obtain the interest area, generally located in the center of the hand image. Finger valley and hand contour detection methods are applied to obtain the point of reference. In order to overcome issues of contactless sensing such as translation, rotation, and variation in scaling the image, the ROI of the valley between the fingers and centroid in [59] has been applied. The ROI images are resized into 128×128 pixels, and the format of the image is updated to BMP. The input images are normalized. The ROI results of the dataset are shown in Figure 3.

2.3. Feature Extraction

Feature extraction is a process of obtaining distinctive features that contain unique information about an object. The features could be extracted by calculating the pixels in a certain area or direction. The feature extraction process plays an important role in palm vein recognition to distinguish an object from others.
There are various approaches to extracting the features of palm veins, such as line, code, and texture features. Texture features are one of the most popular in research. Texture feature extraction involves obtaining features based on the grayscale pixel in the palm vein image and calculating the pixel value or descriptor palm vein image value. The texture feature can provide unique information that distinguishes one palm vein pattern from another. Some of the texture feature methods that are effective for obtaining distinctive features are wavelet feature extraction and histogram of oriented gradient.

2.3.1. Wavelet Feature Extraction

A wavelet feature is a texture feature that is extracted based on wavelet transformation. Wavelet transformation is applied to the palm vein image to obtain four component images. The components consist of approximation images (cA), a horizontal detail image (cH), a vertical detail image (cV), and a diagonal detail image (cD) [60]. If a palm vein image is symbolized as I, the wavelet transformation generates I = {cA, cH, cV, cD}. The extraction of wavelet features is shown in Figure 4. This approximation image is a result of two low-pass filter processes in wavelet transformation. By using the approximation image only, the classification accuracy result can be increased [61].

2.3.2. Histogram of Oriented Gradient Feature Extraction

Based on [62], a HOG feature is a descriptor feature that has been widely used in image processing. The HOG feature mostly improves the accuracy and quality of the recognition process. The increased accuracy is obtained by utilizing a block, i.e., the smallest gradient on an image. Based on the calculated gradient, the HOG feature has some advantages in terms of the robustness of intensity and direction invariance. The area intensity of the block is then normalized by processing the local histogram value with the remaining cells of the assigned block. This normalization process produces better results in lit and shaded conditions. The vertical and horizontal directions of the gradients are calculated mathematically between the pixel on the image and a certainty kernel factor. The vertical direction gradient represents image direction, while the horizontal direction gradient represents image magnitude. Equation (1) shows the calculation gradient |G| of the magnitude based on the intensity of the vertical and horizontal pixels, Ix and Iy. Ix and Iy can be calculated by (2) and (3). On the other hand, the value of the direction image, θ, can be found by (4).
G = I x 2 + I y 2
I x = I D x ;   D x = 1 0 1
I y = I D y ;   D y = 1 0 1
θ = arctan I y I x
These image direction and image magnitude components are then divided into certain blocks to generate a bin of the histogram structure directions. The arrangement of the bin then produces HOG features, as shown in Figure 5.

2.3.3. Hybrid DWT and HOG Feature Extraction

Combining a wavelet-transformed palm vein image with HOG features shows the essential information of a palm vein image. A block diagram of the hybrid wavelet and HOG feature extraction is shown in Figure 6. The input palm vein image is transformed by Haar wavelet to obtain an approximation coefficient. This approximated image is then calculated to acquire gradient magnitude and gradient direction. Based on these gradients, the HOG feature can be aligned. The procedure of the proposed hybrid wavelet and HOG feature extraction method is shown in Table 3. The results of the proposed method are provided as HOG features in Figure 7a–e for the CASIA, Vera, Tongji, PolyU, and PUT datasets, respectively.

2.4. Recognition Based on Convolutional Neural Network

The simulation of palm vein images involves a recognition process that consists of two main processes: enrollment and recognition [63]. The recognition process is carried out to match certain features that are registered and stored in template storage so that the decision module can provide a final decision as to whether or not a person is registered in the database (genuine or impostor) using CNN.
The CNN process includes training and evaluation processes. For the training and evaluation of our proposed method, palm vein images from five datasets containing various numbers of images are used. The images from each dataset were divided into three subsets for training, validation, and testing of 70%, 20%, and 10%, respectively. The simulation was conducted using the Python programming language and the Jupyter Notebook as an integrated development environment (IDE) on the proposed VeinCNN model. Additionally, we adopted the TensorFlow backend with the Keras framework [64]. The parameters of the proposed VeinCNN method were determined to govern the architecture of the network, which were the type and depth of layer, activation function, output shape, kernel size, and number of filters. The layers consist of a convolutional layer, a max-pooling layer, a flattened layer, and a dense layer. To handle the nonlinearity of palm vein images and reduce computational resources, ReLU and sigmoid were used. Since palm vein images consist of detailed lines and edges, small kernel sizes of two and three are more suitable for this simulation. The number of filters was set to 32 and 64 as a power of two so that the simulation process could effectively filter the palm vein images. This proposed method compiled binary cross-entropy as the loss function and root mean square propagation as the optimizer. Early stopping was applied to increase the training efficiency and minimize overfitting.

2.4.1. VeinCNN

CNN is one of several deep learning methods used to recognize images. However, the arrangement of the CNN must be a consideration to obtain a satisfactory performance, which involves the amount of input data and the network structure. A lack of data will lead to overfitting. Furthermore, the layers in the CNN will also affect the performance. In general, the convolution layer is the first, while the fully connected layer is the last. Convolution layers support the CNN to extract the characteristic features and information of an image. The fully connected layer will decide the number of parameters processed at the end of the network. Restricting the network to a maximum number of parameters should be a consideration, since too many parameters will lead to issues related to an increase in computational resources and overfitting [65]. There are several ways to find an agreement between the amount of data and the number of parameters that should be considered in a CNN for palm vein images. The convolution and max-pooling layers must be set into an optimum configuration in palm vein recognition. Therefore, a CNN configuration is developed in this research to verify the palm vein image, called VeinCNN. The developed VeinCNN applies one input layer, four convolution layers, four max-pooling layers, one flattened layer, and two dense layers, as shown in Figure 8.
The input layer is used to gain a palm vein image that is converted into a 128 × 128 size input with three layers. The next layers are the three convolution layers and four max-pooling layers with kernel size 3 and kernel size 2, respectively. The next step is proceeded by the flattened layer and dense layer. A summary of the network structure of VeinCNN is provided in Table 4.

2.4.2. Vein Recognition Using Hybrid DWT-HOG VeinCNN Feature Extraction

The noise in unclear images of veins can randomly affect the acquisition process and interfere with the recognition process. Filtration of palm vein images using the wavelet transformation method could handle this interference and provide clearer images from which gradient features can be extracted by applying HOG feature extraction. Using an orientation of 9, 8 × 8 pixels per cell and 2 × 2 cells per block, optimum images could be obtained for detecting the vein pattern related to orientation and direction. HOG is capable of gaining information features that are more noise resistant and representative. Detailed and complicated palm vein images can potentially be verified automatically using noise-resistant feature extraction and a CNN network with a more compact structure.
The proposed hybrid wavelet and HOG feature extraction method based on VeinCNN is provided by a block diagram in Figure 9, where five palm vein datasets from CASIA, Vera, Tongji, PolyU, and PUT are used, as shown in Table 2. Since this research was conducted on numerous datasets, each dataset is represented by an image in each process in order to provide a simple block diagram. This block diagram represents the performance of palm vein recognition, which is clearly shown by the third step in which the hybridization process is performed.

2.5. Performance Biometric Evaluation

An assessment of the biometric system’s performance was conducted. Accuracy, receiver operating characteristics (ROCs), and equal error rate (EER) are frequently used to evaluate performance. Accuracy is a measurement of the reliability of a biometric system. It compares the total true positives and true negatives with the overall total in a biometric system. The ROC curve is a two-axis representation. Both the true-positive rate and the false-positive rate are included in it. There is an area beneath the curve formed by the curves of those two categories. This is the area under the curve (AUC) score. There is a range of 0 to 1. EER is the name given to the junction of the ROC’s diagonal line and curve. EER displays the potential for the biometric authentication system to determine if a given probability is mistakenly positive or negative [66].

3. Results

In this paper, recognition is conducted using VeinCNN with a combination of wavelet feature extraction and HOG using five specified datasets. The accuracy, EER, and AUC for each dataset represent the recognition results. To investigate the impact of the feature extraction method, the proposed VeinCNN was used to simulate the recognition process without any feature extraction, which is described as the raw data. Then, the simulation-applied wavelet and HOG, respectively, and the hybrid wavelet and HOG feature extractor were applied sequentially to the VeinCNN structure. The sequential process in deep learning tends to preserve more computational resources, such as GPU resources and workload, compared with the parallel process. Table 5 shows a comparison of the results of the feature extractor.
VeinCNN, which is used to process the raw data and is considered the simplest method, provides the benchmark result for comparison with the other feature extractors. The results for the standalone wavelet feature extraction methods in this case show instability. Meanwhile, the HOG feature extraction methods applied to VeinCNN show conflicting results. In general, HOG feature extraction presents better results than wavelet transformation. Remarkably, the best result appeared when the hybrid wavelet and HOG were applied to VeinCNN. The accuracy and AUC increased in some datasets, and the EER decreased in mostly all datasets. The wavelet transform prepared the image, and HOG shows the edge features based on their gradient and magnitude. Table 6 shows a summary of the accuracy, EER, and AUC results for all datasets. The proposed feature extractor results are promising when compared with published works. On the CASIA dataset, the accuracy obtained by this proposed feature extractor is 99.85% compared with 99.25% obtained using HOG alone [56]. On the PUT dataset, the accuracy obtained by this proposed feature extractor is 99.85% compared with 93.92% obtained in a previous work [55].

4. Discussion

4.1. Total Parameter

The recognition performance results of VeinCNN using a combination of wavelet feature extraction and HOG are conducted and compared with several transfer learning methods as a benchmark in CNN research. In general, the parameters that are used in the proposed method are simpler than three existing transfer learning methods: AlexNet, VGG16, and ResNet50.
The application of the Hybrid DWT-HOG VeinCNN Method significantly impacts the total number of parameters in the CNN. The VeinCNN hybrid wavelet and HOG method requires 388,546 parameters. This number is much lower than in other transfer learning methods. It is 72 times lower than AlexNet, 38 times lower than VGG16, and 61 times lower than ResNet50.
The slight differences in the accuracy, AUC, and EER values could be compensated for by the lower computational resources. The low number of total parameters means that less computational resources are required. The comparison result of the performance is shown in the points. Figure 10 shows the difference in the total number of parameters in AlexNet [40], VGG16 [65], ResNet50 [66], and the proposed method.

4.2. Accuracy

Compared with AlexNet, VGG16, and ResNet50, the proposed recognition scheme Hybrid DWT-HOG VeinCNN is capable of obtaining great accuracy on all datasets, which is shown in Table 7. This proposed scheme gained the highest accuracy of 99.85% and 98.15% on the CASIA and PUT datasets, respectively, and gained the lowest accuracy of 85.97% on the PolyU dataset, which is slightly (4.9%) lower than the maximum accuracy achieved with the AlexNet recognition scheme. However, the low accuracy on the PolyU dataset achieved by the proposed recognition scheme is much better than that of existing schemes. Hence, the proposed recognition scheme is capable of performing with great accuracy on palm vein images.

4.3. Area under Curve

The proposed recognition scheme DWT-HOG VeinCNNalso results in an AUC as great as the accuracy on all datasets. The maximum AUC is achieved on the CASIA and PUT datasets with values of 99.85% and 98.15%, while the minimum AUC is attained on the PolyU dataset with an AUC of 85.88%. This value is 4.9% lower than the result attained with AlexNet, similar to results attained for accuracy as shown in Figure 11. This condition shows that the proposed scheme has adequate consistency. The results show that our proposed model attained the highest AUC of 95.3% on the Vera dataset.

4.4. EER

The EER generated by the proposed recognition scheme hybrid DWT-HOG VeinCNN is varied for all datasets, as shown in Table 8. The best EER value, 0.0083, is achieved by the proposed scheme on the CASIA dataset, so this scheme can avoid image misreading. The highest EER gained by this scheme, 0.0083, is 0.5% different from that achieved by AlexNet. Even so, the EER value generated by the proposed hybrid DWT-HOG VeinCNN recognition scheme is more consistent than others, including AlexNet, VGG16, and ResNet50. In fact, the average EER gained by the proposed scheme is 0.0592, 0.02% higher than that of AlexNet. The EER generated by the proposed scheme on the Tongji dataset is in line with that of the AlexNet recognition scheme, which is opposed to VGG16 and ResNet50 recognition schemes. Hence, the proposed recognition scheme hybrid DWT-HOG VeinCNN has the potential to avoid palm vein misreading.

5. Conclusions

This paper proposed a new scheme of CNN to recognize palm veins based on images. The proposed method, called VeinCNN, combines the features of DWT and HOG in order to robustly distinguish vein features.
VeinCNN starts from the fact that the acquisition process can be randomly affected by noise in indistinct images of veins. The recognition process can be affected by noisy images. Utilizing the wavelet transformation method allowed for the filtering of interference in venous images. After this transformation, the image is sharper and better prepared for the extraction of gradient features using HOG feature extraction. HOG features can acquire information properties that are more noise resistant and representative. The palm vein image has the potential to be an object to be automatically confirmed by a noise-resistant feature extraction method and modified CNN network.
The results show that wavelet transformation generally yields inferior results to HOG feature extraction. The best outcome was obtained when VeinCNN was hybridized with wavelet and HOG. Some datasets had an improvement in accuracy and AUC, but nearly all datasets had a decline in EER. The best result was obtained on the CASIA dataset with 99.85% accuracy, 99.80% AUC, and a 0.0083 EER value. Moreover, a total of 388,546 parameters used on the VeinCNN hybrid DWT and HOG method can maintain the results for accuracy, AUC, and EER.
According to the results, the proposed DWT-HOG VeinCNN method is a promising method compared with other recognition methods in transfer learning to obtain satisfactory palm vein recognition.

Author Contributions

Conceptualization, D.G.; methodology, M.W.; software, M.W.; validation, B.B. and D.G.; formal analysis, B.B. and D.G.; investigation, M.W. and R.C.; resources, M.W.; data curation, M.W.: writing—original draft preparation, M.W.; writing—review and editing, B.B. and R.C.; visualization, M.W. and B.B.; supervision, D.G. and B.B.; project administration, B.B.; funding acquisition, D.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Hibah Publikasi Terindeks Internasional (PUTI) Kolaborasi Internasional (KI) (2Q2) number NKB-782/UN2.RST/HKP.05.00/2020 from Universitas Indonesia with Dadang Gunawan (corresponding author) as lead of researchers.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

This manuscript uses image dataset which are from Chinese Academy of Sciences Institute of Automation (CASIA) [50], Hong Kong Polytechnic University (PolyU) [51], PUT [52], Tongji [53], and University of Applied Sciences Western Switzerland and the Idiap Research Institute [54].

Conflicts of Interest

On behalf of all the authors, the corresponding author states that they have no conflicts of interest.

References

  1. Jain, A.K.; Bolle, R.; Pankanti, S. Biometrics: Personal Identification in Network Society; Springer Science+Business Media, Inc.: New York, NY, USA, 2006. [Google Scholar]
  2. Shoniregun, C.A.; Crosier, S. Securing Biometrics Applications; Springer: New York, NY, USA, 2008. [Google Scholar]
  3. Pato, J.N.; Millet, L.I. Biometric Recognition: Challenges and Opportunities; Pato, J.N., Lynette I. Millett, L.I., Eds.; National Academies Press: Washington, DC, USA, 2010. [Google Scholar]
  4. Kono, M.; Ueki, H.; Umemura, S. Near-Infrared Finger Vein Patterns for Personal Identification. Appl. Opt. 2002, 41, 7429–7436. [Google Scholar] [CrossRef]
  5. Mil’shtein, S.; Pillai, A.; Shendye, A.; Liessner, C.; Baier, M. Fingerprint Recognition Algorithms for Partial and Full Fingerprints. In Proceedings of the 2008 IEEE Conference on Technologies for Homeland Security, Waltham, MA, USA, 12–13 May 2008; pp. 449–452. [Google Scholar] [CrossRef]
  6. Bounneche, M.D.; Boubchir, L.; Bouridane, A.; Nekhoul, B.; Ali-Chérif, A. Multi-Spectral Palmprint Recognition Based on Oriented Multiscale Log-Gabor Filters. Neurocomputing 2016, 205, 274–286. [Google Scholar] [CrossRef]
  7. Li, L.; Mu, X.; Li, S.; Peng, H. A Review of Face Recognition Technology. IEEE Access 2020, 8, 139110–139120. [Google Scholar] [CrossRef]
  8. Hofbauer, H.; Jalilian, E.; Sequeira, A.F.; Ferryman, J.; Uhl, A. Mobile NIR Iris Recognition: Identifying Problems and Solutions. In Proceedings of the 2018 IEEE 9th International Conference on Biometrics Theory, Applications and Systems (BTAS), Redondo Beach, CA, USA, 22–25 October 2018; pp. 1–9. [Google Scholar] [CrossRef]
  9. Khemani, A.; Choudhary, A. Intrinsic Biometrics. Int. J. Eng. Res. Technol. 2015, 4, 243–248. [Google Scholar]
  10. Tan, R.; Perkowski, M. Toward Improving Electrocardiogram (ECG) Biometric Verification Using Mobile Sensors: A Two-Stage Classifier Approach. Sensors 2017, 17, 410. [Google Scholar] [CrossRef]
  11. Chai, R.; Naik, G.R.; Ling, S.H.; Nguyen, H.T. Hybrid Brain–Computer Interface for Biomedical Cyber-Physical System Application Using Wireless Embedded EEG Systems. Biomed. Eng. Online 2017, 16, 5. [Google Scholar] [CrossRef]
  12. Nait-Ali, A. Hidden Biometrics: Towards Using Biosignals and Biomedical Images for Security Applications. In Proceedings of the International Workshop on Systems, Signal Processing and their Applications, WOSSPA, Tipaza, Algeria, 9–11 May 2011; pp. 352–356. [Google Scholar] [CrossRef]
  13. Ganz, A.; Witek, B.; Perkins, C., Jr.; Pino-Luey, D.; Resta-Flarer, F.; Bennett, H.; Lesser, J.; Ng, J.; Chiao, F.B. Vein Visualization: Patient Characteristic Factors and Efficacy of a New Infrared Vein Finder Technology. BJA Br. J. Anaesth. 2013, 110, 966–971. [Google Scholar] [CrossRef]
  14. Kumar, A.; Prathyusha, K.V. Personal Authentication Using Hand Vein Triangulation and Knuckle Shape. IEEE Trans. Image Process. 2009, 18, 2127–2136. [Google Scholar] [CrossRef] [PubMed]
  15. Liu, J.; Cui, J.; Xue, D.; Jia, X. Palm-Dorsa Vein Recognition Based on Independent Principle Component Analysis. In Proceedings of the 2011 International Conference on Image Analysis and Signal Processing, Hubei, China, 21–23 October 2011; pp. 660–664. [Google Scholar] [CrossRef]
  16. Zhou, Y.; Kumar, A. Contactless Palm Vein Identification Using Multiple Representations. In Proceedings of the 2010 Fourth IEEE International Conference on Biometrics: Theory, Applications and Systems (BTAS), Washington, DC, USA, 27–29 September 2010; pp. 1–6. [Google Scholar] [CrossRef]
  17. Sarkar, I.; Alisherov, F.; Kim, T.; Bhattacharyya, D. Palm Vein Authentication System: A Review. Int. J. Control Autom. 2010, 3, 27–34. [Google Scholar]
  18. Nayar, G.R.; Bhaskar, A.; Satheesh, L.; Kumar, P.S.; P, A.R. Personal Authentication Using Partial Palmprint and Palmvein Images with Image Quality Measures. In Proceedings of the 2015 International Conference on Computing and Network Communications (CoCoNet), Trivandrum, India, 16–19 December 2015; pp. 191–198. [Google Scholar] [CrossRef]
  19. Hoshyar, A.; Sulaiman, R.; Noori Houshyar, A. Smart Access Control with Finger Vein Authentication and Neural Network. J. Am. Sci. 2011, 7, 192–200. [Google Scholar]
  20. Syarif, M.A.; Ong, T.S.; Teoh, A.B.J.; Tee, C. Enhanced Maximum Curvature Descriptors for Finger Vein Verification. Multimed. Tools Appl. 2017, 76, 6859–6887. [Google Scholar] [CrossRef]
  21. Shahzad, A.; Walter, N.; Malik, A.S.; Saad, N.M.; Meriaudeau, F. Multispectral Venous Images Analysis for Optimum Illumination Selection. In Proceedings of the 2013 IEEE International Conference on Image Processing, Melbourne, VIC, Australia, 15–18 September 2013; pp. 2383–2387. [Google Scholar] [CrossRef]
  22. Gurunathan, V.; Bharathi, S.; Sudhakar, R. Image Enhancement Techniques for Palm Vein Images. In Proceedings of the 2015 International Conference on Advanced Computing and Communication Systems, Coimbatore, India, 5–7 January 2015; pp. 1–5. [Google Scholar] [CrossRef]
  23. Zhou, Y.; Liu, Y.; Feng, Q.; Yang, F.; Huang, J.; Nie, Y. Palm-Vein Classification Based on Principal Orientation Features. PLoS ONE 2014, 9, e112429. [Google Scholar] [CrossRef]
  24. Stanuch, M.; Wodzinski, M.; Skalski, A. Contact-Free Multispectral Identity Verification System Using Palm Veins and Deep Neural Network. Sensors 2020, 20, 5695. [Google Scholar] [CrossRef]
  25. Sasikala, R.; Sandhya, S.; Ravichandran, D.K.; Balsubramaniam, D.D. A Survey on Human Identification Using Palm-Vein Images Using Laplacian Filter. Int. J. Innov. Res. Comput. Commun. Eng. 2016, 4, 6599–6605. [Google Scholar]
  26. Wu, X.; Gao, E.; Tang, Y.; Wang, K. A Novel Biometric System Based on Hand Vein. In Proceedings of the 2010 Fifth International Conference on Frontier of Computer Science and Technology, Changchun, China, 18–22 August 2010; pp. 522–526. [Google Scholar] [CrossRef]
  27. Akintoye, K.A.; Rahim, M.S.M.; Abdullah, A.H. Enhancement of Finger Vein Image Using Multifiltering Algorithm. ARPN J. Eng. Appl. Sci. 2018, 13, 644–648. [Google Scholar]
  28. Raut, S.D.; Humbe, V.T.; Mane, A.V. Development of Biometrie Palm Vein Trait Based Person Recognition System: Palm Vein Biometrics System. In Proceedings of the Intelligent Systems and Information Management (ICISIM), 2017 1st International Conference on IEEE, Aurangabad, India, 5–6 October 2017; pp. 18–21. [Google Scholar]
  29. Elnasir, S.; Shamsuddin, S.M. Palm Vein Recognition Based on 2D-Discrete Wavelet Transform and Linear Discrimination Analysis. Int. J. Adv. Soft Comput. 2014, 6, 43–59. [Google Scholar]
  30. Debnath, L. Wavelet Transforms and Their Applications. Appl. Math. 2003, 48, 78. [Google Scholar] [CrossRef]
  31. Wu, W.; Li, Y.; Zhang, Y.; Li, C. Identity Recognition System Based on Multi-Spectral Palm Vein Image. Electronics 2023, 12, 3503. [Google Scholar] [CrossRef]
  32. Fronitasari, D.; Gunawan, D. Local Descriptor Approach to Wrist Vein Recognition with DVH-LBP Domain Feature Selection Scheme. Int. J. Adv. Sci. Eng. Inf. Technol. 2019, 9, 1025–1032. [Google Scholar] [CrossRef]
  33. Bashar, K.; Murshed, M. Texture Based Vein Biometrics for Human Identification: A Comparative Study. In Proceedings of the 2018 IEEE 42nd Annual Computer Software and Applications Conference (COMPSAC), Tokyo, Japan, 23–27 July 2018; Volume 2, pp. 571–576. [Google Scholar] [CrossRef]
  34. Dalal, N.; Triggs, B. Histograms of Oriented Gradients for Human Detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20-25 June 2005; Volume 1, pp. 886–893. [Google Scholar]
  35. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep Feature Extraction and Classification of Hyperspectral Images Based on Convolutional Neural Networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef]
  36. Suryani, D.; Doetsch, P.; Ney, H. On the Benefits of Convolutional Neural Network Combinations in Offline Handwriting Recognition. In Proceedings of the 2016 15th International Conference on Frontiers in Handwriting Recognition (ICFHR), Shenzhen, China, 23–26 October 2016; pp. 193–198. [Google Scholar] [CrossRef]
  37. Meng, G.; Fang, P.; Zhang, B. Finger Vein Recognition Based on Convolutional Neural Network. In MATEC Web of Conferences; EDP Sciences: Les Ulis, France, 2017; Volume 128, p. 4015. [Google Scholar]
  38. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet Classification with Deep Convolutional Neural Networks. In Proceedings of the International Conference on Neural Information Processing Systems (NIPS), Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  39. Fairuz, S.; Habaebi, M.H.; Elsheikh, E.M.A.; Chebil, A.J. Convolutional Neural Network-Based Finger Vein Recognition Using Near Infrared Images. In Proceedings of the 2018 7th International Conference on Computer and Communication Engineering (ICCCE), Kuala Lumpur, Malaysia, 19–20 September 2018; pp. 453–458. [Google Scholar] [CrossRef]
  40. Liu, W.; Li, W.; Sun, L.; Zhang, L.; Chen, P. Finger Vein Recognition Based on Deep Learning. In Proceedings of the 2017 12th IEEE Conference on Industrial Electronics and Applications (ICIEA), Siem Reap, Cambodia, 18–20 June 2017; pp. 205–210. [Google Scholar] [CrossRef]
  41. Wan, H.; Chen, L.; Song, H.; Yang, J. Dorsal Hand Vein Recognition Based on Convolutional Neural Networks. In Proceedings of the 2017 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Kansas City, MO, USA, 13–16 November 2017; pp. 1215–1221. [Google Scholar] [CrossRef]
  42. Li, X.; Huang, D.; Wang, Y. Comparative Study of Deep Learning Methods on Dorsal Hand Vein Recognition BT-Biometric Recognition; You, Z., Zhou, J., Wang, Y., Sun, Z., Shan, S., Zheng, W., Feng, J., Zhao, Q., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 296–306. [Google Scholar]
  43. Qin, H.; El Yacoubi, M.A.; Lin, J.; Liu, B. An Iterative Deep Neural Network for Hand-Vein Verification. IEEE Access 2019, 7, 34823–34837. [Google Scholar] [CrossRef]
  44. Chantaf, S.; Hilal, A.; Elsaleh, R. Palm Vein Biometric Authentication Using Convolutional Neural Networks BT-Proceedings of the 8th International Conference on Sciences of Electronics, Technologies of Information and Telecommunications (SETIT’18); Bouhlel, M.S., Rovetta, S., Eds.; Springer International Publishing: Cham, Switzerland, 2020; Volume 1, pp. 352–363. [Google Scholar]
  45. Hong, H.G.; Lee, M.B.; Park, K.R. Convolutional Neural Network-Based Finger-Vein Recognition Using NIR Image Sensors. Sensors 2017, 17, 1–21. [Google Scholar] [CrossRef] [PubMed]
  46. Chen, Y.Y.; Hsia, C.H.; Chen, P.H. Contactless Multispectral Palm-Vein Recognition with Lightweight Convolutional Neural Network. IEEE Access 2021, 9, 149796–149806. [Google Scholar] [CrossRef]
  47. Wang, J.; Pan, Z.; Wang, G.; Li, M.; Li, Y. Spatial Pyramid Pooling of Selective Convolutional Features for Vein Recognition. IEEE Access 2018, 6, 28563–28572. [Google Scholar] [CrossRef]
  48. Wang, J.; Yang, K.; Pan, Z.; Wang, G.; Li, M.; Li, Y. Minutiae-Based Weighting Aggregation of Deep Convolutional Features for Vein Recognition. IEEE Access 2018, 6, 61640–61650. [Google Scholar] [CrossRef]
  49. Hou, B.; Yan, R. Convolutional Auto-Encoder Based Deep Feature Learning for Finger-Vein Verification. In Proceedings of the 2018 IEEE International Symposium on Medical Measurements and Applications (MeMeA), Rome, Italy, 11–13 June 2018; pp. 1–5. [Google Scholar] [CrossRef]
  50. Hao, Y.; Sun, Z.; Tan, T.; Ren, C. Multispectral Palm Image Fusion for Accurate Contact-Free Palmprint Recognition. In Proceedings of the Image Processing, 2008. ICIP 2008. 15th IEEE International Conference on, San Diego, CA, USA, 12–15 October 2008; pp. 281–284. [Google Scholar]
  51. Zhang, D.; Guo, Z.; Lu, G.; Zhang, L.; Zuo, W. An Online System of Multispectral Palmprint Verification. IEEE Trans. Instrum. Meas. 2010, 59, 480–490. [Google Scholar] [CrossRef]
  52. Kabacinski, R.; Kowalski, M. Vein Pattern Database and Benchmark Results. Electron. Lett. 2011, 47, 1127–1128. [Google Scholar] [CrossRef]
  53. Zhang, L.; Cheng, Z.; Shen, Y.; Wang, D. Palmprint and Palmvein Recognition Based on DCNN and A New Large-Scale Contactless Palmvein Dataset. Symmetry 2018, 10, 78. [Google Scholar] [CrossRef]
  54. Tome, P.; Marcel, S. On the Vulnerability of Palm Vein Recognition to Spoofing Attacks. In Proceedings of the 2015 International Conference on Biometrics (ICB), Phuket, Thailand, 19–22 May 2015; pp. 319–325. [Google Scholar]
  55. Wulandari, M.; Gunawan, D. On the Performance of Pretrained CNN Aimed at Palm Vein Recognition Application. In Proceedings of the 2019 11th International Conference on Information Technology and Electrical Engineering, ICITEE 2019, Pattaya, Thailand, 10–11 October 2019. [Google Scholar] [CrossRef]
  56. Kuang, H.; Zhong, Z.; Liu, X.; Ma, X. Palm Vein Recognition Using Convolution Neural Network Based on Feature Fusion with HOG Feature. In Proceedings of the 2020 5th International Conference on Smart Grid and Electrical Automation (ICSGEA), Zhangjiajie, China, 13–14 June 2020; pp. 295–299. [Google Scholar] [CrossRef]
  57. Al-Rababah, K.; Mustaffa, M.R.; Doraisamy, S.C.; Khalid, F. Hybrid Discrete Wavelet Transform and Histogram of Oriented Gradients for Feature Extraction and Classification of Breast Dynamic Thermogram Sequences. In Proceedings of the 2021 Fifth International Conference on Information Retrieval and Knowledge Management (CAMP), Kuala Lumpur, Malaysia, 15–16 June 2021; pp. 31–35. [Google Scholar] [CrossRef]
  58. Ristiana, R.; Kusumandari, D.E.; Simbolon, A.I.; Amri, M.F.; Sanhaji, G.; Rumiah, R. A Comparative Study of Thermal Face Recognition Based on Haar Wavelet Transform (HWT) and Histogram of Gradient (HoG). In Proceedings of the 2021 3rd International Symposium on Material and Electrical Engineering Conference (ISMEE), Bandung, Indonesia, 10–11 November 2021; pp. 242–248. [Google Scholar] [CrossRef]
  59. Zhang, L.; Li, L.; Yang, A.; Shen, Y.; Yang, M. Towards Contactless Palmprint Recognition: A Novel Device, a New Benchmark, and a Collaborative Representation Based Identification Approach. Pattern Recognit. 2017, 69, 199–212. [Google Scholar] [CrossRef]
  60. Arora, S.; Brar, Y.S.; Kumar, S. HAAR Wavelet Transform for Solution of Image Retrieval. Int. J. Adv. Comput. Math. Sci. 2014, 5, 27–31. [Google Scholar]
  61. Zou, W.; Li, Y. Image Classification Using Wavelet Coefficients in Low-Pass Bands. In Proceedings of the 2007 International Joint Conference on Neural Networks, Orlando, FL, USA, 12–17 August 2007; pp. 114–118. [Google Scholar] [CrossRef]
  62. Shyla, N.S.J.; Emmanuel, W.R.S. Automated Classification of Glaucoma Using DWT and HOG Features with Extreme Learning Machine. In Proceedings of the 2021 Third International Conference on Intelligent Communication Technologies and Virtual Mobile Networks (ICICV), Tirunelveli, India, 4–6 February 2021; pp. 725–730. [Google Scholar] [CrossRef]
  63. Noh, Z.M.; Ramli, A.R.; Saripan, M.I.; Hanafi, M. Overview and Challenges of Palm Vein Biometric System. Int. J. Biom. 2016, 8, 2–18. [Google Scholar] [CrossRef]
  64. Ketkar, N.; Santana, E. Deep Learning with Python. Master’s Thesis, Apress, Berkeley, CA, USA, 2017; pp. 95–109. [Google Scholar]
  65. Liu, J.; Pan, Y.; Li, M.; Chen, Z.; Tang, L.; Lu, C.; Wang, J. Applications of Deep Learning to MRI Images: A Survey. Big Data Min. Anal. 2018, 1, 1–18. [Google Scholar] [CrossRef]
  66. Sarkar, A.; Singh, B.K. A Review on Performance, Security and Various Biometric Template Protection Schemes for Biometric Authentication Systems. Multimed. Tools Appl. 2020, 79, 27721–27776. [Google Scholar] [CrossRef]
  67. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  68. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
Figure 1. Palm vein recognition with hybrid DWT-HOG VeinCNN.
Figure 1. Palm vein recognition with hybrid DWT-HOG VeinCNN.
Sensors 24 00341 g001
Figure 2. Examples from palm vein datasets: (a) CASIA, (b) Vera, (c) Tongji, (d) PolyU, and (e) PUT.
Figure 2. Examples from palm vein datasets: (a) CASIA, (b) Vera, (c) Tongji, (d) PolyU, and (e) PUT.
Sensors 24 00341 g002
Figure 3. ROI results for (a) CASIA, (b) Vera, (c) Tongji, (d) PolyU, and (e) PUT datasets.
Figure 3. ROI results for (a) CASIA, (b) Vera, (c) Tongji, (d) PolyU, and (e) PUT datasets.
Sensors 24 00341 g003
Figure 4. Wavelet feature extraction process.
Figure 4. Wavelet feature extraction process.
Sensors 24 00341 g004
Figure 5. Histogram of oriented gradient feature extraction.
Figure 5. Histogram of oriented gradient feature extraction.
Sensors 24 00341 g005
Figure 6. Block diagram of hybrid DWT and HOG feature extraction.
Figure 6. Block diagram of hybrid DWT and HOG feature extraction.
Sensors 24 00341 g006
Figure 7. Wavelet and HOG feature extraction for each dataset: (a) CASIA, (b) Vera, (c) Tongji, (d) PolyU, and (e) PUT.
Figure 7. Wavelet and HOG feature extraction for each dataset: (a) CASIA, (b) Vera, (c) Tongji, (d) PolyU, and (e) PUT.
Sensors 24 00341 g007
Figure 8. The structure of VeinCNN.
Figure 8. The structure of VeinCNN.
Sensors 24 00341 g008
Figure 9. Block diagram of the proposed VeinCNN and wavelet feature extraction HOG combination palm vein recognition process.
Figure 9. Block diagram of the proposed VeinCNN and wavelet feature extraction HOG combination palm vein recognition process.
Sensors 24 00341 g009
Figure 10. Total parameter in AlexNet, VGG16, ResNet50, and the proposed method.
Figure 10. Total parameter in AlexNet, VGG16, ResNet50, and the proposed method.
Sensors 24 00341 g010
Figure 11. The AUC results of the transfer-learning-based methods on (a) CASIA, (b) Vera, (c) Tongji, (d) PolyU and (e) PUT dataset [40,65,66].
Figure 11. The AUC results of the transfer-learning-based methods on (a) CASIA, (b) Vera, (c) Tongji, (d) PolyU and (e) PUT dataset [40,65,66].
Sensors 24 00341 g011
Table 1. Comparison with state-of-the-art models.
Table 1. Comparison with state-of-the-art models.
AuthorsDatasetFeature ExtractorMethod
Wang et al. (2018) [47]PolyU-VGG-16
Qin et al. (2019) [43]CASIA
PolyU
-DBN
Wulandari et al. (2019) [55]PUTDWTCNN
Chantaf et al. (2020) [44]Nonpublic-CNN
Kuang et al. (2020) [56]CASIA
PolyU
HOGCNN
Chen et al. (2021) [46]CASIA
PUT
GaborCNN
Wu et al. (2023) [31]Nonpublic
CASIA
Tongji
PolyU
SDSPCA-NPEDistance Feature Matching
ProposedCASIA
Vera
Tongji
PolyU
PUT
Hybrid DWT and HOGVeinCNN
Table 2. Summary of palm vein datasets.
Table 2. Summary of palm vein datasets.
No.DatasetTotal VolunteersTotal ImagesNumber of SessionImage SizeFormatOfficial ROI
1.CASIA20024002 768 × 576 jpgNot Available
2.Vera22022002 480 × 576 pngAvailable
3.Tongji60012,0002 800 × 600 tiffAvailable
4.PolyU50024,0002 352 × 288 jpgAvailable
5.PUT10012003 768 × 1024 bmpNot Available
Table 3. The procedure of the proposed hybrid DWT and HOG feature extraction.
Table 3. The procedure of the proposed hybrid DWT and HOG feature extraction.
Proposed Hybrid Feature Extractor DWT and HOG Procedure
1: X m , n P a l m   V e i n   I m a g e   ( A m , n )
2: C A m 2 , n 2 , C H m 2 , n 2 , C V m 2 , n 2 , C D m 2 , n 2 H a a r   W a v e l e t   T r a n s f o r m   ( X m , n )
3: Y m , n P a l m   V e i n   I m a g e   ( C A m 2 , n 2 )
4: G x , G y G r a d i e n t ( Y m , n )
5: G r a d i e n t M a g n i t u d e M a g n i t u d e G x , G y
6: G r a d i e n t D i r e c t i o n D i r e c t i o n G x , G y
7: H O G F e a t u r e H i s t o g r a m   N o r m a l i z a t i o n G r a d i e n t M a g n i t u d e , G r a d i e n t D i r e c t i o n
8: H O G i m a g e H O G F e a t u r e
Table 4. Summary of the VeinCNN network structure.
Table 4. Summary of the VeinCNN network structure.
LayerTypeActivation FunctionOutput ShapeKernel SizeNumber of Filters
0Input-128 × 128--
12D convReLU126 × 126332
22D max poolingReLU63 × 63232
32D convReLU61 × 61364
42D max poolingReLU30 × 30264
52D convReLU28 × 28364
62D max poolingReLU14 × 14264
72D convReLU12 × 12364
82D max poolingReLU6 × 6264
9Flattened-2304--
10DenseReLU128--
11DenseSigmoid2--
Table 5. The result of accuracy, AUC, and EER comparison methods.
Table 5. The result of accuracy, AUC, and EER comparison methods.
DatasetFeature ExtractorAccuracy (%)AUC (%)EER
CASIARaw data99.6999.600.0167
DWT99.8598.800.0250
HOG99.8599.800.0083
Hybrid DWT and HOG (proposed)99.8599.800.0083
VeraRaw data84.1484.100.1273
DWT81.1681.200.2000
HOG94.7894.600.0636
Hybrid DWT and HOG (proposed)95.5795.100.0545
TongjiRaw data90.3790.400.1383
DWT92.0492.000.0866
HOG94.8194.400.0750
Hybrid DWT and HOG (proposed)94.9194.900.0650
PolyURaw data59.4459.400.3000
DWT58.3360.200.2566
HOG82.5982.600.2066
Hybrid DWT and HOG (proposed)85.8885.900.1467
PUTRaw data93.5293.500.1000
DWT74.0781.300.1330
HOG97.2297.200.0333
Hybrid DWT and HOG (proposed)98.1298.100.0167
Table 6. Summary of accuracy, AUC, and EER for each dataset using the proposed method.
Table 6. Summary of accuracy, AUC, and EER for each dataset using the proposed method.
DatasetAccuracy (%)AUC (%)EER
CASIA99.8599.800.0083
Vera95.5795.100.0545
Tongji94.9194.900.0650
PolyU85.8885.900.1467
PUT98.1298.100.0167
Table 7. The accuracy results of transfer-learning-based methods on datasets.
Table 7. The accuracy results of transfer-learning-based methods on datasets.
MethodsAccuracy on Dataset (%)
CASIAVeraTongjiPolyUPUT
AlexNet [38]96.6090.8693.8073.9897.22
VGG16 [67]99.3891.0478.4370.8397.22
ResNet50 [68]99.6986.3888.4379.1794.44
Proposed hybrid DWT-HOG VeinCNN99.8595.5294.9185.9798.15
Table 8. The EER results of transfer-learning-based methods on datasets.
Table 8. The EER results of transfer-learning-based methods on datasets.
Recognition SchemeEER on Dataset
CASIAVeraTongjiPolyUPUT
AlexNet [38]0.06790.06720.06850.33700.0185
VGG16 [67]0.01230.06720.19630.25280.0555
ResNet50 [68]0.00830.05600.11670.15190.0185
Proposed Hybrid DWT-HOG VeinCNN0.00830.05450.06300.14600.0167
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wulandari, M.; Chai, R.; Basari, B.; Gunawan, D. Hybrid Feature Extractor Using Discrete Wavelet Transform and Histogram of Oriented Gradient on Convolutional-Neural-Network-Based Palm Vein Recognition. Sensors 2024, 24, 341. https://doi.org/10.3390/s24020341

AMA Style

Wulandari M, Chai R, Basari B, Gunawan D. Hybrid Feature Extractor Using Discrete Wavelet Transform and Histogram of Oriented Gradient on Convolutional-Neural-Network-Based Palm Vein Recognition. Sensors. 2024; 24(2):341. https://doi.org/10.3390/s24020341

Chicago/Turabian Style

Wulandari, Meirista, Rifai Chai, Basari Basari, and Dadang Gunawan. 2024. "Hybrid Feature Extractor Using Discrete Wavelet Transform and Histogram of Oriented Gradient on Convolutional-Neural-Network-Based Palm Vein Recognition" Sensors 24, no. 2: 341. https://doi.org/10.3390/s24020341

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop