1. Introduction
Palmprint recognition constitutes a pivotal biometric technology deployed in the identification and verification of individuals, relying on the distinctive patterns inherent in their palmprints. This method, known for its reliability and security, finds extensive applications in diverse fields, including access control, security systems, and forensic investigations [
1]. The palmprint recognition methodology originated with a focus on forensic analysis of latent prints, capitalizing on the intricate and extensive textural features compared to fingerprints. In forensic applications, high-resolution images exceeding 400 dots per inch (dpi) are utilized to capture detailed structural information. In contrast, civil and commercial systems, such as access control, opt for lower resolutions under 150 dpi to balance utility and practicality. The larger surface area covered by palmprints allows for highly discriminative characterization, even in low-quality images. The resolution-field condition trade-off underscores the differing goals: Forensic usage requires definitive one-to-one matching for evidence, whereas access control emphasizes immediate user authentication and system integration. A comprehensive understanding of these contexts and their implications on image quality, feature representation, and matching algorithms is crucial for adapting palmprint recognition to diverse application needs. The process of palmprint recognition entails several fundamental stages: image capture and acquisition, preprocessing for normalization and enhancement, descriptive feature extraction, and finally, pattern matching for classification.
Palmprint image acquisition involves capturing high-quality palmprint images using various devices like cameras, scanners, or smartphones. These images are then subjected to preprocessing techniques, encompassing noise reduction, normalization, and enhancement, to ensure consistent submission despite any restrictions on the availability of materials and/or refined input data. Following preprocessing, relevant features are extracted, like minutiae points specific to an individual’s palm, ridges, and lines. These features are crucial for accurate identification and are obtained through advanced image processing methods [
2].
Palmprint recognition systems possess notable advantages—notably their non-intrusive nature, stability of features over time, and the abundance of unique identifying characteristics on the palm [
3]. However, challenges persist, including variations in illumination, pose, and image quality, necessitating meticulous attention for precise and dependable recognition outcomes.
Sustained research efforts, especially within image processing, machine learning, and deep learning, have significantly enhanced palmprint recognition systems. These advancements have solidified the integration of palmprint recognition as an indispensable component within contemporary biometric security applications [
4].
Despite promising advancements in palmprint recognition, numerous unresolved issues and open challenges persist in the field. These challenges encompass diverse factors such as changes in pose, occlusion, blurring, image resolution, and the synthesis of palmprints [
5,
6]. Successfully addressing these challenges requires substantial efforts aimed at enhancing palmprint acquisition, normalization techniques, and recognition algorithms. Such efforts are crucial for deploying palmprint recognition within forensic analysis, surveillance systems, mobile phone security, and various commercial applications.
To confront and deal with these challenges, this paper provides a meticulous analysis that highlights notable advances that have significantly influenced the development of palmprint biometric recognition. This discussion covers historical evolutions up to the present, with focused attention on anticipating future directions in the field.
In this comprehensive review, our main contributions can be succinctly outlined as follows:
- -
We deliver a timely, thorough, and concise review of the extensive literature on image-based palmprint recognition. This includes an analysis of both contact and contactless palmprint databases, along with the employed evaluation methods. Our evaluation covers 20 databases and more than 60 publications from late 2002 to 2023.
- -
Our goal is to enlighten emerging scholars by highlighting significant advances in the historical context of the field and directing them to relevant references for in-depth exploration.
- -
We present a systematic categorization of palmprint feature extraction techniques, including contemporary methods rooted in deep learning. The purpose of developing this taxonomy is to structure the existing literature on palmprint recognition approaches and provide a coherent framework for understanding the diverse methodologies employed in the field.
- -
We provide an up-to-date and thorough comprehensive survey of both contact and contactless databases utilized in the realm of palmprint recognition. Our methodology involves organizing these databases and building a chronological timeline, showing the evolution of these datasets over time in terms of the number of individuals represented and the number of samples per individual.
- -
We scrutinize the existing deep learning-based methodologies, highlighting their exceptional performance on intricate, unregulated, and extensive datasets. Consequently, our examination offers researchers with an inclusive understanding of deep learning-based techniques, which have significantly transformed palmprint recognition paradigms since the early 2015s.
This paper’s subsequent sections are outlined as follows:
Section 2 delves into the anatomical structure of the palm of the hand. Then,
Section 3 focuses on the palmprint as a distinctive biometric modality. In
Section 4, a comprehensive overview of the general framework for palmprint recognition systems is presented. The challenges inherent in the implementation and the employment of palmprint recognition systems are discussed in
Section 5.
Section 6 introduces palmprint recognition databases and the proposed taxonomy.
Section 7 provides a meticulous classification of palmprint feature extraction methods, detailing the pivotal contributions that have significantly advanced this field. Finally,
Section 8 serves as the conclusion, summarizing key findings and addressing future research directions.
2. What Is a Palmprint?
The palmprint refers to the unique pattern of ridges and valleys found on the inner surface of the hand, excluding the wrist and fingers. Palmprints, akin to fingerprints, are a biometric characteristic unique to each individual. A seminal study by Shu and Zhang in 1998 [
7] explored the viability of palmprints as a means of personal identification, establishing them as a form of physical biometrics. Their findings highlighted distinctive features in palmprints, including major lines (life, heart, and head lines), wrinkles, minutiae, and delta points. Each palmprint is unique, and the surface of the palm provides more information space compared to fingerprints, so it contains a greater amount of information.
In general, the attributes of the palmprint manifest on multiple levels, each discernible in various types of palmprint images. Typically, these characteristics are visible across different image resolutions, with lower resolution, around 100 pixels per inch (ppi) [
8,
9], exhibiting a pronounced texture in which dark lines are of particular significance and visibility. Notably, among these lines, the three widest and longest are termed major lines, constituted by the heart line gathering with the head and life lines, and the remaining lines are referred to as wrinkles [
10], as illustrated in
Figure 1. Therefore, in the case of low-resolution images, the predominant features are the major lines, wrinkles, and texture. Nevertheless, the edges of the palmprint remain imperceptible in images of low resolution. In contrast, visibility can be achieved in the case of images with high resolution, of approximately 500 dpi, which unveils local texture intricacies, including minute creases, ridges, valleys, and minutiae points [
10]. Furthermore, images with very high resolution allow an abundance of certain local particular features related to the palmprint to be visualized, including the pores, which can be seen in resolutions exceeding 500 ppi or even reaching 1000 ppi.
3. Why Palmprint Recognition?
In the field of biometric recognition, facial identification still has limitations due to persistent challenges, such as pose, lighting, and orientation variations [
11]. Conversely, fingerprints have been widely adopted due to their efficiency, although certain populations, such as manual workers and the elderly, may have difficulty with capturing fingerprints. In a networked society, reliable personal authentication remains critical for security [
12]. Compared to other biometric modalities, palmprints have proven to be more effective and acceptable. The palmprint biometric system offers higher accuracy than fingerprints and higher acceptance than facial recognition. With characteristics such as uniqueness, reliability, and security, palmprints have been widely adopted by security agencies, providing a cost-effective and non-intrusive option for developing accurate and efficient biometric systems.
Advanced research in palmprint feature extraction [
13,
14] has been conducted for contactless systems. Contactless palmprint recognition aims to improve usability and privacy. However, the lack of a knuckle guide can lead to variations in palmprint images due to hand movements. Various methods, such as the utilization of texture operators like local binary pattern (LBP) [
15] and Gabor filters [
16], were proposed to overcome these challenges.
Palmprint has advantages over other biometric methods, including iris and fingerprint, in terms of identity matching. Palmprints offer the advantage of easy capture with low-resolution devices, mitigating the high costs associated with other modalities. Moreover, law enforcement agencies have extensively employed palmprints for criminal identification, leveraging their unique and stable characteristics [
17,
18]. These prints encapsulate diverse features like primary lines, minutiae points, ridges, and overall texture. Each feature class contributes significantly to the individuality and discriminative power of a palmprint. This flexibility permits adaptation to the specific security requirements of individuals and organizations.
4. Structure of a Palmprint Recognition System
As delineated in
Figure 2, the palmprint recognition framework structure encompasses four key stages similar to broader biometric architectures—(i) image acquisition, followed by (ii) preprocessing and (iii) feature extraction, and culminating with (iv) classification [
19]. The preprocessing stage aims to enhance image performance and remove extraneous elements. Then, the process moves to the feature extraction stage, which allows features to be elicited from the image of the palm through advanced image analytics. Finally, the image proceeds to the classification stage, where it undergoes classification to match image samples with individuals and identify the closest match in the database to the palmprint used in the test.
4.1. Image Acquisition
Contingent on the imaging apparatus, palmprint acquisition methodologies bifurcate into contact-based palmprint images and contactless ones [
14]. The first category necessitates direct palm-sensor contact; however, the second one involves no direct physical contact. Indeed, in the first case, images are acquired with palms placed on the device and hands guided by positioning markers for the user. Conversely, in the second scenario, images are captured without any physical contact with the device.
Figure 3 illustrates both modes of palmprint image acquisition, with and without contact [
20,
21].
4.2. Preprocessing
Image preprocessing involves denoising and smoothing the region of interest (ROI) in the entered data before deriving significant features within the palmprint images. ROI extraction in palmprint recognition adverts to the process of identifying and isolating the specific area of a palmprint image that contains the most relevant and distinctive features for recognition purposes [
22]. This extraction is critical for accurate feature analysis and comparison within a recognition system. Various techniques are used to extract the ROI, which typically involves locating the central area of the palmprint image where key features like lines, ridges and minutiae points are concentrated.
The ROI extraction process aims to enhance the ability and the fineness of palmprint recognition systems by focusing computational efforts on the most informative part of the palmprint image. This targeted approach ensures that only the relevant features are considered during feature extraction and comparison, resulting in more reliable and accurate recognition results. Proper ROI extraction methods are essential for achieving optimal performance in palmprint recognition systems, making it a fundamental step in the overall recognition process.
Figure 4 highlights an example of the preprocessing module of the palmprint identification system, comprising five essential stages [
23].
4.3. Feature Extraction
Feature extraction captures distinctive features from biometric data to create a unique digital representation of the palmprint. Algorithms transform raw data into discriminative features used for identification or verification. These features must be invariant to irrelevant variation and highlight fundamental characteristics. The following methods are most commonly used in palmprint feature extractison: line-based, subspace learning-based, local direction encoding-based, texture descriptor-based, and deep learning-based methods. This phase forms the core of this article and will be presented in more detail in
Section 7.
4.4. Classification
During recognition, the features derived from the entered palmprint are compared with the features stored in a database. Various matching algorithms, such as Euclidean distance or neural networks, are used to determine the similarity between the input features and the stored templates.
4.5. Evaluation Performances
The valid accuracy and equal error rate (EER) serve as widely accepted metrics for evaluating the performance of biometric systems. These metrics are fundamental in judging the effectiveness of such systems and are commonly utilized in the field. Valid accuracy assesses the overall correctness of the system in authenticating users, indicating its capability to accurately identify legitimate users. Conversely, the EER stand as a pivotal metric in evaluating biometric system performance. It pinpoints the precise operating condition where the false acceptance rate (FAR) and false rejection rate (FRR) converge equally, signifying an equilibrium assessment of the system performance.
The valid accuracy is the percentage of correctly accepted genuine instances or positive matches in the biometric system, which can be obtained through:
where:
True acceptances: the number of instances where the biometric system correctly accepts a genuine user.
Total number of genuine instances: the total number of instances where a genuine user attempts authentication.
The EER serves as an equilibrium point between the two error rates, which allows the FAR and FRR to intersect.
where:
It is important to note that lower EER values indicate better performance in terms of balancing the false acceptance rate (FAR) and false rejection rate (FRR).
In an ideal system, the recognition rate would be 100% and the EER would be 0%. However, in practice, there is often a trade-off between these two metrics, and system designers aim to find a balance that meets the requirements of the specific application.
5. Palmprint Biometric Recognition Challenges
Palmprint recognition presents a number of complex challenges, primarily due to reduced pattern quality, variations in focal length, nonlinear deformation caused by contactless image capture systems, and computational complexity caused by the large size of typical palmprint images. In addition, contactless palmprint research ffaces specific problems [
24]. First, the accuracy of contactless palmprint matching tends to decrease compared to contact images due to more pronounced image variations. This requires the development of advanced matching techniques to improve accuracy. Second, automated recognition of contactless palmprints from entered hands is complex due to dynamic or unstable backgrounds. Existing research addresses this problem by using fixed backgrounds for image acquisition and pixel-wise operators for key point detection.
Palmprint recognition shares several common problems with traditional fingerprint recognition, including the detection of ridges, valleys, and minutiae points [
25]. However, palmprints are larger and more complex, which slows down their recognition at high resolutions. The deformation of fingerprints, especially due to joint variations, is a critical and more complex issue compared to the deformation of palmprints.
Different regions of palmprints exhibit varying qualities and levels of uniqueness. Computational challenges arise from the fact that databases are not always maintained in the same coordinate system during palmprint operations [
26]. This affects minutiae-matching algorithms, which become less effective for palmprints due to their higher density.
All biometric systems face challenges in accuracy, scalability, and usability, and improving accuracy relies on strategies such as the use of multimodal biometric systems [
27]. In the realm of contactless palmprint recognition, challenges include the degradation of matching accuracy and automatic image detection. Advanced approaches and fixed backgrounds are needed to address these issues.
7. Feature Extraction Approaches
The majority of biometric recognition systems using palmprint images extract distinctive features and then compare these features to enrolled models archived in a database. We proposed a categorization of palmprint recognition approaches based on both the type of data employed and the specific strategy utilized for extracting pertinent features. This categorization divides palmprint recognition methods into five overarching classes: line-based, deep learning-based, subspace learning-based, local direction encoding-based, and texture descriptor-based methods (as shown in
Figure 9).
7.1. Line-Based Approaches
Line-based methods focus on the identification and extraction of the main and local lines embedded in a palmprint in an image. These distinctive lines, which include both the prominent main ridges and the finer local ridges, serve as key features to facilitate accurate and efficient recognition processes. By strategically detecting and analyzing these lines, line-based techniques exploit the inherent uniqueness of the palm ridge pattern. The sophisticated ability of these methods to capture the intricate interplay between the major and minor lines enables the creation of highly informative palmprint templates, providing the basis for reliable and discriminatory recognition systems. This meticulous line-oriented approach adds an extra dimension of precision to palmprint recognition, making line-based methods an invaluable tool in the arsenal of biometric security mechanisms.
Li et al. (2002) [
43] proposed a novel approach for palmprint identification by exploiting the power of the Fourier transform to extract and represent spatial frequency features from palmprint images. Prior to feature extraction, the palmprint images are aligned and normalized. Then, the Fourier transform acts as a bridge, seamlessly transforming the palmprint image from the spatial domain, characterized by pixel intensities, into the frequency domain and detecting major lines in the contours. Finally, the retrieved features are used to guide a multi-level search in the database for the best match to the template.
Jia et al. (2008) [
44] proposed a multi-feature-based technique for palmprint recognition that combines primary line (PL) and locality preserving projection (LPP) features. The technique involves extracting the main lines from the query image and comparing them with the main lines present in each image within the training set. Next, it creates a smaller training set consisting of the images with the highest similarity scores. Finally, the technique fuses the similarity scores of the main lines and the LPP features at the decision level and recognizes the query image in the smaller training set.
Jia et al. (2013) [
45] introduced a novel approach for palmprint identification named histogram of oriented lines (HOL). This technique draws inspiration from the widely used histogram of oriented gradients (HOG) technique. Unlike HOG, which primarily captures edge information, HOL delves deeper, specifically targeting and characterizing the prominent lines that define a palmprint’s unique identity through the use of a series of Gabor filters with varying orientations or modified finite random transform (MFRAT). In the matching phase, they use the Euclidean distance as the similarity measure.
Luo et al. (2016) [
46] unveiled a local line directional pattern (LLDP) descriptor based on line direction space for palmprint identification. To capture the line direction features, they used methods such as modified finite radon transform (MFRAT) and the real component of the Gabor filter. For the categorization process, they used Manhattan and Chi-square distances.
Mokni et al. (2017) [
47] proposed an intra-model palmprint recognition method that combines the characteristics of the major line and texture features. First, an elastic shape analysis framework was employed to explore the shape characteristics of the main line. Then, the texture information was explored using fractal analysis. Then, to improve the system accuracy, the important information from the various collected features of the main line shape and the texture pattern were merged. Finally, a random forest classifier was used to identify palmprints after combining the shape analysis-based and fractal-based features.
Gumaei et al. (2018) [
48] proposed the HOG-SGF technique for palmprint identification, which combines histogram of oriented gradients (HOG) features with a steerable Gaussian filter (SGF). First, all palmprint images are preprocessed to segment only the necessary ROI. Then, the palmprint features are extracted using HOG-SGF. In the next phase, the dimensionality of the palmprint features is reduced using an efficient auto-encoder (AE). Finally, the regularized extreme learning machine (RELM) classifier is used for palm identification.
Zhou et al. (2019) [
49] developed a palmprint feature extraction network founded on the double biologically inspired transform (DBIT), aiming to elucidate the mechanisms through which the optical human system perceives palmprints. This network comprises two phases, each applying dual convolutional layers succeeded by sum pooling, rectified linear unit (ReLU) activation, and normalization and combination operations. The first stage activates orientation-selective filters to elicit line and edge responses. The subsequent layer drives rotation-, scale-, and translation-invariant feature maps. Additionally, Pearson correlation and weighted fusion techniques are combined to assess the features’ discriminability and provide palmprint matching.
In order to synthesize the reviewed studies in this subsection,
Table 3 offers a summary overview, outlining the various palmprint feature extraction methodologies spanning the utilized techniques, leveraged datasets, implemented experimental protocols, and key findings.
7.2. Subspace Learning-Based Approaches
Subspace learning-based methods work by extracting and assimilating key features from a palmprint image through the acquisition of a latent subspace guided by a variety of constraints. These methods go beyond traditional feature extraction techniques by dynamically learning and encapsulating the most salient aspects of palmprint patterns into a lower dimensional subspace. By incorporating constraints that include structural, statistical, and contextual information, these techniques meticulously fine-tune their subspace representations to capture the subtle intricacies of individual palmprint variations. This process effectively distills the complexity of palmprint data into a more compact and discriminative form, laying the groundwork for increased recognition accuracy. Subspace learning-based methods emerge as powerful tools to decipher the underlying palmprint data structure, allowing for the creation of highly informative templates that are adept at detecting minute differences while also accounting for broader pattern trends.
Wu et al. (2003) [
50] proposed the Fisherpalm, a palmprint recognition system founded on fisher’s linear discriminant (FLD) analysis. Within this approach, every palmprint is treated as a point within a high-dimensional image space. Palmprints are then mapped from this high-dimensional space to a much lower-dimensional feature space. This transformation enhances the system’s ability to effectively discriminate between the palmprints of different individuals.
Connie et al. (2005) [
2] introduced an automated system for palmprint recognition utilizing a peg-free scanner. Specifically, to deal with projection issues following the ROI extraction, principal component analysis (PCA), independent component analysis (ICA), and fisher discriminant analysis (FDA) were investigated for dimensionality reduction. Additionally, wavelet transform provided complementary multi-resolution texture characterization. The authors concluded that the configuration involving wavelet + FDA, denoted as WFDA, outperforms other configurations in terms of performance.
Hu et al. (2007) [
51] proposed a technique known as 2D locality preserving projection (2DLPP) for palmprint identification. This method focuses on feature extraction and is based on the concept of locality preservation and image matrix projection. To perform the classification, they used a nearest neighbor classifier considering the L2 norm for measuring similarity.
Pan and Ruan (2008) [
52] proposed an approach known as two-dimensional locality preserving projections (I2DLPP) for palmprint identification. This technique focuses on simplifying computational complexity and reducing feature dimensions by employing two main steps: First, it performs a projection of the training space along the row direction using a two-dimensional principal component analysis (2DPCA). Thereafter, I2DLPP is conducted over the resultant compressed columnar direction by employing a nearest neighbor graph. Additionally, the authors propose applying I2DLPP on Gabor-filtered images (I2DLPPG) to further enrich textural characterization identification. Extensive analysis revealed that the input filtered images significantly improved computational efficiency and identification accuracy.
Lu and Tan (2011) [
53] proposed an approach called diagonal discriminant locality preserving projections (Dia-DLPP) for the identification of both faces and palmprints. This approach is crafted to capture in both directions—vertical and horizontal—by cueing discriminant information from data. Uniquely, diagonalized images were integrated during the training and testing phases to enhance the discriminative capabilities. The authors additionally proposed a weighted discriminative variant (W2D-DLPP) that explicitly assigns greater significance to more identity-discriminative pixel clusters when computing projection vectors. The discriminative scores are incorporated into the traditional 2D-DLPP technique, resulting in the refined method W2D-DLPP. This integration of discriminative pixel weighting significantly improves the identification performance of both 2D-DLPP and Dia-DLPP, leading to improved accuracy in face and palmprint identification tasks.
Rida et al. (2018) [
54] introduced a palmprint recognition system that relies on a set of sparse representations (SR). They used two-dimensional principal component analysis (2D-PCA) to build an initial sample dictionary and then used two-dimensional linear discriminant analysis (2D-LDA) to extract discriminative features.
Rida et al. [
55] devised an ensemble framework leveraging the random subspace method (RSM) for contactless palmprint recognition classification. Two-dimensional principal component analysis (2DPCA) was applied to obtain multiple dimensional eigenvector random subspaces. Thereafter, two-dimensional linear discriminant analysis (2DLDA) was conducted within each random 2DPCA projection to retrieve the most discriminative feature subsets. In addition, Euclidean distances with nearest neighbor classifiers were subsequently implemented on each subspace. Ultimately, a nonlinear decision function was constructed, comprising individual classifiers that vote by majority.
Wan et al. (2021) [
56] introduced a feature extraction technique for palmprint recognition termed sparse 2D discriminant local preserving projection (SF2DDLPP) that integrates elasticity into the dimensionality reduction process. Their method first constructs a fuzzy membership matrix using the fuzzy k-nearest neighbors algorithm (FKNN), computed separately for within-class and between-class weight matrices to encode intra-personal and inter-personal variations. Two theorems are subsequently derived to efficiently obtain the generalized eigenfunctions for optimized class separation. Finally, elastic net regularization is utilized to determine the optimal sparse projection matrix.
Zhao et al. (2022) [
57] proposed an innovative approach known as double-cohesion learning-based multi-view and discriminant palmprint recognition (DC_MDPR). This approach effectively exploits the multi-view features of palmprint images along with the inherent data structure. To achieve this goal, they introduced a method called double cohesion, which combines inter-view cohesion and intra-class cohesion. This technique aims to enhance the distinctiveness of multiple features, reduce feature dimensions, and enhance the representation of these features within the same subspace.
Wan et al. (2023) [
58] proposed an approach known as low-rank two-dimensional local discriminant graph embedding (LR-2DLDGE) for the purpose of feature extraction and dimensionality reduction. Initially, the technique uses a graph embedding (GE) framework to capture and preserve essential discriminative information within local neighborhoods of the data. Next, LR-2DLDGE is designed to ensure that data points within the feature space are maximally independent across different classes, thereby enhancing discriminative capabilities. To bolster the method’s resilience against noise and corruption, the approach incorporates an L1 norm constraint and employs low-rank learning techniques.
Table 4 provides a summary of the studies discussed in this subsection, elucidating the employed feature extraction methods, datasets, experimental protocols, and principal findings.
7.3. Local Direction Encoding-Based Approaches
Methods based on local orientation coding focus on extracting and encoding the prevailing orientation information inherent in each pixel of a palmprint image. These methods differ from conventional techniques by focusing on the underlying directional characteristics of the ridge patterns, capturing the nuanced variations in orientation that contribute to the uniqueness of each palmprint. By extracting dominant directions at a local level, these methods reveal the intricate minutiae of the palm surface, going beyond traditional ridge-based representations. This wealth of directional information is then distilled into compact and informative bitwise codes, creating a powerful and discriminative coding scheme. These techniques effectively balance the precision of directional cues with the efficiency of compact coding, resulting in improved recognition performance. In the landscape of biometric authentication, these methods carve out a distinctive niche, offering a fusion of geometric understanding and efficient data representation that contributes to the development of accurate and reliable identity verification systems.
Kumar and Shen (2004) [
59] proposed a palmprint recognition method based on real Gabor function (RGF) filtering. The process starts by normalizing palmprint images. Subsequently, these normalized images are subjected to multi-channel filtering using a set of RGF filters. Distinctive features, referred to as PalmCode, are computed within multiple overlapping concentric bands using each of these filtered images.
Kong et al. (2006) [
60] proposed a feature-level coding method for palmprint recognition. First, palmprint image features are extracted utilizing a bank of elliptic Gabor filters. Then, a feature-level fusion technique is offered to create a single feature known as the fusion code. The normalized Hamming distance between two fusion codes is then used to determine their similarity. Finally, a dynamic threshold is applied for final judgments.
Mansoor et al. [
61] proposed a multi-scale feature encoding strategy fusing contourlet transforms (CTs) and non-subsampled contourlet transforms (NSCTs) for palmprint recognition. This method aims to jointly capture localized texture details alongside global features within palmprint imagery, representing them as a compact and fixed-length palm code. The iterated directional filter banks are introduced to divide the two-dimensional spectrum into small slices. The feature vector is then formed by computing the block-wise directional energy in the transform domains. Finally, for matching, normalized Euclidean distances between vectorial codes quantify palmprint identity similarity.
Zhang et al. (2012) [
62] proposed a novel approach for palmprint identification based on local direction encoding. Specifically, the authors augment native BOCV descriptors with additional “fragile bits” constituting noise-sensitive activations to derive extended BOCV (E-BOCV). A consolidated similarity metric was obtained by synergizing fragile pattern distance (FPD) with Hamming distances to capture the mismatches between two code maps.
Zhang and Gu (2013) [
63] suggested employing a weighted fusion scheme integrating two-phase test sample sparse representation (TPTSR) with competitive coding methods for palmprint identification. First, a competitive coding algorithm is employed to obtain the directional matching score of two images. Thereafter, TPTSR is computed globally to match the score of the two images. Finally, the two scores are added together to categorize the test sample.
Li et al. (2014) [
64] presented directed representations (DRs) for palmprint identification. First, a representation is proposed for an appearance-based technique based on multiple anisotropic filters. Subsequently, the feature extraction and the dimension reduction are guaranteed using the PCA technique. Finally, a compressed sensing classification step is implemented to distinguish between palms of different hands.
Fei et al. (2016) [
65] presented a robust palmprint recognition method, the double-orientation code (DOC) approach. This technique offers a reliable way to represent palmprint orientation features, as shown through an investigation of palmprint orientation-based coding theory. They also introduced a novel nonlinear angular matching score metric for efficient similarity assessment between DOC-encoded palmprints, boosting the overall effectiveness of the technique in palmprint identification.
Xu et al. (2016) [
66] introduced a palmprint identification technique called discriminative and robust competitive code (DRCC), which emphasizes discriminative and robust techniques based on dominant orientation. Their approach combines dominant orientation and lateral codes to capture important orientation features in palmprints. Strategic weighting during orientation extraction improves accuracy while using the same Gabor filters as the conventional method. This innovation holds promise for accurate and efficient palmprint orientation extraction.
Almaghtuf et al. (2020) [
67] proposed a palmprint coding technique known as difference of block means (DBM). To derive the palmprint code, they followed the recommended approach: First, they computed the difference between overlapping block means of identical size within the interest area of the palmprint to extract palm-related information in both the vertical and horizontal directions. Then, vertical and horizontal codes were generated by applying thresholding to the DBM features. Finally, the Hamming distance, which is the average of the vertical and horizontal distances, was used for the matching step.
Liang et al. (2020) [
68] developed a multi-feature palmprint recognition framework predicated on modeling orientation field patterns termed histograms of line mixed distances (HODlm) alongside histograms of response distances (HODr). Subsequently, the multi-feature two-phase sparse representation (MTPSR) was designed to improve the overall matching cost and to allow the handling of palmprint feature recognition.
Table 5 furnishes an overview of the studies discussed in this subsection, outlining the employed feature extraction methodologies, datasets, experimental protocols, and key findings.
7.4. Texture-Based Approaches
Texture-based methods take advantage of the intricate and diverse local features present in palmprint patterns. By exploiting these rich textural features, these methods aim to achieve higher accuracy and reliability in the identification process. Unlike traditional methods that rely solely on global features, texture-based techniques delve into the fine-grained details of the palm surface, capturing an array of minutiae such as ridges, wrinkles, and pores. This meticulous analysis enables the creation of comprehensive and distinctive palmprint templates that facilitate robust and discriminative recognition. As a result, these methods are the cornerstone of biometric authentication systems, where the intricate texture patterns of an individual’s palm provide a wealth of information for secure and accurate identity verification.
Hammami et al. (2014) [
69] employed a technique involving the division of the complete palmprint image into smaller sub-regions. Within each of these sub-regions, they applied the local binary pattern (LBP) operator to capture the texture characteristics. To enhance recognition efficiency and minimize memory consumption, they introduced a selection process, which retained only the most distinctive areas for the identification task. The sequential forward floating selection (SFFS) algorithm was the basic method employed for this purpose.
Raghavendra and Busch (2015) [
70] introduced an innovative and straightforward method for palmprint inspection, exploiting the distributed feature representation extracted from the bank of binarized statistical image features (B-BSIF). The BSIF specifically functions as a texture descriptor akin to the LBP, but its distinctiveness lies in its approach to acquiring filters. Unlike the LBP, which manually defines filters, BSIF filters are learned from real images.
Tamrakar and Khanna (2016) [
71] introduced a palmprint recognition approach. Initially, the region of interest (ROI) is obtained from palmprint images. Subsequently, to mitigate computational costs and noise, first-level decomposition is performed on the ROI by employing the Haar wavelet. Combining this image with its block-wise histograms, which statistically summarize local variations, yields a comprehensive descriptor called the block-wise Gaussian derivative phase pattern histogram (BGDPPH). Having extracted robust features, kernel discriminant analysis (KDA) is applied to refine their discriminative power. Finally, the Euclidean distance is used for classification.
Doghmane et al. (2018) [
72] proposed a local palmprint feature descriptor based on the Gabor wavelet, a local phase quantization (LPQ) descriptor and a spatial pyramid histogram (SPH) descriptor for palmprint image extraction. First, the Gabor wavelet and the LPQ are used to recover invariant blur for multiscale and multi-orientation features. The SPH is then employed for vertical decomposition to concatenate a group of local features into a large histogram called the Gabor LPQ special pyramid histogram (GLSPH) feature for each image. The GLSPH features are then projected into a whitened linear discriminant analysis (WLDA) subspace to reduce their dimensions and make them further discriminative, resulting in the DGLSPH feature. Ultimately, to deal with classification, the K-nearest neighbor (K-NN) classifier is employed in this context.
Zhang et al. (2018) [
73] introduced an innovative approach to palmprint recognition that involves a two-stage process using a combination of weighted adaptive center symmetric local binary pattern (WACS-LBP) and weighted sparse representation-based classification (WSRC). Their methodology first implements WACS-LBP in an initial labeling stage to assign the test sample a limited set of feasible class labels. WSRC is then employed in the ensuing identification phase to determine the final class membership from this reduced label set. Core to their approach is the strategic conversion of the intrinsically complex complete classification problem into a more tractable task through substantial reduction of the number of output classes under consideration at each phase.
El-Tahrouni et al. (2019) [
74] proposed a multispectral palmprint identification method that incorporates Pascal coefficient multispectral local binary pattern (PCMLBP) and pyramid histogram orientation gradient (PHOG) descriptors. They performed two experimental procedures. In the first procedure, only the PCMLBP descriptor was used for feature extraction. In the second procedure, PCMLBP was combined with PHOG, resulting in an improved recognition rate. PCA was then used to reduce the dimensionality of the feature vectors. Finally, random sample linear discriminant analysis (LDA) was utilized for classification.
Attallah et al. (2019) [
75] introduced a palmprint identification approach that involves merging spiral features with LBP filters and selecting the optimal features using minimum redundancy maximum relevance (mRMR). This process starts by partitioning the palmprint image into smaller blocks, akin to meticulously examining individual pieces of a mosaic. Within each block, the researchers delve deeper, analyzing two crucial statistical descriptors: skewness and kurtosis. The Hamming distance is then applied to compute both inter-similarities between different blocks within the same palmprint and intra-similarities between corresponding blocks across different palmprints.
Chaudhary and Srivastava (2020) [
76] proposed an approach for feature extraction in palmprint identification known as two-dimensional cochlear transform (2DCT). This method was designed to efficiently capture distinctive palmprint features. To validate the efficacy of the method, the authors performed comprehensive analyses, including both theoretical and empirical assessments. The theoretical evaluation involved demonstrating the orthogonality properties of the transform, whereas the empirical evaluation considered its performance under various challenging conditions. For the classification step, they adopted the k-nearest neighbors (KNN) algorithm, using the Euclidean distance metric for similarity assessment.
Zhang et al. (2020) [
77] proposed a contactless palmprint identification and recognition technique integrating hierarchical multi-scale complete local binary patterns (HMS-CLBPs) for scale-invariant texture encoding with weighted sparse representation-based classification (WSRC) for pattern matching. Local texture descriptors are first extracted over multi-scale domains to capture fine and coarse palmprint texture. The descriptors are weighted, vectorized, and augmented to construct a highly overcomplete dictionary of training sets. Following the creation of the dictionary, it is utilized to find a sparse representation of the test sample, yielding the sparse coefficients corresponding to each dictionary atom. Palmprint image recognition is then performed by computing the reconstruction residuals between the test sample and its synthesized approximation under each class-specific sub-dictionary.
Amrouni et al. (2022) [
78] introduced a feature extraction approach called multiresolution analysis. First, they applied the discrete wavelet transform (DWT) to an original palmprint image. This application facilitated the creation of multiple image representations, known as sub-bands, each with different resolutions. In addition, they utilized the local texture descriptor known as binarized statistical image features (BSIF) and applied it not only to the original image but also to the sub-bands produced by the DWT at lower resolutions. The resulting histograms from each of these levels were then merged to form a final feature vector.
Table 6 provides a summary of the studies discussed in this subsection, presenting the feature extraction methods used, the datasets employed, the experimental protocols, and the key findings.
7.5. Deep Learning-Based Approaches
In this category, the methods frequently employed convolutional neural networks (CNNs). These networks comprise convolutional layers, pooling layers, and fully connected layers, allowing for simultaneous feature extraction and classification [
79,
80].
Izadpanahkakhk et al. (2018) [
81] proposed a system consisting of three main modules. (i) The Region of Interest Extraction Module (REM) is responsible for extracting palmprint ROIs using a bounding box approach. First, the input images are subjected to a preprocessing step. Then, a transfer learning technique using CNNs is applied to identify the optimal placement of bounding boxes on the palmprint images. This module extracts regions of interest (ROIs) from the palmprint, which serve as input for the subsequent feature extraction module. (ii) In the Feature extraction module (FEM), features are represented using a pre-trained CNN architecture. It applies the learned representations to extract discriminative features from the palmprint ROIs. (iii) The matching module (MM) takes the feature vector generated in the previous step as input and uses a machine learning classifier to perform the recognition task.
Matkowski et al. (2019) [
18] developed an end-to-end deep learning approach named the end-to-end palmprint recognition network (EE-PRnet). This network comprises two fundamental components: ROI localization and alignment network (ROI-LAnet) and feature extraction and recognition network (FERnet). ROI-LAnet is tasked with transforming all input palmprint images into a consistent coordinate system and delineating the ROI containing distinguishing structural information. ROI-LAnet comprises two segments: the first is a pre-trained VGG-16 network with its top layers removed, and the second is a fully connected regression network. FERnet is tasked with extracting and recognizing palmprint features. This network is a self-contained CNN based on a modified VGG-16 architecture.
Chai et al. (2019) [
30] proposed the use of two separate comprehensive CNN systems, named PalmNet and GenderNet. These networks were thoroughly trained to excel in two tasks: palmprint recognition and gender categorization, respectively. Their groundbreaking study not only demonstrated unprecedented performance in biometric recognition but also needed to validate the notion that integrating gender information could improve the accuracy of palmprint identification. To further test this concept, they created two Boost CNN networks, BoostNet-Sequential and BoostNet-Parallel, with the goal of combining the strengths of palmprint recognition and gender categorization.
Genovese et al. (2019) [
82] presented PalmNet, an innovative CNN architecture. In particular, their unsupervised training eliminates the need for class labels. They also proposed a novel Gabor-based technique that uses PCA to refine adaptive filters within the CNN, thereby improving its specificity for palmprint recognition.
Zhao and Zhang (2020) [
83] presented a novel approach to improve multi-scenario palmprint recognition using a versatile framework. It uses deep convolutional networks (DC-NNs), termed deep discriminative representation (DDR), to learn effective features. These features work well for palmprint recognition under various conditions. The key innovation is to train DC-NNs to extract discriminative features from palmprints that include global abstract and local compact attributes. This framework remains effective even with limited training data, providing a new avenue for advancing palmprint recognition in diverse scenarios.
Liu and Kumar (2020) [
24] presented a robust and versatile deep learning framework for contactless palmprint identification. Their approach uses a fully convolutional network to generate complex residual features (RFNs), which imsproves accuracy and generalizability. A distinctive feature is the utilization of a soft-shifted triplet loss function, which enhances the learning of discriminating the features of palmprints. Additionally, they incorporated a contactless palm detector, customized and trained utilizing the prompter CNN model, for effective detection of palmprint regions across diverse backgrounds.
Liu et al. (2021) [
84] introduced an end-to-end deep hashing network tailored for few-shot contactless palmprint identification, termed the similarity metric hashing network (SMHNet). Their framework integrates a structural similarity index (SSIM) module to elicit multi-scale representations encoding both holistic topology and localized texture details. A composite SSIM loss function alongside distance metrics supervises the training process for enhanced inter-class separability. Additionally, a hashing unit learns binary compact codes optimized for efficient storage and fast retrieval demonstrated significant improvement in few-shot recognition of palmprints.
Shen et al. (2022) [
41] introduced a progressive target distribution loss (PTD Loss) function, which is tailored to minimize the gap between positive cross-device sample affinities and negative within-device sample relations. Additionally, the authors established a new cross-device palmprint identification dataset compositing color images sourced from multiple capture platforms.
Shao and Zhong (2022) [
85] developed a deep metric learning paradigm designed for open-set contactless palmprint identification termed weight-based meta-metric learning (W2ML). Their framework strategically partitions the dataset into training and testing subsets without overlap between the two phases. The training set is further divided into multiple tasks, each comprising a support set for representation learning and a query set for few-shot generalization assessment akin to meta-learning by aggregating support sets from the task-specific subspaces into consolidated positive and negative meta-sets. The model is then trained using set-based distances between them. Additionally, hard prototype mining and weighting further enhances discrimination by identifying and prioritizing the most informative samples within each meta-set (from positive and negative). Extensive experiments demonstrated significant gains over conventional approaches, constituting a vital step towards palmprint identification systems.
Türk et al. (2023) [
86] devised a hybrid palmprint recognition framework fusing deep learning and classical machine learning methodologies. Their processing pipeline starts with multiple preprocessing interventions encompassing boundary delineation, binarization, finger exclusion, edge contour extraction, noise filtering, and image thinning to maximize distinctive friction ridge information. The second step focuses on extracting the ROI for palmprint images. Thereafter, a CNN feature extractor learns hierarchical representations encoding both textural and topological traits. Finally, the CNN embeddings are classified using CNN classifier gathering with support vector machine (SVM) classifiers for gallery identity prototypes for palmprint identification.
Table 7 provides a summary of the studies discussed in this subsection, presenting the feature extraction methods used, the datasets employed, the experimental protocols, and the key findings.
7.6. Comparative Analysis
In this subsection, a comprehensive overview of the strengths and weaknesses of various palmprint identification and recognition methods, including line-based, subspace learning-based, local direction encoding-based, texture descriptor-based, and deep learning-based approaches, was presented. A summary of the methods discussed in the previous subsections is presented in
Table 8. The comparative analysis between the five approaches indicates that many approaches demonstrate satisfactory performance with simple and controlled datasets. However, significant disparities in both performance and computational cost arise when dealing with large-scale and unconstrained datasets due to the challenges posed by diverse environmental conditions.
In summary, there is a compelling opportunity to develop novel and real-time models specifically tailored to unconstrained palmprint recognition. Such advancements are essential to enhance overall performance, achieve a certain level of maturity in the field, and facilitate widespread commercial deployment.
8. Future Directions
Unlike more established biometric modalities, palmprint recognition is relatively nascent, necessitating further scrutiny. Challenges and issues that are well explored in facial and fingerprint recognition demand in-depth investigation in the context of palmprint recognition. This section outlines critical research topics requiring comprehensive exploration in future endeavors. Insights on emerging ideas are provided to guide and inspire forthcoming research initiatives.
8.1. Enhancement of Palmprint Imagery through the Application of Generative Adversarial Networks
The utilization of generative adversarial networks (GANs) [
87,
88,
89,
90] in the context of palmprint recognition, specifically applied to tasks such as inpainting, enhancement, deblurring, recolorizing, and segmentation, represents a highly promising paradigm. The inherent architecture of GANs, comprising a generator and a discriminator, proves instrumental in ameliorating palmprint images by proficiently generating authentic data to address instances of missing or occluded information. This versatile application extends to mitigating challenges associated with occlusions, blurring, diminished visual quality, and noise, thereby augmenting the accuracy of palmprint biometric recognition.
It is noteworthy that the computational demands of GAN training are considerable, necessitating meticulous optimization to facilitate real-time deployment. The multifaceted integration of GANs in the palmprint recognition domain, encompassing diverse tasks, presents a comprehensive strategy poised to be strategically leveraged in the imminent future. This strategic utilization holds the potential to substantially enhance the resilience and precision of palmprint recognition systems.
8.2. Enhancing Recognition Rates and Expediting Processing time through the Exploitation of Soft Biometric Attributes
Soft biometrics denotes the utilization of non-intrusive and readily quantifiable attributes for the purpose of biometric identification. Personal characteristics such as gender, ethnicity, age, scars, marks, and tattoos exemplify instances of soft biometric traits [
91,
92,
93,
94]. Within the realm of palmprint biometrics, the integration of soft biometrics is posited to augment recognition accuracy and diminish processing time, achieved through a judicious reduction in dataset inquiries. Through the incorporation of these methodologies, it is posited that a decrease in processing time and an increase in recognition accuracy can be realized in the domain of palmprint biometrics, thereby enhancing the operational efficiency and practical applicability of the system across diverse contexts.
8.3. Utilizing Three-Dimensional Representations to Mitigate Image Acquisition Challenges
Previous palmprint research has predominantly focused on two-dimensional (2D) images, which are susceptible to environmental factors. Recognition methodologies, categorized into line, texture, subspace, and coding approaches, often compromise accuracy due to the lack of depth information in 2D representations. To address this limitation, three-dimensional (3D) palmprints offer promising biometric identification with qualities of uniqueness, stability, and universality [
95,
96,
97,
98]. However, adopting 3D palmprints introduces challenges in data volume and computational complexity. Sparse point clouds from 3D sensors impact mesh resolution and identification performance. Prolonged pre-processing time and compatibility issues with 3D recognition algorithms further complicate matters. Advanced research is crucial for optimized 3D sensors, addressing time efficiency and data volume challenges in palmprint-based recognition. Enhancing 3D sensor capabilities is imperative for efficient and accurate palmprint recognition.
8.4. Exploring Liveness Detection and Mitigating Vulnerability to Spoofing Attacks
Liveness detection and susceptibility to spoofing attacks are significant concerns in palmprint biometrics. Despite the advantages of palmprint recognition, it is not immune to deceptive practices, similar to other biometric modalities. Spoofing attacks, involving the presentation of fake biometric data such as displayed or printed images, pose challenges to verifying the authenticity of presented data, jeopardizing the privacy and security of palmprint biometrics. Despite the issue’s importance, limited attention has been given to spoofing attacks on palmprint biometrics, as evident from the sparse existing studies [
99,
100,
101,
102].
Addressing this gap is crucial and necessitates research in robust presentation attack detection algorithms tailored to palmprint-based recognition systems. The lack of suitable anti-spoofing databases compounds this challenge, underscoring the urgency to fortify the security of palmprint recognition systems against adversarial activities.
9. Conclusions
This study furnishes an exhaustive survey of the palmprint biometrics literature, encompassing benchmark datasets, challenges and impediments, assessment metrics, and predominant techniques. Specifically, a rigorous analysis and comparison of myriad approaches is conducted across five taxonomic categories of feature extraction methodologies. Furthermore, a systematic classification of the diverse palmprint datasets employed for algorithmic development and testing is provided alongside documented performances of experimental results. Additionally, the study delineates outstanding challenges necessitating immediate attention through further inquiry to advance automated palmprint identification systems. Overall, via comprehensive aggregation and the juxtaposition of factors, this taxonomic synthesis aims to edify discernment of comparative virtues and limitations underlying contemporary methodologies. It is expected that this taxonomic survey will serve as an inspiration for the research community and emerging scholars and will encourage further advances in palmprint recognition.