Next Article in Journal
Direct Measurement of Lateral Force Using Dual Cantilevers
Next Article in Special Issue
Improving Fingerprint Verification Using Minutiae Triplets
Previous Article in Journal
Monocular Camera/IMU/GNSS Integration for Ground Vehicle Navigation in Challenging GNSS Environments
Previous Article in Special Issue
Finger Vein Recognition Based on a Personalized Best Bit Map
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Two-Level Evaluation on Sensor Interoperability of Features in Fingerprint Image Segmentation

School of Computer Science and Technology, Shandong University, Jinan 250101, Shandong, China
*
Author to whom correspondence should be addressed.
Sensors 2012, 12(3), 3186-3199; https://doi.org/10.3390/s120303186
Submission received: 9 February 2012 / Revised: 22 February 2012 / Accepted: 22 February 2012 / Published: 7 March 2012
(This article belongs to the Special Issue Hand-Based Biometrics Sensors and Systems)

Abstract

: Features used in fingerprint segmentation significantly affect the segmentation performance. Various features exhibit different discriminating abilities on fingerprint images derived from different sensors. One feature which has better discriminating ability on images derived from a certain sensor may not adapt to segment images derived from other sensors. This degrades the segmentation performance. This paper empirically analyzes the sensor interoperability problem of segmentation feature, which refers to the feature’s ability to adapt to the raw fingerprints captured by different sensors. To address this issue, this paper presents a two-level feature evaluation method, including the first level feature evaluation based on segmentation error rate and the second level feature evaluation based on decision tree. The proposed method is performed on a number of fingerprint databases which are obtained from various sensors. Experimental results show that the proposed method can effectively evaluate the sensor interoperability of features, and the features with good evaluation results acquire better segmentation accuracies of images originating from different sensors.

Graphical Abstract

1. Introduction

Fingerprint segmentation is an important pre-processing step in automatic fingerprint recognition system [1]. A fingerprint image usually consists of two regions: the foreground and the background. The foreground which contains effective ridge information is originated from the contact of a fingertip with the sensor. The noisy area at the borders of the image is called the background. Fingerprint segmentation aims to separate the fingerprint foreground area from the background area. Accurate segmentation is especially important for the reliable extraction of minutiae, and also reduces significantly the time of subsequent processing.

Various fingerprint segmentation methods have been proposed by previous researchers, which can be roughly divided into two types: block-wise methods [28] and pixel-wise methods [912]. Block-wise methods classify the image blocks into foreground and background based on the extracted block-wise features, and pixel-wise methods classify pixels through the analysis of pixel-wise features. According to whether the label information is used, the fingerprint segmentation methods can also be treated as unsupervised [58], supervised [2,3,9,11] and semi-supervised ones [13,14].

Fingerprint images collected by different sensors usually have different characteristics, quality and resolution. However, most fingerprint recognition systems are designed for fingerprints derived from a certain sensor, and when dealing with fingerprints derived from other sensors, the performance of the recognition systems may be significantly affected. Therefore, fingerprint recognition systems encounter a sensor interoperability problem. Sensor interoperability is defined as “the ability of a biometric system to adapt to the raw data obtained from a variety of sensors” [15].

Fingerprint segmentation, a crucial processing step of the fingerprint recognition system, inevitably encounters the sensor interoperability problem. There are mainly two reasons for this [8]. On the one hand, a feature obtained from different sensors may be confused, which results in a block or a pixel being regarded as different categories under views of different sensors. On the other hand, segmentation models trained on one database collected by a certain sensor need to be retrained when dealing with images derived from other sensors.

Much attention has been paid to the sensor interoperability problem of fingerprint segmentation. The works [8,13,16] usually follow two directions: (1) extracting features with interoperability and (2) designing segmentation methods with interoperability. In [16], Ren investigated the feature selection for sensor interoperability and took fingerprint segmentation as a case study. Studies show that features exhibit different sensor interoperability in images derived from various sensors. Variance is found to be an interoperable feature in fingerprint segmentation. In [8], we empirically analyzed the sensor interoperability problem in fingerprint segmentation, and proposed a k-means based segmentation method to address the issue. In [13], Guo proposed a personalized fingerprint segmentation method which learns a special segmentation model for each input fingerprint image and gets over the differences originated from various sensors.

In this paper, we first empirically analyze the sensor interoperability problem of features in fingerprint segmentation. Then a two-level method is proposed to evaluate the sensor interoperability of commonly used features. This method preliminarily evaluates the feature(s) through the first level evaluation based on segmentation error rate, the feature(s) whose segmentation error rate is high and not stable will be eliminated; the remaining candidate feature(s) will participate in a second level evaluation which is based on a decision tree, and the feature or feature set with good sensor interoperability will be selected according to information theory. The effectiveness of the proposed method is validated by experiments performed on a number of fingerprint databases derived from various sensors.

The paper is organized as follows: Section 2 analyzes the sensor inoperability problem of features in fingerprint segmentation through empirical studies. Section 3 proposes our two-level feature evaluation method. Section 4 reports the experimental results. Finally, Section 5 draws conclusions and discusses future work.

2. Sensor Interoperability Problem of Feature in Fingerprint Segmentation

In fingerprint segmentation, features are an important topic and discriminating features usually leads to favorable segmentation performance. There is abundant research on segmentation features, which mainly focuses on defining the discriminating features. The commonly used features in fingerprint segmentation include gray-level features [2,3,9,17,18] (such as gray mean, gray variance, contrast, etc.), texture features [3,5,7,9,19,20] (such as gradient, coherence, Gabor response, etc.), and other features [12,2123] (such as Harris corner point features, polarimetric feature, number of invalidated minutiae, etc.).

Due to the differences between sensing technologies, fingerprint images derived from different sensors usually have different characteristics, resolution, quality and so forth. Therefore, various features have different discriminating abilities on images derived from different sensors. In order to investigate the influence of various sensors on the segmentation feature, we randomly selected fingerprints from a number of open databases and analyzed the feature histograms.

The fingerprints were collected from three open fingerprint databases: FVC2000 [24], FVC2002 [25], and FVC2004 [26]. Each open database contains four sub-databases, where the first three sub-databases are derived from three different types of sensors, and the last sub-database is generated synthetically. The sensors used in the open databases are presented in Table 1. Each database consists of a training set of 80 images and a test set of 800 images.

We randomly select 10 fingerprint images from each real sub-database to construct a database containing 90 images. Each of the 90 images is partitioned into non-overlapping blocks of 8 × 8 pixels, and then all the blocks are manually labeled as the foreground class and the background class. For each block, we extract four features of mean [9], variance [9], contrast [17], and gradient [7]. In the following contents, we will compare four features’ histograms of fingerprints from same sub-database and different sub-databases, respectively.

We first investigate the histograms of fingerprints from same sub-database. A sample result is given in Figure 1 which shows the histograms of mean, variance, contrast, gradient of 10 fingerprints in FVC2000 DB2. We can see that under the view of mean (variance, contrast or gradient), the foreground and background blocks are statistically separable. Actually, the foreground and background blocks of fingerprints from same sub-database can be separated by most segmentation features, and thus for most segmentation features, they have good discriminating abilities and can achieve favorable segmentation performance in images derived from the same sensor.

Then we compare the histograms of fingerprints from same sensor with that of 30 fingerprints from three different sensors. For example, Figure 2 shows the histograms of fingerprints derived from three sub-databases of FVC2000. Compared with Figure 1, for each feature, we can see that the overlapped area of red line and blue line increases, which shows that the foreground and background blocks are not easy to separate. Therefore, it is not difficult to understand that the discriminating abilities of features on images captured by different sensors reduce and the segmentation performances of these images are not satisfactory. The segmentation features suffer from sensor interoperability problem.

Figure 3 provides the features’ histograms of all the 90 fingerprints. In comparison with Figure 2, the overlaps of every feature become more and more complex as the number of different sensors increases. For segmentation features, the more different sensors fingerprints originating from, the more difficult to separate the foreground blocks from the background blocks. This is primarily caused by the sensor interoperability problem.

As the number of different sensors increases, the discriminating abilities of features reduce more or less; however, the degrees of reduction are different for various features. From above three examples, we can see the overlaps of Mean feature become significantly larger as the number of different sensors increases. Therefore, when dealing with fingerprints collected by different sensors, mean feature would not well separates the foreground from background, which results in bad segmentation performance. On the contrary, the change of the gradient histogram is not as obvious. From the gradient histogram in Figure 3, we can see the foreground and background blocks are still statistically separable when fingerprints originated from nine different sensors. Besides, the discriminating abilities of contrast and variance are moderate for fingerprints derived from different sensors.

Feature segmentation, which has an effect on the segmentation performance, is a crucial factor in fingerprint segmentation. Different features exhibit different discriminating abilities on fingerprint images derived from various sensors. Some features can separate well the foreground from the background in images collected by same sensor. However, when dealing with fingerprints collected by different sensors, the discriminating abilities degrade, which results in bad segmentation performance. As mentioned above, a lot of features have been applied to fingerprint segmentation. Which features have good sensor interoperability and adapt to images from various sensors? How to evaluate the sensor interoperability of features and select feature(s) with good sensor interoperability? Can we achieve high segmentation accuracy with selected feature(s) when dealing with fingerprints derived from different sensors? With the above analysis, an effective and robust feature evaluation method should be found to address the sensor interoperability problem of feature in fingerprint segmentation.

3. Two-Level Feature Evaluation Method

In order to evaluate the sensor interoperability of features in fingerprint segmentation, we propose a two-level evaluation method, which is composed of two steps. Firstly, all the candidate features participate in the first level evaluation based on segmentation error rate, and the features whose average segmentation error rate is high and not stable will be eliminated; then the remaining candidate features will participate in the second level evaluation which is based on a decision tree, the features will be evaluated according to information gain ratio. Therefore, the two-level evaluation method selects the feature or feature set with good sensor interoperability.

3.1. Feature Evaluation Based on Segmentation Error Rate

In the first level evaluation, the mean value and variance of the segmentation error rate are used to evaluate features. The segmentation error rate as a criterion for evaluating a feature is given in [2]. The error rate Err is defined as follows:

Err = N err N total = p ( w 0 | w 1 ) + p ( w 1 | w 0 )
where Nerr denotes the number of misclassified blocks, Ntotal is the total number of blocks, w0 represents background class, while w1 represents foreground class.

In this step, we make use of a k-means based segmentation method [8] to acquire the segmentation error rate. Since the k-means based segmentation method avoids adjusting thresholds or retraining classifiers for fingerprints derived from various sensors, it ensures sensor interoperability and small time consumption. Firstly, fingerprint images are divided into non-overlapping blocks of the same size, where for each block all the candidate features are extracted. Then for each candidate feature, the k-means based method is performed on each image to get the segmentation result. In the end, the segmentation error rate is acquired by comparing the segmentation result with the manually segmented image.

The mean value of the error rates in images derived from different sensors is a measure to the segmentation performance of every candidate feature. The variance of the error rates reflects the stability of the segmentation performance. The features which have low mean value and variance of segmentation error rates can better adapt to different sensors. Therefore, the mean value and variance of error rate are used to evaluate features finally in the first level evaluation.

The framework of the first level feature evaluation is presented in Figure 4. The detail steps of first level evaluation are described as follows:

  • Step 1: Suppose there are different sensors Si (i = 1 to n). We randomly select p fingerprint images derived from each sensor. Each fingerprint image is partitioned into non-overlapping blocks of same size, and for each block, q candidate features are extracted. Fj (j = 1 to q) represent the candidate feature. Then the blocks are manually labeled as two classes: foreground blocks and background blocks.

  • Step 2: Take the candidate feature F1 as example. All the blocks are represented by feature F1. The center block of the fingerprint image is initialed as the cluster center of the foreground class, and one border block is regarded as the cluster center of the background one. The k-means algorithm is performed on each image to cluster the blocks into the foreground cluster and the background cluster. The segmentation error rate on each image is acquired by comparing clustering results with manually segmented image.

  • Step 3: For each candidate feature, the segmentation error rate on one sensor is represented by the average error rate on p images derived from this sensor. After that, the mean value and variance of segmentation error rate on all sensors are computed. Again, we still take the candidate feature F1 as example. The segmentation error rate in images derived from sensor Si is named as F1Ei (i = 1 to n). The mean value F1M and variance F1V of segmentation error rates are defined as Equations (2) and (3):

    F 1 M = 1 n i = 1 n F 1 E i
    F 1 V = 1 n i = 1 n ( F 1 E i F 1 M ) 2

  • Step 4: For all the candidate features, the mean value and variance of the segmentation error rates are acquired by repeating Step 2 and Step 3. We name the mean value and variance of candidate feature Fj as FjM and FjV (j = 1 to q), respectively.

  • Step 5: For candidate feature Fj (j = 1 to q), if FjM < Tm and FjV < Tv, Fj is selected. Tm and Tv are empirical thresholds of mean value and variance, which are determined by experiments. In the first level evaluation, we can get the selected feature set {FS1, FS2 ⋯, FSt} (t ≤ q).

3.2. Feature Evaluation Based on Decision Tree

According to the first level evaluation, the features with high segmentation error rate and instability have been eliminated. The remaining features will participate in the second level feature evaluation. The framework of the second level feature evaluation is shown in Figure 5. In this step, a C4.5 decision tree algorithm [27] is performed to do the selection of the feature set {FS1, FS2 …, FSt}. Decision tree learning is one of the most widely used methods for inductive inference [28]. In the structure of a decision tree, leave nodes represent classifications and branches denote conjunctions of features that lead to those classifications. C4.5 algorithm which chooses attribute by computing gain ratio is one of the-state-of-art decision tree induction algorithms. The gain ratio is defined to be the quotient of information gain and split information [28]:

GainRate ( S , A ) = Gain ( S , A ) SplitInformation ( S , A )
where S and A represent sample set and attribute, respectively.

C4.5 can handle continuous attributes and choose the attributes with largest gain ratio as tree nodes. We regard the features as attributes and evaluate the features using gain ratio.

A number of images captured by sensor Si are randomly selected. Each fingerprint image is partitioned into non-overlapping blocks of same size, and for each block, candidate features FS1 to FSt are extracted. In order to perform the following processes, the values of candidate features are normalized into [0, 1] using Min-max normalization. Then the blocks are manually labeled as two classes: foreground blocks and background blocks. For each block, a vector can be acquired as follows:

( FS 1 , FS 2 , FS 3 , , FS t , Label )

All the blocks are used as training samples. A training set is produced by sampling part of blocks without replacement from the training samples. After repeating this k times, k training sets can be acquired. In each training set, a decision tree is trained using C4.5 algorithm. Then two steps will be conducted in each decision tree:

  • Compute contribution rate of features. For each feature, the contribution rate denotes the rate of samples classified into the right category through this feature. In a decision tree, leaf nodes represent the class information, by which how many samples are classified into the right category can be obtained. Each path from the tree root to a leaf node corresponds to a conjunction of feature tests. In each decision tree, we firstly compute the rate of samples classified into the right category by each leaf node. Then we retrieve the path from the tree root to every leaf node. Take feature FS1 as example. If feature FS1 appears in this path, we accumulate the right rate for FS1. It is to be noted that if a feature appears more than once in a path, its right rate won’t be accumulated repeatedly. In a decision tree, the sum of the right rates in all paths is defined to be the contribution rate of feature FS1. For each candidate feature, the contribution rate CRj (j = 1 to n) is the mean value of contribution rates in k training sets.

  • Compute appearance rate of features. For each feature, the appearance rate is defined to represent the quotient of this feature’s total appearance time in all decision trees and the total number of decision trees. It is to be noted that if a feature appears more than once in a decision tree, its appearance time is regarded as one instead of being accumulated repeatedly. We represent appearance rate of candidate feature using ARj (j = 1 to n).

If CRj > Tc and ARj > Ta, the feature FSj is selected. Tc and Ta are empirical thresholds of contribution rate and appearance rate, which are determined experimentally.

According to the two-level evaluation method, we can get the selected feature set {FSs1, FSs2 …, FSsm}, which has good sensor interoperability for images derived from different sensors.

4. Experiments

Experiments are conducted using the fingerprint databases listed on Table 1. The experiments involved two parts: we firstly evaluated the sensor interoperability of features on real sub-databases and selected a feature or feature set with good sensor interoperability, then we verified the selected features by applying them to segment fingerprint images. The experiments are conducted under the WEKA platform [29]. We use decision tree (J48) and support vector machine (LIBSVM) implementations in this platform.

4.1. Feature Evaluation and Results

In this section, segmentation features are evaluated using the two-level feature evaluation method, and the feature or feature set with good sensor interoperability are selected according to the evaluation results. The candidate features are composed of eight commonly used features, that is, mean(M) [9], variance(V) [9], coherence(Coh) [9], contrast(Con) [17], combination of variance and its gradient(VarG) [18], gradient magnitude(GraM) [7], block clusters degree(CluD) [2], and standard deviation of Gabor features(SDG) [19].

We randomly select 10 fingerprint images from each real sub-database. Each image is partitioned into non-overlapping blocks with the same size of 8 × 8 pixels, and for each block, eight candidate features are extracted. Then the blocks are manually labeled as two classes: foreground blocks and background blocks. For each candidate feature, the first level evaluation based on segmentation error rate is performed. The mean value and variance of segmentation error rate of the candidate features are shown in Table 2. In our experiment, we set Tm = 0.2 and Tv = 0.03, mean, combination of variance and its gradient, block clusters degree, and standard deviation of Gabor features are eliminated from the candidate feature set.

The second level evaluation based on a decision tree is performed on the second level candidate feature set {variance, coherence, contrast, gradient magnitude}. In the second level evaluation, two fingerprint images from each real sub-database are selected. Each image is also partitioned into non-overlapping blocks with the same size of 8 × 8 pixels, and for each block, four candidate features are extracted. The values of these four features are normalized into [0, 1] using Min-max normalization. Then the blocks are manually labeled as two classes, and for each block, a vector can be acquired as follows: {variance, coherence, contrast, gradient magnitude, label}

A training set is produced by sampling 10% blocks without replacement from every image. After repeating this 10 times, 10 training sets can be acquired. In each training set, a decision tree is trained using C4.5 algorithm. After assembling the results in 10 training sets, the contribution rate and appearance rate of every second level candidate feature can be obtained, which are shown in Table 3. We set Tc = 0.1 and Ta = 0.8 in our experiment, the contrast and gradient magnitude are selected finally. The feature evaluation procedure and result are presented in Table 4.

4.2. Verification for Selected Features

In this section, we verify the sensor interoperability of the selected features in fingerprint segmentation. Firstly, we compare the selected features with the classic coherence, mean, variance (CMV) [8,9,11] feature combination to show the effectiveness of the evaluation method. Secondly, the segmentation model trained on one database is used to segment fingerprint images in heterogeneous databases and the segmentation performance of the selected features is studied. The segmentation performance is measured by error rate, which is defined in Equation (1).

4.2.1. Comparison to CMV Feature Combination

In the experiment, we compare the selected features with coherence, mean, variance (CMV) feature combination, which is a representative and classic feature combination. For each real sub-database of FVC2000, FVC2002 and FVC2004, 15 images are randomly selected, five images are used to construct a combined database containing 45 images and the remaining 10 images are used for testing. Each fingerprint image is partitioned into non-overlapping blocks of 8 × 8 pixel, and for each block, the selected features and CMV features are extracted, respectively. Then, the blocks are manually labeled as two classes: foreground blocks and background blocks. For each image in combined database, 10% blocks are randomly selected as training samples. The SVM classifiers are trained using selected features and CMV feature combination on the training set, respectively. Then the two classifiers are used to segment other 10 fingerprint images in each real sub-database. The results are shown in Table 5.

Compared with the CMV feature combination, the selected features have lower average segmentation error rates in images derived from different sensors. It is to be noted that the variance of segmentation error rate of the selected features is one third of that of the CMV feature combination. With the above observations we can conclude that the selected features have better segmentation performance for different sensors and the performance is relatively stable. In other words, the selected features exhibit good sensor interoperability.

4.2.2. Heterogeneous Databases Test

In this section, we learn the segmentation model in one sub-database and use it to segment fingerprint images in heterogeneous sub-databases. We randomly select five images from FVC2002DB2 for training and 10 images from each real sub-database for testing. Each fingerprint image is partitioned into non-overlapping blocks of 8 × 8 pixels, and for each block, the selected features are extracted. Then the blocks are manually labeled as two classes: foreground blocks and background blocks. The SVM classifier learnt from FVC2002DB2 is used to segment fingerprint images in all the real sub-databases.

The results are shown in Table 6. The segmentation model which is learnt using the selected features not only has favorable performance in homogenous sub-database but also has lower segmentation error rates in heterogeneous sub-databases. The low mean value and variance of error rate demonstrate the good sensor interoperability of the selected features.

5. Conclusions

This work studied the sensor interoperability problem of feature in fingerprint segmentation. We analyzed this problem by investigating the histograms of features in images derived from different sensors. Then a two-level evaluation method is proposed to evaluate features and select a feature or feature set for sensor interoperable fingerprint segmentation. The proposed method implements feature evaluation by the first level evaluation based on segmentation error rate and the second level evaluation based on a decision tree. The first level evaluation eliminates the features with high segmentation error rates and instability, and then the second level evaluation selects the features with largest information gain rate in images derived from different sensors. Experimental results demonstrate that the proposed method can effectively evaluate and select the features with good sensor interoperability, and thus segmentation accuracies of images collected by different sensors are significantly improved with the selected features. The evaluation on sensor interoperability of features also give an impetus to the application of the segmentation algorithms in the internet environment.

Our future works will focus on finding more creative methods to evaluate the sensor interoperability of features and select the features with good sensor interoperability for fingerprint segmentation. Furthermore, we will also conduct research on finding more discriminating features for various sensors.

Acknowledgments

This work is supported by National Natural Science Foundation of China under Grant No. 61173069 and 61070097, and Natural Science Foundation of Shandong Province under Grant No. ZR2009GM003. The authors would like to thank Shuaiqiang Wang for his helpful comments and constructive advices on structuring the paper. In addition, the authors would particularly like to thank the anonymous reviewers for helpful suggestions.

References

  1. Maltoni, D.; Maio, D.; Jain, A.K.; Prabhakar, S. Handbook of Fingerprint Recognition, 2nd ed; Springer: London, UK; p. 2009.
  2. Chen, X.J.; Tian, J.; Cheng, J.G.; Yang, X. Segmentation of fingerprint images using linear classifier. EURASIP J. Appl. Signal Process 2004, 4, 480–494. [Google Scholar]
  3. Zhu, E.; Yin, J.P.; Hu, C.F.; Zhang, G.M. A systematic method for fingerprint ridge orientation estimation and image segmentation. Pattern Recognit 2006, 39, 1452–1472. [Google Scholar]
  4. Ratha, N.K.; Chen, S.; Jain, A.K. Adaptive flow orientation-based feature extraction in fingerprint images. Pattern Recognit 1995, 28, 1657–1672. [Google Scholar]
  5. Mehtre, B.M.; Murthy, N.N.; Kapoor, S.; Chatterjee, B. Segmentation of fingerprint images using the directional image. Pattern Recognit 1987, 20, 429–435. [Google Scholar]
  6. Mehtre, B.M.; Chatterjee, B. Segmentation of fingerprint images—A composite method. Pattern Recognit 1989, 22, 381–385. [Google Scholar]
  7. Qi, J.; Xie, M. Segmentation of Fingerprint Images Using the Gradient Vector Field. Proceedings of International Conference on Cybernetics and Intelligent Systems, Chengdu, China, 21–24 September 2008; pp. 543–545.
  8. Yang, G.P.; Zhou, G.T.; Yin, Y.L.; Yang, X.K. K-Means based fingerprint segmentation with sensor interoperability. EURASIP J. Adv. Signal Process 2010. [Google Scholar] [CrossRef]
  9. Bazen, A.M.; Gerez, S.H. Segmentation of Fingerprint Images. Proceedings of the Workshop on Circuits Systems and Signal Processing, Veldhoven, The Netherlands, 17–18 November 2001; pp. 276–280.
  10. Wang, L.; Dai, M.; Geng, G.H. Fingerprint Image Segmentation by Energy of Gaussian-Hermite Moments. Proceedings of the 5th Chinese Conference on Biometric Recognition. Advances in Biometric Person Authentication, Guangzhou, China, 13–14 December 2004; pp. 414–423.
  11. Yin, Y.L.; Wang, Y.R.; Yang, X.K. Fingerprint Image Segmentation Based on Quadric Surface Model. Proceedings of 5th International Conference on Audio and Video Based Biometric Person Authentication, Hilton Rye Town, NY, USA, 20–22 July 2005; pp. 647–655.
  12. Wu, C.H.; Tulyakov, S.; Govindaraju, V. Robust Point Based Feature Fingerprint Segmentation Algorithm. Proceedings of the International Conference on Advances in Biometrics, Seoul, Korea, 27–29 August 2007; pp. 1095–1103.
  13. Guo, X.J.; Yin, Y.L.; Shi, Z.C. Personalized Fingerprint Segmentation. Proceeding of 16th International Conference on Neural Information Processing (ICONIP-09), Bangkok, Thailand, 1–5 December 2009; pp. 798–809.
  14. Zhou, G.T.; Yin, Y.L.; Guo, W.J.; Ren, C.X. Fingerprint algorithm based on co-training. J. ShanDong Univ. (Eng. Sci.) 2009, 39, 22–26. [Google Scholar]
  15. Ross, A; Jain, A. Biometric Sensor Interoperability: A Case Study in Fingerprints. Proceedings of International ECCV Workshop on Biometric Authentication, Prague, Czech Republic, 15 May 2004; pp. 134–145.
  16. Ren, C.X.; Yin, Y.L.; Ma, J.; Yang, G.P. Feature Selection for Sensor Interoperability: A Case Study in Fingerprint Segmentation. Proceedings of the IEEE International Conference on Systems, Man and Cybernetics (SMC-09), San Antonio, TX, USA, 11–14 October 2009; pp. 5057–5062.
  17. Wang, S; Zhang, W.W.; Wang, Y.S. New features extraction and application in fingerprint segmentation. Acta Autom. Sinica 2003, 29, 622–626. [Google Scholar]
  18. Fan, D.J.; Sun, B.; Feng, J.F. Adaptive fingerprint segmentation based on variance and its gradient. J. Comput.-Aided Des. Comput. Graph 2008, 20, 742–747. [Google Scholar]
  19. Shen, L.L.; Kot, A.; Koo, W.M. Quality Measures of Fingerprint Images. Proceedings of the Third International Conference on Audio- and Video-Based Biometric Person Authentication, Halmstad, Sweden, 6–8 June 2001; pp. 266–271.
  20. Hu, C.F.; Yin, J.P.; Zhu, E.; Li, Y. A Composite Fingerprint Segmentation Based on Log-Gabor Filter and Orientation Reliability. Proceedings of the 17th International Conference on Image Processing, Hong Kong, China, 26–29 September 2010; pp. 3097–3100.
  21. Chen, C.; Zhang, D.; Zhang, L.; Zhao, Y.Q. Segmentation of Fingerprint Image by Using Polarimetric Feature. Proceedings of the International Conference on Autonomous and Intelligent Systems, Povoa de Varzim, Portugal, 21–23 June 2010; pp. 1–4.
  22. Baig, A.; Bouridane, A.; Kurugollu, F. A Corner Strength Based Fingerprint Segmentation Algorithm with Dynamic Thresholding. Proceedings of the 19th International Conference on Pattern Recognition, Tampa, FL, USA, 8–11 December 2008; pp. 1–4.
  23. Mieloch, K.; Munk, A.; Mihailescu, P. Improved Fingerprint Image Segmentation and Reconstruction of Low Quality Areas. Proceedings of the 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 1241–1244.
  24. FVC2000. Available online: http://bias.csr.unibo.it/fvc2000 (accessed on 14 September 2000).
  25. FVC2002. Available online: http://bias.csr.unibo.it/fvc2002 (accessed on 1 November 2001).
  26. FVC2004. Available online: http://bias.csr.unibo.it/fvc2004 (accessed on 31 March 2003).
  27. Quinlan, J.R. C4.5: Programs for Machine Learning; Morgan Kaufmann: San Francisco, CA, USA, 1993. [Google Scholar]
  28. Mitchell, T.M. Machine Learning; McGraw-Hill: New York, NY, USA, 1997; pp. 52–79. [Google Scholar]
  29. Witten, I.H.; Frank, E. Data Mining: Practical Machine Learning Tools and Techniques, 2nd ed; Morgan Kaufmann: San Francisco, CA, USA, 2005. [Google Scholar]
Figure 1. Features’ histograms of 10 fingerprints in FVC2000 DB2.
Figure 1. Features’ histograms of 10 fingerprints in FVC2000 DB2.
Sensors 12 03186f1 1024
Figure 2. Features’ histograms of fingerprints in three different sub-databases of FVC2000.
Figure 2. Features’ histograms of fingerprints in three different sub-databases of FVC2000.
Sensors 12 03186f2 1024
Figure 3. Features’ histograms of fingerprints in 9 different sub-databases of FVC2000, FVC2002 and FVC2004.
Figure 3. Features’ histograms of fingerprints in 9 different sub-databases of FVC2000, FVC2002 and FVC2004.
Sensors 12 03186f3 1024
Figure 4. Framework of the first level evaluation.
Figure 4. Framework of the first level evaluation.
Sensors 12 03186f4 1024
Figure 5. Framework of the second level evaluation.
Figure 5. Framework of the second level evaluation.
Sensors 12 03186f5 1024
Table 1. FVC fingerprint database sensor list.
Table 1. FVC fingerprint database sensor list.
Database ImageSensor typeSizeResolution
FVC2000 DB1Low-cost Optical Sensor “Secure Desktop Scanner” by KeyTronic300 × 300500 dpi
FVC2000 DB2Low-cost Capacitive Sensor “TouchChip” by ST Microelectronics256 × 364500 dpi
FVC2000 DB3Optical Sensor “DF-90” by Identicator Technology448 × 478500 dpi
FVC2002 DB1Optical Sensor “TouchView II” by Identix388 × 374500 dpi
FVC2002 DB2Optical Sensor “FX2000” by Biometrika296 × 560569 dpi
FVC2002 DB3Capacitive Sensor “100 SC” by Precise Biometrics300 × 300500 dpi
FVC2004 DB1Optical Sensor “V300” by CrossMatch640 × 480500 dpi
FVC2004 DB2Optical Sensor “U.are.U 4000” by Digital Persona328 × 364500 dpi
FVC2004 DB3Thermal sweeping Sensor “FingerChip FCD4B14CB” by Atmel300 × 480512 dpi
Table 2. The mean value and variance of segmentation error rate.
Table 2. The mean value and variance of segmentation error rate.
MVCohConVarGGraMCluDSDG

Mean Value0.6689720.1165340.1354580.1892480.4630090.0679440.3588990.567037
Variance0.0707550.0128470.0036170.0224030.027740.0022120.0337250.01576
Table 3. The contribution rate and appearance rate of second level candidate features.
Table 3. The contribution rate and appearance rate of second level candidate features.
VCohConGraM

Contribution Rate0.02610.01890.98470.1293
Appearance Rate0.90.710.9
Table 4. Feature evaluation procedure and result.
Table 4. Feature evaluation procedure and result.
ProcedureFeature Set

Initial{M, V, Coh, Con, VarG, GraM, CluD, SDG}
First Level Feature Evaluation{V, Coh, Con, GraM}
Second Level Feature Evaluation{Con, GraM}
Table 5. Segmentation error rates comparison.
Table 5. Segmentation error rates comparison.
Database2000 db12000 db22000 db32002 db12002 db22002 db32004 db12004 db22004 db3MeanVariance

Selected Features7.7714%2.2791%1.9038%2.001%1.8319%3.0449%3.729%4.8062%3.5961%3.44%0.000366
CMV7.4939%1.7442%3.0312%2.4753%2.6303%3.0041%0.9637%4.5021%10.3399%4.02%0.000914
Table 6. Segmentation error rates in homogenous and heterogeneous sub-databases.
Table 6. Segmentation error rates in homogenous and heterogeneous sub-databases.
Database2000 db12000 db22000 db32002 db12002 db22002 db32004 db12004 db22004 db3MeanVariance

Selected Features7.7959%2.3333%1.8876%1.9664%1.8361%3.0449%4.2020%4.8002%3.3990%3.4739%0.000374

Share and Cite

MDPI and ACS Style

Yang, G.; Li, Y.; Yin, Y.; Li, Y.-S. Two-Level Evaluation on Sensor Interoperability of Features in Fingerprint Image Segmentation. Sensors 2012, 12, 3186-3199. https://doi.org/10.3390/s120303186

AMA Style

Yang G, Li Y, Yin Y, Li Y-S. Two-Level Evaluation on Sensor Interoperability of Features in Fingerprint Image Segmentation. Sensors. 2012; 12(3):3186-3199. https://doi.org/10.3390/s120303186

Chicago/Turabian Style

Yang, Gongping, Ying Li, Yilong Yin, and Ya-Shuo Li. 2012. "Two-Level Evaluation on Sensor Interoperability of Features in Fingerprint Image Segmentation" Sensors 12, no. 3: 3186-3199. https://doi.org/10.3390/s120303186

Article Metrics

Back to TopTop