Next Article in Journal
Comparing Relational and Ontological Triple Stores in Healthcare Domain
Previous Article in Journal
Acknowledgement to Reviewers of Entropy in 2016
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Face Detection Based on Skin Color Segmentation Using Fuzzy Entropy

by
Francisco A. Pujol
1,*,
Mar Pujol
2,
Antonio Jimeno-Morenilla
1 and
María José Pujol
3
1
Department of Computer Technology, University of Alicante, Alicante 03690, Spain
2
Department of Computer Science and Artificial Intelligence, University of Alicante, Alicante 03690, Spain
3
Department of Applied Mathematics, University of Alicante, Alicante 03690, Spain
*
Author to whom correspondence should be addressed.
Entropy 2017, 19(1), 26; https://doi.org/10.3390/e19010026
Submission received: 14 November 2016 / Revised: 15 December 2016 / Accepted: 22 December 2016 / Published: 11 January 2017
(This article belongs to the Section Information Theory, Probability and Statistics)

Abstract

:
Face detection is the first step of any automated face recognition system. One of the most popular approaches to detect faces in color images is using a skin color segmentation scheme, which in many cases needs a proper representation of color spaces to interpret image information. In this paper, we propose a fuzzy system for detecting skin in color images, so that each color tone is assumed to be a fuzzy set. The Red, Green, and Blue (RGB), the Hue, Saturation and Value (HSV), and the YCbCr (where Y is the luminance and Cb,Cr are the chroma components) color systems are used for the development of our fuzzy design. Thus, a fuzzy three-partition entropy approach is used to calculate all of the parameters needed for the fuzzy systems, and then, a face detection method is also developed to validate the segmentation results. The results of the experiments show a correct skin detection rate between 94% and 96% for our fuzzy segmentation methods, with a false positive rate of about 0.5% in all cases. Furthermore, the average correct face detection rate is above 93%, and even when working with heterogeneous backgrounds and different light conditions, it achieves almost 88% correct detections. Thus, our method leads to accurate face detection results with low false positive and false negative rates.

1. Introduction

In recent years, there has been intensive research to develop complex security systems involving biometric features. In general, a biometric system is a pattern recognition system that makes a personal identification by determining the authenticity of a specific physiological or behavioral characteristic possessed by the user. One of the most popular biometric techniques to identify users is face recognition. In general, an automatic face recognition system is based on extracting a set of features from the user’s face, either geometric characteristics or some information of textures and shapes of the different elements that constitute a human face. Before a recognition algorithm is applied, faces must have been detected in an image using some detection method. Therefore, detecting a face is the first step for the recognition process to be efficiently completed.
Early face detection algorithms, which used heuristic and anthropometric techniques, appeared in the 1970s [1]. These methods often failed, since they were very sensitive to changes in lighting conditions or image scale; moreover, they only detected a face in well-defined environments, such as high-contrast images with a white background. In order to detect faces in complex environments, the algorithm had to be totally re-designed.
New security requirements and some technology advances, such as the massive popularization of multimedia applications in telecommunication systems, allowed face detection algorithms to be re-visited, and newer and more efficient systems have been developed in the last few years, including low-level analysis, active shape models, feature analysis, etc. [2,3].
There are many approaches to face detection, such as knowledge- and appearance-based methods, feature-invariant algorithms or template-matching techniques; for further details, see [2,4,5]. In this work, we shall focus on a pre-processing stage for detecting faces in color images. To do this, we assume that the problem consists of detecting skin regions in the image, and then, the segmented skin clusters will be further processed to validate that they belong to face regions.
Recently, skin color detection has been intensively researched in computer vision, becoming very popular as a powerful cue for detecting, segmenting and tracking faces and hands [6,7,8]. One of the main requirements for a face detection system is that it can handle the variations that occur in natural image sequences. These variations can be in the image conditions (size, location, lighting, orientation, point of view), in the face appearance (glasses, beard, mustache, makeup, expression) or in the image contents (complexity of the background, occlusion by other objects, number of faces). Human skin-color is one of the most robust face features regarding the variations mentioned above and is used as the core for the color segmentation model developed in this paper.
The main goal of skin color modeling is to extract a set of decision rules that efficiently find the pixels in an image that belong to human skin and eliminate those that represent any other elements in a scene. There are several possibilities to classify these methods; for instance, in [9], two categories for modeling skin color distributions were defined: physical-based approaches and statistical approaches. The first one describe the formation of images; e.g., evaluating how the appearance of skin color varies according to light changes in a scene [10]. Focusing on the statistical approaches, in [2], three categories for modeling skin color distributions were defined: explicitly-defined skin region, non-parametric skin distribution and parametric skin distribution models, either adaptive or non-adaptive. We will use this classification as a basis for our review.
In general, the first method to develop a skin classifier consists of defining the bounding limits of the regions corresponding to a color belonging to the skin, by means of some numerical (and often empirical) rules. In other terms, we define a set of threshold values in order to classify each pixel as “skin” or “non-skin” by evaluating the color for every pixel in a new input image. It is clear that the efficiency of the color segmentation stage highly depends on the selected color space. Thus, RGB is the most commonly-used basis for color descriptions, though it has the negative aspect that each of the coordinates (red, violet and blue) is subject to luminance effects from the lighting intensity of the environment, which may result in a misclassification of skin and non-skin pixels. That is to say, color spaces with separated luminance and chrominance components, like HSV or YCbCr, seem to be more appropriate for face detection; thus, the HSV color space is widely used for skin color detection, due to its relation to how humans perceive colors. On the other hand, YCbCr is useful in compression applications, but the specification of colors is somewhat unintuitive. Many other proposals using different color spaces can be found in the literature [11,12,13]. The main advantage of using these techniques is the simplicity of the skin detection rules, which permit a fast classification of pixels. The major drawback is the difficulty in finding both robust color spaces and reliable decision rules for a convenient skin detection. Nevertheless, as shown in [4], if an optimum skin detector is designed for every color space, their performance will be almost the same.
On the other hand, there is a set of techniques that estimate the distribution of skin color by means of a training phase, without deriving an explicit model. These methods are often referred to as non-parametric skin models. Generally, they create a skin probability map (SPM) [14,15], which consists of assigning a probability to each point in different color spaces. The illumination conditions are often unknown in an arbitrary image, so the variation in skin colors is much less constrained than in the controlled setup. Again, the selection of an appropriate color space is essential to achieve optimal results. Many non-parametric methods have been proposed for skin color modeling: histogram lookup tables (LUT) [16,17], SPMs using Bayesian theory [18,19] or neural networks [20,21], among others.
Non-parametric methods achieve low computation times for both training and classifying skin pixels, and they do not depend on the skin distribution shape or on the selection of the color space, either. However, they require a huge storage capacity, and the training must be performed with a representative dataset.
Finally, the searching for a more compact skin model representation for certain applications, as well as the need of having a generalization of the training data have resulted in the development of the so-called parametric skin distribution models. These algorithms consist of building a decision rule where an explicit skin distribution model is obtained, using different features, such as mean values, covariance matrices, mixtures of Gaussians or elliptic boundary models [22,23]. Parametric methods can also be fast, as they make it possible to interpolate data and to generalize the training phase for incomplete data. In spite of this, their performance highly depends on the skin distribution shape, and many parametric models ignore the static non-skin color. Subsequently, this fact leads to high false positive rates compared with non-parametric methods.
Many recent works [2,4,24] are based on the underlying idea of representing the skin color in an optimal color space (such as RGB, HSV or YCbCr) by means of the so-called skin cluster. Color information is an efficient tool for identifying facial areas if the skin color model can be properly adapted for different lighting environments. As a consequence, our proposal consists of taking the main advantages of explicitly-defined skin models, i.e., the simplicity of their decision rules, and combining them with the main benefits of non-parametric skin models, that is to say, low computation times for training and classifying. To do this, our method includes a fuzzy model for detecting skin in color images. The completed experiments (see Table 1, where bold text represents the system with better results) show that our method gives accurate results and that using a fuzzy approach to skin color detection allows the system to obtain a correctly-detected face even when working with uncontrolled environments and different light conditions. This model is explained in the following section.
This paper is organized as follows: In Section 2, we will define some terms first, and afterwards, the process of designing the fuzzy systems for detecting skin in RGB, HSV and YCbCr color spaces will be widely described. The face detection algorithm is also outlined at the end of Section 2. Then, the experiments on (i) segmenting skin and (ii) detecting faces are explained and discussed in Section 3, where a comparison with some explicitly-defined skin segmentation algorithms and some SPM methods is also included. Finally, we shall conclude with a discussion of the results in Section 4, and some remarks on our work are summarized in Section 5.

2. Development of a Fuzzy Face Detection System

As described before, automated face detection mainly consists of identifying the regions of the image containing a face or a set of faces. In this section, we will first develop a fuzzy system to extract skin from color images, taking into account different color spaces. Clearly, not all of the detected skin regions would contain faces: some of them will belong to other skin areas, such as the neck, hands, etc., while others will correspond to objects with similar colors to the skin. Thus, a method to detect where faces are in the segmented images must be developed, as well. Figure 1 shows an overview of our face detection algorithm.
Our approach involves two phases: in the first one, the system is trained. Thus, all of the parameters needed for the definition of the fuzzy rules for skin color detection are estimated here. Then, the second step implies an extensive set of experiments with different databases for validating the proposed model. In order to achieve our goals, let us define some terms first.

2.1. Definitions

Definition 1.
A region R in an image I is said to be connexif there is a neighborhood relation between any pixel p R .
Definition 2.
A connex region R F in an image I contains a face F when the set of the main features of F, ϕ , is included in R F .
The extracted features ϕ may include information about geometrical features (position of the eyes, mouth, etc.), texture information, color spaces, etc. In our case, ϕ will consider the face area, shape and skin color detection.
It is clear that any image I may contain several faces and that a connex region R i may not contain any face. Let us assume that a face F i is defined as the i-th rectangular connex region R F i containing the distinctive features of a face, ϕ , and that is limited by the pixels p 1 i , p 2 i , located at the top and bottom vertexes of the region.
In this work, we shall focus on face detection in color images. A color space is a method by which color can be specified, created and visualized.
Definition 3.
A color vector c C , where C is a color space, of a pixel p is defined as an l-tuple of color components c ( p ) = { c 1 ( p ) , c 2 ( p ) , , c l ( p ) } , where c i ( p ) , for i = 1 , 2 , , l , may have N different values.
For instance, for the RGB color space, l = 3 and c ( p ) = { r p , g p , b p } , where { r p , g p , b p } [ 0 , 255 ] .
After the face detection process is completed, the system creates an output binary image O where every pixel is labeled as foreground, i.e., face pixel, or background, i.e., non-face pixel. Therefore, a face detection problem is a segmentation problem, in which the main goal is to find image regions containing a face. This can be stated as follows:
Definition 4.
Let I be an image of size W = n × m pixels. Let Γ I be a set of faces in I, formed by a set of k non-overlapped regions, R F i , each one containing a face: Γ I = { R F 1 , R F 2 , , R F k } . A face detector is a function that converts pixels p j , for j = 1 , 2 , , W , in the original image I into a binary output image O as follows:
O ( p j ) = 1 if   p j   belongs   to   a   face 0 if   p j   does   not   belong   to   a   face
This definition can be used for any face detection scheme, such as knowledge-based, template matching or feature-based methods [5]. Due to the fact there are some situations (such as lighting conditions, face orientation or similar colors between faces and background) that may not achieve an optimum segmentation when applying a face detection scheme, some post-processing filter must be applied to obtain the set Γ I . For our proposals, face detection will be performed using a skin color segmentation algorithm in a convenient color space C and a subsequent computation of the areas of the segmented connex regions R i Γ I .

2.2. Fuzzy Sets and Skin Detection

In Section 1, we have outlined the main approaches to skin color detection. As we know, in any image segmentation scheme, there is a high level of uncertainty for a classifier to automatically obtain an optimum segmentation [25]. This fact can be also extended to face detection and, in particular, to skin color segmentation. Thus, we find that applying fuzzy theory can be a convenient way to obtain good detection rates, since a fuzzy set-theoretic model provides a mechanism to represent and manipulate uncertainty within an image.
Color image segmentation using fuzzy classification is a pixel-based segmentation method. This method assigns a color class to each pixel of an input image by applying a set of fuzzy rules on it. We can use this approach to achieve our goal: a pixel can be classified as “skin” or “non-skin” according to a set of fuzzy rules extracted from a training stage using different color spaces. To do this, each color plane will be considered as a fuzzy set, so that the skin detection is performed through fuzzy functions representing the membership degree of each pixel to the different classes. This is accomplished as follows:
Definition 5.
Given a color image I of size W = n × m pixels, where each pixel is defined using a color vector c in a color space C, so that c ( p ) = { c 1 ( p ) , c 2 ( p ) , , c l ( p ) } , p I , the histogram of C , Ψ ( C ) , is defined as a b × l array Ψ ( C ) = { f 1 , f 2 , , f l } , such that each f i is the frequency vector of the color component c i , for i = 1 , 2 , , l , using b bins, on the image I.
As a result, the value of each bin is the number of pixels in image I having the color c i . If Ψ ( C ) is normalized by W, then Ψ ( C ) takes the color space C into the interval [0, 1]; that is, Ψ ( C ) represents the probability distribution of each color c i to be present in image I. According to Zadeh’s theory [26], a fuzzy set is a pair ( A , m ) where A is a set and m: A [ 0 , 1 ] . This can be applied to the color histogram, where the fuzzy set can be defined as the pair ( C , Ψ ) , where C is the color space and Ψ: C [ 0 , 1 ] is the normalized histogram. For each c C , Ψ ( c ) is the grade of membership of c, so that c ( C , Ψ ) c C and Ψ ( c ) 0 .
The research on skin color face detection is mostly based on the popular RGB, YCbCr and HSV color space models [27]. In addition, some previous works have modeled these color spaces by means of fuzzy sets and fuzzy relations. Thus, the RGB system was used to represent skin colors [28], where a fuzzy model is extracted after asking people to classify human faces according to their skin color; then, a modified RGB fuzzy skin detector was presented in [29], based on the work of [30], trying to improve the skin detection performance. For the HSV space and its variants, the hue component can be defined by means of a fuzzy representation in order to take into account the non-uniformity of the colors’ distribution. Consequently, the authors in [31] proposed to represent colors with trapezoidal or triangular fuzzy subsets, associating colors with fuzzy sets; fuzzy linguistic hierarchies with different numbers of labels, depending on the desired granularity, were used in [32]. Then, a fuzzy classifier to detect the presence of faces in small windows, with an HSV color model to detect skin, was presented in [33]. Finally, fuzzy representations of YCbCr for modeling skin can be found in [34,35].
As shown, since RGB, YCbCr and HSV are the most common color spaces models for segmenting skin, we shall use these color systems in this paper, as well. In order to use our fuzzy approach, we must calculate first the normalized histogram for the considered color spaces, RGB, HSV and YCbCr:
  • Ψ R G B ( C ) = { Ψ ( R ) , Ψ ( G ) , Ψ ( B ) } ;
  • Ψ H S V ( C ) = { Ψ ( H ) , Ψ ( S ) , Ψ ( V ) } ;
  • Ψ Y C b C r ( C ) = { Ψ ( Y ) , Ψ ( C b ) , Ψ ( C r ) } .
Since the skin detection will be used as a pre-processing task for detecting a face in an image, the training has been performed using a set of 200 images from the XM2VTS database [36], extracting only the skin information, using different ethnic groups and changing lighting conditions. The XM2VTS face database contains eight recordings of 295 subjects each, acquired over a period of four months. The background in this set of images is homogeneous, with a good contrast for detecting skin. A group of images from the training set is shown in Figure 2.
The results after obtaining Ψ R G B ( C ) are shown in Figure 3.
From Figure 3a, the membership functions μ SKIN i ( c i ) for the skin color in each plane can be modeled using a bell-shaped function, such that:
μ SKIN i ( c i ) = β i e ( c i α i ) 2 2 σ i 2
where i = { R , G , B } ; { c R , c G , c B } [ 0 , 255 ] ; β i = max Ψ R G B ( c i ) , σ i 2 is the variance of each fuzzy set c i and α i = arg max c i Ψ R G B ( c i ) . The results of the model for the skin pixels are shown in Figure 3b.
Finally, for the background pixels, i.e., the non-skin pixels in the image, let us consider a variation of the model introduced in [37], which identifies the fuzziness in the transition region between the object (in this case, the skin) and the background classes. Thus, the membership value of a point to the object is determined by applying an S-function and a Z-function to the each color plane, so that:
μ NSKIN i S ( c i ) = 0 for c i a S i 2 c i a S i γ S i a S i 2 for a S i c i b S i 1 2 c i γ S i γ S i a S i 2 for b S i c i γ S i 1 for γ S i c i
μ NSKIN i Z ( c i ) = 1 for c i a Z i 1 2 c i a Z i γ Z i a Z i 2 for a Z i c i b Z i 2 c i a Z i γ Z i a Z i 2 for b Z i c i γ Z i 0 for γ Z i c i
where i = { R , G , B } ; { c R , c G , c B } [ 0 , 255 ] and { a S i , b S i , γ S i , a Z i , b Z i , γ Z i } are the cross-over points that determine the shape of the membership functions for the non-skin pixels.
Accordingly, the same process has been adopted for the HSV and YCbCr color spaces. Although many previous works claim that color information in the HSV system is mainly contained either only in the hue component [38] or in the hue and saturation components [39], from our experiments, it became clear that the recognition of skin is also influenced by the intensity (value) of the image. This fact makes us consider all three color planes for achieving an accurate color skin detection. The results of calculating the histogram for each color plane in HSV are shown in Figure 4a. Note that intensity values are between zero and 255.
In the same way, for the YCbCr space, the two chrominance components are usually used to extract skin clusters, whereas the luminance component Y is discarded. However, practically speaking, the skin-tone color is non-linearly dependent on luminance [35], so for our proposals, we have also taken into account all three components, whose values are between zero and 255. The histograms are shown in Figure 5a, and their Gaussian approximations can be seen in Figure 5b.
As in the RGB case, each plane ( H , S and V) and ( Y , C b and C r ) can be modeled, respectively, using a bell-shaped function (see Figure 4b and Figure 5b) defined as stated in Equation (2), where i = { R , G , B , H , S , V , Y , C b , C r } ; β i = max Ψ k ( c i ) , α i = arg max c i Ψ k ( c i ) ; { c R , c G , c B , c H , c S , c V , c Y , c C b , c C r } [ 0 , 255 ] ; k refers to RGB, HSV or YCbCr color systems, and σ i 2 is the variance of each fuzzy set c i .
Finally, for the non-skin pixels in the image, we will apply the S- and Z-functions defined by Equations (3) and (4), where i = { R , G , B , H , S , V , Y , C b , C r } .
Now, given an input image I, for any pixel p I , its color components are fuzzified, according to the parameters defined in Equations (2)–(4) and using some of the three color spaces considered above. In order to determine these parameters, a maximum entropy criterion will be used.
Shannon entropy [40] has been widely used in information theory. It is a measure of the uncertainty associated with a random variable. Specifically, Shannon entropy quantifies the expected value of the information contained in a message [41]. As a result, Shannon entropy is an extremely powerful approach in image segmentation [42,43,44]. In general terms, the selection of an appropriate threshold using Shannon entropy consists of finding an optimum threshold by maximizing the entropy function. In [45], Cheng et al. used the concept of fuzzy c-partition and the maximum entropy principle to select threshold values for gray-level images. When it is used for multilevel thresholding segmentation, the fuzzy c-partition entropy measures the quantity of extracted information in the image segmentation [46]. A larger value of entropy indicates that more information is extracted in this process. Taking this methodology into account, the skin color detection process is a segmentation process, as images are divided into a suitable number of fuzzy sets to represent skin and non-skin pixels. From the study developed in this section, it becomes clear that a fuzzy three-partition entropy approach for each color plane will segment images, giving as a result skin and non-skin regions.
Let p SKIN i , p NSKIN i S , p NSKIN i Z be the probabilities of the three fuzzy sets resulting from Equations (2)–(4), defined as follows:
p SKIN i = c i = 0 255 Ψ k ( c i ) μ SKIN i ( c i ) , p NSKIN i S = c i = 0 255 Ψ k ( c i ) μ NSKIN i S ( c i ) , p NSKIN i Z = c i = 0 255 Ψ k ( c i ) μ NSKIN i Z ( c i )
Then, the entropy of the fuzzy three-partition can be calculated as:
H ( a Z i , b Z i , γ Z i , a S i , b S i , γ S i ) = p SKIN i log ( p SKIN i ) p NSKIN i S log ( p NSKIN i S ) p NSKIN i Z log ( p NSKIN i Z )
The optimal combination of the parameters in Equation (6) can be computed by the maximum entropy criterion, that is the entropy H ( a Z i , b Z i , γ Z i , a S i , b S i , γ S i ) will be maximized when selecting parameters ( a Z i , b Z i , γ Z i , a S i , b S i , γ S i ) in a proper way. This process must be repeated for each plane and each color system used in the system.
After the fuzzy inference system has been defined, skin-candidate regions are detected. That is, we must remark that the segmented regions have the same color as that of the skin. The following process is to assure that these regions actually belong to a face or not. This fact will be discussed in the following subsection.

2.3. Face Detection

After the segmentation process, the system creates an output binary image where pixels are segmented as skin or non-skin pixels. The skin region localization process is achieved by applying an eight-connected blob algorithm [47]. Then, after identifying skin regions, the algorithm only accepts those having an appropriate size over the total image area. Let I , O be the input and output images, both of size W = n × m pixels; let R k be a region segmented as skin in O. The system will consider R k as a face region R F k if the following relationship is accomplished:
A k W p
where the area of region R k , A k , represents the number of skin pixels in R k . Hence, regions above threshold p are assumed to contain human faces.

3. Results

Let us show now the results of some experiments completed for our model and compare it with some other existing methods.
First, the values of the parameters in order to maximize the entropy H ( a Z i , b Z i , γ Z i , a S i , b S i ) must be found. After performing the experiments with the training set defined in Figure 2, we found that the optimum values are: a S i = γ Z i = arg max c i Ψ R G B ( c i ) ; a Z i = 0 . 5 γ Z i ; γ S i = 1 . 5 a S i ; and b S i , b Z i correspond to the membership value of c i = b S i (or b Z i ) in μ NSKIN i S ( c i ) , μ NSKIN i Z ( c i ) that is equal to 0.5. The results for the models of the fuzzy skin and non-skin classes for the RGB color space are shown in Figure 6.
The results for the models of the fuzzy skin and non-skin classes for the HSV and YCbCr color spaces are shown in Figure 7.

3.1. Results for the Proposed Fuzzy Systems

The tests have been performed with three different databases:
  • The XM2VTS face database [36], described before in Section 2.1;
  • The VALID database [48], which consists of five recording sessions of 106 subjects over a period of one month. One session is recorded in a studio with controlled lighting and no background noise, and the other four sessions are recorded in office-type scenarios;
  • The color FERET database [49], which is composed of 11,338 face images of 994 different persons, taken from different angles of view, over the course of 15 sessions between 1993 and 1996. For our experiments, we used the sets of images fa (images with frontal view) and fb (frontal images from the same individuals but with different facial expression), corresponding to 843 individuals with frontal images only.
For all of the experiments in this section, we have used a set of 200 images from each database. Our set was taken from persons of different ethnicities: Caucasian, Asian and African.
According to the fuzzy skin detector defined before, Figure 8, Figure 9 and Figure 10 show the results for different images in all three databases. The images include different genders and ethnic groups and, for the VALID database, complex backgrounds and changing lighting conditions.
From these results, we must point out that the proposed method leads to very accurate results for the XM2VTS database, which has both good illumination conditions and high contrast between the foreground and the background. In the case of the VALID database, the results of the skin detector are also quite precise, in spite of the unfavorable conditions for achieving good quality results. Some errors appear mainly due to background colors that are very similar (or even identical) to skin colors; in addition, the lighting conditions will affect the final result. This effect is also noticeable for the images in the color FERET database, where there is a great variability in the illumination conditions of the images. However, the general results are good, and after applying a post-processing stage for detecting a face from the skin segmentation results, most of these errors can be corrected, as we will discuss in Section 3.3.

3.2. Comparison with Other Methods

Let us now compare our system with some other existing algorithms for skin color detection. First, an explicitly-defined skin region algorithm for the RGB color space is used; in [2], a pixel ( R 0 , G 0 , B 0 ) is classified as skin if:
( a ) R 0 > 95 G 0 > 40 B 0 > 20 , and ( b ) max ( R 0 , G 0 , B 0 ) min ( R 0 , G 0 , B 0 ) > 15 , and ( c ) R 0 G 0 > 15 R 0 > G 0 R 0 > B 0
We also found similar proposals for HSV and YCbCr color systems. In [50], it was found that, as we stated before, the recognition of skin color is influenced by the intensity (value) and saturation of the image; consequently, the distribution of intensity, saturation and hue should be defined for accurate skin detection. After training 200 pictures, including different ethnic groups (Asian, Caucasian and African), they obtained the following rules for a pixel ( H 0 , S 0 , V 0 ) to be classified as skin, where the three components are normalized:
( a ) 0 . 1 < S 0 < 0 . 68 , and ( b ) 0 . 4 V 0 0 . 014 H 0 V 0 + 0 . 062 0 . 01 , 0 . 13 V 0 0 . 25 , or ( c ) 0 . 4 V 0 0 . 014 H 0 0 . 67 V 0 0 . 014 , 0 . 25 V 0 0 . 38 , or ( d ) V 0 0 . 34 0 . 03 H 0 0 . 67 V 0 0 . 014 , 0 . 38 V 0 0 . 46 , or
( e ) V 0 0 . 34 0 . 03 H 0 V 0 0 . 31 0 . 009 , 0 . 46 V 0 0 . 6 , or ( f ) 0 . 91 V 0 0 . 14 H 0 V 0 0 . 31 0 . 009 , 0 . 6 V 0 0 . 76 , or ( g ) 0 . 91 V 0 0 . 14 H 0 1 . 17 V 0 0 . 0082 , 0 . 76 V 0 0 . 91 , or ( h ) V 0 0 . 91 0 . 0041 H 0 1 . 17 V 0 0 . 0082 , 0 . 91 V 0 1
Finally, in [51], a pixel ( Y 0 , C b 0 , C r 0 ) is classified as skin if:
( a ) 60 < Y 0 < 255 , and ( b ) 100 < Cb 0 < 125 , and ( c ) 135 < Cr 0 < 170
On the other hand, a skin probability map (SPM) approach to each one of the color spaces is also taken into account. For this method, given skin and non-skin histogram models, a skin pixel classifier is defined using the Bayes’ theorem. Then, the inequality P ( s k i n | c ) T can be used as a skin detection rule, where T is a selected threshold value and P ( s k i n | c ) is the probability of observing skin, given a concrete c color value (see [14,23] for further details). We performed a training with the same images used in the development of the fuzzy systems, and obtained, as a result, the probability distribution for all three considered color systems.
Figure 11, Figure 12 and Figure 13 show the results for some images for our proposal and the existing skin segmentation methods named before. In each case, for the SPM models, we show the threshold T that minimizes the number of false positives and false negatives.
Finally, we have manually extracted the ideal skin segmentation for all of the images and compared it with the segmented results of all of the methods implemented. Therefore, in Table 2, we show the results of the false positive (FP) and false negative (FN) rates, as well as the correct skin detection rate for the images in the experiments, for all of the algorithms considered. Our method obtains the better results for each color system, as shown in bold text in Table 2.

3.3. Results on Face Detection

As mentioned in Section 2.3, after defining a skin color detection system, we will need to determine if the resulting pixels belong to a face or not. To do this, we proposed a threshold p relating the area of the segmented skin regions and the number of pixels in the input image. After a training process, we found p = 0 . 3 as the best threshold in order to identify if a skin region belongs to a face. Let us consider first well-illuminated images with a homogeneous background, such as the ones in the XM2VTS database; we can see an example in Figure 14. Here, the results after the skin segmentation process and the subsequent face detector are shown; in order to compare different methods, we have used the ones that obtained the best results in Table 2 for each kind of algorithm. Therefore, the explicit rules and our fuzzy proposal for the RGB color system are taken into account, as well as the skin probability map for the HSV system.
As expected, the results show that when working with well-illuminated images and homogeneous backgrounds, faces are detected properly for all of the algorithms considered. There are some false positive skin pixels belonging to the shirt region, but they are isolated in small regions whose areas are below threshold p, so these regions are not considered as face areas, and consequently, the face is detected correctly.
Let us explain now a more complex example, where the Valid database is utilized for the tests. As mentioned before, most images in the Valid database were acquired in a noisy “real-world” office environment with no control of illumination, so face detection algorithms must deal with tougher conditions than in both the XM2VTS and FERET databases. Again, the results after the skin segmentation process and the subsequent face detector are shown in Figure 15.
The results of the Valid database are highly dependent on the conditions where images were taken, since a great amount of images in this database are recorded in office-type scenarios. For the example shown, we can clearly appreciate the differences between the results of all three algorithms, and what is more, it becomes clear how the face is not well detected for the first two cases. Thus, for the explicit RGB system, although the algorithm extracts in a proper way the facial skin pixels, some pixels in the shirt are also considered as belonging to skin and connected to the face region (i.e., they are false positives); this way, they cannot be separated from the face, and as a consequence, this results in a bad face detection. Accordingly, an incorrect face detection is also obtained when using the SPM for the HSV color system. In this case, even a bigger portion of the shirt is considered as skin, leading again to a bad detection.
On the contrary, for our fuzzy RGB detector, the pixels of the face are correctly segmented and classified as belonging to the same region. The area of this region is above the threshold ratio p, so this region is considered as a face R F . Moreover, false positive skin pixels, such as the ones in the shirt, are isolated from the face, and the areas of these regions are below threshold p, so these regions are not considered as belonging to a face, which results in a very accurate face detection.
To sum up, Table 3 shows the results on face detection for all of the images used for testing the system from the three databases. In addition, a comparison between our detector with other related works is shown in Table 4. Thus, the seminal work of Viola and Jones [52] is selected first. Moreover, convolutional neural networks (CNN) are becoming extremely popular in pattern recognition in the last few years. Recently, Li et al. [53] built a CNN cascade for fast face detection, and the results obtained from this method are considered in Table 4, as well.
The best results on face detection using the mentioned dataset are obtained for our fuzzy RGB detector (as shown in bold text in Table 4), with an average of 93.53% correct detections; even for the Valid database, it achieves almost 88% correct face detections, whereas the explicit RGB algorithm has about 79%, and the SPM for the HSV color system only gets about 62% correct detections. Finally, our system improves the results of the Viola and Jones face detector, which is the de facto standard in many applications of face detection and achieves similar results to the CNN cascade proposal. Consequently, we believe that the developed fuzzy skin detector is a reliable face detector.

4. Discussion

According to the results of the experiments, it seems clear that our fuzzy proposal for segmenting skin achieves good detection rates in all three color systems considered, whilst false positives and false negatives are kept at low levels. Moreover, our approach can classify faces from different ethnic groups, giving an accurate skin segmentation result. Thus, the best global results for all of the algorithms implemented in terms of correct classified pixels (96.16%) are for the fuzzy RGB detector, followed by the other two fuzzy proposals (the fuzzy HSV detector and the fuzzy YCbCr detector achieve 95.82% and 94.87% correct segmented pixels, respectively). This gives an improvement of about 5% compared with the explicit and SPM detection algorithms.
Regarding the false positive errors, we must recall that this rate measures the proportion of actual negative pixels that are incorrectly identified as skin pixels; this rate is critical, because a face can be misdetected due to a high number of false positive pixels, such as hair, clothes and background colors that are similar to skin. In this case, the fuzzy YCbCr detector gives the lowest FP rate (only 0.48%), followed by the other two fuzzy detectors, which are slightly above 0.5%. Again, these rates are better than the other implemented algorithms. Moreover, considering the false negative errors, where skin pixels are not identified properly, the fuzzy RGB detector has the lowest rate (3.31%), followed another time by the other two fuzzy proposals (3.62% for the fuzzy HSV detector and 4.65% for the fuzzy YCbCr method). These results improve the ones from the other algorithms about 4%–5% in most cases.
When working with good illumination conditions and homogeneous backgrounds, such as the case of the images in the XM2VTS database, the results obtained for all of the algorithms are good, with correct detection rates reaching nearly 95%. For this database, the fuzzy RGB detector is the best one, with 98.12% correct classification of pixels. Besides, images in the color FERET database have generally homogeneous backgrounds and uniform lighting conditions, but sometimes the colors in the background are similar to the ones in human skin, which leads to the highest false positive rates for all three databases (see Figure 8, Figure 9 and Figure 10 and Figure 13). Nevertheless, the correct detection results for our fuzzy proposals are higher than 95%, with FP rates below 1% in all of the cases.
On the other hand, working in unconstrained environments, such as the ones we can find in the VALID database, gives, as expected, the worst segmentation results. Thus, the explicit and SPM algorithms hardly arrive at 85% correct results. However, all of the three fuzzy methods proposed have detection rates above 93% for this database, and as a consequence, they behave reasonably well for complex backgrounds and changing illumination environments, giving very accurate results.
To sum up, we have verified that our proposal of using fuzzy expert systems for skin color face detection gets precise results even when dealing with uncontrolled scenarios and changing lighting conditions, making it a suitable technique for segmenting skin in different environment conditions.

5. Conclusions

Face recognition systems have become a popular tool for identifying and verifying the identity of users in recent years. In order to complete successfully the recognition process, an initial, robust face detection algorithm must be implemented. Among the methods for face detection, color skin segmentation algorithms have become important in the last few years, providing a robust tool for detecting faces.
This article has presented a review of different skin detection algorithms on color images, having studied the main features of the most common color spaces: RGB, HSV and YCbCr. A proposal of fuzzy skin color detectors for these color spaces has been discussed throughout this work. Thus, each color plane has been modeled using different fuzzy sets for the skin and non-skin classes and the inference system and defuzzifying process have been explained. The results of the experiments show that our fuzzy algorithms give the better results regarding the classification of pixels as skin (between 94%–96%), improving about 5% both the explicit rules and the SPM methods. Considering face detection, the fuzzy RGB system achieves almost 94% of correct detections, being robust even when dealing with uncontrolled environments and changing light conditions. Therefore, our method leads to very accurate face detection results with low false positive and false negative rates.
Future works aim at extending the fuzzy system to other color spaces, such as TSL or YIQ. An adaptive training of the skin model can be used to improve the classification of skin regions for our fuzzy method. We are also focusing on working with images where more than one face appears, adapting the face detection algorithm accordingly.

Acknowledgments

This work has been supported by the Ministerio de Economía y Competitividad (Spain), Project TIN2013-40982-R. Project co-financed with FEDER funds.

Author Contributions

Francisco A. Pujol and Mar Pujol conceived of and designed the experiments. Antonio Jimeno-Morenilla performed the experiments. María José Pujol formalized the fuzzy entropy model and analyzed the data. Francisco A. Pujol and Mar Pujol wrote the paper. All authors have read and approved the final manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hjelmås, E.; Low, B.K. Face Detection: A Survey. Comput. Vis. Image Underst. 2001, 83, 236–274. [Google Scholar] [CrossRef]
  2. Vezhnevets, V.; Sazonov, V.; Andreeva, A. A Survey on Pixel-Based Skin Color Detection Techniques. Proc. Graphicon 2003, 3, 85–92. [Google Scholar]
  3. Zhang, C.; Zhang, Z. A Survey of Recent Advances in Face Detection; Technical Report; Microsoft Research: Redmond, WA, USA, 2010. [Google Scholar]
  4. Albiol, A.; Torres, L.; Delp, E.J. Optimum color spaces for skin detection. ICIP 2001, 1, 122–124. [Google Scholar]
  5. Yang, M.H.; Kriegman, D.J.; Ahuja, N. Detecting faces in images: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 34–58. [Google Scholar] [CrossRef]
  6. Kakumanu, P.; Makrogiannis, S.; Bourbakis, N. A survey of skin-color modeling and detection methods. Pattern Recogn. 2007, 40, 1106–1122. [Google Scholar] [CrossRef]
  7. Khan, R.; Hanbury, A.; Stöttinger, J.; Bais, A. Color based skin classification. Pattern Recogn. Lett. 2012, 33, 157–163. [Google Scholar] [CrossRef]
  8. Sun, H.M. Skin detection for single images using dynamic skin color modeling. Pattern Recogn. 2010, 43, 1413–1420. [Google Scholar] [CrossRef]
  9. Zheng, H.; Daoudi, M.; Jedynak, B. Blocking Adult Images Based on Statistical Skin Detection. Electron. Lett. Comput. Vis. Image Anal. 2004, 4, 1–14. [Google Scholar]
  10. Kelly, W.; Donnellan, A.; Molloy, D. Screening for Objectionable Images: A Review of Skin Detection Techniques. In Proceedings of the 2008 International Conference Machine Vision and Image Processing, Coleraine, Ireland, 3–5 September 2008; pp. 51–158.
  11. Chen, H.S.; Wang, T.M.; Chen, S.H.; Liu, J.S. Skin-color correction method based on hue template mapping for wide color gamut liquid crystal display devices. Color Res. Appl. 2011, 36, 335–348. [Google Scholar] [CrossRef]
  12. Tomaz, F.; Candeias, T.; Shahbazkia, H. Improved Automatic Skin Detection in Color Images. In Proceedings of the VIIth Digital Image Computing: Techniques and Applications, Sydeny, Australia, 10–12 December 2003; pp. 419–428.
  13. Xiang, F.H.; Suandi, S.A. Fusion of Multi Color Space for Human Skin RegionSegmentation. Int. J. Inf. Electron. Eng. 2013, 3, 172–174. [Google Scholar] [CrossRef]
  14. Brand, J.; Mason, J.S.; Roach, M.; Pawlewski, M. Enhancing face detection in color images using a skin probability map. In Proceedings of the 2001 International Symposium on Intelligent Multimedia, Video and Speech Processing, Hong Kong, China, 4 May 2001; pp. 344–347.
  15. Xu, J.; Zhang, X. A Real-Time Hand Detection System during Hand over Face Occlusion. Int. J. Mult. Ubiquitous Eng. 2015, 10, 287–302. [Google Scholar] [CrossRef]
  16. Liu, Q.; Peng, G.Z. A robust skin color based face detection algorithm. In Proceedings of the 2010 International Asia Conference on Informatics in Control, Automation and Robotics, Wuhan, China, 6–7 March 2010; pp. 525–528.
  17. Lü, W.; Huang, J. Skin detection method based on cascaded AdaBoost classifier. J. Shanghai Jiaotong Univ. 2012, 17, 197–202. [Google Scholar] [CrossRef]
  18. Ma, Z.; Leijon, A. Bayesian Estimation of Beta Mixture Models with Variational Inference. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 2160–2173. [Google Scholar] [PubMed]
  19. Shirali-Shahreza, S.; Mousavi, M.E. A New Bayesian Classifier for Skin Detection. In Proceedings of the 3rd International Conference on Innovative Computing Information and Control, Dalian, China, 18–20 June 2008.
  20. Hwang, C.L.; Lu, K.D. The Segmentation of Different Skin Colors Using the Combination of Graph Cuts and Probability Neural Network. In Advances in Computational Intelligence; Springer: Berlin/Heidelberg, Germany, 2011; pp. 25–33. [Google Scholar]
  21. Phung, S.L.; Chai, D.; Bouzerdoum, A. A universal and robust human skin color model using neural networks. In Proceedings of the 2001 International Joint Conference on Neural Networks, Washington, DC, USA, 15–19 July 2001; pp. 2844–2849.
  22. Wimmer, M.; Radig, B.; Beetz, M. A Person and Context Specific Approach for Skin Color Classification. In Proceedings of the 18th International Conference on Pattern Recognition, Hong Kong, China, 20–24 August 2006; pp. 39–42.
  23. Yang, M.H.; Ahuja, N. Gaussian mixture model for human skin color and its applications in image and video databases. In Proceedings of the Storage and Retrieval for Image and Video Databases VII, San Jose, CA, USA, 26–29 January 1999; pp. 458–466.
  24. Yang, J.; Fu, Z.; Tan, T.; Hu, W. Skin color detection using multiple cues. In Proceedings of the 17th International Conference on Pattern Recognition, Cambridge, UK, 23–26 August 2004; pp. 632–635.
  25. Zhang, L.; Ji, Q. A Bayesian Network Model for Automatic and Interactive Image Segmentation. IEEE Trans. Image Process. 2011, 20, 2582–2593. [Google Scholar] [CrossRef] [PubMed]
  26. Zadeh, L.A. Fuzzy sets. Inf. Control 1965, 8, 338–353. [Google Scholar] [CrossRef]
  27. Kale, S.; Saradadevi, M. Active Facial Tracking. Asian J. Comput. Sci. Inf. Technol. 2013, 1, 60–63. [Google Scholar]
  28. Boaventura, I.A.G.; Volpe, V.M.; Silva, I.N.D.; Gonzaga, A. Fuzzy Classification of Human Skin Color in Color Images. In Proceedings of the 2006 IEEE International Conference on Systems, Man and Cybernetics, Taipei, Taiwan, 8–11 October 2006; pp. 5071–5075.
  29. Selamat, A.; Maarof, M.A.; Chin, T.Y. Fuzzy Mamdani Inference System Skin Detection. In Proceedings of the Ninth International Conference on Hybrid Intelligent Systems, Shenyang, China, 12–14 August 2009; pp. 57–62.
  30. Kovac, J.; Peer, P.; Solina, F. Human skin color clustering for face detection. In Proceddings of the IEEE Region 8, Eurocon 2003. Computer as a Tool, Ljubljana, Slovenia, 22–24 September 2003; pp. 144–148.
  31. Truck, I.; Akdag, H.; Borgi, A. A symbolic approach for colorimetric alterations. In Proceedings of the 2nd International Conference in Fuzzy Logic and Technology, Leicester, UK, 5–7 September 2001; pp. 105–108.
  32. Herrera, F.; Martinez, L. A model based on linguistic 2-tuples for dealing with multigranular hierarchical linguistic contexts in multi-expert decision-making. IEEE Trans. Syst. Man Cybern. 2001, 31, 227–234. [Google Scholar] [CrossRef] [PubMed]
  33. Bhatia, A.; Srivastava, S.; Agarwal, A. Face Detection Using Fuzzy Logic and Skin Color Segmentation in Images. In Proceedings of the 3rd International Conference on Emerging Trends in Engineering and Technology, Goa, India, 19–21 November 2010; pp. 225–228.
  34. Hmid, M.B.; Jemaa, Y.B. Fuzzy classification, image segmentation and shape analysis for Human face detection. In Proceedings of the 8th international Conference on Signal Processing, Nice, France, 10–13 December 2006.
  35. Iraji, M.S.; Tosinia, A. Skin color segmentation in YCBCR color space with adaptive fuzzy neural network (Anfis). Int. J. Image Graph. Signal Process. 2012, 4, 35–41. [Google Scholar] [CrossRef]
  36. Messer, K.; Matas, J.; Kittler, J.; Luettin, J.; Maitre, G. XM2VTSDB: The extended M2VTS database. In Proceedings of the Second International Conference on Audio and Video-Based Biometric Person Authentication, Washington, DC, USA, 22–24 March 1999; pp. 965–966.
  37. Murthy, C.A.; Pal, S.K. Fuzzy thresholding: Mathematical framework, bound functions and weighted moving average technique. Pattern Recogn. Lett. 1990, 11, 197–206. [Google Scholar] [CrossRef]
  38. Nusirwan, A.; Kit, C.W.; John, S. RGB-H-CbCr Skin Colour Model for Human Face Detection. In Proceedings of the MMU International Symposium on Information&Communications Technologies, Petaling Jaya, Malaysia, 16–17 November 2006; pp. 579–584.
  39. Chitra, S.; Balakrishnan, G. Comparative study for two color spaces HSCbCr and YCbCr in skin color detection. Appl. Math. Sci. 2012, 6, 4229–4238. [Google Scholar]
  40. Shannon, C.E. A Mathematical Theory of Communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  41. Wu, Y.; Zhou, Y.; Saveriades, G.; Agaian, S.; Noonan, J.P.; Natarajan, P. Local Shannon entropy measure with statistical tests for image randomness. Inf. Sci. 2013, 222, 323–342. [Google Scholar] [CrossRef]
  42. Abdel-Khalek, S.; Ben Ishak, A.; Omer, O.A.; Obada, A.S. A two-dimensional image segmentation method based on genetic algorithm and entropy. Optik 2017, 131, 414–422. [Google Scholar] [CrossRef]
  43. Susan, S.; Kumar, A. Auto-segmentation using mean-shift and entropy analysis. In Proceedings of the 3rd International Conference on Computing for Sustainable Global Development, New Delhi, India, 16–18 March 2016; pp. 292–296.
  44. Bhowmik, M.; Sarkar, A.; Das, R. Shannon entropy based fuzzy distance norm for pixel classification in remote sensing imagery. In Proceedins of the Third International Conference on Computer, Communication, Control and Information Technology, Hooghly, India, 7–8 Febuary 2015.
  45. Cheng, H.D.; Chen, J.R.; Li, J. Threshold selection based on fuzzy c-partition entropy approach. Pattern Recogn. 1998, 31, 857–870. [Google Scholar] [CrossRef]
  46. Yin, S.; Zhao, X.; Wang, W.; Gong, M. Efficient multilevel image segmentation through fuzzy entropy maximization and graph cut optimization. Pattern Recogn. 2014, 47, 2894–2907. [Google Scholar] [CrossRef]
  47. Di Stefano, L.; Bulgarelli, A. A simple and efficient connected components labeling algorithm. In Proceedings of the 1999 International Conference on Image Analysis and Processing, Venice, Italy, 27–29 September 1999; pp. 322–327.
  48. Fox, N.A.; O’Mullane, B.A.; Reilly, R.B. VALID: A new practical audio-visual database, and comparative results. In Proceedings of the 2005 International Conference on Audio-and Video-Based Biometric Person Authentication, Hilton Rye Town, NY, USA, 20–22 July 2005; pp. 777–786.
  49. Phillips, P.J.; Moon, H.; Rizvi, S.A.; Rauss, P.J. The FERET evaluation methodology for face-recognition algorithms. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1090–1104. [Google Scholar] [CrossRef]
  50. Zhang, X.N.; Jiang, J.; Liang, Z.H.; Liu, C.L. Skin color enhancement based on favorite skin color in HSV color space. IEEE Trans. Consum. Electr. 2010, 56, 1789–1793. [Google Scholar] [CrossRef]
  51. Huang, D.Y.; Lin, C.J.; Hu, W.C. Learning-based face detection by adaptive switching of skin color models and AdaBoost under varying illumination. J. Inf. Hiding Multimed. Signal Process. 2011, 2, 204–216. [Google Scholar]
  52. Viola, P.; Jones, M.J. Robust Real-Time Face Detection. Int. J. Comput. Vis. 2004, 57, 137–154. [Google Scholar] [CrossRef]
  53. Li, H.; Lin, Z.; Shen, X.; Brandt, J.; Hua, G. A Convolutional Neural Network Cascade for Face Detection. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 8–10 June 2015; pp. 5325–5334.
Figure 1. A representation of our fuzzy face detection system.
Figure 1. A representation of our fuzzy face detection system.
Entropy 19 00026 g001
Figure 2. Training database used for the creation of the fuzzy skin detector system.
Figure 2. Training database used for the creation of the fuzzy skin detector system.
Entropy 19 00026 g002
Figure 3. Creation of a fuzzy system for the RGB model. (a) Computation of Ψ R G B ( C ) for the training set; (b) Gaussian approximation of Ψ R G B ( C ) for each color plane.
Figure 3. Creation of a fuzzy system for the RGB model. (a) Computation of Ψ R G B ( C ) for the training set; (b) Gaussian approximation of Ψ R G B ( C ) for each color plane.
Entropy 19 00026 g003
Figure 4. Creation of a fuzzy system for the HSV model. (a) Computation of Ψ H S V ( C ) for the training set; (b) Gaussian approximation of Ψ H S V ( C ) for each color plane.
Figure 4. Creation of a fuzzy system for the HSV model. (a) Computation of Ψ H S V ( C ) for the training set; (b) Gaussian approximation of Ψ H S V ( C ) for each color plane.
Entropy 19 00026 g004
Figure 5. Creation of a fuzzy system for the YCbCr model. (a) Computation of Ψ Y C b C r ( C ) for the training set; (b) Gaussian approximation of Ψ Y C b C r ( C ) for each color plane.
Figure 5. Creation of a fuzzy system for the YCbCr model. (a) Computation of Ψ Y C b C r ( C ) for the training set; (b) Gaussian approximation of Ψ Y C b C r ( C ) for each color plane.
Entropy 19 00026 g005
Figure 6. Modeling the skin and non-skin classes for the RGB color space.
Figure 6. Modeling the skin and non-skin classes for the RGB color space.
Entropy 19 00026 g006
Figure 7. Modeling the skin and non-skin classes. (a) HSV color space; (b) YCbCr color space.
Figure 7. Modeling the skin and non-skin classes. (a) HSV color space; (b) YCbCr color space.
Entropy 19 00026 g007
Figure 8. Results of the RGB fuzzy skin detector for different images. First row: images from the XM2VTS and the VALID databases. Second row: images from the color FERET database.
Figure 8. Results of the RGB fuzzy skin detector for different images. First row: images from the XM2VTS and the VALID databases. Second row: images from the color FERET database.
Entropy 19 00026 g008
Figure 9. Results of the HSV fuzzy skin detector for different images. First row: images from the XM2VTS and the VALID databases. Second row: images from the color FERET database.
Figure 9. Results of the HSV fuzzy skin detector for different images. First row: images from the XM2VTS and the VALID databases. Second row: images from the color FERET database.
Entropy 19 00026 g009
Figure 10. Results of the YCbCr fuzzy skin detector for different images. First row: images from the XM2VTS and the VALID databases. Second row: images from the color FERET database.
Figure 10. Results of the YCbCr fuzzy skin detector for different images. First row: images from the XM2VTS and the VALID databases. Second row: images from the color FERET database.
Entropy 19 00026 g010
Figure 11. Sample results from the XM2VTS database. (a) original image; (b) results for explicit rules in RGB; (c) results for explicit rules in HSV; (d) results for explicit rules in YCbCr; (e) results for SPM in RGB, T = 0.12; (f) results for SPM in HSV, T = 0.06; (g) results for SPM in YCbCr, T = 0.04; (h) results for our fuzzy model in RGB; (i) results for our fuzzy model in HSV; (j) results for our fuzzy model in YCbCr.
Figure 11. Sample results from the XM2VTS database. (a) original image; (b) results for explicit rules in RGB; (c) results for explicit rules in HSV; (d) results for explicit rules in YCbCr; (e) results for SPM in RGB, T = 0.12; (f) results for SPM in HSV, T = 0.06; (g) results for SPM in YCbCr, T = 0.04; (h) results for our fuzzy model in RGB; (i) results for our fuzzy model in HSV; (j) results for our fuzzy model in YCbCr.
Entropy 19 00026 g011
Figure 12. Sample results from the Valid database. (a) original image; (b) results for explicit rules in RGB; (c) results for explicit rules in HSV; (d) results for explicit rules in YCbCr; (e) results for SPM in RGB, T = 0.20; (f) results for SPM in HSV, T = 0.21; (g) results for SPM in YCbCr, T = 0.23; (h) results for our fuzzy model in RGB; (i) results for our fuzzy model in HSV; (j) results for our fuzzy model in YCbCr.
Figure 12. Sample results from the Valid database. (a) original image; (b) results for explicit rules in RGB; (c) results for explicit rules in HSV; (d) results for explicit rules in YCbCr; (e) results for SPM in RGB, T = 0.20; (f) results for SPM in HSV, T = 0.21; (g) results for SPM in YCbCr, T = 0.23; (h) results for our fuzzy model in RGB; (i) results for our fuzzy model in HSV; (j) results for our fuzzy model in YCbCr.
Entropy 19 00026 g012
Figure 13. Sample results from the color FERET database. (a) original image; (b) results for explicit rules in RGB; (c) results for explicit rules in HSV; (d) results for explicit rules in YCbCr; (e) results for SPM in RGB, T = 0.09; (f) results for SPM in HSV, T = 0.05; (g) results for SPM in YCbCr, T = 0.025; (h) results for our fuzzy model in RGB; (i) results for our fuzzy model in HSV; (j) results for our fuzzy model in YCbCr.
Figure 13. Sample results from the color FERET database. (a) original image; (b) results for explicit rules in RGB; (c) results for explicit rules in HSV; (d) results for explicit rules in YCbCr; (e) results for SPM in RGB, T = 0.09; (f) results for SPM in HSV, T = 0.05; (g) results for SPM in YCbCr, T = 0.025; (h) results for our fuzzy model in RGB; (i) results for our fuzzy model in HSV; (j) results for our fuzzy model in YCbCr.
Entropy 19 00026 g013
Figure 14. Results for face detection in the XM2VTS database. (a) Original image; (b) result of the explicit RGB skin color classifier; (c) RGB face detection; (d) result of the SPM HSV skin color classifier; (e) SPM HSV face detection; (f) result of the fuzzy RGB skin color classifier; (g) fuzzy RGB face detection.
Figure 14. Results for face detection in the XM2VTS database. (a) Original image; (b) result of the explicit RGB skin color classifier; (c) RGB face detection; (d) result of the SPM HSV skin color classifier; (e) SPM HSV face detection; (f) result of the fuzzy RGB skin color classifier; (g) fuzzy RGB face detection.
Entropy 19 00026 g014
Figure 15. Results for face detection in the Valid database. (a) Original image; (b) result of the explicit RGB skin color classifier; (c) RGB face detection; (d) result of the SPM HSV skin color classifier; (e) SPM HSV face detection; (f) result of the fuzzy RGB skin color classifier; (g) fuzzy RGB face detection.
Figure 15. Results for face detection in the Valid database. (a) Original image; (b) result of the explicit RGB skin color classifier; (c) RGB face detection; (d) result of the SPM HSV skin color classifier; (e) SPM HSV face detection; (f) result of the fuzzy RGB skin color classifier; (g) fuzzy RGB face detection.
Entropy 19 00026 g015
Table 1. A summary of the face detection results with some color skin segmentation algorithms. SPM, skin probability map (Bold text represents the system with better results).
Table 1. A summary of the face detection results with some color skin segmentation algorithms. SPM, skin probability map (Bold text represents the system with better results).
AlgorithmIncorrect (%)Correct (%)
Explicit RGB rules12.0387.97
SPM HSV22.2977.71
Our proposal (in the RGB system)6.4793.53
Table 2. Comparison between algorithms.
Table 2. Comparison between algorithms.
AlgorithmXM2VTSVALID
FP (%)FN (%)Correct (%)FP (%)FN (%)Correct (%)
Explicit RGB rules0.125.7594.140.4014.4785.13
Explicit HSV rules0.475.1294.412.1912.9184.91
Explicit YCbCr rules0.554.1295.330.8015.1484.06
SPM RGB1.653.4394.926.4017.5376.06
SPM HSV0.816.3992.800.9813.9785.04
SPM YCbCr0.236.0693.710.599.3990.02
Fuzzy RGB0.221.6698.120.645.7093.66
Fuzzy HSV0.182.1297.700.915.6293.47
Fuzzy YCbCr0.183.3196.520.606.3893.02
Algorithm FERET Global
FP (%)FN (%)Correct (%)FP (%)FN (%)Correct (%)
Explicit RGB rules3.142.4594.411.227.5691.22
Explicit HSV rules6.384.8288.803.017.6189.38
Explicit YCbCr rules1.697.9390.391.019.0689.93
SPM RGB2.655.4491.913.578.8087.63
SPM HSV6.300.4193.292.706.9290.38
SPM YCbCr7.106.1686.742.647.2090.16
Fuzzy RGB0.752.5696.690.543.3196.16
Fuzzy HSV0.603.1396.270.563.6295.82
Fuzzy YCbCr0.674.2795.050.484.6594.87
Table 3. Comparison of face detection with different color skin segmentation algorithms.
Table 3. Comparison of face detection with different color skin segmentation algorithms.
AlgorithmXM2VTSVALID
Incorrect (%)Correct (%)Incorrect (%)Correct (%)
Explicit RGB rules5.8294.1621.0678.94
SPM HSV10.1089.9037.4362.57
Fuzzy RGB1.9898.0212.5687.44
AlgorithmFERETGlobal
Incorrect (%)Correct (%)Incorrect (%)Correct (%)
Explicit RGB rules9.2190.7912.0387.97
SPM HSV19.3380.6722.2977.71
Fuzzy RGB4.8795.136.4793.53
Table 4. Comparison of other face detection methods.
Table 4. Comparison of other face detection methods.
AlgorithmIncorrect (%)Correct (%)
Viola and Jones [52]7.6392.37
Li et al. [53]6.8793.13
Fuzzy RGB6.4793.53

Share and Cite

MDPI and ACS Style

Pujol, F.A.; Pujol, M.; Jimeno-Morenilla, A.; Pujol, M.J. Face Detection Based on Skin Color Segmentation Using Fuzzy Entropy. Entropy 2017, 19, 26. https://doi.org/10.3390/e19010026

AMA Style

Pujol FA, Pujol M, Jimeno-Morenilla A, Pujol MJ. Face Detection Based on Skin Color Segmentation Using Fuzzy Entropy. Entropy. 2017; 19(1):26. https://doi.org/10.3390/e19010026

Chicago/Turabian Style

Pujol, Francisco A., Mar Pujol, Antonio Jimeno-Morenilla, and María José Pujol. 2017. "Face Detection Based on Skin Color Segmentation Using Fuzzy Entropy" Entropy 19, no. 1: 26. https://doi.org/10.3390/e19010026

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop