Abstract
Skin detection is an important process in many applications like hand gesture recognition, face detection and ego-vision systems. This paper presents a new skin detection method based on a dynamic generation of the skin cluster range in the YCbCr color space, by taking into account the lighting conditions. The method is based on the identification of skin color clusters in the YCb and YCr subspaces. The experimental results, carried out on two publicly available databases, show that the proposed method is robust against illumination changes and achieves satisfactory results in terms of both qualitative and quantitative performance evaluation parameters.
You have full access to this open access chapter, Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
Skin detection is an important issue in color image processing, which has been extensively studied over the years. It is a useful technique for the detection, segmentation and tracking of human skin in images or video streams. The interest in skin detection algorithms derives from their applicability to a wide range of applications such as gesture recognition, video surveillance, human computer interaction, ego-vision systems, human activity recognition [4, 6, 21], hand gestures detection and tracking [7, 8, 23, 31], nude images and video blocking [5, 26], feature extraction for content-based image retrieval [20], and age estimation [22].
Skin detection is a process that allows the extraction of candidate skin pixels in an image. In most cases, skin detection is performed by using pixel based techniques: a pixel is classified as a skin or non-skin pixel, independently from its neighbors, and only by using pixel color information. In addition, region based skin segmentation methods make use of extra information, for example spatial arrangement or texture information on the pixels detected in the skin detection process, to determine the boundaries of human skin regions [16, 17]. Therefore, a good pixel based method for skin detection can narrow the computational cost of the next process of segmentation, and, moreover, can improve the results of the segmentation. The main issue is to achieve a satisfactory skin detection under uncontrolled lighting conditions, since many applications, for example egovision systems, require the detection of skin human regions, both indoors and outdoors, with high or low illumination conditions. Most approaches use specific color spaces to de-correlate chromatic components from luminance, since they are less sensitive to lighting conditions [9, 12]. However, some studies [10, 14] have shown that the luminance component plays an important role in skin detection and so it should not be rejected.
In the present work, an explicit skin cluster method, in the YCbCr colour space, is proposed. The method takes into account the illumination changes of the examined image, and tries to minimize both false positives and false negatives. It results to be computationally efficient for real-time applications.
The rest of the paper is organized as follows: in Sect. 2, a description of the related work for skin detection is presented; Sect. 3 describes the proposed approach; in Sect. 4, some results and comparative evaluations performed on two publicly available databases are reported; finally, in Sect. 5 conclusions are drawn.
2 Related Work
Many recent surveys describe the various skin detection approaches [13, 18, 30]. Many approaches have been proposed for the skin colour detection; they include linear classifiers [9, 11, 12, 19, 25, 27], Bayesian [10, 15] or Gaussian classifiers [11, 29], and artificial neural network [1, 2, 28].
It has been demonstrated that the human skin colour can be modelled in many colour spaces [3, 24].
In [19], heuristic rules in the RGB color space are used to detect skin pixels; these rules depending on the image illumination; other methods adopt linear or non-linear transformations of the RGB colour space, such that other colour spaces can be generated. In particular, those colour spaces that separate the luminance and the chrominance components, are the most commonly used in skin colour detection approaches. This is the case of HSV and YCbCr, which are a non- linear and a linear transformation of the RGB colour space, respectively. Concerning to the HSV colour space, Hue (H) and Saturation (S) are the chrominance components, while Value (V) is the luminance component. Most of the methods that work in this colour space ignore the luminance component, since it does not result to be discriminant [12, 25]; however, also some methods that include the luminance in the process of skin detection have been presented [27]. Concerning to the YCbCr colour space the chrominance components Cb and Cr are obtained by subtracting the luminance component Y from blue and from red, respectively. Also in this case, some approaches ignore the luminance component [9], while others take it into account [14]. However, it has been demonstrated that skin colour is non-linearly dependent on the luminance component in different colour spaces, thus the luminance component should be included in the skin detection process [11, 14].
Explicit cluster methods are based on the definition of colour rules, in particular on the definition of a colour range for skin pixels [9, 25], or on the definition of a shape for skin pixel distribution (e.g., rectangle and ellipse). In [14], two different skin cluster models, that take into account the luminance component, have been proposed. In the first model, the skin clusters are determined by two central curves, one for the YCr and one for the YCb subspace, and by their spreads in the respective subspaces. In the second model, a single skin cluster is represented by an ellipse in a transformed CbCr subspace.
In this paper, a new approach, that works in the YCbCr colour space is proposed. In particular, taking into account the illumination conditions of the examined image, a dynamic cluster for the YCb and YCr subspaces is computed.
3 The Proposed Approach
As already shown in [14], the distribution of skin pixels in the YCb and YCr subspaces presents a trapezoidal shape (see Fig. 1a), differently from distribution of skin and non-skin pixels (see Fig. 1b).
Moreover, we experimentally observed that the size and shape of these trapezia change depending on the lighting conditions. In particular, we have observed that:
-
for images in high illumination conditions, the bases of the two trapezia in the YCb and YCr subspaces representing the skin colour clusters are larger than those associated with the skin colour clusters in low illumination conditions;
-
the positions of the vertices of the trapezia change according to the illumination conditions of the examined image;
-
for skin pixels, the minimum value of Cr (in the following Cr min ) and the maximum value of Cb (in the following Cb max ) are practically fixed at the values 133 and 128, respectively, as reported in [9], while the maximum value of Cr and the minimum value of Cb strongly change with the illumination conditions;
-
for a skin pixel, the values of the Cr and Cb generally satisfy the following conditions:
With reference to the Fig. 2, the vertices A and D of the larger basis of the trapezium related to the YCr skin subspace are given by (Y min , Cr min ) and (Y max , Cr min ), where Y min  = 0, Y max  = 255 and Cr min  = 133. The same applies to the vertices E and H of the larger basis of the trapezium related to the YCb skin subspace that are given by (Y min , Cb max ) and (Y max , Cb max ), with Cb max  = 128. Concerning to the vertices B and C of the shorter basis of the trapezium associated with the YCr skin subspace, they are set to (Y 0, Cr max ) and (Y 1, Cr max ). Taking into account the histogram of the pixels with values of Cr in the range [133,183], Cr max is set to the maximum of Cr, associated with at least a 10% of image pixels. So, Y 0 and Y 1 values are set as the 5th percentile and the 95th percentile of the Y component, respectively, considering all the pixels of the image with Cr = Cr max . The same process is applied to find the vertices F and G with coordinates (Y 2, Cb min ) and (Y 3, Cb min ) respectively, of the shorter basis of the trapezium associated with the YCb skin subspace.
Set a Y value, a point on the upper border of the trapezium in the YCr subspace will have coordinates (Y, T Cr (Y)), while a point on the lower bound of the trapezium in the YCb subspace, will have coordinates (Y, T Cb (Y)). T Cr (Y) and T Cb (Y) are given by:
where \( d_{Cr} = Cr_{max} - Cr_{min} \)
where \( d_{Cb} = Cb_{max} - Cb_{min} \)
Finally, we classify a pixel as skin pixel, if it satisfies the following two conditions:
4 Results and Comparison
The proposed approach has been compared with the method described in [9], which also works in the YCbCr colour space but with a fixed colour range, and with the method presented in [14], considering the both formulations of the skin cluster models. The approach has been tested on the Hand Gesture Recognition (HGR) database [17], containing 1,558 skin images of human hand and arm postures taken with different lighting conditions and on the Compaq database [15], a large database that consists of 4,675 colour images, containing skin images in unconstrained illumination and background conditions. Some qualitative results, for our approach and for the methods with which we compared, are shown in Figs. 3 and 4, for some selected images of the HGR and Compaq databases. Starting from a qualitative analysis of the results, it is clear that the method proposed in [9] generally obtains good results, but in some cases, many false positives are found (see row 3 in Fig. 3 and rows 1 and 3 in Fig. 4); in fact, the pixels belonging to regions of eyes and mouth or of background are generally wrongly detected as skin pixels. Concerning to the method in [14], in the YCbCr formulation, also in this case, many false positives are found, particularly in presence of high or low illumination conditions (see all the results of the Fig. 3 and rows 2 and 5 in Fig. 4); the skin detection performance improves on the case of its formulation in the transformed CbCr subspace, but in some cases many false negative are detected (see row 1 in Fig. 3 and rows 1 and 4 in Fig. 4).
Moreover, quantitative results in terms of F-measure are reported in Table 1, for all the analysed approaches. The proposed approach outperforms the other methods in terms of F-measure.
Finally, the computational cost of the proposed approach has been estimated. The performance of the algorithm has been estimated on a PC equipped with an Intel Xeon E5-2623 at 3 GHz, and with 16 GB RAM. For an image with a size of 320 × 480, the execution time is, on average, 8 ms.
5 Conclusion
We have presented a new approach for skin detection in the YCbCr color space. The method shows some robustness to variations in illumination conditions, because the skin cluster range in the YCbCr color space is defined dynamically, taking into account the luminance component. In particular, two clusters are found, one in the YCb subspace and one in the YCr subspace.
The performance of the method has been tested on two publicly available databases, producing satisfactory results both qualitatively and in terms of quantitative performance evaluation parameters such as F-measure. The results of a comparative analysis are promising. With respect to methods based on fixed cluster ranges, the proposed one provide adequate results also on images acquired in low or high illumination conditions.
References
Al-Mohair, H.K., Mohamad-Saleh, J., Suandi, S.A.: Human skin color detection: a review on neural network perspective. Int. J. Innov. Comput. Inf. Control 8(12), 8115–8131 (2012)
Al-Mohair, H.K., Saleh, J.M., Suandi, S.A.: Hybrid human skin detection using neural network and k-means clustering technique. Appl. Soft Comput. 33, 337–347 (2015)
Albiol, A., Torres, L., Delp, E.J.: Optimum color spaces for skin detection. In: ICIP, vol. 1, pp. 122–124 (2001)
Baraldi, L., Paci, F., Serra, G., Benini, L., Cucchiara, R.: Gesture recognition in ego-centric videos using dense trajectories and hand segmentation. In: 2014 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 702–707. IEEE (2014)
Basilio, J.A.M., Torres, G.A., Pérez, G.S., Medina, L.K.T., Meana, H.M.P.: Explicit image detection using YCbCr space color model as skin detection. In: Applications of Mathematics and Computer Engineering, pp. 123–128 (2011)
Betancourt, A.: A sequential classifier for hand detection in the framework of egocentric vision. In: 2014 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 600–605. IEEE (2014)
Brancati, N., Caggianese, G., Frucci, M., Gallo, L., Neroni, P.: Experiencing touchless interaction with augmented content on wearable head mounted displays in cultural heritage applications. Personal and Ubiquitous Computing https://dx.doi.org/10.1007/s00779-016-0987-8
Brancati, N., Caggianese, G., Frucci, M., Gallo, L., Neroni, P.: Robust fingertip detection in egocentric vision under varying illumination conditions. In: 2015 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), pp. 1–6. IEEE (2015)
Chai, D., Ngan, K.N.: Face segmentation using skin-color map in videophone applications. IEEE Trans. Circ. Syst. Video Technol. 9(4), 551–564 (1999)
Cheddad, A., Condell, J., Curran, K., McKevitt, P.: A skin tone detection algorithm for an adaptive approach to steganography. Sig. Process. 89(12), 2465–2478 (2009)
Garcia, C., Tziritas, G.: Face detection using quantized skin color regions merging and wavelet packet analysis. IEEE Trans. Multimedia 1(3), 264–277 (1999)
Guo, J.M., Liu, Y.F., Chang, C.H., Nguyen, H.S.: Improved hand tracking system. IEEE Trans. Circ. Syst. Video Technol. 22(5), 693–701 (2012)
Hjelmås, E., Low, B.K.: Face detection: A survey. Comput. Vis. Image Underst. 83(3), 236–274 (2001)
Hsu, R.L., Abdel-Mottaleb, M., Jain, A.K.: Face detection in color images. IEEE Trans. Pattern Anal. Mach. Intell. 24(5), 696–706 (2002)
Jones, M.J., Rehg, J.M.: Statistical color models with application to skin detection. Int. J. Comput. Vision 46(1), 81–96 (2002)
Kawulok, M.: Fast propagation-based skin regions segmentation in color images. In: 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), pp. 1–7. IEEE (2013)
Kawulok, M., Kawulok, J., Nalepa, J.: Spatial-based skin detection using discriminative skin-presence features. Pattern Recogn. Lett. 41, 3–13 (2014)
Khan, R., Hanbury, A., Stöttinger, J., Bais, A.: Color based skin classification. Pattern Recogn. Lett. 33(2), 157–163 (2012)
Kovac, J., Peer, P., Solina, F.: Human Skin Color Clustering For Face Detection, vol. 2. IEEE (2003)
Kruppa, H., Bauer, Martin, A., Schiele, B.: Skin patch detection in real-world images. In: Gool, L. (ed.) DAGM 2002. LNCS, vol. 2449, pp. 109–116. Springer, Heidelberg (2002). doi:10.1007/3-540-45783-6_14
Li, C., Kitani, K.M.: Pixel-level hand detection in ego-centric videos. In: 2013 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3570– 3577. IEEE (2013)
Luu, K., Bui, T.D., Suen, C.Y., Ricanek, K.: Spectral regression based age determination. In: 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 103–107. IEEE (2010)
Nalepa, J., Grzejszczak, T., Kawulok, M.: Wrist localization in color images for hand gesture recognition. Man-Machine Interactions 3, 79–86 (2014). Springer
Shin, M.C., Chang, K.I., Tsap, L.V.: Does colorspace transformation make anydifference on skin detection? In: Proceedings of the Sixth IEEE Workshop on Applications of Computer Vision 2002 (WACV 2002), pp. 275–279. IEEE (2002)
Sobottka, K., Pitas, I.: A novel method for automatic face segmentation, facial feature extraction and tracking. Sig. Process. Image Commun. 12(3), 263–281 (1998)
Stöttinger, J., Hanbury, A., Liensberger, C., Khan, R.: Skin paths for contextual flagging adult videos. In: Bebis, G., Boyle, R., Parvin, B., Koracin, D., Kuno, Y., Wang, J., Pajarola, R., Lindstrom, P., Hinkenjann, A., Encarnação, Miguel, L., Silva, Cláudio, T., Coming, D. (eds.) ISVC 2009. LNCS, vol. 5876, pp. 303–314. Springer, Heidelberg (2009). doi:10.1007/978-3-642-10520-3_28
Tsekeridou, S., Pitas, I.: Facial feature extraction in frontal views using biometric analogies. In: 9th European Signal Processing Conference (EUSIPCO 1998), pp. 1–4. IEEE (1998)
Yogarajah, P., Condell, J., Curran, K., McKevitt, P., Cheddad, A.: A dynamic threshold approach for skin tone detection in colour images. Int. J. Biometrics 4(1), 38–55 (2011)
Zaidan, A., Ahmad, N.N., Karim, H.A., Larbani, M., Zaidan, B., Sali, A.: Image skin segmentation based on multi-agent learning bayesian and neural network. Eng. Appl. Artif. Intell. 32, 136–150 (2014)
Zhao, W., Chellappa, R., Phillips, P.J., Rosenfeld, A.: Face recognition: A literature survey. ACM Comput. Surv. (CSUR) 35(4), 399–458 (2003)
Zhu, Y., Xu, G., Kriegman, D.J.: A real-time approach to the spotting, representation, and recognition of hand gestures for human–computer interaction. Comput. Vis. Image Underst. 85(3), 189–208 (2002)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this paper
Cite this paper
Brancati, N., De Pietro, G., Frucci, M., Gallo, L. (2017). Dynamic Colour Clustering for Skin Detection Under Different Lighting Conditions. In: Krasnoproshin, V., Ablameyko, S. (eds) Pattern Recognition and Information Processing. PRIP 2016. Communications in Computer and Information Science, vol 673. Springer, Cham. https://doi.org/10.1007/978-3-319-54220-1_3
Download citation
DOI: https://doi.org/10.1007/978-3-319-54220-1_3
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-54219-5
Online ISBN: 978-3-319-54220-1
eBook Packages: Computer ScienceComputer Science (R0)