Next Article in Journal
An Automatic Identification Method for the Blink Artifacts in the Magnetoencephalography with Machine Learning
Next Article in Special Issue
Assessment of Periodontium Temperature Changes under Orthodontic Force by Using Objective and Automatic Classifier
Previous Article in Journal
Deep and Meaningful E-Learning with Social Virtual Reality Environments in Higher Education: A Systematic Literature Review
Previous Article in Special Issue
Early Detection of Alzheimer’s Disease Using Polar Harmonic Transforms and Optimized Wavelet Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Method in Predicting Hypertension Using Facial Images

1
Clinical Medicine Division, Korea Institute of Oriental Medicine (KIOM), 1672, Yuseong-daero, Yuseong-gu, Daejeon 34054, Korea
2
Korean Convergence Medicine, University of Science and Technology, 217, Gajeong-ro, Yuseong-gu, Daejeon 34113, Korea
3
Future Medicine Division, Korea Institute of Oriental Medicine (KIOM), 1672, Yuseong-daero, Yuseong-gu, Daejeon 34054, Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(5), 2414; https://doi.org/10.3390/app11052414
Submission received: 9 February 2021 / Revised: 2 March 2021 / Accepted: 6 March 2021 / Published: 9 March 2021
(This article belongs to the Special Issue Medical Artificial Intelligence)

Abstract

:
Hypertension has been a crucial public health challenge among adults. This study aimed to develop a novel method for non-contact prediction of hypertension using facial characteristics such as facial features and facial color. The data of 1099 subjects (376 men and 723 women) analyzed in this study were obtained from the Korean Constitutional Multicenter Study of Korean medicine Data Center (KDC) at the Korea Institute of Oriental Medicine (KIOM). Facial images were collected and facial variables were extracted using image processing techniques. Analysis of covariance (ANCOVA) and Least Absolute Shrinkage and Selection Operator (LASSO) were performed to compare and identify the facial characteristic variables between the hypertension group and normal group. We found that the most distinct facial feature differences between hypertension patients and normal individuals were facial shape and nose shape for men in addition to eye shape and nose shape for women. In terms of facial colors, cheek color in men, as well as forehead and nose color in women, were the most distinct facial colors between the hypertension groups and normal individuals. Looking at the AUC value, the prediction power for women is better than men. In conclusion, we managed to explore and identify the facial characteristics variables related to hypertension. This study may provide new evidence in the validity of predicting hypertension using facial characteristics.

1. Introduction

Hypertension, also known as high blood pressure in layman’s terms, is one of the chronic metabolism diseases with a prevalence of over a billion people in the world. In South Korea, hypertension accounted for 11.3% of the entire deaths per 100,000 populations in 2017 [1]. The number of people suffering from chronic diseases who visited the hospital in 2015 was 14.39 million, with hypertension being the highest number (5.71 million) among them [2]. In addition, the total medical cost for hypertension treatment was assessed to be 13.4%, which accounts for approximately 2.5 billion US dollars [3]. This means that monitoring and regulating hypertension is fundamental to reduce the general burden of disease in Korean society as well as to improve the quality of life.
Hypertension is usually asymptomatic in the early stages and is mainly identified through targeted or opportunistic screening in primary care. The suboptimal blood pressure (BP) control in conjunction with the ignorance of hypertension has been the stumbling block in hypertension management, resulting in the necessity of routine BP screening. Although the golden standard for hypertension diagnosis is cuff-based BP measurements at present, developments of self-screening by utilizing new technologies have the potential to increase the early detection of hypertension.
Apart from the classic BP measurement, recent technologies have provided new potential for the evaluation of various parameters such as vital signals, speech analysis, heart rate, and step count [4]. Several studies have utilized signal-derived BP measurement such as photoplethysmogram and electrocardiogram signals as the indicators of BP [5,6,7]. A recent study also proposed that BP can estimate using acoustic characteristics [8]. Such novel strategies are also gradually being developed to be compatible with smartphone health applications and wearable devices.
The face is the most observed region in a person’s lifetime and perception of a person’s health according to their facial appearance is a day-to-day occurrence. Numerous attempts have been made by researchers to find the association between facial appearance and health for decades [9]. Recent studies have been focusing on facial shape to investigate the elements of faces that affect the perception of health [10]. A study on craniofacial anthropometry found that the face has many manifestations associated with increased susceptibility to specific diseases and any distinct changes in face can be an indication of a health issue [11]. There are also studies that are associated with facial shape with the diagnosis of diseases where morphological characteristics in facial shape can be related to the presence of diabetes and hypertension [12,13,14]. Current approaches also showed that skin color and texture contribute to facial health. Moreover, many studies have been conducted to elucidate the association between skin color and hypertension [15,16].
In this study, we proposed that facial characteristics such as facial features or facial color can be used for hypertension screening in a non-contact, unconscious, and labor-free manner. Facial characteristics that are extracted from facial images through image processing techniques may be indicative of hypertension and have the potential to play a daily role in self-screening of hypertension. This study aimed to develop a novel method by determining the possibility of predicting hypertension using facial characteristics.

2. Materials and Methods

2.1. Data Collection

Data analyzed in this study were obtained from the Korean Constitutional Multicenter Study of Korean medicine Data Center (KDC) at the Korea Institute of Oriental Medicine (KIOM). For the analysis of facial characteristics, 1099 subjects (376 men and 723 women) were included. All the subjects who were eligible met the following inclusion criteria: (1) age ranging from 18 to 90 years and (2) being healthy and not suffering from chronic diseases. Subjects who have undergone plastic surgery or facial reconstruction surgery due to any circumstances were excluded. All subjects were required to sign an informed consent form, and this study was approved by the KIOM Institutional Review Board (I-0910/02-001).
Data on facial images were acquired in a standard environment. Facial images were taken with a neutral expression using a digital camera (DSLR Nikon D5300 digital camera, Nikon Corporation, Thailand, 2013) with an 85-mm lens under bilateral illumination at a fixed subject-camera distance of 1.6 m. A color chart used for color correction was attached to a photographic ruler that was placed approximately 1 cm below the chin. The image resolution was selected as 2992 × 2000 pixels with a total of 24.16 million pixel count.
The first image was the full face frontal view where subjects were asked to look directly into the camera with a neural expression. Their heads were positioned so that the central point of the two pupils and the two points (cross point between the facial contour and upper auricular perimeters) are on the same line horizontally. The photographic ruler was to be placed below the chin in vertical alignment. The second image is the profile view where the subject’s face is turned approximately 90 degrees from the front. Subjects were requested to look straight ahead with a neural expression, whereas the photographic ruler will be placed below the chin in vertical alignment with the nose. Only one side of their face and the eye on the far side were shown. The central point of the pupil from the side and point of upper auricular were also on the same horizontal line.

2.2. Data Extraction

The data were collected by well-trained technicians under standard operating procedures on the general characteristics of subjects and facial characteristic variables. Facial images were collected and facial feature landmarks were extracted through using image processing techniques reproduced from our previous study [17,18]. Due to the sensitivity on illumination changes, the red (R), green (G), and blue (B) values of each pixel in the complexion regions were extracted and converted into L*, a*, b* values, where they express color numerically; L* values represent lightness, while a* and b* values represent green-red and blue-yellow color components.
Facial landmarks and facial color areas used for analysis were illustrated in Figure 1. For the facial landmarks of frontal view, the numberings shown were first marked on the right side of the face. Using the similar landmarks numberings on the right side of the face, 1 was added for the landmarks’ numberings on the left side of the face. For instance, landmark number 10 was the upper eyebrow point on the right and landmark number 110 was the upper eyebrow point on the left. Descriptions of the facial landmarks were presented in Table 1. The variables that consist of two or more landmark numbers were defined using length, ratio, width, or angle as listed in Table 2. For example, FD_10_110 indicated the distance between the upper eyebrow point on the right and between the upper eyebrow points on the left in the frontal picture.

2.3. Group Classification

For the diagnosis of hypertension, we referred to the 2018 Korean society of hypertension guideline for the management of hypertension where hypertension is defined using a threshold of >140/90 mmHg [19]. On the other hand, physician-diagnosed hypertension defined the subjects who were diagnosed with hypertension at least once by a doctor but did not take any medications and/or were currently taking antihypertension medicine to control their BP. Hence, the hypertension group in this study included subjects with systolic BP (SBP) readings of ≥140 mmHg and/or diastolic BP (DBP) readings of ≥90 mmHg as well as physician-diagnosed hypertension, regardless of whether the subjects did or did not take medications. Those subjects with BP readings of <120 mmHg and <80 mmHg, including those who had fully recovered from hypertension through treatment, were classified in the normal group.

2.4. Statistical Analysis

IBM SPSS Statistics 23.0 for Windows (IBM, Armonk, New York, NY, USA) and R version 3.3.3 (Another Canoe, The R Foundation for Statistical Computing, Vienna, Austria, 2017) were used for all the analyses in this study.
Two statistical tests, independent two-sample t-tests and analysis of covariance (ANCOVA), were performed using IBM SPSS Statistics 23.0 for Windows (IBM, Armonk, New York, NY, USA) at a significant level of 0.05. A t-test was used to test the differences between the hypertension group and normal group in general characteristics such as age, height, weight, BMI, SBP, and DBP for both genders. The statistical results of t-tests were presented as mean (standard deviation). The ANCOVA, with age and BMI as covariates, were performed in each men and women to compare the differences of facial characteristic variables between the hypertension group and normal group. The statistical results of ANCOVA were presented as adjusted mean (standard error) by age and BMI.
Additionally, statistical analysis using the Least Absolute Shrinkage and Selection Operator (LASSO) and receiver operating characteristic (ROC) curve analysis were con-ducted using R version 3.3.3 (Another Canoe, The R Foundation for Statistical Computing, Vienna, Austria, 2017). As combined facial characteristic variables are correlated, LASSO, as a variable subset selection method in logistic regression, was applied for selecting reliable combined indices while assessing the association between hypertension and combined facial characteristics variables. The LASSO penalty selects some of the predictors and discards the others if they are correlated [20,21,22,23]. In this study, the tuning parameter λ in the objective function of LASSO was obtained using the R function cv.glmnet, which is the function used to perform cross-validation for tuning selection [23,24].
The values of the area under the ROC curve (AUC) and the confidence intervals of AUC to evaluate the performance of prediction models were calculated using 5-fold cross-validation. In 5-fold cross-validation, the data set is divided into five subsets and the test method is repeated five times. Each time 1 of the 5 subset is used for a test set, and the other 4 subsets are training sets. The average accuracy across all 5 tests is computed. In this study, we obtained the estimates and confidence intervals of AUC using the R function ci.cvAUC, which calculates the influence curve-based confidence intervals for cross-validated AUC estimates [25].

3. Results

3.1. General Characteristics of the Subjects

The general characteristics of the subjects are summarized to compare the differences between the hypertension group and the normal group for both men and women in general.
The number of subjects in this research study was 1099 in total. There are 173 subjects in the hypertension group and 203 subjects in the normal group of men and 221 subjects in the hypertension group and 502 subjects in the normal group of women. Significant differences between the hypertension group and the normal group were observed in most of the general characteristics in both men and women, except for the pulse as shown in Table 3.
The mean age of men subjects was 43.66 ± 15.84 years in the normal group and 55.30 ± 11.97 years in hypertension group, whereas the mean age of women subjects was 44.93 ± 14.15 years in normal group and 61.64 ± 11.62 years in hypertension group. In both men and women, age, body weight, and BMI, SDP, and DBP were significantly higher in the hypertension group than in the normal group.

3.2. Differences in Facial Characteristics among the Subject Groups

After adjustment for age and BMI, there were 10 variables in facial features that showed significant differences for both genders, whereas there were three facial color variables for men and six facial color variables for women had significant differences in the facial characteristics between hypertension and normal group. Out of the 10 facial feature variables in both men and women, six facial feature variables in men were defined by the facial shape and five facial feature variables in women were defined by eyes (Table 4 and Table 5). For facial color variables, all three variables in men were cheek color variables, and three out of five variables in women were forehead color variables (Figure 2).

3.3. Association between Hypertension and Facial Characteristics

The LASSO method was performed to select the facial characteristic variables that could be used as hypertension predictors. In this study, several facial characteristic variables were selected by LASSO analysis. In general, age and BMI were selected for both men and women. All facial characteristics variables selected by LASSO were shown in Table 6 and Table 7. Illustrative figures of facial characteristics were shown in Figure 3.

3.4. Accuracy of Prediction Models

Both ROC and AUC metrics were used for evaluating the prediction model’s performance. ROC analysis is used for evaluating the accuracy of the prediction model that classifies subjects into hypertension or non-hypertension, whereas AUC summarized the prediction accuracy. In addition, 5-fold cross-validation was also performed to obtain the estimates and confidence intervals for the AUC.
The AUC (95%CI) value in the women is 0.827 (0.794–0.860) The AUC value for men is slightly lower than total women with the AUC (95% CI) value of 0.706 (0.652–0.760) as shown in Table 8. The ROC curves for LASSO using 5-fold cross-validation were shown in Figure 4.

4. Discussion

In our study, we utilized the facial landmarks to enable us to investigate the difference in the facial characteristics of hypertension patients and normal subjects. Facial landmarks extracted from the automatic extraction program will provide us with the facial feature characteristics of our subjects, whereas CIELAB color space values will provide us with the subjects’ facial color characteristics. At the same time, this will enable us to make use of facial recognition technology in medical research.
In this study, facial characteristics, as well as demographic characteristics, were used as variables to predict hypertension. By using the facial landmarks and CIELAB color system, our study revealed the role of facial characteristics in predicting the of hypertension. Our study obtained several interesting results after we statistically analyzed the associations between facial characteristics indices and hypertension together with demographic characteristics.
We found that facial shape width (jaw area) and nose length in men, in addition to eye shape and nose length in women, were the most distinct facial feature differences between hypertension patients and normal individuals. For facial colors, cheek color in men, as well as forehead and nose color in women, were the most distinct facial colors between the hypertension groups and normal individuals.
To find the predictors for hypertension using facial characteristics, we further analyzed our results using the LASSO method and evaluate the accuracy of the hypertension prediction model. According to our results, facial angle, eye length and angle, mouth ratio, forehead shape, nose tilting angle, forehead color, and cheek color variables were selected by the predicting model. As this study is in the exploratory stage, we only manage to explore all the variables related to hypertension. A larger and more balanced sample size is needed to formulate a predicting model thoroughly.
Looking at the AUC values, the case model generated for women is the better case model with the AUC (95%CI) value of 0.827 (0.794–0.860). In comparison to women, men had a lower AUC value (AUC (95%CI) value of 0.706 (0.652–0.760)), and we assumed that it may be partly due to the low sample size. In general, both case models of men and women in the prediction of hypertension can be considered fairly well. This indicates that there might be a possibility in predicting hypertension using facial characteristics.
This study examined the possibility of predicting hypertension in a non-contact method. We believed that this study can be further developed into a total unconscious screening of hypertension on a daily life basis. For example, the daily self-screening of hypertension can be performed simply by looking at the mirror which could analyze the facial characteristics of the individual. Such method can provide an alternative method to patients who were unable to monitor hypertension using labor-intensive conventional methods on a daily basis. This can also minimize health care costs through early detection of hypertension and reach a target population with minimum inconvenience to patients. At the same time, it can ensure that diagnoses of hypertension are not overlooked or discovered too late.
There are several limitations to this study. First, the sample size of the hypertension group and the normal group is unbalanced, which might introduce bias. Second, data from this study were collected by different technicians because the subjects were recruited from multiple sites across the country. Although all technicians were well-trained and followed a standard procedure, inconsistency may nonetheless exist in the data collection due to individual differences. In addition, there might be several confounding factors for hypertension required to generalize the study that we may not able to consider. Therefore, further studies should include balanced sample size, and the consistency in data collection should be increased. These studies should also include other known factors for hypertension.
We believed that this research work will contribute to the development of the study of hypertension in four important ways: first, by providing a wide perspective on the issues pertinent to the screening methods of hypertension; second, by critically examining the possibility of bringing a simpler approach on hypertension screening using facial characteristics; third, by analyzing both facial features and facial colors of hypertension individuals through various methods, allowing a meaningful comparison between facial characteristics of hypertension individuals and non-hypertension individuals; and, fourth, by formulating a prediction model that enables a non-contact and unconscious screening of hypertension on a daily life basis.

5. Conclusions

In conclusion, the exploration of the possibility of predicting hypertension using facial characteristics can be considered accomplished. We found that facial characteristics variables were associated with hypertension. Although more studies and research are still needed to fully develop such technology, this study provided a shred of new evidence that the prediction of hypertension using facial characteristics could be valid and advanced into an effective screening program. In the future, we aimed to recruit more subjects and conduct more research to create a real-time measurement platform based on this technology with the hope to contribute to the field of big data as well as the new era of artificial intelligence.

Author Contributions

Conceptualization: S.L. and J.-H.D. Methodology: L.A. and M.H.Y. Data Curation: M.H.Y. Writing—Original Draft: L.A. Writing—Review and Editing: S.L. Visualization: L.A. Supervision: S.L. Project Administration: S.L. and J.-H.D. Funding Acquisition: J.-H.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Collection of Clinical Big Data and Construction of Service Platform for Developing Korean Medicine Doctor with Artificial Intelligence research project (KSN2013110) and supported by the Bio & Medical Technology Development Program of the National Research Foundation of Korea (NRF-2015M3A9B6027139).

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Institutional Review Board of Korea Institute of Oriental Medicine (I-0910/02-001).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study. Written informed consent has been obtained from the patient(s) to publish this paper.

Data Availability Statement

The data used to support the findings of this study are available from the corresponding author upon request.

Acknowledgments

We would like to thank Jong Yeol Kim for his support in this project.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Korea Statistical Information Service (KOSIS). Statistics for Number of Deaths due to Hypertension. Available online: www.kosis.kr (accessed on 23 January 2021).
  2. Lee, S.W.; Lee, H.-Y.; Ihm, S.H.; Park, S.H.; Kim, T.H.; Kim, H.C. Status of hypertension screening in the Korea National General Health Screening Program: A questionnaire survey on 210 screening centers in two metropolitan areas. Clin. Hypertens 2017, 23, 23. [Google Scholar] [CrossRef]
  3. National Health Insurance Corporation (NHIS). Medical Service Usage Statistics by Region. Available online: http://www.nhis.or.kr/bbs7/boards/B0075 (accessed on 23 January 2021).
  4. Tsoi, K.; Yiu, K.; Lee, H.; Cheng, H.-M.; Wang, T.-D.; Tay, J.-C.; Teo, B.W.; Turana, Y.; Soenarta, A.A.; Sogunuru, G.P.; et al. Applications of artificial intelligence for hypertension management. J. Clin. Hypertens 2021. [Google Scholar] [CrossRef] [PubMed]
  5. Monte-Moreno, E. Non-invasive estimate of blood glucose and blood pressure from a photoplethysmograph by means of machine learning techniques. Artif. Intell. Med. 2011, 53, 127–138. [Google Scholar] [CrossRef] [PubMed]
  6. Esmaelpoor, J.; Moradi, M.H.; Kadkhodamohammadi, A. A multistage deep neural network model for blood pressure estimation using photoplethysmogram signals. Comput. Biol. Med. 2020, 120, 103719. [Google Scholar] [CrossRef]
  7. Miao, F.; Wen, B.; Hu, Z.; Fortino, G.; Wang, X.P.; Liu, Z.D.; Tang, M.; Li, Y. Continuous blood pressure measurement from one-channel electrocardiogram signal using deep-learning techniques. Artif. Intell. Med. 2020, 108, 101919. [Google Scholar] [CrossRef] [PubMed]
  8. Ankışhan, H. Blood pressure prediction from speech recordings. Biomed. Signal Process. Control 2020, 58, 101842. [Google Scholar] [CrossRef]
  9. Stephen, I.D.; Hiew, V.; Coetzee, V.; Tiddeman, B.P.; Perrett, D.I. Facial shape analysis identifies valid cues to aspects of physiological health in Caucasian, Asian, and African populations. Front. Psychol. 2017, 8. [Google Scholar] [CrossRef] [Green Version]
  10. Stephen, I.D.; Law Smith, M.J.; Stirrat, M.R.; Perrett, D.I. Facial skin coloration affects perceived health of human faces. Int. J. Primatol. 2009, 30, 845–857. [Google Scholar] [CrossRef] [Green Version]
  11. O’Higgins, P.; Bastir, M.; Kupczik, K. Shaping the human face. Int. Congr. Ser. 2006, 1296, 55–73. [Google Scholar] [CrossRef]
  12. Nunes, L.; Souza de Jesus, A.; Augusto Casotti, C.; Araújo, E. Geometric morphometrics and face shape characteristics associated with chronic disease in the elderly. Biosci. J. 2018, 34, 1035–1046. [Google Scholar] [CrossRef] [Green Version]
  13. Demayo, C.; Torres, M.; Veña, C. Face shapes of diabetics and non-diabetics described using geometric morphometrics. Internet J. Endocrinol. 2009, 6, 1–12. [Google Scholar] [CrossRef]
  14. Solon, C.C.; Torres, M.A.; Demayo, C. Analyzing shape of faces of hypertensive and non-hypertensive males using geometric morphometric methods. J. Med. Bioeng. 2013, 2, 126–130. [Google Scholar] [CrossRef]
  15. Sweet, E.; McDade, T.W.; Kiefe, C.I.; Liu, K. Relationships between skin color, income, and blood pressure among African Americans in the CARDIA study. Am. J. Public Health 2007, 97, 2253–2259. [Google Scholar] [CrossRef]
  16. Gravlee, C.C.; Dressler, W.W.; Bernard, H.R. Skin color, social classification, and blood pressure in Southeastern Puerto Rico. Am. J. Public Health 2005, 95, 2191–2197. [Google Scholar] [CrossRef]
  17. Do, J.H.; Ku, B.; Jang, J.S.; Kim, H.; Kim, J.Y. Analysis of Sasang constitutional types using facial features with compensation for photographic distance. Integr. Med. Res. 2012, 1, 26–35. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Ang, L.; Kim, J.Y.; Lee, J. Analysis of facial features according to Sasang types between native Japanese and native Korean populations. Evid Based Complement Altern. Med. 2018, 2018, 8. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Larson, S.; Cho, M.-C.; Tsioufis, K.; Yang, E. Korean society of hypertension guideline for the management of hypertension: A comparison of American, European, and Korean blood pressure guidelines. Eur. Heart J. 2020, 41, 1384–1386. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Meier, L.; Van De Geer, S.; Bühlmann, P. The group lasso for logistic regression. J. R. Stat. Soc. Ser. B Stat Methodol. 2008, 70, 53–71. [Google Scholar] [CrossRef] [Green Version]
  21. Friedman, J.; Hastie, T.; Tibshirani, R.; Simon, N.; Narasimhan, B.; Qian, J. Lasso and Elastic-Net Regularized Generalized Linear Models. Available online: https://glmnet.stanford.edu/ (accessed on 26 February 2021).
  22. Tibshirani, R. Regression Shrinkage and Selection via the Lasso. J. R. Stat. Soc. Ser. B Stat. Methodol. 1996, 58, 267–288. [Google Scholar] [CrossRef]
  23. Hastie, T.; Qian, J. Glmnet Vignette. Available online: https://web.stanford.edu/~hastie/glmnet/glmnet_alpha.html (accessed on 26 February 2021).
  24. Friedman, J.; Hastie, T.; Tibshirani, R. Regularization Paths for Generalized Linear Models via Coordinate Descent. J. Stat. Softw. 2010, 33, 1–22. [Google Scholar] [CrossRef] [Green Version]
  25. LeDell, E.; Petersen, M.; van der Laan, M. Computationally efficient confidence intervals for cross-validated area under the ROC curve estimates. Electron. J. Stat. 2015, 9, 1525, 1583–1607. [Google Scholar] [CrossRef]
Figure 1. Facial features’ landmarks and facial color areas used for analysis.
Figure 1. Facial features’ landmarks and facial color areas used for analysis.
Applsci 11 02414 g001
Figure 2. Illustrative figures of facial characteristics for both genders.
Figure 2. Illustrative figures of facial characteristics for both genders.
Applsci 11 02414 g002
Figure 3. Illustration of selected variables by LASSO for both genders.
Figure 3. Illustration of selected variables by LASSO for both genders.
Applsci 11 02414 g003
Figure 4. ROC curves for LASSO using 5-fold cross-validation in the analysis of the combination of facial feature and facial color. Black lines show individual ROC curves for 5-fold cross-validation, and the red line shows the average of individual ROC curves.
Figure 4. ROC curves for LASSO using 5-fold cross-validation in the analysis of the combination of facial feature and facial color. Black lines show individual ROC curves for 5-fold cross-validation, and the red line shows the average of individual ROC curves.
Applsci 11 02414 g004
Table 1. Desciptions of facial feature landmarks.
Table 1. Desciptions of facial feature landmarks.
Frontal FaceProfile Face
NumberNameNumberName
10Upper eyebrow point6Hairline point
17Upper eyelid point7Mid forehead point
18Inner eye angle point9Lower forehead point
20Inner pupil point10Eyebrow point
21Pupil point12Top nasal point
22Outer pupil point13Nasal point
25Outer eye lid point14Nose apex point
26Lower eye lid point21Under nose point
33Horizontal facial end point22Upper lip point
36Outer nostril point25Mid lip point
37Inner nostril point32Upper chin point
38Philtrum point33Lower chin point
43Lower jaw angle point36Hyoid point
47Hairline point71Forehead 1st point
50Mid lip point72Forehead 2nd point
52Glabella point73Forehead 3rd point
53Right eye horizontal end point77Forehead horizontal point
80Upper lip point84Nasal inclination staring point
81Septum point87Nasal inclination 1st point
9010 cm starting point88Nasal inclination 2nd point
9110 cm end point9010 cm starting point
94Right nose horizontal end point9110 cm end point
Table 2. Descriptions on facial landmarks’ variables [18].
Table 2. Descriptions on facial landmarks’ variables [18].
VariablesDescription
FD_n1_n2 [or PD_n,_n2]The length between two points in the frontal/profile picture
FDH_n,_n2 [or PDH_n1_n2]The horizontal length between two points in the frontal/profile picture
FDV_n1_n2 [or PDV_n1_n2]The vertical length between two points in the frontal/profile picture
FDL_n1_n2_n3
[or PDL_n1_n2_n3]
The length between the point n1 and segments n2, n3
FHD_n1_n2_n3_n4
[or PHD_n1_n2_n3_n4]
FDH_n1_n2/FD_n3_n4
[or PDH_n1_n2/PD_n3_n4]
FDH_n1_n2_n3_n4
[or PDH_n1_n2_n3_n4]
FD_n1_n2/FDH_n3_n4
[or PD_n1_n2/PDH_n3_n4]
FDD_n1_n2_n3_n4
[or PDD_n1_n2_n3_n4]
FD_n1_n2/FD_n3_n4
[or PD_n1_n2/PD_n3_n4]
FVD_n1_n2_n3_n4
[or PVD_n1_n2_n3_n4]
FDV_n1_n2/FD_n3_n4
[or PDV_n1_n2/PD_n3_n4]
FVV_n1_n2_n3_n4
[or PVV_n1_n2_n3_n4]
FDV_n1_n2/FDV_n3_n4
[or PDV_n1_n2/PDV_n3_n4]
FVH_n1_n2_n3_n4
[or PVH_n1_n2_n3_n4]
FDV_n1_n2/FDH_n3_n4
[or PDV_n1_n2/PDH_n3_n4]
FA_n1_n2 [or PA_n1_n2]The angle that the straight-line vector (n1, n2) makes with the horizontal line in the frontal/profile image
FAs_n1_n2 [or PA_n1_n2]180—The angle that the straight-line vector (n1, n2) makes with the horizontal line in the frontal/profile image
FAi_n1_n2 [or PAi_n1_n2]The angle that the straight-line vector (n1, n2) makes with the horizontal line in the frontal/profile image * (−1)
FAis_ n1_n2 [or PAis_n1_n2]180—The angle that the straight-line vector (n1, n2) makes with the horizontal line in the frontal/profile image
FA_n1_n2_n3
[or PA_n1_n2_n3 ]
The angle formed by the three points n1, n2, and n3 in the frontal/profile image
FArea02The area of the face defined using points 53, 94, 194, and 153
FArea03The area of the face defined using points 94, 43, 143, and 194
F: frontal view variables; P: profile view variables.
Table 3. General characteristics of subjects.
Table 3. General characteristics of subjects.
Men (n = 376)p-ValueWomen (n = 723)p-Value
Normal (n = 262)Hypertension (n = 114)Normal (n = 560)Hypertension (n = 163)
Age [yrs]43.66 ± 15.8455.30 ± 11.97<0.001 **44.93 ± 14.1561.64 ± 11.62<0.001 **
Height [cm]171.11 ± 6.41169.34 ± 6.170.012 *158.68 ± 5.63154.98 ± 5.900.002 **
Weight [kg]69.35 ± 10.4572.26 ± 10.270.013 *57.05 ± 8.3259.48 ± 8.82<0.001 **
BMI [kg/m2]23.64 ± 2.9725.17 ± 2.93<0.001 **22.66 ± 3.1124.74 ± 3.19<0.001 **
SBP [mmHg]119.81 ± 15.01131.24 ± 15.75<0.001 **113.45 ± 15.01127.72 ± 16.29<0.001 **
DBP [mmHg]77.42 ± 11.7283.32 ± 11.78<0.001 **73.23 ± 10.8180.58 ± 10.920.039 *
BMI, Body mass index; SBP, systolic blood pressure; DBP, diastolic blood pressure. Data are represented in the mean ± SD (standard deviation). * p < 0.05; ** p < 0.01. p-values were obtained from t-tests between the normal group and the hypertension group.
Table 4. Differences in facial characteristic variables in men.
Table 4. Differences in facial characteristic variables in men.
Variables DescriptionsVariablesMen (n = 376)
Normal (n = 262)Hypertensive (n = 114)p-Value
Frontal faceMiddle face length and ratioFDV_52_5077.51 ± 0.3276.28 ± 0.490.043
FVD_52_50_53_1530.51 ± 0.000.50 ± 0.000.002
FVD_52_81_53_1530.32 ± 0.000.31 ± 0.000.003
Middle face angle and areaFA_117_125_143121.06 ± 0.38122.62 ± 0.600.033
FA_18_43_5057.00 ± 0.2255.93 ± 0.340.010
FA_18_94_5060.65 ± 0.2759.16 ± 0.420.004
Ratio of eye variableFHD_25_125_53_1530.68 ± 0.000.67 ± 0.000.048
Profile faceNose width and areaPDH_12_1420.66 ± 0.1419.97 ± 0.220.012
PDV_12_2151.82 ± 0.2450.86 ± 0.370.035
PD_12_2152.20 ± 0.2451.20 ± 0.380.031
Facial colorCheck colorChRD_b20.53 ± 0.2519.53 ± 0.390.036
ChLU_L60.03 ± 0.3461.53 ± 0.540.024
ChLW_L57.50 ± 0.3759.04 ± 0.570.029
Data are represented as adjusted mean ± SE (standard error). p-values were calculated using ANCOVA with a significant level of p < 0.05.
Table 5. Differences in facial characteristic variables in women.
Table 5. Differences in facial characteristic variables in women.
Variable DescriptionsVariablesWomen (n = 723)
Normal (n = 560)Hypertensive (n = 163)p-Value
Frontal faceFace contour angleFAs_153_19488.43 ± 0.1189.04 ± 0.230.020
Eye length and ratioFD_17_269.27 ± 0.068.81 ± 0.110.001
FD_117_1269.36 ± 0.069.02 ± 0.120.011
FDH_17_26_18_250.29 ± 0.000.28 ± 0.000.007
FDD_17_26_52_810.21 ± 0.000.20 ± 0.000.003
FDD_17_26_52_500.13 ± 0.000.12 ± 0.000.002
Profile faceForehead shapePAi_71_7278.55 ± 0.2577.30 ± 0.490.029
Nose length and areaPDH_14_2113.29 ± 0.0712.96 ± 0.130.026
PA_14_2145.92 ± 0.1947.12 ± 0.380.007
PA_12_14_21107.81 ± 0.22109.02 ± 0.430.017
Facial colorForehead colorFhR_L69.97 ± 0.2468.62 ± 0.470.015
FhR_a10.24 ± 0.1110.77 ± 0.210.030
FhL_a10.07 ± 0.1010.74 ± 0.210.006
FhW_a10.15 ± 0.1010.76 ± 0.200.009
Cheek colorChRU_L65.07 ± 0.2163.49 ± 0.420.002
ChRW_L61.61 ± 0.2260.21 ± 0.430.005
Nose colorNose_L70.22 ± 0.2468.85 ± 0.480.016
Nose_a11.87 ± 0.1112.55 ± 0.220.009
Data are represented as adjusted mean ± SE (standard error). p-values were calculated using ANCOVA with a significant level of p < 0.05.
Table 6. Variables selected using a model based on LASSO in men.
Table 6. Variables selected using a model based on LASSO in men.
Variable DescriptionsSelected Variables
BMI
AGE
Frontal faceFacial width, ratio, and angle FD_53_153
FVD_52_81_53_153
FA_117_125_143
FA_18_94_50
Eye length, ratio, and angleFD_117_126
FHD_18_118_53_153
FHD_25_125_53_153
FAis_125_117
Mouth ratioFVV_80_50_81_50
Profile faceForehead shapePAi_9_7
PDV_9_12
Nose horizontal distance and tilting anglePDH_12_14
PA_87_21
Facial colorForehead colorFhR_a
Cheek colorChRD_b
ChLU_L
Table 7. Variables selected using a model based on LASSO in women.
Table 7. Variables selected using a model based on LASSO in women.
Variable DescriptionsSelected Variables
BMI
AGE
Frontal faceFacial shape ratio and angleFVV_47_52_81_50
FAs_153_194
FA_118_125_143
FA_17_25_43
FA_117_125_143
FA_18_25_94
Eye length and ratioFD_17_26
FDD_17_26_52_50
Mouth ratioFVV_80_50_52_50
Profile faceForehead shapePAi_71_72
PA_10_12
Nose tilting anglePA_14_21
PA_12_14_21
Face widthPDH_44_53
Facial colorForehead colorFhL_a
Cheek colorChRU_L
Table 8. Calculated AUC values.
Table 8. Calculated AUC values.
GendernNumber of Selected Variables AUC (95% CI)Grading
Men376180.706 (0.652–0.760)Fair
Women723180.827 (0.794–0.860)Good
AUC academic point system: 0.90–1 = excellent (A); 0.80–0.90 = good (B); 0.70–0.80 = fair (C); 0.60–0.70 = poor (D); 0.50–0.60 = fail (F).
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ang, L.; Yim, M.H.; Do, J.-H.; Lee, S. A Novel Method in Predicting Hypertension Using Facial Images. Appl. Sci. 2021, 11, 2414. https://doi.org/10.3390/app11052414

AMA Style

Ang L, Yim MH, Do J-H, Lee S. A Novel Method in Predicting Hypertension Using Facial Images. Applied Sciences. 2021; 11(5):2414. https://doi.org/10.3390/app11052414

Chicago/Turabian Style

Ang, Lin, Mi Hong Yim, Jun-Hyeong Do, and Sanghun Lee. 2021. "A Novel Method in Predicting Hypertension Using Facial Images" Applied Sciences 11, no. 5: 2414. https://doi.org/10.3390/app11052414

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop