The importance of the visual component of cosmetic products’ effects and aesthetic treatments goes beyond protocoling. Artificial intelligence (AI) and machine learning advances reshape the future of aesthetics and anti-ageing domains.1–11 Such digital innovations are recognized within cosmetic dermatology and are used to complement traditional skin evaluation methods. Our appearance has a significant impact on our self-perception and how we are perceived by others,12 both of which contribute significantly to our life satisfaction, overall well-being,13 and confidence.14
Clinical assessment, sensorial data, biophysical evaluation, and imaging methods15 are used as criteria to assess the skin condition and measure the effect of interventions.
In the aesthetic field, expert-level grading is quite well established, but it still utilizes a limited number of categories for grading and can inherit the biases of the expert.16,17 In practice, it is challenging to arrange patient evaluation by sensorial and biophysical measurements since it requires their physical attendance at a dedicated research center. The same applies to laboratory photography. However, as our smartphones have become an increasingly versatile computing device and an advanced imaging system in our pocket,18 we can anticipate a rapid rise in high-scale delivery of computer vision-based selfie diagnostics to patients (Table 1).
Table 1. -
Comparison of Automation Potential and Application of AI Algorithms with Regard to Different Skin Analysis Methods
Group of Methods |
Data Type |
Automation Potential with AI Methods |
Potential to Deliver Tech at Scale to End-Users |
Limitations |
Sensors |
Time series |
Medium |
Medium |
Results might be hard to reproduce19
|
Biophysical |
Time series |
Medium |
Low |
Requires special equipment Datasets have a limited number of data points |
Clinical evaluation |
Classes of severity |
Medium |
High |
Expert grading might be biased17,20
|
Imaging |
Images |
High |
High |
Requires data standardization21
|
Although selfie photographs are an essential attribute of social media, it remains difficult to capture strictly standardized images.16 Data quality is crucial to the performance of computer vision algorithms. Therefore, dedicated methods for image-quality control should be applied to achieve the adequate level of accuracy in the analysis.22 As an example, in Haut.AI’s research group, we use our LIQA (live image-quality assurance) software, which was developed in-house to analyze image quality live, as the image is being taken, and provide instant feedback to the user.
Once we have captured a standardized, high-quality image and processed it with computer vision methods, we get predictions from the algorithms’ models. The nature of these predictions or scores varies depending on the type of algorithm. Generally, the main tasks in biomedical imaging are object detection, segmentation of the region of interest, classification and ranking.23,24 For object detection, the output can be binary (object detected or not detected) or quantitative (counting the number of objects, eg, pigmentation spots). For segmentation, the algorithm tries to provide a mask of the target object, whereas in classification, it assigns a class or several classes to an image. The ranking model orders the images based on a partial ranking obtained from an annotation.
The presentation of the results of computer vision-based systems can be done using absolute value. Although these scores will be very useful for statistical research, they will not necessarily be grasped by the doctor and the patient.
To organize the data and format it in a suitable manner so scores from different subjects or different conditions can be compared, we need to apply a normalization technique. Normalization allows data to be presented across a specific range, for example, between 0 and 100. Normalization transforms the scores’ absolute values to relative values.
The absolute scores are calculated on a diverse dataset of a relatively large size and can be normalized to relative based on the extreme distribution of values in the dataset. This way, we will be able to compare results with the general population presented in the dataset. The normalization can be repeated as new high-quality image data is collected.
AI-based face analysis technology should aid both the doctor and the patient’s understanding of the patient’s skin and face condition, acting as an assistant for the doctor and an educator for the patient. In a study of the perception of using AI to interpret radiology data,25 it was found that the main benefits to patients were actionable information, second opinion, and preparing for the consultation. Patients’ perceptions and acceptance of using AI technology to interpret their photographic data will largely depend on how comprehensive, useful, and personalized it will be.
The differences in skin properties between genders26,27 are well-known. In this work, the author analyzes the distribution of skin parameter scores in the Beauty.AI dataset, which constitutes an anonymized collection of skin data derived from selfies. The aim of this study is to analyze the difference in the distribution of skin parameters in men and women of different ages and explore whether the same scoring scales for different age and gender groups can be used.
METHODS
Image Processing
Image-quality analysis and standardization were performed by Haut.AI’s Image Quality software.28 First, the system checked the position of the face and the head’s rotation angle on the x (yaw), y (pitch) and z (roll) axes based on 400+ predicted facial keypoints. Then, the system standardized the face image by aligning the center of the face, calculated from the geometry of the facial features. Next, the following image-quality metrics were analyzed: face illumination, background illumination, presence of shadows and glares on the face, presence of visual noise, image resolution, and the resolution of the bounding box that included only the face area. Finally, the system eliminated all pixels not related to the skin to anonymize the image and remove all artifacts that could affect the calculation of skin parameters by computer vision algorithms.
To calculate the skin parameters for the face, we utilized Haut.AI’s Skin Metrics Report software.28 This computer vision-based software was trained on a total of three million data points, including synthetic and real image data. It utilizes neural network algorithms and computer vision methods, and it calculates 136 face and skin attributes. The following skin parameters were selected for analysis in this study: Sagging, Dark Circles, Eye Bags, Wrinkles, Pores, Uniformness, Acne, Pigmentation, Redness, and Translucency.
Dataset
The author used the private Beauty.AI dataset,29 which constitutes a collection of fully anonymized skin data derived from selfie pictures. Beauty.AI is a research project established in 2015 with the objective to develop algorithms for evaluating human appearance. Participants submitted their photos using Beauty.AI’s project mobile application and consented to the processing of those images for the evaluation of the algorithms. As no protected health information was accessed and all photographs were fully anonymized, institutional review board approval was not required.
The Beauty.AI dataset includes 17,700 selfies captured with smartphone cameras. From this dataset, the images of 433 subjects whose selfie photographs had sufficient image quality were selected. The images contained key facial landmarks, adequate illumination and head position, no face occlusion, no distortion and no image noise. The inclusion criteria were set to filter out photographs that lacked at least one quality requirement. The individual typology angle30 of subjects in the dataset ranged from −42° to 60°. The subjects were 18–67 years of age, with a mean age of 29.2 years. The age information was provided by the subjects. The dataset included 244 female and 189 male subjects.
Results Visualization
To visualize the results observed in the dataset, the boxplot function from the matplotlib31 plotting library for the Python32 programming language was used.
Statistical Analysis
To compare the skin metrics for different subject groups, the Mann–Whitney U test32,33 from the scipy34 analytics library for the Python programming language was used.
RESULTS
All participants were divided into six groups: women 18–30 years old (154 subjects), men 18–30 years old (123 subjects), women 30–45 years old (72 subjects), men 30–45 years old (45 subjects), women 45 years old and older (18 subjects), and men 30–45 years old and older (21 subjects). The gender classes were balanced, whereas the age classes were biased toward younger ages.
Haut.AI’s Skin Metrics Report software provides the numeric output for the skin parameters on a scale of 0 to 100, where 0 is the lowest score observed in the dataset and 100 is the best score observed in the dataset. The observed distribution in the dataset is presented in Figure, Supplemental Digital Content 1, which shows the distribution of the skin parameters for the face observed in the dataset in different age and gender groups. Color references for groups: light pink for women 18–30 years old, light blue for men 18–30 years old, magenta for women 30–45 years old, blue for men 30–45 years old, fuschia for women 45 years old and older, and dark blue for men 45 years old and older. The yellow line plot on the graphs connects the median values for every age group for each of criteria, https://links.lww.com/PRS/F373.
Gender Differences
The summarized results are presented in Table 2. First, we compared the results for the all-female subject groups and all-male subject groups. We found that the difference in the distribution was significant for all parameters except for dark circles and acne when comparing all women and men. The Sagging, Wrinkles, Uniformness, Pores, Pigmentation, and Redness scores were higher in women compared to men. However, the Eye Bags and Translucency scores were higher for men than for women.
Table 2. -
Mann–Whitney U Test Results for Face Parameter Scores in Different Gender Groups of the Same Age
*
Parameter |
Median Diff.F-M All Ages |
p
|
Median Diff. F-M <30 Years Old |
p
|
Median Diff. F-M 30–45 Years Old |
p
|
Median Diff. F-M 45+ Years Old |
p
|
Sagging |
18.5 |
<0.001
|
15.0 |
<0.001
|
18.0 |
<0.001
|
30.0 |
<0.001
|
Dark circles |
0.0 |
0.11 |
1.0 |
0.31 |
−1.0 |
0.25 |
6.5 |
0.08 |
Eye bags |
−5.0 |
<0.001
|
−4.0 |
<0.001
|
−3.5 |
0.02
|
−2.5 |
0.09 |
Wrinkles |
2.0 |
<0.001
|
1.0 |
<0.001
|
4.5 |
0.05
|
4.0 |
0.49 |
Pores |
12.0 |
<0.001
|
12.0 |
<0.001
|
14.0 |
<0.001
|
17.0 |
0.04
|
Uniformness |
10.0 |
<0.001
|
10.0 |
<0.001
|
5.0 |
0.07 |
20.0 |
0.01
|
Acne |
0.0 |
0.34 |
1.5 |
0.12 |
−2.0 |
0.03
|
−2.0 |
0.05
|
Pigmentation |
1.0 |
0.02
|
1.0 |
0.01
|
0.0 |
0.35 |
0.0 |
0.21 |
Redness |
3.0 |
0.01
|
5.0 |
0.02
|
1.0 |
0.33 |
5.5 |
0.19 |
Translucency |
−4.0 |
<0.001
|
−3.5 |
0.02
|
−2.0 |
0.03
|
−1.5 |
0.27 |
*The p values in bold font indicate statistical significance (P < 0.05).
The same results were observed when comparing the 18- to 30-year-old female and male groups.
In the 30- to 45-year-old groups, the Sagging, Wrinkles, and Pores scores were higher for women than for men. It was observed that Eye Bags, Acne, and Translucency scores were higher for men. For other parameters, the difference was not statistically significant.
In the groups where subjects were older than 45 years old, it was only in the Sagging, Pores, Uniformness and Acne parameters that the difference in scores was statistically significant. For women, the Sagging, Pores, and Uniformness scores were noticeably higher.
Age Groups Difference for Women
The results for differences in the parameter scores were analyzed for the different female (Table 3) and male (Table 4) age groups. We observed a statistically significant difference in the scores for the Sagging, Eye Bags, Wrinkles, Pores and Uniformness parameters in the group of women 18−30 years old when compared with the group of women 30−45 years old.
Table 3. -
Mann−Whitney U Test Results for Face Parameter Scores for Women of Different Age Groups
*
Parameter |
Median F |
Median F <30 |
Median F 30−45 |
Median F 45+ |
p, 30 vs 30-45 |
p, 30 vs 45+ |
p, 30−45 vs 45+ |
Sagging |
77 |
79 |
71 |
68 |
0.03
|
0.14 |
0.34 |
Dark circles |
61 |
62 |
60 |
63.5 |
0.13 |
0.47 |
0.31 |
Eye Bags |
57 |
60 |
53.5 |
50.5 |
<0.001
|
<0.001
|
0.05
|
Wrinkles |
97 |
97 |
95.5 |
76 |
<0.001
|
<0.001
|
<0.001
|
Pores |
90 |
92 |
89 |
85 |
0.01
|
0.03
|
0.34 |
Uniformness |
55 |
55 |
50 |
45 |
0.01
|
0.01
|
0.13 |
Acne |
97 |
97 |
96 |
98 |
0.39 |
0.05 |
0.03
|
Pigmentation |
97 |
97 |
96 |
95 |
0.07 |
0.01
|
0.06 |
Redness |
83 |
85 |
82 |
79.5 |
0.09 |
0.03
|
0.12 |
Translucency |
36 |
36 |
35 |
41.5 |
0.26 |
0.06 |
0.02
|
*The p values in bold font indicate statistical significance (P < 0.05).
Table 4. -
Mann−Whitney U Test Results for Face Parameter Scores in Men of Different Age Groups
*
Parameter |
Median, M |
Median, M <30 |
Median, M 30−45 |
Median, M >45 |
p, 30 vs 30−45 |
p, 30 vs 45+ |
p, 30−45 vs 45+ |
Sagging |
59 |
64 |
53 |
38 |
<0.001
|
<0.001
|
0.37 |
Dark circles |
61 |
61 |
61 |
57 |
0.12 |
0.02
|
0.17 |
Eye bags |
62 |
64 |
57 |
53 |
0.01
|
<0.001
|
0.11 |
Wrinkles |
95 |
96 |
91 |
72 |
0.01
|
<0.001
|
0.02
|
Pores |
78 |
80 |
75 |
68 |
0.02
|
0.01
|
0.19 |
Uniformness |
45 |
45 |
45 |
25 |
0.32 |
<0.001
|
<0.001
|
Acne |
97 |
95.5 |
98 |
100 |
0.01
|
<0.001
|
0.01
|
Pigmentation |
96 |
96 |
96 |
95 |
0.46 |
0.31 |
0.35 |
Redness |
80 |
80 |
81 |
74 |
0.44 |
0.03
|
0.04
|
Translucency |
40 |
39.5 |
37 |
43 |
0.48 |
0.14 |
0.15 |
*The p values in bold font indicate statistical significance (P < 0.05).
A statistically significant difference was observed in the same parameters when comparing the group 18−30 years old with the group older than 45 years old. It was also observed for the Pigmentation and Redness parameters.
When comparing the female 30−45 years old group with the group older than 45 years old, a statistically significant difference was observed for the Eye Bags, Wrinkles, Acne, and Translucency parameters. In all cases, when we observed statistically significant differences, the values of the parameters declined with age.
Age Groups Difference for Men
A statistically significant difference was observed in the group of men 18−30 years old when compared with the group 30−45 years old for the Sagging, Eye Bags, Wrinkles, Pores, and Acne parameters.
When comparing the group of men 18–30 years old with the group of men older than 45 years old, a statistically significant difference was observed for the same parameters mentioned above, as well as for Dark Circles, Uniformness and Redness.
When comparing the 30- to 45-year-old group with the group older than 45 years old, a statistically significant difference was identified for the Wrinkles, Uniformness, Acne, and Redness parameters. For most of the skin parameters, the scores declined with age when statistically significant results were observed.
DISCUSSION
Gender-specific scales for selected skin parameters have been discussed and suggested by several researchers in the past.26,35 However, based on the author’s previously published work,10,11 the accuracy of AI systems—including those for skin analysis—can benefit from population-specific algorithms and biomarkers that have also been confirmed using other research groups.36,37
Of all the skin parameters reviewed in this study, the least variability for both genders was observed for Dark Circles, Eye Bags, Acne, Pigmentation, and Redness. This finding suggests that the same scale can be used for men and women of different ages to compare the results for different individuals.
Subjects from different age groups tend to display considerable variation in their Sagging, Wrinkles, Pores, Uniformness, Redness, and Translucency features.
These observations indicate that the variability between individuals is large within different age and gender groups. It is suggested that the individual’s age and gender group should be taken into consideration when developing scales. In practice, it means that the absolute scores for skin parameters will be normalized to the minimal and maximum scores observed in each demographic group.
Facial aging is characterized by change in the face shape, texture, and color of the skin.38,39 Uniformness reflects how flawless the skin is, and it is represented by a combination of color and texture. The change in the Sagging parameter reflects changes in the skin’s elasticity, fat, muscle tissues, and the facial skeleton.40 A deterioration in the Pores parameter can indicate a loss of skin elasticity. The Pearson correlation coefficients between the chronological age and the Uniformness, Sagging, and Pores parameters were equal to −0.42, −0.4, −0.35, respectively.
The Pearson correlation coefficient between the chronological age and the Wrinkle score was equal to −0.56. It is notable that the difference in the median value for the 30- to 45-years old group and 45 years old and older group was the greatest for the Wrinkle parameter in both the male and female groups.
For most of the skin parameter criteria utilized in this study, we observed significant variability among the subjects. This kind of variability between individuals can be explained by the heterogeneity of the studied population. The sample covers people of different races, geography, and climatic conditions.41,42 It is also worth emphasizing the possibility of a genetic predisposition in terms of the pace of aging43,44 and the difference in lifestyles among individuals.45,46 It is also suggested in some previous studies that people who look younger than their chronological age could also have better health overall.47
Computer vision methods can become a powerful tool to measure the effectiveness of treatments administered topically and surgical interventions. Combined with mobile delivery, image-based face and skin analysis will allow remote patient screening and monitoring. The main value for the patient can be the second opinion, remote prescreening, and a greater understanding of the skin’s current condition.
The benefits for the doctor are the ability to save time by providing remote consultations, the preliminary report on the patient’s current skin profile, and being more informed when selecting the appropriate procedure. Selfie-based mobile image analysis is a cost-efficient tool since it does not require any expensive equipment. Software requires a lower level of investment compared to hardware. Another benefit of mobile image-based analysis is that the software analysis calculations can be done in a secure cloud environment and the results can be retrieved instantly within 2–3 seconds. The accessibility of selfies allows a larger number of data points to be collected, providing more insights into the skin’s dynamics. The implementation of selfie-based analysis as a new instrument will require an investment in patient education. First of all, the criteria of what images are appropriate for the analysis should be clearly communicated to the patient.
In this work, the author did not discuss the development of different Fitzpatrick skin type-specific and ethnicity-specific groups.48–51 However, the author is planning to extend their research to these areas.
Use of gender-specific and age-specific scales allows the patients’ scores to be compared with their reference groups. As a result, realistic expectations can be set about the skin results that can be achieved within a patient’s relevant demographic group. These realistic goals allow positive reinforcement and inspire patient optimism, both of which play important roles in the treatment process.52
REFERENCES
1. Hogarty DT, Su JC, Phan K, et al. Artificial intelligence in dermatology—where we are and the way to the future: A review. Am J Clin Dermatol. 2020;21:41–47.
2. Eapen BR. Artificial intelligence in dermatology: A practical introduction to a paradigm shift. Indian Dermatol Online J. 2020;11:881–889.
3. Liang X, Yang X, Yin S, et al. Artificial intelligence in plastic surgery: Applications and challenges. Aesthetic Plast Surg. 2021;45:784–790.
4. Chowdhry A. How My Beauty Matches is using machine learning to disrupt the $445 billion beauty market. Forbes, August 14, 2017. Available at:
https://www.forbes.com/sites/amitchowdhry/2017/08/14/how-my-beauty-matches-is-using-machine-learning-to-disrupt-the-445-billion-beauty-industry/. Accessed August 30, 2022.
5. Lambros V. Facial aging: A 54-year, three-dimensional population study. Plast Reconstr Surg. 2020;145:921–928.
6. Flament F, Jacquet L, Ye C. Artificial intelligence analysis of over half a million European and Chinese women reveals striking differences in the facial skin ageing process. J Eur Acad Dermatol Venereol. 2022;36:1136–1142.
7. Meron E, Thaysen M, Angeli S, et al. Meeting report: Aging research and drug discovery. Aging (Albany NY) 2022;14:530–543.
8. Galkin F, Mamoshina P, Kochetov K, et al. DeepMAge: A methylation aging clock developed with deep learning. Aging Dis. 2021;12:1252–1262.
9. Zhavoronkov A, Mamoshina P. Deep aging clocks: The emergence of AI-based biomarkers of aging and longevity. Trends Pharmacol Sci. 2019;40:546–549.
10. Bobrov E, Georgievskaya A, Kiselev K, et al. PhotoAgeClock: Deep learning algorithms for development of non-invasive visual biomarkers of aging. Aging (Albany NY). 2018;10:3249–3259.
11. Georgievskaya A, Tlyachev T, Kiselev K, et al. 14086 A new multimodal age prediction image analysis method from hands images of different age groups by neural network model. J Am Acad Dermatol. 2020;83(6S):AB18.
12. Spisak BR, Grabo AE, Arvey RD, et al. The age of exploration and exploitation: Younger-looking leaders endorsed for change and older-looking leaders endorsed for stability. Leadersh Q 2014;25:805–816.
13. Linn MW, Hunter K. Perception of age in the elderly. J Gerontol. 1979;34:46–52.
14. Hong J, Koo B, Koo J. The psychosocial and occupational impact of chronic skin disease. Dermatol Ther. 2008;21:54–59.
15. Serup J, Jemec GBE, Grove GL. Handbook of Non-Invasive Methods and the Skin. 2nd ed. Boca Raton, Fla. CRC Press; 2006.
16. Nkengne A, Bertin C, Stamatas GN, et al. Influence of facial skin attributes on the perceived age of Caucasian women. J Eur Acad Dermatol Venereol. 2008;22:982–991.
17. Dehon H, Brédart S. An “other-race” effect in age estimation from faces. Perception. 2001;30:1107–1113.
18. Holland P. The world in your pocket: How smartphones will get smarter in 2022. CNET; 2022. Available at
https://www.cnet.com/tech/mobile/the-world-in-your-pocket-how-smartphones-get-smarter-in-2022/.
19. Alanen E, Nuutinen J, Nicklén K, et al. Measurement of hydration in the stratum corneum with the MoistureMeter and comparison with the corneometer. Skin Res Technol. 2004;10:32–37.
20. Wright DB, Sladden B. An own gender bias and the importance of hair in face recognition. Acta Psychol (Amst). 2003;114:101–114.
21. Henderson JL, Larrabee WF Jr, Krieger BD. Photographic standards for facial plastic surgery. Arch Facial Plast Surg. 2005;7:331–333.
22. Sang J, Lei Z, Li SZ. Face image quality evaluation for ISO/IEC standards 19794-5 and 29794-5. Adv Biometrics. 2009;229–238.
23. Schlett T, Rathgeb C, Henniger O, Galbally J, Fierrez J, Busch C. Face image quality assessment: A literature survey. ACM Comput Surv. 2020.
24. Esteva A, Chou K, Yeung S, et al. Deep learning-enabled medical computer vision. NPJ Digit Med. 2021;4:5.
25. Zhang Z, Citardi D, Wang D, et al. Patients’ perceptions of using artificial intelligence (AI)-based technology to comprehend radiology imaging data. Health Informatics J. 2021;27:146045822110112.
26. Tur E, Maibach HI. Gender and Dermatology. Cham, Switzerland: Springer; 2018.
27. Flament F, Amar D, Feltin C, Bazin R. Evaluating age-related changes of some facial signs among men of four different ethnic groups. Int J Cosmet Sci. 2018;40:502–515.
28. Haut.AI. Computer vision for skincare. Available at
https://www.haut.ai.
29. Beauty.AI. Beauty.AI 2.0. Available at:
http://beauty.ai. Accessed August 30, 2022.
30. Chardon A, Cretois I, Hourseau C. Skin colour typology and suntanning pathways. Int J Cosmet Sci. 1991;13:191–208.
31. Hunter J. D. Matplotlib: A 2D graphics environment. Comput Sci Eng 2007;9:90–95.
32. van Rossum G, Python Development Team. Python Tutorial: Release 3.6.4. Suwanee, Ga.: 12th Media Services, 2018.
33. Mann HB, Whitney DR. On a test of whether one of two random variables is stochastically larger than the other. Ann Math Statist. 1947;18:50–60.
34. Virtanen P, Gommers R, Oliphant T, et al. SciPy 1.0: Fundamental algorithms for scientific computing in Python. Nat Methods, 2020;17:261–272.
35. Chien AL, Qi J, Cheng N, et al. Perioral wrinkles are associated with female gender, aging, and smoking: Development of a gender-specific photonumeric scale. J Am Acad Dermatol. 2016;74:924–930.
36. Mamoshina P, Kochetov K, Putin E, et al. Population specific biomarkers of human aging: A big data study using South Korean, Canadian, and Eastern European patient populations. J Gerontol A Biol Sci Med Sci. 2018;73:1482–1490.
37. Moskalev A. Biomarkers of Human Aging. Cham, Switzerland: Springer Nature; 2019.
38. Nkengne A, Bertin C. Aging and facial changes-documenting clinical signs, part 2: Methods of documentation of facial changes. Skinmed. 2016;14:429–435.
39. Nkengne A, Bertin C. Aging and facial changes–documenting clinical signs, part 1: Clinical changes of the aging face. Skinmed. 2013;11:281–286.
40. Mendelson B, Wong CH. Changes in the facial skeleton with aging: implications and clinical applications in facial rejuvenation. Aesthetic Plast Surg. 2020;44:1151–1158.
41. Engebretsen KA, Johansen JD, Kezic S, et al. The effect of environmental humidity and temperature on skin barrier function and dermatitis. J Eur Acad Dermatol Venereol. 2016;30:223–249.
42. Singh B, Maibach H. Climate and skin function: an overview. Skin Res Technol. 2013;19:207–212.
43. Naval J, Alonso V, Herranz MA. Genetic polymorphisms and skin aging: the identification of population genotypic groups holds potential for personalized treatments. Clin Cosmet Investig Dermatol. 2014;7:207–214.
44. Zhan Y, Hägg S. Association between genetically predicted telomere length and facial skin aging in the UK Biobank: A Mendelian randomization study. Geroscience. 2021;43:1519–1525.
45. Krutmann J, Bouloc A, Sore G, et al. The skin aging exposome. J Dermatol Sci. 2017;85:152–161.
46. Molina-García M, Granger C, Trullàs C, et al. Exposome and skin: Part 1. Bibliometric analysis and review of the impact of exposome approaches on dermatology. Dermatol Ther (Heidelb). 2022;12:345–359.
47. Uotinen V, Rantanen T, Suutama T. Perceived age as a predictor of old age mortality: A 13-year prospective study. Age Ageing. 2005;34:368–372.
48. Bazin R, Doublet E. Skin Aging Atlas: Caucasian Type. Paris: Med’Com; 2007.
49. Bazin R, Flament F, Giron F. Skin Aging Atlas: Volume 3, Afro-American Type. Paris: Med’Com; 2012.
50. Bazin R, Flament F. Skin Aging Atlas: Volume 2, Asian Type. Paris: Med’Com; 2010.
51. Bazin R, Flament F, Rubert V. Skin Aging Atlas: Volume 4, Indian Type. Paris: Med’Com; 2015.
52. Sarcu D, Adamson P. Psychology of the facelift patient. Facial Plast Surg. 2017;33:252–259.