The present application relates to a skin diagnosis system and method based on image analysis using deep learning.
The present application claims the priorities of Korean Patent Application No. 10-2021-0136568 filed on Oct. 14, 2021, and Korean Patent Application No. 10-2022-0131817 filed on Oct. 13, 2022, the entire contents of which is incorporated herein by reference.
In addition to medical use, skin diagnosis is actively used in the cosmetics industry and related research fields. With the development of online markets and personalized services, a demand for methods to perform the skin diagnosis more easily and accurately is increasing. Meanwhile, various existing skin diagnosis and measurement methods have high accuracy, but have limited measurement items and require expensive equipment. Existing image analysis-based skin diagnosis methods are relatively inexpensive and can measure various items with a single image, but have low accuracy and are difficult or expensive to obtain detailed quantitative data.
Further, in order to avoid the difficulty of achieving accuracy and data labeling issues, the existing skin diagnosis methods have not provided detection location and quantitative data for symptoms or have realized the symptoms into simple grading judgments, which caused a problem of failing to provide actual symptom location and quantitative information for each item.
Furthermore, since the existing prior technologies do not include a procedure for removing personal identification information contained in a face image that is used for the skin diagnosis, there is a problem in that management or protection of the personal information is insufficient during analysis processing or data collection.
Other prior technologies only define a concept of the skin diagnosis using an image by artificial intelligence (AI), and do not include technical definitions for actual detailed operation or diagnosis, or limit image collection and diagnosis equipment to a smartphone, a smart device, etc., whereby the scope of applying the prior technologies is limited because they do not take into account a possible combination of collection environment and separate diagnostic service structure in various imaging devices.
The present application is intended to solve the above-mentioned problems and provides a diagnostic technology capable of recognition and integrated analysis, such as facial skin aging, by applying artificial intelligence (AI) technology to skin aging evaluation and skin lesion evaluation. Specifically, by creating an artificial neural network model through deep learning, symptoms according to each item in the skin image can be diagnosed and analyzed to quantitatively derive detailed location for each subject, shape, and degree of severity of actual skin symptoms. It is possible to analyze the number, average size or shape of each symptom, the overall area or intensity of the symptom, etc., and the obtained information can be used to provide visualization information that emphasizes the diagnosis in the skin image.
An image analysis-based skin diagnosis method using deep learning performed by a processor according to an aspect of the present application comprises the steps of: acquiring a face image of a subject by photographing a target skin; deriving shape or location information of a facial structure by recognizing feature points capable of identifying an individual in the acquired face image: de-identifying the face image on the basis of the shape or location information of the facial structure such that personal information of an analysis target cannot be identified; and visualizing and providing skin diagnosis results for items corresponding to artificial neural network models and symptom locations for each item by inputting the de-identified face images into a plurality of the artificial neural network models, respectively, wherein the items may include at least one of diagnoses for wrinkles, pigmentation, pores, erythema, and aging.
In an embodiment, the step of de-identifying the face image may include: separating feature points and a background from the face image and removing a skin shape having the feature points: dividing the face image from which the feature points have been removed into a plurality of patches; and rearranging the plurality of patches by randomizing them.
In an embodiment, the feature points may be at least one of eyebrows, eyes, nose, and lips.
In an embodiment, each of the plurality of artificial neural network models is learned using a plurality of training samples as learning data, and the plurality of training samples may include transformations using a data augmentation technique.
In an embodiment, the data augmentation technique may include at least one of random crop, blur, and flip processing.
In an embodiment, the artificial neural network model for diagnosing the wrinkles may be learned by adjusting parameters including the total number of a subject with detected wrinkles, an estimate of intensity (depth) compared to the surrounding undetected area in the subject with detected wrinkles, the total area with detected wrinkles, the length and width of the detected wrinkles, and may output analysis results for at least one among the number of wrinkles, intensity (depth) of wrinkles, wrinkle area, wrinkle length, wrinkle width, distribution for each intensity, area, length or width of wrinkles, and wrinkle score.
In an embodiment, the artificial neural network model for diagnosing the pigmentation may be learned by adjusting parameters including the total number of a subject with detected pigmentation, an estimate of intensity compared to the surrounding undetected area in the subject with detected pigmentation, the total area with detected pigmentation, and may output analysis results for at least one among the number of pigmentation, intensity of pigmentation, pigmentation area, distribution for each intensity, area, length or width of pigmentation, and pigmentation score.
In an embodiment, the artificial neural network model for diagnosing the pores may be learned by adjusting parameters including the total number of a subject with detected pores, an estimate of intensity (depth) compared to the surrounding undetected area in the subject with detected pores, the total area with detected pores, pore length and pore width, and may output analysis results for at least one among the number of pores, intensity (depth) of pores, pore size, pore area, pore length, pore width, pore sagging (length to width ratio), distribution for each intensity, area, length, width or sagging of pores, and pore score.
In an embodiment, the artificial neural network model for diagnosing the erythema may be learned by adjusting parameters including the total number of a subject with detected erythema, an estimate of intensity compared to the surrounding undetected area in the subject with detected erythema, the total area with detected erythema, and may output analysis results for at least one among the number of erythema, intensity of erythema, erythema area, distribution for each intensity or area of erythema, and erythema score.
In an embodiment, the artificial neural network model for diagnosing the aging may predict age for facial aging or facial skin aging estimated from a face image by inputting at least one of the de-identified face image, the output result of a single artificial neural network model, and the value that integrates the output result of a plurality of artificial neural network models.
In an embodiment, each of the plurality of artificial neural network models may be an encoder-decoder structural model based on U-net model.
In an embodiment, each of the plurality of artificial neural network models may be learned in the form of an imageNet pre-trained weight based on ResNet.
In an embodiment, the present method may further comprise the step of evaluating at least one of antioxidant efficacy and whitening efficacy for a specific product based on skin diagnosis results for the above items.
In an embodiment, the present method may further comprise the step of, before inputting the de-identified face image into the plurality of artificial neural network models, obtaining information about the subject's skin concerns and lifestyle through a questionnaire, wherein the result of skin diagnose is the one resulting from the subject's skin concerns and lifestyle.
In an embodiment, the present method may further comprise the step of recommending a specific product tailored to the subject's skin concerns and lifestyle or providing beauty eating habit, based on the result of skin diagnose.
According to another aspect of the present application, a skin diagnosis system based on image analysis using deep learning comprises: an imaging unit that acquires a face image by photographing a target skin: a face detection model that derives shape or location information of a facial structure by recognizing feature points capable of identifying an individual in the acquired face image: a de-identification model that de-identifies the face image on the basis of the shape or location information of the facial structure such that personal information of an analysis target cannot be identified; and a plurality of artificial neural network models for each of at least one item among diagnoses for wrinkles, pigmentation, pores, erythema, and aging, wherein the plurality of artificial neural network models receive the de-identified face image as input, and visualize and provide skin diagnosis results for items corresponding to the artificial neural network models and symptom locations for each item.
A skin diagnosis system and method based on image analysis using deep learning according to an aspect of the present application are separately composed of a face detection model that finds location of a face, a de-identification model that leaves only the skin area in a face image and performing random arrangement to de-identify personal information, and a model for each diagnostic symptom including diagnoses for wrinkles, pigmentation, pores, erythema, and aging. Therefore, optimization and function separation or adjustment can be performed for each characteristic, which enables overlapping diagnosis and efficient system management and operation.
Skin diagnosis can be performed even with a partial skin image that has been de-identified from the acquired face image. Since the diagnosis system is designed not to utilize, store, and/or process personal information, system operation and data collection can proactively respond to management without leaking the personal information.
The images used in the system to diagnose skin condition are not limited to images captured from a professional photography equipment, a general camera, or a smartphone, and are designed to derive the characteristics of each skin diagnosis item according to the image condition so that the diagnosis system can be used for various purposes by minimizing the impact on a type or form of the image.
In long-term research on skin aging, such as anti-aging research, that requires long-term follow-up analysis, and evaluation of the efficacy of anti-aging products, and evaluation of the efficacy of whitening cosmetics that require analysis of subtle change in skin pigment, evaluation of skin change and the efficacy of cosmetics can be quantitatively performed. Therefore, various skin diagnostic items can be performed quickly and at low cost without a separate skin measurement equipment or analysis preparation, thereby providing advantageous characteristics in the fields such as long-term follow-up or large-scale skin research.
The effects of the present application are not limited to the effects mentioned above, and other effects not mentioned will be clearly understood by those skilled in the art from the description of the claims.
In order to disclose the technical solutions of the embodiments of the present application or the prior art more clearly, drawings necessary to describe the embodiments are briefly introduced below. It should be understood that the drawings below are intended only to illustrate the embodiments of the present specification and are not intended to limit them. Also, for clarity of explanation, some elements to which various variations such as exaggeration and omission are applied may be illustrated in the drawings below.
The terminology used herein is only intended to refer to specific embodiments and is not intended to limit the present application. As used herein, singular forms also include plural forms unless the phrases clearly indicate the contrary. As described in the specification, the meaning of “comprising” refers to specify a certain characteristic, area, integer, step, operation, item and/or component, and does not exclude the presence or addition of another characteristic, area, integer, step, operation, item and/or component.
Although not defined differently, all the terms including technical and scientific terms used herein have the same meaning as those generally understood by a person who has an ordinary knowledge in the technical field to which the present application belongs. The terms defined in commonly used dictionaries are further interpreted as having meanings consistent with related technical literature and currently disclosed contents, and are not interpreted in ideal or very formal meanings unless defined otherwise.
Hereinafter, embodiments of the present application will be reviewed in detail with reference to the drawings.
Referring to
In this specification, the artificial neural network model may include a deep learning model, wherein the deep learning model may be in the form of artificial neural networks stacked in multiple layers. The deep learning model automatically learns the features of each image by learning a large amount of data in a deep neural network consisting of a network of the multiple layers, and through this, trains the network in the manner of minimizing errors in the objective function, that is, the prediction accuracy.
In this specification, the deep learning model may use, for example, CNN (Convolutional Neural Network), DHN (Deep Hierarchical Network), CDBN (Convolutional Deep Belief Network), DDN (Deconvolutional Deep Network), etc., but a variety of deep learning models are available now or in the future.
The skin diagnosis system 1 according to embodiments may be entirely hardware, entirely software, or may have aspects that are partly hardware and partly software. For example, the system may collectively refer to a hardware equipped with data processing capabilities and an operating software for driving the hardware. In this specification, the terms such as “unit,” “system,” and “device” are intended to refer to a combination of a hardware and a software driven by the hardware. For example, the hardware may be a data processing device that includes CPU (Central Processing Unit), GPU (Graphics Processing Unit), or other processor. Further, the software may refer to an executing process, object, executable file, thread of execution, program, etc.
For example, the concepts (e.g., operation of the components 10 to 17) of the skin diagnosis system 1 (such as the VISIA-CR® system from Canfield Company (Fairfield, NJ, USA)) may be integrated into various computing devices (e.g., a smart device including a smartphone, a computer, etc.) as an imaging system for skin analysis, a software for the computer, or a web for customer. The software may be designed to process captured images of a target skin, and may diagnose and evaluate the target skin in the images and further provide the operation results to the user.
The imaging unit 10 is a component that photographs the target skin, and may be implemented with various imaging devices that can be used in the present technical field. In an embodiment, the imaging unit 10 is configured to acquire a face image by photographing an image of the target skin. For example, the imaging unit 10 may be implemented with VISIA-CR® from Canfield Company (Fairfield, NJ, USA).
The imaging unit 10 may provide image data of the target skin to other components 11 to 17. Additionally, the imaging unit 10 may be further configured to perform operation of extracting a partial area of the target skin's image, for example, a cropping processing operation.
It will be apparent to those skilled in the art that the skin diagnosis system 1 may comprise other components which are not explicitly described herein. For example, the skin diagnosis system may comprise other hardware elements necessary for the operations described herein, including a network interface, an input device for data entry, an output device for display, printing or other data presentation, and a storage device such as memory.
Referring to
Referring to
Referring to
The wrinkle model that diagnoses wrinkles can output analysis results for at least one among the number of wrinkles, intensity (depth) of wrinkles, wrinkle area, wrinkle length, wrinkle width, distribution for each intensity, area, length or width of wrinkles, and wrinkle score, by inputting the de-identified face image. The pigmentation model that diagnoses pigmentation can output analysis results for at least one among the number of pigmentation, intensity of pigmentation, pigmentation area, distribution for each intensity or area of pigmentation, and pigmentation score. The pore model that diagnoses pores can output analysis results for at least one among the number of pores, intensity (depth) of pores, pore size, pore area, pore length, pore width, pore sagging (length to width ratio), distribution for each intensity, area, length, width or sagging of pores, and pore score. The erythema model that diagnoses erythema can output analysis results for at least one among the number of erythema, intensity of erythema, erythema area, distribution for each intensity or area of erythema, and erythema score. Each of the wrinkle score, the pigmentation score, the pore score, and the erythema score means a scored numeral value for the skin condition of each individual skin item based on pre-collected data, and is calculated by considering single or multiple combinations of data such as the intensity and total area of wrinkles, pigmentation, pores, and erythema, respectively. The analysis results output from the wrinkle model, pigmentation model, pore model, and erythema model can be visually displayed in the area corresponding to the symptom of the item. In addition, the aging diagnosis model that diagnoses facial aging can output analysis results on the facial aging and the facial skin aging by inputting the de-identified or identified face images.
Referring to
Referring to
Therefore, the skin diagnosis can be performed even with a partial skin image that has been de-identified from the acquired face image through the de-identification model. Since the diagnosis system is designed not to utilize, store, and/or process personal information, system operation and data collection can proactively respond to management without leaking the personal information.
Referring to
The plurality of training samples may be modified using data augmentation technique. The data augmentation is a technique that increases adaptability to various data environments and development performance by transforming data sets for the artificial intelligence learning into various realistic shapes to secure data diversity and quantity beyond the actual collected data.
In an embodiment, the data augmentation technique may include at least one of random crop, blur, and flip processing. Through the data augmentation technique, general and consistent results can be obtained even in various environments, thereby being capable of exerting the effect of enabling meaningful operation even in untrained devices or environments.
Referring to
The ResNET model is a multi-stage artificial neural network model of high performance developed by Microsoft. The issue of accuracy degradation that occurs when forming the multi-stage neural network to improve the performance is solved by utilizing a difference between input and output at each stage for learning, called residual learning. The imageNet is a large-scale database for developing a software that recognizes objects in the images, wherein the pre-trained weight refers to a condition of the initial artificial intelligence model learned to assign weights to a specific condition based on this data. The mIoU (mean Intersection over Union) is one of the methodologies for evaluating digital image analysis models, and is a method of quantifying a match rate of the results by dividing the number of common pixels between a target and a predicted result by the total number of pixels. The Cross-Entropy loss is one of the methodologies for evaluating digital analysis models, and uses an entropy-based loss function to evaluate a difference between probability distribution of target data and probability distribution of predicted model result data. The Focal loss is one of the methodologies for evaluating digital analysis models, and is an evaluation method that uses weighting parameters to focus on samples that are difficult to analyze by reducing the impact of easy-to-analyze sample data on overall prediction model learning.
Due to the nature of a skin, in most cases, the skin is in an environment (imbalanced data) with many general skin areas and few lesion areas. Therefore, for smooth learning, a learning method is used that measures the frequency of each label and reflects it when calculating loss. As the diagnostic model is separated for each item, individual characteristics are optimized and overlapping diagnosis for each item is possible. Further, by applying the de-identification model to a separate face image, security issues, for example, personal information protection for diagnostic information such as personal identification data that are difficult to remove with only the face detection model can be addressed.
Referring to
A pigmentation model for diagnosing pigmentation may be learned by adjusting parameters including the total number of a subject with detected pigmentation, an estimate of intensity compared to the surrounding undetected area in the subject with detected pigmentation, the total area with detected pigmentation, and may output analysis results for the number of pigmentation, intensity of pigmentation, pigmentation area, distribution for each intensity or area of pigmentation, and pigmentation score.
A pore model for diagnosing pores may be learned by adjusting parameters including the total number of a subject with detected pores, an estimate of intensity (depth) compared to the surrounding undetected area in the subject with detected pores, the total area with detected pores, pore length and pore width, and may output analysis results for the number of pores, depth of pores, pore area, pore length, pore width, pore sagging (length to width ratio), distribution for each intensity, area, length, width or sagging of pores, and pore score.
An erythema model for diagnosing erythema may be learned by adjusting parameters including the total number of a subject with detected erythema, an estimate of intensity compared to the surrounding undetected area in the subject with detected erythema, the total area with detected erythema, and may output analysis results for the number of erythema, intensity of erythema, erythema area, distribution for each intensity or area of erythema, and erythema score.
An aging diagnosis model may output analysis results for facial aging or facial skin age estimated from a face image.
Further, the face detection model may output a facial size, which is a size of detected skin area, after de-identifying the face image.
Referring to
Further, the skin diagnosis system 1 separately consists of a de-identified model for processing individual information and a model for each diagnostic symptom including diagnoses for wrinkles, pigmentation, pores, and erythema, so that optimization and function separation or adjustment can be performed for each characteristic, which enables overlapping diagnosis and efficient system management and operation.
Each of
Referring to
Referring to
Referring to
In case of pores, the number/average size/total area of pores showed an increasing correlation with age, while the intensity (depth) decreased. Referring to
Referring to
In an experimental example, as a result that the skin diagnosis was performed on images of long-term skin tracking research results of a retinol product that had proven anti-aging efficacy, it was confirmed during the 4-year follow-up period that the group that used the product showed a significant skin improvement effect, while the pigmentation and the number of wrinkles increased or remained in the group that did not use the product. This made it possible to provide more detailed and quantified evaluation results with the same results as separate clinical measurement evaluation results.
Referring to
Referring to
Based on the skin diagnosis results, the skin diagnosis system 1 may recommend a specific product tailored to the subject's skin concerns and lifestyle or provide beauty eating habits.
In an embodiment, the skin diagnosis system 1 provides analysis results of the subject's skin condition for each item of pigmentation, pores, erythema, and wrinkles. Further, information obtained through the questionnaire is used to provide comments on the subject's skin concerns and provide tips related to cleansing or lifestyle. In addition to a customized skin care method according to the subject's skin condition, cosmetics suitable for each skin condition can be recommended.
Since such a skin diagnosis system and method based on image analysis using deep learning are separately composed of the recognition model of facial location and feature points, the de-identification model of personal information, and a model for each diagnostic symptom including diagnoses for wrinkles, pigmentation, pores, erythema, and aging, optimization and function separation or adjustment can be performed for each characteristic, which enables overlapping diagnosis and efficient system management and operation.
The skin diagnosis can be performed even with a partial skin image that has been de-identified from the acquired face image. Since the diagnosis system is designed not to utilize, store, and/or process personal information, system operation and data collection can proactively respond to management without leaking the personal information.
The images used in the system to diagnose skin condition are not limited to images captured from a professional photography equipment, a general camera, or a smartphone, and are designed to derive the characteristics of each skin diagnosis item according to the image condition, so that the diagnosis system can be used for various purposes by minimizing the impact on a type or form of the image.
In long-term research on skin aging, such as anti-aging research, that requires long-term follow-up analysis, evaluation of the efficacy of anti-aging products, and evaluation of the efficacy of whitening cosmetics that require analysis of subtle change in skin pigment, evaluation of skin change and the efficacy of cosmetics can be quantitatively performed. Therefore, various skin diagnostic items can be performed quickly and at low cost without a separate skin measurement equipment or analysis preparation, thereby providing advantageous characteristics in the fields such as long-term follow-up or large-scale skin research.
The operation of the image analysis-based skin diagnosis system and method according to the embodiments described above may be at least partially implemented by a computer program and recorded on a computer-readable recording medium. For example, it may be implemented together with a program product consisting of the computer-readable medium having a program code, which can be executed by a processor to perform any or all steps, operations, or processes described.
The computer-readable recording medium includes all types of the recording devices for storing data that can be read by a computer. The computer-readable recording medium includes, for example, ROM, RAM, CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, etc. Additionally, the computer-readable recording medium may be distributed across a computer system connected to a network, and may be stored and executed in a distributed manner by a computer-readable code. Further, a functional program, a code, and a code segment for implementing these embodiments can be easily understood by those skilled in the art to which these embodiments belong.
As described above, the present application is described with reference to the embodiments shown in the drawings, but these are merely illustrative examples, and those skilled in the art will understand that various modifications and variations of the embodiments can be carried out therefrom. However, such modifications should be considered to be within the technical protection scope of the present application. Therefore, the true technical protection scope of the present application should be determined by the technical spirit of the attached claims.
A skin diagnosis system and method based on image analysis using deep learning according to an aspect of the present application are separately of a face detection model that finds location of a face, a de-identification model that leaves only the skin area in a face image and performing random arrangement to de-identify personal information, and a model for each diagnostic symptom including diagnoses for wrinkles, pigmentation, pores, erythema, and aging. Therefore, optimization and function separation or adjustment can be performed for each characteristic, which enables overlapping diagnosis and efficient system management and operation.
Number | Date | Country | Kind |
---|---|---|---|
10-2021-0136568 | Oct 2021 | KR | national |
10-2022-0131817 | Oct 2022 | KR | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/KR2022/015586 | 10/14/2022 | WO |