The present disclosure relates to the field of computer vision and machine learning, specifically, an automatic scalp sebum classification system and a method for automatic classification of scalp sebum using a visual transformation model.
This disclosure primarily pertains to the field of computer vision and machine learning. More specifically, it involves a system and method for automatically classifying scalp sebum properties using a visual transformation model.
Traditional methods for determining scalp sebum properties, such as dry, neutral, or oily, usually rely on manual inspection and evaluation. These methods are not just time-consuming, but also prone to human error and variability due to subjective judgement. Furthermore, the mental state or professional knowledge of the evaluator may affect the manual method, leading to inconsistent results.
Some existing computer vision techniques have been used to automate this process. However, these methods often struggle to handle subtle feature differences that are pivotal for accurate sebum classification. Additionally, the data used for training and validation in these methods often suffer from class imbalance, requiring data augmentation techniques to improve model performance.
Moreover, privacy issues have been raised for the use of high-resolution scalp images, especially when these images are stored or transmitted for analysis.
Therefore, there is a demand for a more reliable, efficient, and privacy-protecting method for automatic classification of scalp sebum. This method ought to handle subtle feature differences and class imbalance in the data, while ensuring the privacy of the scalp images being analyzed.
This disclosure introduces a system and method for automatic classification of scalp sebum using a visual transformation model. The goal of this innovation is to overcome the limitations of existing methods and provide a solution that is both effective and reliable. Traditional methods typically rely on manual inspection, which is not just time-consuming, but also prone to human error and subjectivity. This innovation automates this process, thereby eliminating these issues.
The system captures a high-resolution image of a part of the scalp, then divides these images into non-overlapping blocks. Each block is transformed into a one-dimensional vector. Learnable positional embeddings are added to these vectors to retain spatial information lost in the flattening process. These enhanced vectors are then input into a visual transformation model, which processes the data and outputs a classification label indicating scalp sebum properties, such as dry, neutral, or oily.
A distinctive feature of this innovation is the optional data augmentation module, which performs random rotation and position swapping of the blocks. This not just enhances the performance of the model, addressing the issue of class imbalance, but also adds a layer of privacy protection.
By automating the classification of scalp sebum properties and combining advanced machine learning techniques, this innovation provides a comprehensive and effective solution to the challenges faced by existing methods. It effectively handles subtle feature differences, resolves class imbalance in training data, and ensures the privacy of the scalp images being analyzed.
The objects, spirits, and advantages of the preferred embodiments of the present disclosure will be readily understood by the accompanying drawings and detailed descriptions, wherein:
Referring to
Once the image is captured, step S120 is executed, and the input image is divided into non-overlapping blocks. This division process is carried out by the block extraction module 120, which divides the high-resolution scalp image into smaller, more manageable parts. Each block represents a specific area of the scalp, and this block-based approach allows for more detailed and localized analysis of scalp sebum properties later.
After block extraction, step S130 is executed, and the data augmentation module 130 performs random rotation and position swapping of the blocks. This step introduces variability into the data, which helps to enhance the robustness of the artificial intelligence model by exposing it to a wider range of data scenarios. In addition, the random rotation and position swapping of the blocks also serve as a privacy protection measure. By scrambling the original arrangement of the blocks, the automatic scalp sebum classification system 100 makes it difficult to reconstruct the original scalp image, thereby protecting individual privacy.
Then, step S140 is executed, and each block is converted into a one-dimensional vector by the vector transformation module 140. This transformation process involves flattening the two-dimensional block into a one-dimensional pixel value array. The resulting vector retains the pixel intensity information of the block, but its format is more suitable for processing by subsequent components of the system.
Next, step S150 is executed, and positional information is added to the one-dimensional vector through learnable embedding vectors. This step is carried out by the positional embedding module 150, which adds a layer of spatial information to the vector. The learnable embedding vectors encode the relative position of the block 11 in the original scalp image, allowing the automatic scalp sebum classification system 100 to retain some spatial context.
It is worth noting that the positional embedding module 150 can have various possible configurations. Referring to
Next, step S160 is executed, and the enhanced positional embedding vectors are input into the visual transformation model 160. The visual transformation model 160 is an artificial intelligence model designed for image analysis tasks. It contains a plurality of layers of self-attention mechanisms and feedforward neural networks, which work together to analyze the input vectors and extract meaningful features related to scalp sebum properties.
Finally, step S170 is executed, and the visual transformation model 160 outputs a sebum classification label. This label indicates the sebum properties of the scalp part represented by the input image, and can be one of the following three categories: dry, neutral, or oily. The classification label serves as the final output of the automatic scalp sebum classification system 100, providing a concise indication of the analyzed scalp sebum properties.
Next, we will provide a more detailed description of the aforementioned automatic scalp sebum property classification system and method. Please continue to refer to
Once a high-resolution image of a part of the scalp has been captured, it is processed by the block extraction module 120. As shown in step S120, the block extraction module 120 is responsible for dividing the image into non-overlapping blocks 11 (as shown in
The design of the block extraction process is to ensure that each block 11 contains enough information for subsequent analysis. To this end, the blocks 11 are extracted in such a way that they cover the full range of the scalp part depicted in the image, without overlapping with other blocks. This non-overlapping arrangement ensures that each area of the scalp is analyzed once, preventing redundancy in the analysis. In addition, the non-overlapping arrangement also helps to maintain the spatial integrity of the original image, as each block 11 corresponds to a specific and distinct area of the scalp.
In summary, the process of capturing a high-resolution image of a part of the scalp and dividing these images into non-overlapping blocks 11 ensures that the automatic scalp sebum property classification system 100 can obtain detailed and localized visual information about the scalp, which is then used to determine sebum properties in a reliable and efficient manner.
After extracting blocks 11 from the high-resolution scalp image 10, the automatic scalp sebum property classification system 100 uses the data augmentation module 130 to perform random rotation and position swapping of the blocks 11 (as shown in
First, the random rotation and position swapping of the blocks 11 help to enhance the robustness of the visual transformation model 160. In machine learning, robustness refers to the ability of a model to maintain its performance in the face of changes in input data. By introducing randomness, the automatic scalp sebum property classification system 100 exposes the visual transformation model 160 to a wider range of data scenarios, thereby encouraging the visual transformation model 160 to learn more general and robust representations of scalp sebum properties. This is particularly beneficial in addressing the problem of class imbalance, where some sebum property categories may be underrepresented in the training data. By randomly rotating and swapping the positions of the blocks, the automatic scalp sebum property classification system 100 effectively increases the diversity of training data for each category, helping to mitigate the impact of class imbalance and improve the overall performance of the model.
Second, the random rotation and position swapping of the blocks 11 serve as a privacy protection measure. In this implementation, privacy protection refers to the protection of the personal identity of the scalp image being analyzed. By scrambling the original arrangement of the blocks 11, it becomes difficult to reconstruct the original scalp image from the processed data. This means that even if someone obtains the data without authorization, they cannot identify the individual from the scrambled blocks. This feature is particularly valuable in applications where personal privacy is paramount, such as in medical diagnosis or personal care applications.
In summary, the data augmentation module 130 enhances the robustness of the visual transformation model 160 and protects personal privacy by performing random rotation and position swapping of the blocks 11. By performing random rotation and position swapping of the blocks 11, the system can handle subtle feature differences and class imbalance in the data, while ensuring the privacy of the scalp images being analyzed.
After the data augmentation process, step S140 is executed, and the automatic scalp sebum classification system 100 continues to convert each block 11 into a one-dimensional vector. This conversion is performed by the vector transformation module 140, which transforms the two-dimensional block 11 into a one-dimensional pixel value array. The transformation process involves flattening the block 11, which basically involves rearranging the pixel values from a two-dimensional grid into a one-dimensional sequence. This process retains the pixel intensity information of the block 11, but presents it in a format more suitable for processing by subsequent components of the automatic scalp sebum classification system 100.
Once the block 11 has been converted into a one-dimensional vector, step S150 is executed, and the automatic scalp sebum classification system 100 adds positional information to these one-dimensional vectors through learnable embedding vectors. This step is performed by the positional embedding module 150, which is designed to encode the relative position of the block 11 in the original scalp image. The learnable embedding vectors are basically a set of parameters learned during the training process of the visual transformation model. These parameters represent the spatial relationship between the blocks 11, thus allowing the system to retain some spatial context, despite the flattening process.
The process of adding positional information to a one-dimensional vector involves a series of operations, please also refer to
Then, as shown in step S230, the position embedding is added to the one-dimensional vector representing the block 11. This addition operation effectively combines the pixel intensity information of the block with the spatial information encoded in the position embedding. The resulting vector now contains pixel intensity and position information, and is then prepared to be input into the visual transformation model 160 for further processing.
In summary, the process of converting each block 11 into a one-dimensional vector and adding positional information through learnable embeddings is a pivotal step in the workflow of the automatic scalp sebum classification system 100. This process ensures that the visual transformation model 160 receives comprehensive input data containing pixel intensity and spatial information, thereby achieving a more accurate and global perception of scalp sebum properties.
Additionally, one implementation of the structure of the positional embedding module 150 is shown in
In
In summary, the positional embedding module 150 can extract useful features from the one-dimensional vector of the block 11, and transform these features into a form that allows the model to learn and predict better.
Referring to
The visual transformation model 160 contains a plurality of layers of self-attention mechanisms 162 and feedforward neural networks 164. The self-attention mechanisms 162 allow the visual transformation model 160 to focus on different parts of the input data based on their relevance to the task at hand. In this implementation, the self-attention mechanisms 162 allow the visual transformation model 160 to focus on different blocks 11 of the scalp image 10, giving more attention to blocks that contain more information for the sebum classification task.
Each layer of the self-attention mechanism 162 in the visual transformation model 160 operates by calculating the weighted sum of the input vectors, where the weights are determined by the relevance of each vector to the other vectors. This relevance is quantified using a measure called attention scores, which is calculated based on the similarity between vectors. The attention scores are then used to weight the input vectors, allowing the visual transformation model 160 to focus more on vectors with higher scores.
The feedforward neural networks 164 in the visual transformation model 160 are used to transform the weighted sum of the input vectors into a higher-level representation. These feedforward neural networks 164 consist of a plurality of layers of neurons, each layer performing a linear transformation on its input, followed by a non-linear activation function. The output of the feedforward neural networks 164 is a set of feature vectors, capturing high-level patterns in the input data.
The visual transformation model 160 processes the input vectors in a sequential manner, passing them through a plurality of layers of self-attention mechanisms 162 and feedforward neural networks 164. At each layer, the visual transformation model 160 updates the feature vectors based on the information extracted from the previous layer. This iterative process allows the visual transformation model 160 to gradually refine its understanding of the input data, leading to more accurate sebum classification results.
Finally, the visual transformation model 160 outputs a sebum classification label 166 based on the final set of feature vectors. This sebum classification label 166 indicates the sebum properties of the scalp part represented by the input image, and can be one of the following three categories: dry, neutral, or oily. The sebum classification label 166 serves as the final output of the automatic scalp sebum classification system 100, providing a concise indication of the analyzed scalp sebum properties.
In summary, the two main technical improvements of the automatic scalp sebum classification system 100 contribute greatly to its performance and privacy protection: the random rotation and position swapping of blocks 11, and the use of learnable position embeddings.
The random rotation and position swapping of blocks 11 is performed by the data augmentation module 130. The data augmentation module 130 introduces a degree of randomness into the data, which has two main purposes. First, it enhances the robustness of the visual transformation model 160 by exposing it to a wider range of data scenarios. This is particularly beneficial in addressing the problem of class imbalance, where some sebum property categories may be underrepresented in the training data. By randomly rotating and swapping the positions of the blocks 11, the automatic scalp sebum classification system 100 effectively increases the diversity of training data for each category, helping to mitigate the impact of class imbalance and improve the overall performance of the model.
Second, the random rotation and position swapping of the blocks 11 serve as a privacy protection measure. By scrambling the original arrangement of the blocks 11, the automatic scalp sebum classification system 100 makes it difficult to reconstruct the original scalp image, thereby protecting individual privacy. This feature is particularly valuable in applications where personal privacy is paramount, such as in medical diagnosis or personal care applications.
Another notable technical improvement in the automatic scalp sebum classification system 100 is the use of learnable embedding vectors. This feature is implemented by the positional embedding module 150, which adds a layer of spatial information to the one-dimensional vector representing each block 11, allowing the automatic scalp sebum classification system 100 to retain some spatial context. This is particularly beneficial for the visual transformation model 160, as it allows the visual transformation model 160 to adapt to specific spatial relationships in the data, thereby achieving more accurate classification.
In conclusion, the random rotation and position swapping of blocks 11, and the use of learnable embedding vectors, are the two main technical improvements in the automatic scalp sebum classification system 100 of this disclosure. These improvements enhance its performance and protect privacy. These features ensure that the automatic scalp sebum classification system 100 can handle subtle feature differences and class imbalance in the data, while ensuring the privacy of the scalp images being analyzed.
Number | Date | Country | Kind |
---|---|---|---|
202311571500.6 | Nov 2023 | CN | national |