CN115035068A - Cross-domain self-photographing face pockmark grading image classification method capable of self-adapting skin color - Google Patents

Cross-domain self-photographing face pockmark grading image classification method capable of self-adapting skin color Download PDF

Info

Publication number
CN115035068A
CN115035068A CN202210680710.8A CN202210680710A CN115035068A CN 115035068 A CN115035068 A CN 115035068A CN 202210680710 A CN202210680710 A CN 202210680710A CN 115035068 A CN115035068 A CN 115035068A
Authority
CN
China
Prior art keywords
domain
image
network
self
skin color
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210680710.8A
Other languages
Chinese (zh)
Inventor
谭敏
王瑞瑞
俞俊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Dianzi University
Original Assignee
Hangzhou Dianzi University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hangzhou Dianzi University filed Critical Hangzhou Dianzi University
Priority to CN202210680710.8A priority Critical patent/CN115035068A/en
Publication of CN115035068A publication Critical patent/CN115035068A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/90Determination of colour characteristics
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/56Extraction of image or video features relating to colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/776Validation; Performance evaluation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30088Skin; Dermal
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • G06T2207/30201Face
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/03Recognition of patterns in medical or anatomical images

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Evolutionary Computation (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • Artificial Intelligence (AREA)
  • Medical Informatics (AREA)
  • Databases & Information Systems (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Mathematical Physics (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Quality & Reliability (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention provides a cross-domain self-portrait face pockmark grading image classification method adaptive to skin color. The method comprises the following steps: 1. between the source domain and the target domain, the domain offset is narrowed down with cross-domain data enhancement against the generative network model. 2. And constructing two gating network self-adaptive learning optimal sample weights. An expert gating network self-adaptive learning optimal characteristic weight and a skin color gating network self-adaptive learning optimal skin color weight are constructed. 3. And aligning the sample characteristics between the source domain and the target domain by utilizing a multi-core maximum mean difference method so as to reduce the domain deviation between the source domain and the target domain. 4. And establishing a multi-task end-to-end deep learning model according to the steps, training the whole network on a specific data set, and testing the performance of the final model on a test set. The method can adaptively learn the most appropriate sample weight distribution aiming at a specific data set, and has strong practicability and universality.

Description

Cross-domain self-photographing face pockmark grading image classification method capable of self-adapting skin color
Technical Field
The invention relates to the field of image classification, in particular to a cross-domain selfie face pockmark classification image classification method adaptive to skin color.
Background
With the progress of the times and the rapid development of medical cosmetology, people begin to pay more and more attention to the health condition of their skin, and skin care gradually becomes a hot topic. Acne, commonly known as comedo, is the most common skin disorder. Acne develops in about 80% of the adolescents, with acne symptoms lasting to the end of adulthood in about 3% of men and 12% of women. Due to the influence of the new coronavirus epidemic situation, people have to wear a mask to protect. This makes acne related diseases more severe and occur at more different age stages. However, acne also leaves scars and pigmentation, often leading to inferior and depressed mood. Therefore, a large number of acne patients are in urgent need of more accurate treatment. The severity of acne is critical to the dermatologist making an accurate, normative treatment regimen.
However, when a dermatologist ranks the severity of acne on the skin of a patient, misdiagnosis may occur due to factors such as subjectivity and lack of experience, which may lead to the condition of the patient to be worsened. With the development of deep learning technology, various intelligent auxiliary diagnosis algorithms based on deep learning break through and innovate continuously, and obtain considerable scores in various traditional diagnosis tasks and emerging tasks of medical image analysis, such as medical image classification, detection, segmentation, registration, image-based guided therapy and intervention and the like, so that remarkable potential is shown.
Different from the natural image field, a series of public complete large-scale labeled data sets, such as MNIST, CIFAR, ImageNet and the like, are provided. The labeling process of the acne image data set is high in specificity, the labeled data set is relatively scarce, and due to the existence of the face privacy problem, large-scale complete available data resources are few and few. Meanwhile, the change of the image background, the illumination and the skin color of the patient can bring the change of the data field. Under the condition, the method utilizes a domain self-adaptive method to dig out the information with identifiability from small sample data or a small amount of marked data, and is an effective way for conforming to the characteristics of the neighborhood resources of the current acne image.
Domain adaptation, which essentially fits the distribution differences between different data domains, generally speaking, assumes that the task between different domains is the same. In machine learning problems, it is generally assumed that the test data and the training data have the same distribution. But if this assumption is not verified, the performance of the model on the test set may be significantly degraded. In computer vision applications, such distribution differences (domain shifts) are common in real life and may be the result of changes in conditions (e.g., changes in background, light intensity, etc.). In the classification task, the method guides a classifier which learns the label-free data in the target domain by using the learned knowledge of the label data existing in one or more related source domains. In real life, most of data in a target domain is label-free or the data volume needing to be labeled is too large, and a large amount of manual work is involved. Thus, in this case, a domain-adaptive technique may be selected to build the target model.
Disclosure of Invention
The invention provides a cross-domain selfie face pockmark grading image classification method adaptive to skin color. The traditional pox classification method detects the existing pox firstly, then calculates the number of the detected pox, and classifies the image according to the number. However, the method is different from the traditional variola classifying method, the cross-domain self-adaptation and the image classification are fused in a uniform deep neural network, an end-to-end deep learning model is completed, and the model can directly complete the variola rating and classifying task of the self-shot image and simultaneously complete the cross-domain self-adaptation task of a source domain and a target domain. In the aspect of cross-domain self-adaption tasks of a source domain and a target domain, a confrontation generation network model is utilized to realize style conversion between a source domain sample and a target domain sample, so that domain offset between the source domain sample and the target domain sample is reduced, and cross-domain self-adaption of the source domain and the target domain is realized; in the aspect of image classification of self-photographing image pox classification, a cross-domain self-adaptive module and a self-adaptive skin color sensitive module are utilized to enable the model to obtain a better effect on a related data set.
A self-adaptive skin color cross-domain self-photographing face pockmark grading image classification method comprises the following steps:
step (1): image data pre-processing
Because the sizes of the images in the data set are different, in order to adapt to a deep learning framework, the sizes of the images need to be changed before the model begins to train, and the sizes of the images in the data set are unified.
Step (2): cross-domain data enhancement
The invention introduces a publicly available data set as a source domain to help the learning of the classifier in the target domain. Due to the fact that the samples of the source domain data and the target domain data have different sample illumination, dissimilar sample background and the like, domain deviation exists between the source domain data and the target domain data. Thus, in both the source domain and the target domain, the domain offset is narrowed down with cross-domain data enhancement against the generative network model.
And (3): construction of self-adaptive skin color gating classification model
The invention considers that the skin color change has certain influence on the classification of the pox, for example, under the condition of fair skin, similar phenomena such as acne, allergy and the like occur, and the classifier can identify the acne better so as to provide a more accurate result of skin quality evaluation; in contrast, in the case of black skin, acne is not easily identifiable and it is more difficult for the classifier to make a correct assessment. Therefore, an adaptive skin color gating classification model is constructed, and the skin color gating classification model can adaptively learn a weight vector related to the skin color of the sample according to the sample, so that the classification model has better performance. In the source domain data set and the target domain data set, skin colors are divided into five categories, namely white, neutral, tan, brown and black according to an Individual Type Angle (ITA) index.
And (4): feature space alignment
Because of the domain offset between the source domain data and the target domain data, the respective real images and generated images in the source domain and target domain will also have domain offsets, for which reason a feature-based domain alignment penalty is designed to make up for the domain gap. A multi-core maximum mean difference method is used in the present method.
And (5): model training and testing
And establishing a multitask end-to-end deep learning model according to the steps. And training the network parameters on a specific data set through a back propagation algorithm until the whole network model converges, and then testing the performance of the final model on a test set.
The image data preprocessing in the step (1) comprises the following specific steps:
because the sizes of all images in the data set are different, all the images are uniformly adjusted to a certain fixed size by a bilinear interpolation method. And then randomly cutting the image after the size adjustment to obtain image data with the size of 256 × 256. And finally, carrying out normalization processing on the image.
The cross-domain data enhancement in the step (2) comprises the following specific processes:
2-1. the source field is a collection comprising n images, each image having a corresponding skin quality class label, denoted I s ={(x i ,y i ) I is more than or equal to 1 and less than or equal to n; the target domain is a collection of m images, each image having a corresponding skin quality class label, denoted as
Figure BDA0003696205070000031
Wherein y is i E.g. Y and
Figure BDA0003696205070000032
are respectively an image x i And
Figure BDA0003696205070000033
class label of (1), (2), (…), N c Denotes the category label space, N c Is the total number of categories.
Figure BDA0003696205070000034
Is an image generator that converts samples in the source domain to a sample style with the target domain, the collection of generated images in the target domain being denoted as
Figure BDA0003696205070000035
Figure BDA0003696205070000036
Is an image generator that converts samples in the target domain into a sample style having a source domain, the collection of generated images in the source domain being denoted as
Figure BDA0003696205070000037
The specific formula is defined as follows:
Figure BDA0003696205070000038
wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0003696205070000041
Figure BDA0003696205070000042
refers to a loss function; | | non-woven hair 1 The representation takes a 1 norm.
2-2, respectively constructing a pair of image discriminators in the image space and the feature space and recording the pair of image discriminators
Figure BDA0003696205070000043
And
Figure BDA0003696205070000044
and a pair of feature discriminators
Figure BDA0003696205070000045
And
Figure BDA0003696205070000046
Figure BDA0003696205070000047
for discriminating whether a sample in the source domain that has passed through the image discriminator network is a real image or a generated image;
Figure BDA0003696205070000048
for discriminating whether a sample in the target domain that has passed through the image discriminator network is a real image or a generated image.
Figure BDA0003696205070000049
The system is used for judging whether the features extracted by the classification network in the source domain come from real image samples or generated image samples;
Figure BDA00036962050700000410
for determining whether the features extracted via the classification network in the target domain are from real image samples or from generated image samples.
The discrimination loss in the image space and the feature space is specifically shown in formula 3:
Figure BDA00036962050700000411
wherein
Figure BDA00036962050700000412
And representing a classification network, s represents a source domain, t represents a target domain, and d takes the values of s and t.
Figure BDA00036962050700000413
Wherein l i Real label representing ith sample image. When used to calculate the discriminant loss of image space
Figure BDA00036962050700000414
When the temperature of the water is higher than the set temperature,
Figure BDA00036962050700000415
when used to compute the discriminant loss of the feature space
Figure BDA00036962050700000416
When the temperature of the water is higher than the set temperature,
Figure BDA00036962050700000417
ψ (x) is an intermediate parameter variable, see equation 2.
Constructing the self-adaptive gating classification model in the step (3), wherein the specific process is as follows:
the invention designs an expert gating network and a skin color gating network to adaptively learn the optimal sample weights of the expert network and the sub-network respectively so as to solve the skin color label noise. This approach also guarantees a reasonable contribution of each sample to all sub-networks. In particular, the number of sub-networks is set differently than the number of skin tone categories to break the one-to-one mapping between the number of sub-networks and the number of skin tone categories.
Firstly, a sample obtains corresponding eigenvectors through a plurality of parallel expert networks, meanwhile, the sample can self-adaptively learn a gating weight vector through the expert gating networks, and the gating weight vector is multiplied by the eigenvectors to obtain the final weight combined eigenvector. Then the feature vector of the weight combination passes through a plurality of sub-networks simultaneously, and correspondingly outputs features extracted by the sub-networks; and multiplying the features extracted by the sub-network with the skin color ratio weight vector obtained by the sample through the skin color gating network in a self-adaptive manner to obtain the final combined feature.
In the method, a source domain and a target domain both have a classification network, so that the adaptive gating network has three structures, namely a classification network only in the source domain, a classification network only in the target domain and a classification network combined in the source domain and the target domain. The specific definition is as follows:
Figure BDA0003696205070000051
Figure BDA0003696205070000052
wherein, N' t Indicating the number of sub-networks, N e Which represents the number of the expert networks,
Figure BDA0003696205070000053
denotes the ith subnet full connection layer, M denotes the aggregation layer, E j Denotes the jth expert network, G j Represents N e The jth element, W, of the dimensional weight vector i Is N' t The ith element of the weight vector is dimensioned.
Figure BDA0003696205070000054
Where τ (x) represents the skin tone class label of sample x, e τ(x) Is N' t The τ (x) -th element of the dimensional weight vector is 1;
Figure BDA0003696205070000055
a skin tone category label representing a real image to which the image sample is generated,
Figure BDA0003696205070000056
is N' t The second of the dimensional weight vector
Figure BDA0003696205070000057
Each element is 1; γ and γ' represent the weighting parameters of the real image and the generated image, respectively.
The feature space alignment in the step (4) comprises the following specific processes:
in the method, a multi-kernel maximum mean difference penalty is used in the source domain and the target domain for the real image samples and the generated image samples in the respective domains, respectively. Thus, a cross-domain generated image from another domain is more compatible with a real image of a particular domain, ensuring the plausibility of sharing the same classification model on the real image and the generated image. The maximum mean difference loss for polynuclear is specifically defined as follows:
Figure BDA0003696205070000058
wherein
Figure BDA0003696205070000061
Is a weight parameter that controls the source domain,
Figure BDA0003696205070000062
is a weight parameter of the control target domain; phi is a unit of s Is a mapping function in the source domain, phi t Is a mapping function in the target domain;
Figure BDA0003696205070000063
2 norm is taken; e denotes expectation. The maximum mean difference penalty for multiple cores may be applied to a single domain or to two domains with shared or unshared mapping functions.
The step (5) of constructing the multitask deep learning model specifically refers to optimizing classification loss, image generation loss and domain alignment loss on a specific data set after an end-to-end frame is established according to the steps (2), (3) and (4). And in the training process, parameters of the confrontation generation network model and the classification model are updated simultaneously to obtain a final model, and the training effect is tested on the test set.
The invention has the beneficial effects that:
based on the thought of cross-domain image generation and gated network adaptive learning, a classification model for classifying the severity of face pox of a self-timer is provided. By introducing a publicly available source domain data set as an auxiliary data set, a deep migration learning-based vaccinia severity classification framework is proposed. An expert gating network and a skin color gating network model are designed to adaptively learn the correlation between the label and the skin color of the feature space.
Drawings
FIG. 1 is a schematic flow diagram of the process of the present invention.
Fig. 2 is a schematic diagram of a network framework constructed in the method of the present invention.
Figure 3 experiment of the combination of the number of expert networks and the number of subnetworks.
Details of the embodiments
The present invention will be further described with reference to FIGS. 1 and 2. The invention provides a cross-domain selfie face pockmark grading method adaptive to skin color. The method comprises the following steps: 1. between the source domain and the target domain, the domain offset is narrowed down with cross-domain data enhancement against the generative network model. 2. And constructing two gating network self-adaptive learning optimal sample weights. An expert gating network self-adaptive learning optimal characteristic weight is constructed, and a skin color gating network self-adaptive learning optimal skin color weight is constructed. 3. And aligning the sample characteristics between the source domain and the target domain by utilizing a multi-core maximum mean difference method so as to reduce the domain deviation between the source domain and the target domain. 4. And establishing a multi-task end-to-end deep learning model according to the steps, training the whole network on a specific data set, and testing the performance of the final model on a test set. The method can adaptively learn the most appropriate sample weight distribution aiming at a specific data set, and has strong practicability and universality.
The invention specifically realizes the following steps:
the first step is as follows:
we used the Acne04 dataset as the source domain and the three datasets Acnehdu, Acnehdu P and AcnePGP as the target domains, respectively, for verifying our classification model for self-portrait face pox severity rating. When the model is trained, firstly, the size of the images in the four data sets is adjusted to 286 × 286 by a bilinear interpolation method, then each image is randomly cut into 256 × 256, and finally, the pixel values of the images are normalized. When the model is tested, the data processing procedure is the same as that of training.
The second step is that:
in the method, an Acne04 data set is a source domain data set introduced by us, and three data sets of Acnehdu, Acnehdu P and AcnePGP are respectively used as target domain data sets. The following procedures are described with the Acne04 and Acnehdu datasets as examples.
First, a sample image in the source domain data set Acne04, denoted as x s Through a generator network
Figure BDA0003696205070000071
To obtain a reaction product of s Corresponding generated image, denoted x s2t (ii) a Second, sample image x in the target domain dataset Acnehdu t Through a generator network
Figure BDA0003696205070000072
To obtain a reaction product of t Corresponding generated image, denoted x t2s . In the source domain x s And x t2s All participate in the training of classification models in the source domain; in the target domain x t And x s2t Will participate in the training of the classification model in the target domain.
Then, x in the source domain s And x t2s Pass through discriminator D s In the target domain x t And x s2t Pass through discriminator D t 。D s And D t It is discriminated whether the sample is a real image or a generated image. Meanwhile, a pair of discriminators with the same function is also arranged in the feature space extracted by the classification network.
The third step:
we first tested different structures in three data sets and the results are shown in table 1 below. We have found that the optimal gating network structure is different in different target domain datasets due to the presence of domain offsets. At the same time, we evaluated the number of two important parametric expert networks (N) in the gated classification network e ) And number of subnetworks (N' t ) The results are shown in Table 2. We found 1) optimal N' t May be different from the number of skin tone categories and the number of sub-networks (N' t ) The larger the more readily better performance is obtained; 2) optimal number of expert networks (N) e ) Tendency toFrom less than the number of optimal sub-networks (N' t ) This means that the relevance and effectiveness of the underlying feature space is shared across multiple skin tones. The optimal number of subnetworks and expert networks for a particular data set is adaptively learned using a gating network in our method.
TABLE 1 ablation experiments on gated classification network modules of different structures
Figure BDA0003696205070000073
Figure BDA0003696205070000081
The fourth step:
we impose a multinuclear maximum mean difference (MK-MMD) penalty on the generated and real image datasets to achieve cross-domain alignment. On acneggp, we first tested different numbers of nuclei, K, and alignment structures using gaussian nuclei under a model with cross-domain data enhancement. The results are shown in table 2:
TABLE 2 Multi-core maximum mean Difference loss model experiment
Type (B) Only in the source domain Only in the target domain Federated sharing Federated unshared
K=3 0.872 0.849 0.847 0.857
K=5 0.857 0.766 0.857 0.852
K=8 0.864 0.766 0.766 0.849
In particular, we build MK-MMD penalties on only the source domain, only the target domain, the dual domain sharing the mapping, and the dual domain not sharing the mapping, respectively. We observed that the alignment structure had a significant impact on the performance results, and to achieve better performance we applied the best alignment structure for each dataset under the three gaussian kernels.

Claims (5)

1. A self-adaptive skin color cross-domain self-photographing face pockmark grading image classification method is characterized by comprising the following steps:
step (1): preprocessing image data, and unifying the sizes of images in a data set;
step (2): enhancing cross-domain data, introducing a publicly available data set as a source domain to help the learning of a classifier in a target domain;
and (3): constructing a self-adaptive skin color gating classification model;
and (4): aligning the feature space, and designing domain alignment loss based on features to make up domain difference;
and (5): training and testing a model; establishing a multi-task end-to-end deep learning model, and training network parameters on a specific data set through a back propagation algorithm until the whole network model converges.
2. The method for classifying facial pox-classified images through self-adaptive skin color cross-domain self-timer face according to claim 1, wherein said cross-domain data enhancement of step (2) is performed as follows:
2-1. the source field is a collection comprising n images, each image having a corresponding skin quality class label, denoted I s ={(x i ,y i ) I is more than or equal to 1 and less than or equal to n; the target domain is a collection of m images, each image having a corresponding skin quality class label, denoted as
Figure FDA0003696205060000011
Wherein y is i E.g. Y and
Figure FDA0003696205060000012
are respectively an image x i And
Figure FDA0003696205060000013
class label of (1), (2), (…), N c Denotes the category label space, N c Is the total number of categories;
Figure FDA0003696205060000014
is an image generator that converts samples in the source domain to a sample style with the target domain, the collection of generated images in the target domain being denoted as
Figure FDA0003696205060000015
Figure FDA0003696205060000016
Is an image generator that converts samples in the target domain into a sample style having a source domain, the collection of generated images in the source domain being denoted as
Figure FDA0003696205060000017
The specific formula is defined as follows:
Figure FDA0003696205060000018
wherein x is i ∈I s
Figure FDA0003696205060000019
Refers to a loss function; | | non-woven hair 1 The expression is taken as 1 norm;
2-2, respectively constructing a pair of image discriminators in the image space and the feature space and recording the pair of image discriminators
Figure FDA00036962050600000110
And
Figure FDA00036962050600000111
and a pair of feature discriminators
Figure FDA00036962050600000112
And
Figure FDA00036962050600000113
for discriminating whether a sample in the source domain that has passed through the image discriminator network is a real image or a generated image; for discriminating whether a sample in the target domain that has passed through the image discriminator network is a real image or a generated image;
Figure FDA00036962050600000114
the system is used for judging whether the features extracted by the classification network in the source domain come from real image samples or generated image samples;
Figure FDA0003696205060000023
for judging extraction through classified network in target domainWhether the feature is from a real image sample or a generated image sample;
the discrimination loss in the image space and the feature space is specifically shown as the following formula:
Figure FDA0003696205060000021
wherein
Figure FDA0003696205060000024
Representing a classification network, wherein s represents a source domain, t represents a target domain, and d takes the values of s and t;
Figure FDA0003696205060000022
wherein l i A real label representing the ith sample image; when used to calculate the discriminant loss of image space
Figure FDA0003696205060000025
When the temperature of the water is higher than the set temperature,
Figure FDA0003696205060000026
when used to compute the discriminant loss of the feature space
Figure FDA0003696205060000027
When the temperature of the water is higher than the set temperature,
Figure FDA0003696205060000028
ψ (x) is an intermediate parameter variable.
3. The method for classifying facial pox-classified images of self-adaptive skin color cross-domain self-timer face according to claim 2, wherein said step (3) of constructing an adaptive gating classification model comprises the following steps:
designing an expert gating network and a skin color gating network to respectively and adaptively learn the optimal sample weights of the expert network and the sub-network so as to solve skin color label noise; and the number of sub-networks set is different from the number of skin color categories, thereby destroying the one-to-one mapping between the number of sub-networks and the number of skin color categories;
firstly, a sample obtains corresponding eigenvectors through a plurality of parallel expert networks, meanwhile, the sample can self-adaptively learn a gating weight vector through the expert gating networks, and the gating weight vector is multiplied by the eigenvectors to obtain a final weight combined eigenvector;
then, the feature vector of the weight combination passes through a plurality of sub-networks simultaneously, and correspondingly outputs features extracted by the sub-networks; and multiplying the features extracted by the sub-network with the skin color ratio weight vector obtained by the sample through the skin color gating network in a self-adaptive manner to obtain the final combined feature.
4. The method for classifying facial pox-classified images of self-adaptive skin color cross-domain self-timer face according to claim 3, characterized by adding the following technical characteristics:
there is a classification network in both the source domain and the target domain, so the adaptive skin color gating network has three structures, namely a classification network only in the source domain, a classification network only in the target domain, and a classification network combined in the source domain and the target domain, which are specifically defined as follows:
Figure FDA0003696205060000031
Figure FDA0003696205060000032
wherein, N' t Indicating the number of subnetworks, N e Which is indicative of the number of expert networks,
Figure FDA0003696205060000035
represents the ith sub-network full connection layer, M represents the aggregation layer,E j Representing the jth expert network, G j Represents N e The jth element, W, of the dimensional weight vector i Is N' t The ith element of the weight vector is dimensioned.
Figure FDA0003696205060000033
Where τ (x) represents the skin tone category label of sample x, e τ(x) Is N' t The τ (x) -th element of the dimensional weight vector is 1;
Figure FDA0003696205060000036
a skin tone category label representing a real image to which the image sample is generated,
Figure FDA0003696205060000038
is N' t The second of the dimensional weight vector
Figure FDA0003696205060000037
Each element is 1; γ and γ' represent weight parameters of the real image and the generated image, respectively.
5. The method for classifying facial pox images of self-portrait across domains according to claim 3 or 4, wherein said step (4) of said image feature space alignment comprises the following steps:
using multi-kernel maximum mean difference loss for real image samples and generated image samples in respective domains in a source domain and a target domain respectively; thus, a cross-domain generated image from another domain is more compatible with a real image of a particular domain, ensuring the reasonableness of sharing the same classification model on the real image and the generated image; the maximum mean difference loss for polynuclear is specifically defined as follows:
Figure FDA0003696205060000034
wherein
Figure FDA0003696205060000039
Is a weight parameter that controls the source domain,
Figure FDA00036962050600000310
is a weight parameter of the control target domain; phi is a unit of s Is a mapping function in the source domain, phi t Is a mapping function in the target domain;
Figure FDA0003696205060000041
2 norm is taken; e represents expectation; the maximum mean difference penalty for multiple cores may be applied to a single domain or to two domains with shared or unshared mapping functions.
CN202210680710.8A 2022-06-15 2022-06-15 Cross-domain self-photographing face pockmark grading image classification method capable of self-adapting skin color Pending CN115035068A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210680710.8A CN115035068A (en) 2022-06-15 2022-06-15 Cross-domain self-photographing face pockmark grading image classification method capable of self-adapting skin color

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210680710.8A CN115035068A (en) 2022-06-15 2022-06-15 Cross-domain self-photographing face pockmark grading image classification method capable of self-adapting skin color

Publications (1)

Publication Number Publication Date
CN115035068A true CN115035068A (en) 2022-09-09

Family

ID=83124669

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210680710.8A Pending CN115035068A (en) 2022-06-15 2022-06-15 Cross-domain self-photographing face pockmark grading image classification method capable of self-adapting skin color

Country Status (1)

Country Link
CN (1) CN115035068A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115440346A (en) * 2022-11-07 2022-12-06 四川大学华西医院 Acne grading method, system, equipment and storage medium based on semi-supervised learning
CN116310008A (en) * 2023-05-11 2023-06-23 深圳大学 Image processing method based on less sample learning and related equipment
CN117274266A (en) * 2023-11-22 2023-12-22 深圳市宗匠科技有限公司 Method, device, equipment and storage medium for grading acne severity
CN117435916A (en) * 2023-12-18 2024-01-23 四川云实信息技术有限公司 Self-adaptive migration learning method in aerial photo AI interpretation

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115440346A (en) * 2022-11-07 2022-12-06 四川大学华西医院 Acne grading method, system, equipment and storage medium based on semi-supervised learning
CN115440346B (en) * 2022-11-07 2023-02-24 四川大学华西医院 Acne grading method, system, equipment and storage medium based on semi-supervised learning
CN116310008A (en) * 2023-05-11 2023-06-23 深圳大学 Image processing method based on less sample learning and related equipment
CN116310008B (en) * 2023-05-11 2023-09-19 深圳大学 Image processing method based on less sample learning and related equipment
CN117274266A (en) * 2023-11-22 2023-12-22 深圳市宗匠科技有限公司 Method, device, equipment and storage medium for grading acne severity
CN117274266B (en) * 2023-11-22 2024-03-12 深圳市宗匠科技有限公司 Method, device, equipment and storage medium for grading acne severity
CN117435916A (en) * 2023-12-18 2024-01-23 四川云实信息技术有限公司 Self-adaptive migration learning method in aerial photo AI interpretation
CN117435916B (en) * 2023-12-18 2024-03-12 四川云实信息技术有限公司 Self-adaptive migration learning method in aerial photo AI interpretation

Similar Documents

Publication Publication Date Title
CN115035068A (en) Cross-domain self-photographing face pockmark grading image classification method capable of self-adapting skin color
Li et al. Multi-site fMRI analysis using privacy-preserving federated learning and domain adaptation: ABIDE results
TWI772805B (en) Method for training generative adversarial network, method for generating image, and computer-readable storage medium
CN106295694B (en) A kind of face identification method of iteration weight set of constraints rarefaction representation classification
Cheng et al. Skin lesion classification using relative color features
Cardoso et al. Towards an intelligent medical system for the aesthetic evaluation of breast cancer conservative treatment
Kumar et al. Discovery radiomics with CLEAR-DR: interpretable computer aided diagnosis of diabetic retinopathy
CN112395442B (en) Automatic identification and content filtering method for popular pictures on mobile internet
CN110263801A (en) Image processing model generation method and device, electronic equipment
CN113312965B (en) Face unknown spoofing attack living body detection method and system
Maglogiannis et al. Characterization of digital medical images utilizing support vector machines
Nightingale et al. Perceptual and computational detection of face morphing
Fang et al. Fairness in face presentation attack detection
WO2020051232A1 (en) Decentralized biometric identification and authentication network
JP7152810B2 (en) Disease diagnosis system and method using multicolor model and neural network
CN111914617A (en) Face attribute editing method based on balanced stack type generation countermeasure network
Kanawong Computer-aided tongue image diagnosis and analysis
CN117974607A (en) Tongue image recognition method based on federal learning
Bansal et al. Image Correction and Identification of Ishihara Test Images for Color Blind Individual
Marulkar et al. Nail Disease Prediction using a Deep Learning Integrated Framework
CN115482595B (en) Specific character visual sense counterfeiting detection and identification method based on semantic segmentation
Alheeti et al. A Novel Detection System for Human Retina Based on Fuzzification Neural Network
Szandała et al. PRISM: principal image sections mapping
Pourreza Open-Set Source-Free Domain Adaptation in Fundus Images Analysis
Wachtler et al. Computational modeling of color vision

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination