CN110287930B - Wrinkle classification model training method and device - Google Patents
Wrinkle classification model training method and device Download PDFInfo
- Publication number
- CN110287930B CN110287930B CN201910586506.8A CN201910586506A CN110287930B CN 110287930 B CN110287930 B CN 110287930B CN 201910586506 A CN201910586506 A CN 201910586506A CN 110287930 B CN110287930 B CN 110287930B
- Authority
- CN
- China
- Prior art keywords
- wrinkle
- layer
- face image
- processing
- classification model
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
- G06V10/267—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
- G06V40/171—Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Multimedia (AREA)
- General Health & Medical Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Human Computer Interaction (AREA)
- Image Analysis (AREA)
Abstract
The embodiment of the application provides a wrinkle classification model training method and device, at least one wrinkle is marked on a face image sample with a wrinkle label, and the classification problem of the wrinkles is converted into a wrinkle segmentation problem, so that the marked at least one wrinkle can be used as a supervision signal to classify the wrinkles of the face image, and the accuracy of the classification result is improved.
Description
Technical Field
The application relates to the technical field of image processing, in particular to a wrinkle classification model training method and device.
Background
In some application scenarios, it is desirable to identify whether a facial image has wrinkles. In the related art, a Convolutional Neural Network (CNN) model is generally used as a classification model. However, since the face image generally has interference features such as eye shape and eyelash, the accuracy of classifying whether the face image contains wrinkles by CNN is low.
Disclosure of Invention
In view of the above, an objective of the present application is to provide a method and an apparatus for training a wrinkle classification model to at least partially improve the classification accuracy of the wrinkle classification model.
In order to achieve the above purpose, the embodiments of the present application adopt the following technical solutions:
in a first aspect, an embodiment of the present application provides a wrinkle classification model training method, configured to train a pre-constructed wrinkle classification model, where the wrinkle classification model includes a feature extraction part and an upsampling part; the feature extraction section includes a plurality of first processing layers, and the up-sampling section includes a plurality of second processing layers respectively corresponding to the plurality of first processing layers; the method comprises the following steps:
acquiring a plurality of face image samples and a mask image of each face image sample; a sample label is preset in each face image sample, the sample label is a wrinkle-containing label or a wrinkle-free label, and a mask image of the face image sample provided with the wrinkle-containing label comprises a contour of at least one wrinkle in the face image sample;
inputting each face image sample into the wrinkle classification model respectively; inputting the feature map output by the first processing layer at the highest layer in the feature extraction part into a full connection layer, and calculating the classification loss of the feature map output by the full connection layer and the sample label of the face image sample to obtain a first calculation result; calculating the segmentation loss of the feature map output by at least one second processing layer of the up-sampling part and the mask image of the face image sample to obtain at least one second calculation result;
and adjusting parameters of the wrinkle classification model according to the first calculation result and the second calculation result respectively to realize the training of the wrinkle classification model.
In a second aspect, an embodiment of the present application provides a wrinkle classification model training device, configured to train a pre-constructed wrinkle classification model, where the wrinkle classification model includes a feature extraction part and an upsampling part; the feature extraction section includes a plurality of first processing layers, and the up-sampling section includes a plurality of second processing layers respectively corresponding to the plurality of first processing layers; the device comprises:
the system comprises an acquisition module, a processing module and a display module, wherein the acquisition module is used for acquiring a plurality of face image samples and mask images of each face image sample; a sample label is preset in each face image sample, the sample label is a wrinkle-containing label or a wrinkle-free label, and a mask image of the face image sample provided with the wrinkle-containing label comprises a contour of at least one wrinkle in the face image sample;
the computing module is used for respectively inputting each face image sample into the wrinkle classification model; inputting the feature map output by the first processing layer at the highest layer in the feature extraction part into a full connection layer, and calculating the classification loss of the feature map output by the full connection layer and the sample label of the face image sample to obtain a first calculation result; calculating the segmentation loss of the feature map output by at least one second processing layer of the up-sampling part and the mask image of the face image sample to obtain at least one second calculation result;
and the parameter adjusting module is used for adjusting parameters of the wrinkle classification model according to the first calculation result and the second calculation result respectively so as to realize the training of the wrinkle classification model.
Compared with the prior art, the embodiment of the application provides a wrinkle classification model training method and device, the classification problem of wrinkles is converted into a segmentation problem, at least one wrinkle marked in a mask image can be used as a supervision signal to classify the wrinkles of a face image, and the classification accuracy can be improved through fewer samples.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present application, the drawings that are required to be used in the embodiments will be briefly described below, it should be understood that the following drawings only illustrate some embodiments of the present application and therefore should not be considered as limiting the scope, and for those skilled in the art, other related drawings can be obtained from the drawings without inventive effort.
Fig. 1 is a schematic block diagram of an image processing apparatus according to an embodiment of the present disclosure;
fig. 2 is a schematic structural diagram of a wrinkle classification model according to an embodiment of the present application;
fig. 3 is a schematic flowchart of a wrinkle classification model training method according to an embodiment of the present application;
fig. 4 is a functional block diagram of a wrinkle classification model training device according to an embodiment of the present application.
Icon: 10-an image processing device; 11-a processor; 12-a machine-readable storage medium; 20-wrinkle classification model; 21-a feature extraction section; 22-an upsampling part; 40-a wrinkle classification model training device; 41-an acquisition module; 42-a calculation module; 43-parameter adjustment module.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present application clearer, the technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are some embodiments of the present application, but not all embodiments. The components of the embodiments of the present application, generally described and illustrated in the figures herein, can be arranged and designed in a wide variety of different configurations.
Thus, the following detailed description of the embodiments of the present application, presented in the accompanying drawings, is not intended to limit the scope of the claimed application, but is merely representative of selected embodiments of the application. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
It should be noted that: like reference numbers and letters refer to like items in the following figures, and thus, once an item is defined in one figure, it need not be further defined and explained in subsequent figures.
Referring to fig. 1, fig. 1 is a block diagram of an image processing apparatus 10 according to an embodiment of the present disclosure. The image processing apparatus 10 may be any apparatus having an image processing function, such as a server, a Personal Computer (PC), or the like.
The image processing apparatus 10 includes a processor 11 and a machine-readable storage medium 12, and the processor 11 and the machine-readable storage medium 12 are communicatively connected by a system bus. The machine-readable storage medium 12 stores machine-executable instructions that, when executed by the processor 11, may implement the wrinkle classification model training method described below in this embodiment.
Additionally, the machine-readable storage medium 12 may also store a pre-constructed wrinkle classification model.
It is to be noted that the image processing apparatus shown in fig. 1 is merely illustrative, and the image processing apparatus may include more or less components than those shown in fig. 1, or may have a completely different configuration.
Referring to fig. 2, an architectural diagram of a wrinkle classification model 20 provided in the embodiment of the present application is exemplarily shown.
The wrinkle classification model 20 includes a feature extraction section 21 and an up-sampling section 22, and the feature extraction section 21 includes a plurality of first processing layers, such as L1, L2, L3, L4, and L5 shown in fig. 2. Each first processing layer comprises at least one preset layer group, and the preset layer group comprises a convolution (Conv) layer and an activation (Relu) function which are connected in sequence.
The up-sampling part 22 includes a plurality of second processing layers, such as L5, L6, L7, L8, and L9 shown in fig. 2, corresponding to the plurality of first processing layers, respectively. Here, L5 may be regarded as the first processing layer at the highest level in the feature extraction section 21, or may be regarded as the second processing layer at the highest level in the upsampling section 22. L5 as the first treated layer corresponds to L5 as the second treated layer, L6 corresponds to L4, L7 corresponds to L3, L8 corresponds to L2, and L9 corresponds to L1.
Each second processing layer comprises at least one preset layer group. It should be noted that the convolutional layers and the activation functions in different preset layer groups may have different parameters or the same parameters, and the embodiment is not limited thereto.
Optionally, in this embodiment, in the feature extraction part, a normalization layer (BatchNorm) may be further disposed between the convolution layer of each preset layer group and the activation function, and is used to perform normalization processing on the feature map output by the convolution layer.
Alternatively, the wrinkle classification model may be implemented based on an improved UNet model.
Referring to fig. 3, fig. 3 is a schematic flowchart illustrating a wrinkle classification model training process according to an embodiment of the present application, where the method can be applied to the image processing apparatus shown in fig. 1. The individual steps of the method will be described in detail below.
In step S31, a plurality of face image samples and a mask image for each face image sample are acquired.
Wherein, a sample label can be preset for each face image sample according to whether the face image sample contains wrinkles or not. Specifically, if one face image sample contains wrinkles, a wrinkle-containing label may be added thereto, and if one face image sample does not contain wrinkles, a wrinkle-free label may be added thereto. In practice, a specific field may be set for each face image sample, and when the value of the specific field is data a (e.g., 0), it indicates that there is a wrinkle label, and when the value of the specific field is data b (e.g., 1), it indicates that there is no wrinkle label.
In this embodiment, for a face image sample provided with a wrinkle-containing tag, the mask image thereof includes at least one wrinkle in the face image sample, for example, the wrinkle with the largest depth in the face image sample, so that the number of marked wrinkles can be reduced and the classification accuracy can be improved better.
In this embodiment, the mask image of one face image sample has the same size as that of the face image sample. Also, for a mask image including a wrinkle contour, the value of the pixel at the position where the wrinkle contour is located may be 255, and the values of the pixels at other positions may be 0 (i.e., black).
Step S32, inputting each face image sample into the wrinkle classification model respectively.
In this embodiment, each face image sample input into the wrinkle classification model is first processed sequentially by each first processing layer of the feature extraction part, and for the feature map output by each first processing layer, the feature map may be pooled (pool) and then input into the next first processing layer connected to the first processing layer. For the feature map output by each first processing layer, one copy of the feature map may be copied (copy), and the copied feature map may be cut (crop) and input to the second processing layer corresponding to the first processing layer.
Further, as for the feature map output from the first processing layer (for example, L5) or each second processing layer at the highest layer in the feature extraction section, the feature map may be up-sampled by using the mask image of the face image sample as a supervision signal and then input to the next connected second processing layer. The upsampling may be implemented by an upsampling layer or a deconvolution layer, and the embodiment is not limited thereto.
In the above case, each second processing layer will receive the upsampled feature map and the cut feature map. In this embodiment, each second processing layer may perform stitching (concat) on the received upsampled feature map and the cut feature map, and then process the stitching result and output a corresponding feature map.
Optionally, in this embodiment, when the feature map output by the first processing layer or the second processing layer is up-sampled by using the mask image of any facial image sample as the supervision signal, the mask image may be down-sampled to the same scale as the feature map. Specifically, the downsampling may be performed by using a maxporoling (maximum pooling) algorithm, and the wrinkle contour in the mask image and the black pixels around the wrinkle contour may be prevented from being averaged out, compared with the upsampling performed by using an average pooling algorithm in the related art.
Step S33, inputting the feature map output by the first processing layer at the highest layer in the feature extraction part into a full link layer, and calculating a classification loss (loss) between the feature map output by the full link layer and the sample label of the face image sample to obtain a first calculation result.
Specifically, the output of the full-link layer indicates the classification result of the wrinkle classification model on the face image sample, a specific classification loss function may be used to calculate the classification loss of the feature map output by the full-link layer and the sample label of the face image sample, and the obtained function value is the first calculation result. Alternatively, the classification loss function may be, for example, a Softmax (flexible maximum) function.
By providing the full link layer, the wrinkle classification model may also be applied to classification problems, in the case of being applied to segmentation problems.
Optionally, in this embodiment, step 33 may further include the following sub-steps:
and processing the feature graph output by the first processing layer at the highest layer in the feature extraction part through a global pooling layer, and inputting a processing result into the full connection layer.
In this way, the number of parameters can be reduced without affecting the accuracy of the processing result, thereby improving the training speed and the data processing speed when the wrinkle classification model is actually used.
Step S34, calculating a segmentation loss between the feature map output by the at least one second processing layer of the upsampled part and the mask image of the face image sample, and obtaining at least one second calculation result.
In detail, in this embodiment, the segmentation loss (loss) of the feature map output by each second processing layer except the lowest layer (e.g., L9 shown in fig. 2) in the upsampled part and the mask image of the face image sample may be calculated separately, and a plurality of the second calculation results may be obtained. In the implementation process, a preset segmentation function may be used to calculate the segmentation loss, and the obtained function value is one of the second calculation results.
By calculating segmentation loss at multiple locations in the wrinkle classification model (i.e., setting up multiple layers of supervisory signals), classification accuracy can be improved more effectively.
Alternatively, in this embodiment, before calculating the segmentation loss of the feature map output by each second processing layer except the lowest layer in the upsampling part and the mask image of the face image sample, the mask image of the face image sample may be downsampled into an image with the same scale as the feature map output by the second processing layer. Wherein the downsampling may be performed using the maxpoling algorithm described above.
Step S35, adjusting parameters of the wrinkle classification model according to the first calculation result and the second calculation result, respectively, to implement training of the wrinkle classification model.
In the implementation process, when the first calculation result is obtained, the hyper-parameter of the wrinkle classification model can be adjusted according to the first calculation result, so that the function value of the corresponding classification loss function is reduced. And adjusting the hyper-parameters of the wrinkle classification model according to a second calculation result every time the second calculation result is obtained so as to reduce the function value of the segmentation loss function outputting the second calculation result.
And (4) repeatedly adjusting the hyper-parameters of the wrinkle classification model according to the flow, so as to realize the training of the wrinkle classification model.
In the design, mask images of various scales are used as supervision signals to conduct wrinkle classification training on face image samples, so that a wrinkle classification model can learn more accurate wrinkle features in the training process, and the influence of interference features of the face on the wrinkle features, such as the influence of eyelashes on eye wrinkles, the influence of hair on head raising lines, the influence of the shape of the nose or the mouth corner on statuary lines and the like, can be at least partially filtered. Therefore, the classification precision of the wrinkle classification model is improved, namely the generalization capability of the wrinkle classification model is improved.
Referring to fig. 4, fig. 4 is a functional block diagram of a wrinkle classification model training device 40 according to an embodiment of the present application. The wrinkle classification model training device 40 comprises at least one functional module that may be stored in the machine-readable storage medium 12 in the form of machine-executable instructions. Specifically, the wrinkle classification model training device 40 includes an obtaining module 41, a calculating module 42, and a parameter adjusting module 43.
The obtaining module 41 is configured to obtain a plurality of face image samples and a mask image of each face image sample.
In this embodiment, a sample label is preset in each face image sample, where the sample label is specifically a wrinkle-containing label or a wrinkle-free label, and a mask image of the face image sample with the wrinkle-containing label includes a contour of at least one wrinkle in the face image sample.
The computing module 42 is configured to input each facial image sample into the wrinkle classification model; inputting the feature map output by the first processing layer at the highest layer in the feature extraction part into a full connection layer, and calculating the classification loss of the feature map output by the full connection layer and the sample label of the face image sample to obtain a first calculation result; and calculating the segmentation loss of the characteristic graph output by at least one second processing layer of the up-sampling part and the mask image of the face image sample to obtain at least one second calculation result.
Optionally, the calculating module 42 may be specifically configured to calculate a segmentation loss between the feature map output by each second processing layer except the lowest layer in the upsampled part and the mask image of the face image sample, respectively, to obtain a plurality of second calculation results.
Optionally, the calculation module 42 may be further specifically configured to, before calculating a segmentation loss between the feature map output by each second processing layer except the lowest layer in the upsampling part and the mask image of the facial image sample, downsample the mask image of the facial image sample into an image with the same scale as the feature map output by the second processing layer.
The parameter adjusting module 43 is configured to adjust parameters of the wrinkle classification model according to the first calculation result and the second calculation result, respectively, so as to implement training of the wrinkle classification model.
The detailed description of the above functional modules may refer specifically to the explanation of the corresponding steps above.
In summary, the present application provides a method and an apparatus for training a wrinkle classification model, which acquire a plurality of facial image samples and their respective mask images, wherein each sample is provided with a wrinkle-containing label or a wrinkle-free label, and the facial image sample provided with the wrinkle-containing label includes a contour of at least one wrinkle in the sample; inputting each sample into a wrinkle classification model, inputting a feature map output by the highest layer in a feature extraction part of the model into a full connection layer, and calculating the classification loss of the feature map output by the full connection layer and a sample label of the sample to obtain a first calculation result; calculating the segmentation loss of a characteristic graph output by at least one second processing layer of the upsampling part of the model and the mask image of the sample to obtain at least one second calculation result; and adjusting parameters of the model according to the obtained first calculation result and the second calculation result respectively to realize the training of the model. In this way, the problem of classifying the wrinkles is converted into a segmentation problem, at least one wrinkle marked in the mask image can be used as a supervision signal to classify the wrinkles of the face image, and the classification accuracy can be improved through fewer samples.
In the embodiments provided in the present application, it should be understood that the disclosed apparatus and method may be implemented in other ways. The above-described apparatus embodiments are merely illustrative, and for example, the flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of apparatus, methods and computer products according to embodiments of the present application. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
It is noted that, herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other identical elements in a process, method, article, or apparatus that comprises the element.
The above description is only for the specific embodiments of the present application, but the scope of the present application is not limited thereto, and any person skilled in the art can easily conceive of the changes or substitutions within the technical scope of the present application, and shall be covered by the scope of the present application. Therefore, the protection scope of the present application shall be subject to the protection scope of the claims.
Claims (8)
1. A wrinkle classification model training method is used for training a pre-constructed wrinkle classification model and is characterized in that the wrinkle classification model comprises a feature extraction part and an up-sampling part; the feature extraction section includes a plurality of first processing layers, and the up-sampling section includes a plurality of second processing layers respectively corresponding to the plurality of first processing layers; the method comprises the following steps:
acquiring a plurality of face image samples and a mask image of each face image sample; a sample label is preset in each face image sample, the sample label is a wrinkle-containing label or a wrinkle-free label, and a mask image of the face image sample provided with the wrinkle-containing label comprises a contour of at least one wrinkle in the face image sample;
inputting each face image sample into the wrinkle classification model respectively; inputting the feature map output by the first processing layer at the highest layer in the feature extraction part into a full connection layer, and calculating the classification loss of the feature map output by the full connection layer and the sample label of the face image sample to obtain a first calculation result;
respectively calculating the feature map output by each second processing layer except the lowest layer in the up-sampling part and the segmentation loss of the mask image of the face image sample to obtain a plurality of second calculation results;
when a first calculation result is obtained, the hyperparameter of the wrinkle classification model is adjusted according to the first calculation result so as to reduce the function value of the corresponding classification loss function, and when a second calculation result is obtained, the hyperparameter of the wrinkle classification model can be adjusted according to the second calculation result so as to reduce the function value of the segmentation loss function outputting the second calculation result, thereby realizing the training of the wrinkle classification model.
2. The method according to claim 1, wherein the step of inputting the feature map output from the first processing layer at the highest layer in the feature extraction part into a full connection layer comprises:
and processing the feature graph output by the first processing layer at the highest layer in the feature extraction part through a global pooling layer, and inputting a processing result into the full connection layer.
3. The method of claim 1, further comprising:
and before calculating the segmentation loss of the feature map output by each second processing layer except the lowest layer in the upsampling part and the mask image of the face image sample, downsampling the mask image of the face image sample into an image with the same scale as the feature map output by the second processing layer.
4. The method of claim 3, wherein the step of down-sampling the mask image of the facial image sample into an image having the same scale as the feature map output by the second processing layer comprises:
and downsampling the mask image of the face image sample by adopting a Maxboosting algorithm to obtain an image with the same scale as the feature map output by the second processing layer.
5. The method according to claim 1 or 2, characterized in that the method further comprises:
processing each face image sample input into the wrinkle classification model by using the feature extraction part, cutting a feature graph output by each first processing layer, inputting the feature graph into a second processing layer corresponding to the first processing layer, pooling the feature graphs and inputting the feature graphs into a next first processing layer connected with the first processing layer;
for the feature map output by the first processing layer or each second processing layer at the highest layer in the feature extraction part, the mask image of the face image sample is used as a supervision signal to perform up-sampling on the feature map and then input into the next second processing layer;
and each second processing layer splices the received feature map subjected to the upsampling and the cut feature map, processes the splicing result and outputs the processed result.
6. A wrinkle classification model training device is used for training a pre-constructed wrinkle classification model and is characterized in that the wrinkle classification model comprises a feature extraction part and an up-sampling part; the feature extraction section includes a plurality of first processing layers, and the up-sampling section includes a plurality of second processing layers respectively corresponding to the plurality of first processing layers; the device comprises:
the system comprises an acquisition module, a processing module and a display module, wherein the acquisition module is used for acquiring a plurality of face image samples and mask images of each face image sample; a sample label is preset in each face image sample, the sample label is a wrinkle-containing label or a wrinkle-free label, and a mask image of the face image sample provided with the wrinkle-containing label comprises a contour of at least one wrinkle in the face image sample;
the computing module is used for respectively inputting each face image sample into the wrinkle classification model; inputting the feature map output by the first processing layer at the highest layer in the feature extraction part into a full connection layer, and calculating the classification loss of the feature map output by the full connection layer and the sample label of the face image sample to obtain a first calculation result;
respectively calculating the feature map output by each second processing layer except the lowest layer in the up-sampling part and the segmentation loss of the mask image of the face image sample to obtain a plurality of second calculation results;
and the parameter adjusting module is used for adjusting the hyperparameter of the wrinkle classification model according to the first calculation result when the first calculation result is obtained so as to reduce the function value of the corresponding classification loss function, and adjusting the hyperparameter of the wrinkle classification model according to the second calculation result every time a second calculation result is obtained so as to reduce the function value of the segmentation loss function outputting the second calculation result and realize the training of the wrinkle classification model.
7. The apparatus according to claim 6, wherein the computing module is specifically configured to process, by using a global pooling layer, the feature map output by the first processing layer at the highest layer in the feature extraction part, and output the processing result to the full-connection layer.
8. The apparatus according to claim 6, wherein the computing module is further configured to, before computing a segmentation loss between the feature map output by each second processing layer except the lowest layer in the upsampled part and the mask image of the facial image sample, downsample the mask image of the facial image sample into an image with the same scale as the feature map output by the second processing layer.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910586506.8A CN110287930B (en) | 2019-07-01 | 2019-07-01 | Wrinkle classification model training method and device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910586506.8A CN110287930B (en) | 2019-07-01 | 2019-07-01 | Wrinkle classification model training method and device |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110287930A CN110287930A (en) | 2019-09-27 |
CN110287930B true CN110287930B (en) | 2021-08-20 |
Family
ID=68021547
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910586506.8A Active CN110287930B (en) | 2019-07-01 | 2019-07-01 | Wrinkle classification model training method and device |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110287930B (en) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110738678B (en) * | 2019-10-18 | 2022-05-31 | 厦门美图宜肤科技有限公司 | Face fine line detection method and device, electronic equipment and readable storage medium |
CN112101185B (en) * | 2020-09-11 | 2024-05-28 | 深圳数联天下智能科技有限公司 | Method for training wrinkle detection model, electronic equipment and storage medium |
KR102495889B1 (en) * | 2022-07-13 | 2023-02-06 | 주식회사 룰루랩 | Method for detecting facial wrinkles using deep learning-based wrinkle detection model trained according to semi-automatic labeling and apparatus for the same |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108985181A (en) * | 2018-06-22 | 2018-12-11 | 华中科技大学 | A kind of end-to-end face mask method based on detection segmentation |
CN109191476A (en) * | 2018-09-10 | 2019-01-11 | 重庆邮电大学 | The automatic segmentation of Biomedical Image based on U-net network structure |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2016197303A1 (en) * | 2015-06-08 | 2016-12-15 | Microsoft Technology Licensing, Llc. | Image semantic segmentation |
CN109087306A (en) * | 2018-06-28 | 2018-12-25 | 众安信息技术服务有限公司 | Arteries iconic model training method, dividing method, device and electronic equipment |
CN109544545A (en) * | 2018-11-30 | 2019-03-29 | 济南浪潮高新科技投资发展有限公司 | A kind of salt mine intelligent detecting method and system based on convolutional neural networks |
CN109685776B (en) * | 2018-12-12 | 2021-01-19 | 华中科技大学 | Pulmonary nodule detection method and system based on CT image |
CN109840471B (en) * | 2018-12-14 | 2023-04-14 | 天津大学 | Feasible road segmentation method based on improved Unet network model |
CN109934796A (en) * | 2018-12-26 | 2019-06-25 | 苏州雷泰医疗科技有限公司 | A kind of automatic delineation method of organ based on Deep integrating study |
CN109741343B (en) * | 2018-12-28 | 2020-12-01 | 浙江工业大学 | T1WI-fMRI image tumor collaborative segmentation method based on 3D-Unet and graph theory segmentation |
CN109712165B (en) * | 2018-12-29 | 2022-12-09 | 安徽大学 | Similar foreground image set segmentation method based on convolutional neural network |
CN109840477B (en) * | 2019-01-04 | 2020-11-24 | 苏州飞搜科技有限公司 | Method and device for recognizing shielded face based on feature transformation |
CN109919961A (en) * | 2019-02-22 | 2019-06-21 | 北京深睿博联科技有限责任公司 | A kind of processing method and processing device for aneurysm region in encephalic CTA image |
-
2019
- 2019-07-01 CN CN201910586506.8A patent/CN110287930B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108985181A (en) * | 2018-06-22 | 2018-12-11 | 华中科技大学 | A kind of end-to-end face mask method based on detection segmentation |
CN109191476A (en) * | 2018-09-10 | 2019-01-11 | 重庆邮电大学 | The automatic segmentation of Biomedical Image based on U-net network structure |
Also Published As
Publication number | Publication date |
---|---|
CN110287930A (en) | 2019-09-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108492271B (en) | Automatic image enhancement system and method fusing multi-scale information | |
CN110287930B (en) | Wrinkle classification model training method and device | |
EP3923233A1 (en) | Image denoising method and apparatus | |
CN111476719B (en) | Image processing method, device, computer equipment and storage medium | |
CN104484658A (en) | Face gender recognition method and device based on multi-channel convolution neural network | |
CN110879982A (en) | Crowd counting system and method | |
CN111783779B (en) | Image processing method, apparatus and computer readable storage medium | |
DE112020000448T5 (en) | CAMERA SELF CALIBRATION NETWORK | |
JP2018072938A (en) | Number-of-targets estimation device, number-of-targets estimation method, and program | |
Sakurai et al. | Plant Growth Prediction using Convolutional LSTM. | |
CN110059666A (en) | A kind of attention detection method and device | |
US20220261670A1 (en) | Methods of Processing and Generating Image Data in a Connectionist Network | |
EP4227856A1 (en) | System and method for defect detection | |
CN114565539B (en) | Image defogging method based on online knowledge distillation | |
CN111626379B (en) | X-ray image detection method for pneumonia | |
CN112818821A (en) | Human face acquisition source detection method and device based on visible light and infrared light | |
CN114299375B (en) | Underwater group target detection method fusing YOLOv4 and deformable convolution | |
CN116664677A (en) | Sight estimation method based on super-resolution reconstruction | |
CN114492634A (en) | Fine-grained equipment image classification and identification method and system | |
CN112767331B (en) | Image anomaly detection method based on zero sample learning | |
CN116420163A (en) | Identification system, identification method, program, learning method, learning-completed model, distillation model, and learning data set generation method | |
CN114078149A (en) | Image estimation method, electronic equipment and storage medium | |
CN114418064B (en) | Target detection method, terminal equipment and storage medium | |
CN115170897A (en) | Image processing method based on mask region convolution neural network and application thereof | |
US20240282077A1 (en) | Image processing apparatus, training apparatus, image processing method, training method, and non-transitory computer-readable storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
TR01 | Transfer of patent right | ||
TR01 | Transfer of patent right |
Effective date of registration: 20211209 Address after: 361100 568, No. 942, tonglong Second Road, torch high tech Zone (Xiang'an) Industrial Zone, Xiang'an District, Xiamen City, Fujian Province Patentee after: Xiamen Meitu Yifu Technology Co.,Ltd. Address before: B1f-089, Zone C, Huaxun building, software park, torch high tech Zone, Xiamen City, Fujian Province Patentee before: XIAMEN HOME MEITU TECHNOLOGY Co.,Ltd. |