WO2021052159A1 - Adversarial transfer learning-based face beauty prediction method and device - Google Patents

Adversarial transfer learning-based face beauty prediction method and device Download PDF

Info

Publication number
WO2021052159A1
WO2021052159A1 PCT/CN2020/112528 CN2020112528W WO2021052159A1 WO 2021052159 A1 WO2021052159 A1 WO 2021052159A1 CN 2020112528 W CN2020112528 W CN 2020112528W WO 2021052159 A1 WO2021052159 A1 WO 2021052159A1
Authority
WO
WIPO (PCT)
Prior art keywords
feature
face
beauty prediction
prediction model
face beauty
Prior art date
Application number
PCT/CN2020/112528
Other languages
French (fr)
Chinese (zh)
Inventor
翟懿奎
项俐
甘俊英
麦超云
曾军英
应自炉
Original Assignee
五邑大学
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 五邑大学 filed Critical 五邑大学
Publication of WO2021052159A1 publication Critical patent/WO2021052159A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Definitions

  • the invention relates to the field of image processing, in particular to a method and device for predicting the beauty of a face based on anti-migration learning.
  • Face beauty prediction technology is widely used in the field of photography.
  • deep learning technology the application of deep learning technology to facial beauty prediction technology makes facial beauty prediction results more accurate and more in line with people's cognition.
  • single-task learning ignores the association between tasks, and multi-task learning adds unnecessary combinations to the deep learning network, which increases the redundancy of deep learning tasks, and also increases the burden of network training, which seriously affects The efficiency of classification and recognition.
  • the purpose of the present invention is to solve at least one of the technical problems existing in the prior art, to provide a method and device for predicting the beauty of a face based on multi-task migration, and to reduce the amount of calculation through the similarity measurement.
  • the method for predicting the beauty of a face based on anti-migration learning includes the following steps:
  • the main task is the task of predicting the beauty of the face
  • the auxiliary task is the task of identifying the beauty factors of the face, N>A;
  • measuring the similarity between the N auxiliary tasks and the main task, and obtaining A auxiliary tasks with the highest similarity includes the following steps:
  • the first face beauty prediction model includes a first preprocessing layer for preprocessing face images connected in sequence, a first feature sharing layer for extracting shared image features, and A first independent feature extraction layer that extracts independent features from shared image features, and a first classification layer;
  • the second face beauty prediction model includes a second pre-processing layer connected in sequence, a second feature sharing layer, and a second independent The feature extraction layer is used to fuse independent features and the geometric features and texture features of the corresponding facial beauty prediction task, and the second classification layer.
  • the A first face beauty prediction model is used as the source domain and the second face beauty prediction model is used as the target domain, and the counter network is pre-trained to find the source domain relative to the target domain Migrating the universal feature parameters to the second face beauty prediction model specifically includes the following steps:
  • Mapping step map the source feature to the target feature space to obtain the pseudo target feature T(G(x s ));
  • Distinguish step distinguish the source of the target feature and the pseudo-target feature and calculate the error through the loss function
  • Optimization step use the regularization term r(G(x s ), T(G(x s ))) to measure the distance between the source feature and the pseudo-target feature, and then combine the error to optimize the mapping of the source feature to the target feature space;
  • mapping step Repeat the mapping step, the distinguishing step and the optimization step until the source domain and the target domain are adapted to obtain the common feature parameters
  • Migration step Migrate the general feature parameters to the second face beauty prediction model.
  • the above-mentioned face beauty prediction method based on anti-migration learning has at least the following beneficial effects: find the highest correlation among the auxiliary tasks of identifying face factors through the similarity measure, and construct the first face beauty prediction based on this.
  • the model is pre-trained; the training cost of pre-training is reduced, the deviation caused by the auxiliary tasks with unrelated factors to the first face beauty prediction model is reduced, and the negative migration is avoided.
  • the general feature parameters formed after the pre-training of the confrontation network are migrated to the second face beauty prediction model to achieve the final face beauty prediction. Through the confrontation migration learning, the calculation amount of the second face beauty prediction model training is reduced, and the training time is compressed. Use fewer training images to obtain a more accurate model effect.
  • a face beauty prediction device based on anti-transfer learning includes:
  • the similarity measurement module is used to measure the similarity between N auxiliary tasks and the main task to obtain A auxiliary tasks with the highest similarity.
  • the main task is the task of predicting the beauty of the face
  • the auxiliary task is the task of identifying the beauty factors of the face. N>A;
  • the first model establishment module is used to establish A first face beauty prediction model corresponding to A auxiliary tasks with the highest similarity
  • the second model building module is used to build a second face beauty prediction model for face beauty prediction
  • the parameter migration module is configured to use A of the first face beauty prediction model as the source domain and the second face beauty prediction model as the target domain to find the common feature parameters of the source domain relative to the target domain through pre-training of the confrontation network , Migrating the general feature parameters to the second face beauty prediction model;
  • the calculation module is used to input the face image to be tested to the second face beauty prediction model that is retrained to output the face beauty prediction result.
  • the similarity measurement module includes:
  • the feature expression acquisition module is used to construct a fully-supervised specific network for the N auxiliary tasks and the main task respectively and perform training to obtain the feature expression E s (I) of each task;
  • the tightness measurement module is used to construct a migration network between N auxiliary tasks and the main task, and measure the task tightness between the N auxiliary tasks and the main task, and the task tightness is calculated as follows: Where I is the input, D is the data set, f t (I) is the true value of the t-th input I, L t is the loss between the true value and the predicted value, and E I ⁇ D represents the expected value;
  • the normalization processing module is used to normalize the loss of the migration network through the analytic hierarchy process to obtain the correlation matrix
  • the optimization processing module is used to optimize the incidence matrix to obtain A auxiliary tasks with the highest similarity.
  • the first face beauty prediction model includes a first preprocessing layer for preprocessing face images that are sequentially connected, a first feature sharing layer for extracting shared image features, and A first independent feature extraction layer that extracts independent features from shared image features, and a first classification layer;
  • the second face beauty prediction model includes a second pre-processing layer connected in sequence, a second feature sharing layer, and a second independent The feature extraction layer is used to fuse independent features and the geometric features and texture features of the corresponding facial beauty prediction task, and the second classification layer.
  • the parameter migration module includes:
  • the mapping module is used to map the source feature to the target feature space to obtain the pseudo target feature T(G(x s ));
  • the distinguishing module is used to distinguish the source of the target feature and the pseudo-target feature and calculate the error through the loss function
  • the optimization module is used to use the regularization term r(G(x s ), T(G(x s ))) to measure the distance between the source feature and the pseudo-target feature, and then combine the error to optimize the source feature to the target feature space Mapping
  • the parameter acquisition module is used to acquire general characteristic parameters when both the source domain and the target domain are adapted;
  • the migration sub-module is used to migrate the general feature parameters to the second face beauty prediction model.
  • the above-mentioned facial beauty prediction device based on anti-migration learning has at least the following beneficial effects: the similarity measurement module finds the highest correlation among the auxiliary tasks of identifying facial factors, and the pre-training module builds the first
  • the face beauty prediction model is pre-trained; the training cost of pre-training is reduced, and the deviation caused by the auxiliary tasks with irrelevant factors to the first face beauty prediction model is avoided, and negative migration is avoided.
  • the parameter migration module migrates the general migration parameters formed after the pre-training of the confrontation network to the second face beauty prediction model, and realizes the final face beauty prediction. Through the confrontation migration learning, the calculation amount is reduced, the training time is compressed, and the calculation module is used more. Fewer training images can get a more accurate model effect.
  • FIG. 1 is a step diagram of a face beauty prediction method based on anti-migration learning according to an embodiment of the present invention
  • FIG. 2 is a specific step diagram of step S10
  • FIG. 3 is a schematic diagram of a face beauty prediction method based on anti-migration learning according to an embodiment of the present invention
  • FIG. 4 is another schematic diagram of a face beauty prediction method based on anti-migration learning according to an embodiment of the present invention.
  • FIG. 5 is a specific step diagram of step S30;
  • FIG. 6 is a structural diagram of an apparatus for predicting face beauty based on anti-migration learning according to an embodiment of the present invention.
  • Figure 7 is a structural diagram of the parameter migration module.
  • an embodiment of the present invention provides a face beauty prediction method based on anti-migration learning, which includes the following steps:
  • Step S10 measure the similarity between the N auxiliary tasks and the main task, and obtain A auxiliary tasks with the highest similarity, where the main task is the task of predicting the beauty of the face, and the auxiliary task is the task of identifying the beauty factors of the face, N>A;
  • Step S20 Establish A first face beauty prediction model corresponding to A secondary tasks with the highest similarity, and establish a second face beauty prediction model for face beauty prediction;
  • Step S30 Use A of the first face beauty prediction model as the source domain and the second face beauty prediction model as the target domain. Through pre-training of the confrontation network to find the common feature parameters of the source domain relative to the target domain, Transferring the general feature parameters to the second face beauty prediction model;
  • Step S40 Input the face image to be tested to the second trained face beauty prediction model to output the face beauty prediction result.
  • the similarity measure is used to find the most relevant among the auxiliary tasks of recognizing face factors, and based on this, the first face beauty prediction model is constructed for pre-training; the training cost of pre-training is reduced, Reduce the bias caused by the auxiliary tasks with irrelevant factors to the first face beauty prediction model, and avoid negative transfer.
  • the auxiliary tasks of multiple recognition of facial factors mainly include recognition of facial expression, age, gender, skin color, eye size, eye distance, ear shape, nose size, nose bridge height, lip shape, etc.
  • Transfer learning is to improve the learning of new tasks by transferring knowledge from related tasks that have been learned.
  • Migrating the general feature parameters formed after the adversarial network pre-training to the second face beauty prediction model can effectively reduce the calculation amount of the second face beauty prediction model training, compress the training time, and achieve more accurate results with fewer training images The effect of the model.
  • step S10 specifically includes the following steps:
  • Step S11 Construct a fully-supervised specific network for each of the N auxiliary tasks and the main task and perform training to obtain the feature expression E s (I) of each task.
  • Each specific network has an encoder and a decoder, and all codes are The decoders all have the same ResNet50 structure, and the decoders correspond to different tasks;
  • Step S12. Construct a migration network between N auxiliary tasks and the main task, and measure the task tightness between the N auxiliary tasks and the main task.
  • the task tightness calculation method is: Where I is the input, D is the data set, f t (I) is the true value of the t-th input I, L t is the loss between the true value and the predicted value, and E I ⁇ D represents the expected value;
  • Step S13 Normalize the loss of the migration network through the analytic hierarchy process to obtain the correlation matrix; specifically, for each task pair (i, j) of the source task pointing to the target task, the test set is taken out by the leave-out method after the migration ; For each task, construct a matrix W t , and then use the Laplace smoothing method to control the output result of the matrix W t within the range [0.001, 0.999], and then transform to obtain an incidence matrix, which reflects the similarity between tasks Probability.
  • the calculation method of each element w′ i, j in W′ t is:
  • Step S14 Perform optimization processing on the incidence matrix to obtain A auxiliary tasks with the highest similarity, that is, obtain a subgraph selection problem based on the incidence matrix.
  • the first face beauty prediction model corresponding to each auxiliary task includes a first preprocessing layer 11 for preprocessing face images connected in sequence, and a first feature sharing layer for extracting shared image features. 12.
  • the first face beauty prediction model of all auxiliary tasks can be combined into one, multiple first preprocessing layers 11 are combined into one, and multiple first feature sharing layers 12 are combined into one.
  • the shared image features of all tasks are placed in the first feature sharing layer 12; after the first feature sharing layer 12, multiple first independent feature extraction layers 13 and multiple first classification layers 14 are connected corresponding to different tasks.
  • the second face beauty prediction model includes a second preprocessing layer 21, a second feature sharing layer 22, and a second independent feature extraction layer 23, which are connected in sequence, and are used to fuse independent features with geometric features and textures corresponding to face beauty prediction tasks.
  • the first independent feature extraction layer 13 and the second independent feature extraction layer 23 both include 1 convolutional layer, 1 BN layer, 1 activation function layer, and 1 pooling layer that are connected in sequence, and the first classification layer 14
  • Both the second classification layer 25 and the second classification layer 25 include two fully connected layers.
  • step S30 specifically includes the following steps:
  • Step S32 Mapping step: map the source feature to the target feature space to obtain the pseudo target feature T(G(x s ));
  • Step S33 distinguishing step: the discriminator distinguishes the source of the target feature and the pseudo-target feature and calculates the error through the loss function; wherein the loss function can be a commonly used loss function, such as a cross entropy loss function or a mean square error loss function;
  • Step S34 optimization step: use the regularization term r(G(x s ), T(G(x s ))) to measure the distance between the source feature and the pseudo-target feature, and then combine the error to optimize the source feature to the target feature space
  • the mapping ;
  • Step S35 repeat the mapping step, the distinguishing step and the optimization step until the source domain and the target domain are adapted to obtain the common feature parameters; among them, the source domain and the target domain are adapted to each other, that is, the error of the loss function calculation is less than the set Threshold; the smaller the error, the closer the source domain to the target domain, the better the migration effect;
  • Step S36 Migration step: Migrate the general feature parameters to the second face beauty prediction model.
  • the confrontation network is used to reduce the distribution difference between the source domain and the target domain, that is, the distribution difference between the first face beauty prediction model and the second face beauty prediction model; to achieve the purpose of knowledge transfer and reuse from the source domain to the target domain.
  • the overall migration of the general feature parameters of the first face beauty prediction model to the second face beauty prediction model is specifically as follows: each of the first feature sharing layer 12, the first independent feature extraction layer 13, and the first classification layer 14 The parameters are correspondingly migrated to the second feature sharing layer 22, the second independent feature extraction layer 23, and the second classification layer 25; the entire second face beauty prediction model accepts the parameters of the migration learning, and the overall model is optimized.
  • the above face beauty prediction method measures the similarity of multiple auxiliary tasks, selects the auxiliary tasks with high similarity to the main task, builds a pre-training network model based on this, and transfers the parameters of the pre-trained network model to the main task face
  • the network model of the main task is optimized to avoid the negative transfer caused by useless parameters caused by the training of irrelevant auxiliary tasks, which can greatly reduce the amount of training, improve the efficiency of classification and recognition, and improve the accuracy of classification and recognition.
  • an apparatus for predicting face beauty based on anti-transfer learning, applying the above-mentioned face beauty prediction method includes:
  • the similarity measurement module 100 is used to measure the similarity between N auxiliary tasks and the main task to obtain A auxiliary tasks with the highest similarity, where the main task is the task of predicting the beauty of the face, and the auxiliary task is the task of identifying the factors of the beauty of the face , N>A;
  • the first model building module 210 is used to build A first face beauty prediction models corresponding to A auxiliary tasks with the highest similarity;
  • the second model building module 220 is used to build a second face beauty prediction model used for face beauty prediction
  • the parameter migration module 300 is configured to use A of the first face beauty prediction model as the source domain and the second face beauty prediction model as the target domain to find the common features of the source domain relative to the target domain through adversarial network pre-training Parameters, transferring the general feature parameters to the second face beauty prediction model;
  • the calculation module 400 is used to input the face image to be tested to the second face beauty prediction model that is retrained to output the face beauty prediction result.
  • the similarity measurement module is used to find the most relevant among the auxiliary tasks of identifying face factors, and to construct the first face beauty prediction model for pre-training; reducing the training cost of pre-training , To reduce the bias caused by the auxiliary tasks with irrelevant factors to the first face beauty prediction model, and avoid negative migration.
  • the parameter migration module 300 migrates the general feature parameters formed after pre-training to the second face beauty prediction model, and realizes the final face beauty prediction, reduces the calculation amount through migration learning, compresses the training time, and reaches the calculation module 400 to use less The training images to obtain a more accurate model effect.
  • the similarity measurement module 100 includes:
  • the feature expression acquisition module 110 is used to construct a fully-supervised specific network for the N auxiliary tasks and the main task respectively and perform training to obtain the feature expression E s (I) of each task;
  • the tightness measurement module 120 is used to construct a migration network between N auxiliary tasks and the main task, and measure the task tightness between the N auxiliary tasks and the main task.
  • the calculation method of the task tightness is: Where I is the input, D is the data set, f t (I) is the true value of the t-th input I, L t is the loss between the true value and the predicted value, and E I ⁇ D represents the expected value;
  • the normalization processing module 130 is used to normalize the loss of the migration network through the analytic hierarchy process to obtain an incidence matrix
  • the optimization processing module 140 is used for optimizing the incidence matrix to obtain A auxiliary tasks with the highest similarity.
  • the first face beauty prediction model includes a first preprocessing layer 11 for preprocessing face images, a first feature sharing layer 12 for extracting shared image features, and a first feature sharing layer 12 for extracting features from the shared image.
  • the second face beauty prediction model includes a second preprocessing layer 21, a second feature sharing layer 22, and a second independent feature extraction layer 23 connected in sequence ,
  • a feature fusion layer 24 and a second classification layer 25 used to fuse independent features and geometric features and texture features of the corresponding facial beauty prediction task.
  • the parameter migration module 300 includes:
  • the mapping module 330 is used to map the source feature to the target feature space to obtain the pseudo target feature T(G(x s ));
  • the distinguishing module 340 is used to distinguish the source of the target feature and the pseudo-target feature and calculate the error through the loss function;
  • the optimization module 350 is configured to use the regularization term r(G(x s ), T(G(x s ))) to measure the distance between the source feature and the pseudo target feature, and then combine the error to optimize the source feature to the target feature space
  • the mapping ;
  • the parameter acquisition module 360 is used to acquire general characteristic parameters when both the source domain and the target domain are adapted;
  • the migration sub-module 370 is configured to migrate the general feature parameters to the second face beauty prediction model.
  • the parameter migration module 300 correspondingly migrates the respective parameters of the first feature sharing layer 12, the first independent feature extraction layer 13, and the first classification layer 14 to the second feature sharing layer 22, the second independent feature extraction layer 23, and the second feature sharing layer.
  • Classification layer 25 the respective parameters of the first feature sharing layer 12, the first independent feature extraction layer 13, and the first classification layer 14 to the second feature sharing layer 22, the second independent feature extraction layer 23, and the second feature sharing layer.
  • the above-mentioned facial beauty prediction device measures the similarity of multiple auxiliary tasks, selects the auxiliary tasks with high similarity to the main task, and builds a parameter migration network model based on the confrontation network based on this, and transfers the parameters to the general features generated by the network model
  • the parameters are migrated to the main task face beauty recognition network model, so that the main task network model is optimized, and the negative transfer caused by useless parameters caused by the training of irrelevant auxiliary tasks can be avoided, which can greatly reduce the amount of training and improve the efficiency of classification and recognition. Improve classification and recognition accuracy.
  • Another embodiment of the present invention provides a storage medium that stores executable instructions, and the executable instructions enable a processor connected to the storage medium to perform facial beauty prediction methods based on the above-mentioned anti-migration learning-based face beauty prediction method.
  • the image is processed to obtain the beautiful face recognition result.

Abstract

An adversarial transfer learning-based face beauty prediction method and device, the method comprising the following steps: screening the most relevant auxiliary tasks from among a plurality of face factor recognition auxiliary tasks by means of similarity measurement, and on this basis, constructing a first face beauty prediction model; transferring general feature parameters formed after by pre-training an adversarial network to a second face beauty prediction model; and inputting a face image to be measured to perform recognition. The present method reduces the training cost of pre-training, reduces the negative transfer caused by auxiliary tasks with unrelated factors, and also reduces the calculation complexity for retraining the second face beauty prediction model, so that a more accurate model is obtained using fewer training images.

Description

基于对抗迁移学习的人脸美丽预测方法及装置Face beauty prediction method and device based on anti-migration learning 技术领域Technical field
本发明涉及图像处理领域,特别是基于对抗迁移学习的人脸美丽预测方法及装置。The invention relates to the field of image processing, in particular to a method and device for predicting the beauty of a face based on anti-migration learning.
背景技术Background technique
人脸美丽预测技术在拍照领域得到广泛的应用。同时,随着深度学习技术的发展,将深度学习技术应用到人脸美丽预测技术上使人脸美预测结果更精确,更符合人们的认知。但单任务学习忽略了任务之间的关联,而多任务学习又使不必要的组合添加至深度学习网络中,增加了深度学习任务的冗余度,也加重了网络训练的负担,严重影响到分类识别的效率。Face beauty prediction technology is widely used in the field of photography. At the same time, with the development of deep learning technology, the application of deep learning technology to facial beauty prediction technology makes facial beauty prediction results more accurate and more in line with people's cognition. However, single-task learning ignores the association between tasks, and multi-task learning adds unnecessary combinations to the deep learning network, which increases the redundancy of deep learning tasks, and also increases the burden of network training, which seriously affects The efficiency of classification and recognition.
发明内容Summary of the invention
本发明的目的在于至少解决现有技术中存在的技术问题之一,提供基于多任务迁移的人脸美丽预测方法及装置,通过相似性度量减少计算量。The purpose of the present invention is to solve at least one of the technical problems existing in the prior art, to provide a method and device for predicting the beauty of a face based on multi-task migration, and to reduce the amount of calculation through the similarity measurement.
本发明解决其问题所采用的技术方案是:The technical solutions adopted by the present invention to solve its problems are:
本发明的第一方面,基于对抗迁移学习的人脸美丽预测方法,包括以下步骤:In the first aspect of the present invention, the method for predicting the beauty of a face based on anti-migration learning includes the following steps:
度量N个辅任务与主任务间的相似性,得到A个相似性最高的辅任务,其中主任务为人脸美丽预测任务,辅任务为识别人脸美丽因素的任务,N>A;Measure the similarity between N auxiliary tasks and the main task, and obtain A auxiliary tasks with the highest similarity. The main task is the task of predicting the beauty of the face, and the auxiliary task is the task of identifying the beauty factors of the face, N>A;
建立对应A个相似度最高的辅任务的A个第一人脸美丽预测模型以及建立用于人脸美丽预测的第二人脸美丽预测模型;Establish A first face beauty prediction model corresponding to A secondary tasks with the highest similarity, and establish a second face beauty prediction model for face beauty prediction;
将A个所述第一人脸美丽预测模型作为源域和所述第二人脸美丽预测模型作为目标域通过对抗网络预训练以寻找源域相对目标域的通用特征参数,将所述通用特征参数迁移至所述第二人脸美丽预测模型;Taking A of the first face beauty prediction model as the source domain and the second face beauty prediction model as the target domain, through adversarial network pre-training to find the common feature parameters of the source domain relative to the target domain, the common feature The parameters are transferred to the second face beauty prediction model;
输入待测人脸图像至再训练好的第二人脸美丽预测模型输出人脸美丽预测结果。Input the face image to be tested to the retrained second face beauty prediction model to output the face beauty prediction result.
根据本发明的第一方面,所述度量N个辅任务与主任务间的相似性,得到A个相似性最高的辅任务包括以下步骤:According to the first aspect of the present invention, measuring the similarity between the N auxiliary tasks and the main task, and obtaining A auxiliary tasks with the highest similarity includes the following steps:
对N个辅任务与主任务分别构建全监督的特定网络并进行训练得到每个任务的特征表达E s(I); Construct a fully-supervised specific network for the N auxiliary tasks and the main task respectively and perform training to obtain the characteristic expression E s (I) of each task;
构建N个辅任务与主任务间的迁移网络,并度量N个辅任务与主任务间的任务紧密度, 所述任务紧密度的计算方式为:
Figure PCTCN2020112528-appb-000001
式中I是输入,D是数据集,f t(I)是第t个输入I的真实值,L t是真实值与预测值之间的损失,E I∈D表示期望值;
Construct a migration network between the N auxiliary tasks and the main task, and measure the task tightness between the N auxiliary tasks and the main task, and the task tightness is calculated as follows:
Figure PCTCN2020112528-appb-000001
Where I is the input, D is the data set, f t (I) is the true value of the t-th input I, L t is the loss between the true value and the predicted value, and E I∈D represents the expected value;
通过层次分析法对迁移网络的损失归一化处理得到关联矩阵;Use the analytic hierarchy process to normalize the loss of the migration network to obtain an incidence matrix;
对关联矩阵作最优化处理以得到A个相似性最高的辅任务。Optimize the incidence matrix to get A auxiliary tasks with the highest similarity.
根据本发明的第一方面,所述第一人脸美丽预测模型包括依次连接的用于预处理人脸图像的第一预处理层,用于提取共享图像特征的第一特征共享层,用于从共享图像特征中提取独立特征的第一独立特征提取层,以及第一分类层;所述第二人脸美丽预测模型包括依次连接的第二预处理层,第二特征共享层,第二独立特征提取层,用于融合独立特征和对应人脸美丽预测任务的几何特征、纹理特征的特征融合层,以及第二分类层。According to the first aspect of the present invention, the first face beauty prediction model includes a first preprocessing layer for preprocessing face images connected in sequence, a first feature sharing layer for extracting shared image features, and A first independent feature extraction layer that extracts independent features from shared image features, and a first classification layer; the second face beauty prediction model includes a second pre-processing layer connected in sequence, a second feature sharing layer, and a second independent The feature extraction layer is used to fuse independent features and the geometric features and texture features of the corresponding facial beauty prediction task, and the second classification layer.
根据本发明的第一方面,所述将A个所述第一人脸美丽预测模型作为源域和所述第二人脸美丽预测模型作为目标域通过对抗网络预训练以寻找源域相对目标域的通用特征参数,将所述通用特征参数迁移至所述第二人脸美丽预测模型具体包括以下步骤:According to the first aspect of the present invention, the A first face beauty prediction model is used as the source domain and the second face beauty prediction model is used as the target domain, and the counter network is pre-trained to find the source domain relative to the target domain Migrating the universal feature parameters to the second face beauty prediction model specifically includes the following steps:
提取步骤:提取对应输入第一人脸美丽预测模型的人脸图像的源特征f s=G(x s)和对应输入第二人脸美丽预测模型的人脸图像的目标特征f t=G(x t); Extraction step: extract the source feature f s =G(x s ) corresponding to the face image input to the first face beauty prediction model and the target feature f t =G( x t );
映射步骤:将源特征映射到目标特征空间,得到伪目标特征T(G(x s)); Mapping step: map the source feature to the target feature space to obtain the pseudo target feature T(G(x s ));
区分步骤:区分目标特征和伪目标特征的来源并通过损失函数计算误差;Distinguish step: distinguish the source of the target feature and the pseudo-target feature and calculate the error through the loss function;
优化步骤:利用正则化项r(G(x s),T(G(x s)))测量源特征和伪目标特征之间的距离后结合所述误差优化源特征至目标特征空间的映射; Optimization step: use the regularization term r(G(x s ), T(G(x s ))) to measure the distance between the source feature and the pseudo-target feature, and then combine the error to optimize the mapping of the source feature to the target feature space;
重复映射步骤、区分步骤和优化步骤直至源域与目标域两者域适应以得到通用特征参数;Repeat the mapping step, the distinguishing step and the optimization step until the source domain and the target domain are adapted to obtain the common feature parameters;
迁移步骤:将所述通用特征参数迁移至所述第二人脸美丽预测模型。Migration step: Migrate the general feature parameters to the second face beauty prediction model.
根据本发明的第一方面,所述特征融合层融合几何特征、纹理特征和独立特征的融合方式为求和,计算方式为F fusion=[F CNN,G,H],式中,F fusion为融合特征,F CNN为独立特征,G为几何特征,H为纹理特征。 According to the first aspect of the present invention, the fusion method of the feature fusion layer to fuse geometric features, texture features, and independent features is summation, and the calculation method is F fusion =[F CNN ,G,H], where F fusion is Fusion features, F CNN is an independent feature, G is a geometric feature, H is a texture feature.
上述基于对抗迁移学习的人脸美丽预测方法至少具有以下的有益效果:通过相似性度度量从多个识别人脸因素的辅任务中找到相关性最高的,并以此构建第一人脸美丽预测模型进行预训练;减少预训练的训练成本,减少存在不相关因素的辅任务对第一人脸美丽预测模型 造成的偏差,避免带来负迁移。将对抗网络预训练后形成的通用特征参数迁移至第二人脸美丽预测模型实现最终的人脸美丽预测,通过对抗迁移学习减少第二人脸美丽预测模型训练的计算量,压缩训练时间,达到利用更少的训练图像获得更精准的模型的效果。The above-mentioned face beauty prediction method based on anti-migration learning has at least the following beneficial effects: find the highest correlation among the auxiliary tasks of identifying face factors through the similarity measure, and construct the first face beauty prediction based on this. The model is pre-trained; the training cost of pre-training is reduced, the deviation caused by the auxiliary tasks with unrelated factors to the first face beauty prediction model is reduced, and the negative migration is avoided. The general feature parameters formed after the pre-training of the confrontation network are migrated to the second face beauty prediction model to achieve the final face beauty prediction. Through the confrontation migration learning, the calculation amount of the second face beauty prediction model training is reduced, and the training time is compressed. Use fewer training images to obtain a more accurate model effect.
本发明的第二方面,基于对抗迁移学习的人脸美丽预测装置,包括:In the second aspect of the present invention, a face beauty prediction device based on anti-transfer learning includes:
相似性度量模块,用于度量N个辅任务与主任务间的相似性,得到A个相似性最高的辅任务,其中主任务为人脸美丽预测任务,辅任务为识别人脸美丽因素的任务,N>A;The similarity measurement module is used to measure the similarity between N auxiliary tasks and the main task to obtain A auxiliary tasks with the highest similarity. The main task is the task of predicting the beauty of the face, and the auxiliary task is the task of identifying the beauty factors of the face. N>A;
第一模型建立模块,用于建立对应A个相似度最高的辅任务的A个第一人脸美丽预测模型;The first model establishment module is used to establish A first face beauty prediction model corresponding to A auxiliary tasks with the highest similarity;
第二模型建立模块,用于建立用于人脸美丽预测的第二人脸美丽预测模型;The second model building module is used to build a second face beauty prediction model for face beauty prediction;
参数迁移模块,用于将A个所述第一人脸美丽预测模型作为源域和所述第二人脸美丽预测模型作为目标域通过对抗网络预训练以寻找源域相对目标域的通用特征参数,将所述通用特征参数迁移至所述第二人脸美丽预测模型;The parameter migration module is configured to use A of the first face beauty prediction model as the source domain and the second face beauty prediction model as the target domain to find the common feature parameters of the source domain relative to the target domain through pre-training of the confrontation network , Migrating the general feature parameters to the second face beauty prediction model;
测算模块,用于输入待测人脸图像至再训练好的第二人脸美丽预测模型输出人脸美丽预测结果。The calculation module is used to input the face image to be tested to the second face beauty prediction model that is retrained to output the face beauty prediction result.
根据本发明的第二方面,所述相似性度量模块包括:According to the second aspect of the present invention, the similarity measurement module includes:
特征表达获取模块,用于对N个辅任务与主任务分别构建全监督的特定网络并进行训练得到每个任务的特征表达E s(I); The feature expression acquisition module is used to construct a fully-supervised specific network for the N auxiliary tasks and the main task respectively and perform training to obtain the feature expression E s (I) of each task;
紧密度度量模块,用于构建N个辅任务与主任务间的迁移网络,并度量N个辅任务与主任务间的任务紧密度,所述任务紧密度的计算方式为:
Figure PCTCN2020112528-appb-000002
式中I是输入,D是数据集,f t(I)是第t个输入I的真实值,L t是真实值与预测值之间的损失,E I∈D表示期望值;
The tightness measurement module is used to construct a migration network between N auxiliary tasks and the main task, and measure the task tightness between the N auxiliary tasks and the main task, and the task tightness is calculated as follows:
Figure PCTCN2020112528-appb-000002
Where I is the input, D is the data set, f t (I) is the true value of the t-th input I, L t is the loss between the true value and the predicted value, and E I∈D represents the expected value;
归一化处理模块,用于通过层次分析法对迁移网络的损失归一化处理得到关联矩阵;The normalization processing module is used to normalize the loss of the migration network through the analytic hierarchy process to obtain the correlation matrix;
最优化处理模块,用于对关联矩阵作最优化处理以得到A个相似性最高的辅任务。The optimization processing module is used to optimize the incidence matrix to obtain A auxiliary tasks with the highest similarity.
根据本发明的第二方面,所述第一人脸美丽预测模型包括依次连接的用于预处理人脸图像的第一预处理层,用于提取共享图像特征的第一特征共享层,用于从共享图像特征中提取独立特征的第一独立特征提取层,以及第一分类层;所述第二人脸美丽预测模型包括依次连接的第二预处理层,第二特征共享层,第二独立特征提取层,用于融合独立特征和对应人脸美丽预测任务的几何特征、纹理特征的特征融合层,以及第二分类层。According to the second aspect of the present invention, the first face beauty prediction model includes a first preprocessing layer for preprocessing face images that are sequentially connected, a first feature sharing layer for extracting shared image features, and A first independent feature extraction layer that extracts independent features from shared image features, and a first classification layer; the second face beauty prediction model includes a second pre-processing layer connected in sequence, a second feature sharing layer, and a second independent The feature extraction layer is used to fuse independent features and the geometric features and texture features of the corresponding facial beauty prediction task, and the second classification layer.
根据本发明的第二方面,所述参数迁移模块包括:According to the second aspect of the present invention, the parameter migration module includes:
第一提取模块,用于提取对应输入第一人脸美丽预测模型的人脸图像的源特征f s=G(x s); The first extraction module is used to extract the source feature f s =G(x s ) of the face image corresponding to the input first face beauty prediction model;
第二提取模块,用于提取对应输入第二人脸美丽预测模型的人脸图像的目标特征f t=G(x t); The second extraction module is used to extract the target feature f t =G(x t ) of the face image corresponding to the input to the second face beauty prediction model;
映射模块,用于将源特征映射到目标特征空间,得到伪目标特征T(G(x s)); The mapping module is used to map the source feature to the target feature space to obtain the pseudo target feature T(G(x s ));
区分模块,用于区分目标特征和伪目标特征的来源并通过损失函数计算误差;The distinguishing module is used to distinguish the source of the target feature and the pseudo-target feature and calculate the error through the loss function;
优化模块,用于利用正则化项r(G(x s),T(G(x s)))测量源特征和伪目标特征之间的距离后结合所述误差优化源特征至目标特征空间的映射; The optimization module is used to use the regularization term r(G(x s ), T(G(x s ))) to measure the distance between the source feature and the pseudo-target feature, and then combine the error to optimize the source feature to the target feature space Mapping
参数获取模块,用于当源域与目标域两者域适应时,获取通用特征参数;The parameter acquisition module is used to acquire general characteristic parameters when both the source domain and the target domain are adapted;
迁移子模块,用于将所述通用特征参数迁移至所述第二人脸美丽预测模型。The migration sub-module is used to migrate the general feature parameters to the second face beauty prediction model.
根据本发明的第二方面,所述特征融合层对几何特征、纹理特征和独立特征的融合的计算方式为F fusion=[F CNN,G,H],式中,F fusion为融合特征,F CNN为独立特征,G为几何特征,H为纹理特征。 According to the second aspect of the present invention, the calculation method of the fusion of geometric features, texture features, and independent features by the feature fusion layer is F fusion =[F CNN ,G,H], where F fusion is the fusion feature, and F CNN is an independent feature, G is a geometric feature, and H is a texture feature.
上述基于对抗迁移学习的人脸美丽预测装置至少具有以下的有益效果:通过相似性度度量模块从多个识别人脸因素的辅任务中找到相关性最高的,且预训练模块以此构建第一人脸美丽预测模型进行预训练;减少预训练的训练成本,减少存在不相关因素的辅任务对第一人脸美丽预测模型造成的偏差,避免带来负迁移。参数迁移模块将对抗网络预训练后形成的通用迁移参数迁移至第二人脸美丽预测模型,并实现最终的人脸美丽预测,通过对抗迁移学习减少计算量,压缩训练时间,达到测算模块利用更少的训练图像获得更精准的模型的效果。The above-mentioned facial beauty prediction device based on anti-migration learning has at least the following beneficial effects: the similarity measurement module finds the highest correlation among the auxiliary tasks of identifying facial factors, and the pre-training module builds the first The face beauty prediction model is pre-trained; the training cost of pre-training is reduced, and the deviation caused by the auxiliary tasks with irrelevant factors to the first face beauty prediction model is avoided, and negative migration is avoided. The parameter migration module migrates the general migration parameters formed after the pre-training of the confrontation network to the second face beauty prediction model, and realizes the final face beauty prediction. Through the confrontation migration learning, the calculation amount is reduced, the training time is compressed, and the calculation module is used more. Fewer training images can get a more accurate model effect.
附图说明Description of the drawings
下面结合附图和实例对本发明作进一步说明。The present invention will be further explained below with reference to the drawings and examples.
图1是本发明实施例基于对抗迁移学习的人脸美丽预测方法的步骤图;FIG. 1 is a step diagram of a face beauty prediction method based on anti-migration learning according to an embodiment of the present invention;
图2是步骤S10的具体步骤图;FIG. 2 is a specific step diagram of step S10;
图3是本发明实施例基于对抗迁移学习的人脸美丽预测方法的原理图;FIG. 3 is a schematic diagram of a face beauty prediction method based on anti-migration learning according to an embodiment of the present invention;
图4是本发明实施例基于对抗迁移学习的人脸美丽预测方法的另一原理图;FIG. 4 is another schematic diagram of a face beauty prediction method based on anti-migration learning according to an embodiment of the present invention;
图5是步骤S30的具体步骤图;FIG. 5 is a specific step diagram of step S30;
图6是本发明实施例基于对抗迁移学习的人脸美丽预测装置的结构图;FIG. 6 is a structural diagram of an apparatus for predicting face beauty based on anti-migration learning according to an embodiment of the present invention;
图7是参数迁移模块的结构图。Figure 7 is a structural diagram of the parameter migration module.
具体实施方式detailed description
本部分将详细描述本发明的具体实施例,本发明之较佳实施例在附图中示出,附图的作用在于用图形补充说明书文字部分的描述,使人能够直观地、形象地理解本发明的每个技术特征和整体技术方案,但其不能理解为对本发明保护范围的限制。This section will describe the specific embodiments of the present invention in detail. The preferred embodiments of the present invention are shown in the accompanying drawings. The function of the accompanying drawings is to supplement the description of the text part of the manual with graphics, so that people can understand the present invention intuitively and vividly. Each technical feature and overall technical solution of the invention cannot be understood as a limitation of the protection scope of the present invention.
在本发明的描述中,如果有描述到第一、第二只是用于区分技术特征为目的,而不能理解为指示或暗示相对重要性或者隐含指明所指示的技术特征的数量或者隐含指明所指示的技术特征的先后关系。In the description of the present invention, if the first and second are described for the purpose of distinguishing technical features, they cannot be understood as indicating or implying relative importance or implicitly indicating the number of indicated technical features or implicitly indicating The precedence of the indicated technical characteristics.
本发明的描述中,除非另有明确的限定,设置、安装、连接等词语应做广义理解,所属技术领域技术人员可以结合技术方案的具体内容合理确定上述词语在本发明中的具体含义。In the description of the present invention, unless otherwise clearly defined, terms such as setting, installation, and connection should be understood in a broad sense, and those skilled in the art can reasonably determine the specific meaning of the above terms in the present invention in combination with the specific content of the technical solution.
参照图1,本发明的一个实施例,提供了基于对抗迁移学习的人脸美丽预测方法,包括以下步骤:Referring to Fig. 1, an embodiment of the present invention provides a face beauty prediction method based on anti-migration learning, which includes the following steps:
步骤S10、度量N个辅任务与主任务间的相似性,得到A个相似性最高的辅任务,其中主任务为人脸美丽预测任务,辅任务为识别人脸美丽因素的任务,N>A;Step S10, measure the similarity between the N auxiliary tasks and the main task, and obtain A auxiliary tasks with the highest similarity, where the main task is the task of predicting the beauty of the face, and the auxiliary task is the task of identifying the beauty factors of the face, N>A;
步骤S20、建立对应A个相似度最高的辅任务的A个第一人脸美丽预测模型以及建立用于人脸美丽预测的第二人脸美丽预测模型;Step S20: Establish A first face beauty prediction model corresponding to A secondary tasks with the highest similarity, and establish a second face beauty prediction model for face beauty prediction;
步骤S30、将A个所述第一人脸美丽预测模型作为源域和所述第二人脸美丽预测模型作为目标域通过对抗网络预训练以寻找源域相对目标域的通用特征参数,将所述通用特征参数迁移至所述第二人脸美丽预测模型;Step S30: Use A of the first face beauty prediction model as the source domain and the second face beauty prediction model as the target domain. Through pre-training of the confrontation network to find the common feature parameters of the source domain relative to the target domain, Transferring the general feature parameters to the second face beauty prediction model;
步骤S40、输入待测人脸图像至再训练好的第二人脸美丽预测模型输出人脸美丽预测结果。Step S40: Input the face image to be tested to the second trained face beauty prediction model to output the face beauty prediction result.
在该实施例中,通过相似性度度量从多个识别人脸因素的辅任务中找到相关性最高的,并以此构建第一人脸美丽预测模型进行预训练;减少预训练的训练成本,减少存在不相关因素的辅任务对第一人脸美丽预测模型造成的偏差,避免带来负迁移。多个识别人脸因素的辅任务主要包括对表情、年龄、性别、肤色、眼睛大小、眼距、耳朵形状、鼻子大小、鼻梁高度、唇形等进行识别。In this embodiment, the similarity measure is used to find the most relevant among the auxiliary tasks of recognizing face factors, and based on this, the first face beauty prediction model is constructed for pre-training; the training cost of pre-training is reduced, Reduce the bias caused by the auxiliary tasks with irrelevant factors to the first face beauty prediction model, and avoid negative transfer. The auxiliary tasks of multiple recognition of facial factors mainly include recognition of facial expression, age, gender, skin color, eye size, eye distance, ear shape, nose size, nose bridge height, lip shape, etc.
迁移学习是通过从已学习的相关任务中转移知识来改善新任务的学习。将对抗网络预训练后形成的通用特征参数迁移至第二人脸美丽预测模型能有效减少第二人脸美丽预测模型 训练的计算量,压缩训练时间,达到利用更少的训练图像获得更精准的模型的效果。Transfer learning is to improve the learning of new tasks by transferring knowledge from related tasks that have been learned. Migrating the general feature parameters formed after the adversarial network pre-training to the second face beauty prediction model can effectively reduce the calculation amount of the second face beauty prediction model training, compress the training time, and achieve more accurate results with fewer training images The effect of the model.
参照图2,进一步,步骤S10具体包括以下步骤:Referring to FIG. 2, further, step S10 specifically includes the following steps:
步骤S11、对N个辅任务与主任务分别构建全监督的特定网络并进行训练得到每个任务的特征表达E s(I),每个特定网络均有一个编码器和一个解码器,所有编码器均具有相同的ResNet50结构,而解码器则对应不同的任务; Step S11. Construct a fully-supervised specific network for each of the N auxiliary tasks and the main task and perform training to obtain the feature expression E s (I) of each task. Each specific network has an encoder and a decoder, and all codes are The decoders all have the same ResNet50 structure, and the decoders correspond to different tasks;
步骤S12、构建N个辅任务与主任务间的迁移网络,并度量N个辅任务与主任务间的任务紧密度,任务紧密度的计算方式为:
Figure PCTCN2020112528-appb-000003
式中I是输入,D是数据集,f t(I)是第t个输入I的真实值,L t是真实值与预测值之间的损失,E I∈D表示期望值;
Step S12. Construct a migration network between N auxiliary tasks and the main task, and measure the task tightness between the N auxiliary tasks and the main task. The task tightness calculation method is:
Figure PCTCN2020112528-appb-000003
Where I is the input, D is the data set, f t (I) is the true value of the t-th input I, L t is the loss between the true value and the predicted value, and E I∈D represents the expected value;
步骤S13、通过层次分析法对迁移网络的损失归一化处理得到关联矩阵;具体地,对于每个源任务指向目标任务的任务对(i,j),在迁移后通过留出法取出测试集;对于每一个任务构建成一个矩阵W t,再借助拉普拉斯平滑方法将矩阵W t的输出结果控制在范围[0.001,0.999]内,然后转化得到关联矩阵,关联矩阵反应任务间的相似概率。W′ t中每个元素w′ i,j的计算方式为:
Figure PCTCN2020112528-appb-000004
Step S13: Normalize the loss of the migration network through the analytic hierarchy process to obtain the correlation matrix; specifically, for each task pair (i, j) of the source task pointing to the target task, the test set is taken out by the leave-out method after the migration ; For each task, construct a matrix W t , and then use the Laplace smoothing method to control the output result of the matrix W t within the range [0.001, 0.999], and then transform to obtain an incidence matrix, which reflects the similarity between tasks Probability. The calculation method of each element w′ i, j in W′ t is:
Figure PCTCN2020112528-appb-000004
步骤S14、对关联矩阵作最优化处理以得到A个相似性最高的辅任务,即基于关联矩阵求得子图表选择(subgraph selection)的问题。Step S14: Perform optimization processing on the incidence matrix to obtain A auxiliary tasks with the highest similarity, that is, obtain a subgraph selection problem based on the incidence matrix.
参照图3,进一步,对应每个辅任务的第一人脸美丽预测模型包括依次连接的用于预处理人脸图像的第一预处理层11,用于提取共享图像特征的第一特征共享层12,用于从共享图像特征中提取独立特征的第一独立特征提取层13,以及第一分类层14。参照图4,在其他实施例中,所有辅任务的第一人脸美丽预测模型能组合成一个,将多个第一预处理层11合并为一个以及将多个第一特征共享层12合并为一个,所有任务的共享图像特征放置在该第一特征共享层12中;第一特征共享层12后对应不同任务连接多个第一独立特征提取层13、多个第一分类层14。Referring to FIG. 3, further, the first face beauty prediction model corresponding to each auxiliary task includes a first preprocessing layer 11 for preprocessing face images connected in sequence, and a first feature sharing layer for extracting shared image features. 12. A first independent feature extraction layer 13 for extracting independent features from shared image features, and a first classification layer 14. 4, in other embodiments, the first face beauty prediction model of all auxiliary tasks can be combined into one, multiple first preprocessing layers 11 are combined into one, and multiple first feature sharing layers 12 are combined into one. One, the shared image features of all tasks are placed in the first feature sharing layer 12; after the first feature sharing layer 12, multiple first independent feature extraction layers 13 and multiple first classification layers 14 are connected corresponding to different tasks.
第二人脸美丽预测模型包括依次连接的第二预处理层21,第二特征共享层22,第二独 立特征提取层23,用于融合独立特征和对应人脸美丽预测任务的几何特征、纹理特征的特征融合层24,以及第二分类层25。其中,第一独立特征提取层13和第二独立特征提取层23均包括依次连接的1个卷积层、1个BN层、1个激活函数层和1个池化层,第一分类层14和第二分类层25均包括两个全连接层。The second face beauty prediction model includes a second preprocessing layer 21, a second feature sharing layer 22, and a second independent feature extraction layer 23, which are connected in sequence, and are used to fuse independent features with geometric features and textures corresponding to face beauty prediction tasks. The feature fusion layer 24 of features, and the second classification layer 25. Wherein, the first independent feature extraction layer 13 and the second independent feature extraction layer 23 both include 1 convolutional layer, 1 BN layer, 1 activation function layer, and 1 pooling layer that are connected in sequence, and the first classification layer 14 Both the second classification layer 25 and the second classification layer 25 include two fully connected layers.
参照图5,进一步,步骤S30具体包括以下步骤:5, further, step S30 specifically includes the following steps:
步骤S31、提取步骤:生成器提取对应输入第一人脸美丽预测模型的人脸图像的源特征f s=G(x s)和对应输入第二人脸美丽预测模型的人脸图像的目标特征f t=G(x t); Step S31. Extraction step: The generator extracts the source feature f s =G(x s ) corresponding to the face image input to the first face beauty prediction model and the target feature corresponding to the face image input to the second face beauty prediction model f t =G(x t );
步骤S32、映射步骤:将源特征映射到目标特征空间,得到伪目标特征T(G(x s)); Step S32. Mapping step: map the source feature to the target feature space to obtain the pseudo target feature T(G(x s ));
步骤S33、区分步骤:判别器区分目标特征和伪目标特征的来源并通过损失函数计算误差;其中损失函数可为常用的损失函数,例如交叉熵损失函数或均方误差损失函数;Step S33, distinguishing step: the discriminator distinguishes the source of the target feature and the pseudo-target feature and calculates the error through the loss function; wherein the loss function can be a commonly used loss function, such as a cross entropy loss function or a mean square error loss function;
步骤S34、优化步骤:利用正则化项r(G(x s),T(G(x s)))测量源特征和伪目标特征之间的距离后结合所述误差优化源特征至目标特征空间的映射; Step S34, optimization step: use the regularization term r(G(x s ), T(G(x s ))) to measure the distance between the source feature and the pseudo-target feature, and then combine the error to optimize the source feature to the target feature space The mapping;
步骤S35、重复映射步骤、区分步骤和优化步骤直至源域与目标域两者域适应以得到通用特征参数;其中,源域与目标域两者域适应,即损失函数计算的误差小于设定的阈值;误差越小,表示源域与目标域越接近,迁移效果越好;Step S35, repeat the mapping step, the distinguishing step and the optimization step until the source domain and the target domain are adapted to obtain the common feature parameters; among them, the source domain and the target domain are adapted to each other, that is, the error of the loss function calculation is less than the set Threshold; the smaller the error, the closer the source domain to the target domain, the better the migration effect;
步骤S36、迁移步骤:将所述通用特征参数迁移至所述第二人脸美丽预测模型。Step S36. Migration step: Migrate the general feature parameters to the second face beauty prediction model.
采用对抗网络减少源域与目标域的分布差异,即第一人脸美丽预测模型与第二人脸美丽预测模型的分布差异;达到从源域到目标域知识迁移复用的目的。The confrontation network is used to reduce the distribution difference between the source domain and the target domain, that is, the distribution difference between the first face beauty prediction model and the second face beauty prediction model; to achieve the purpose of knowledge transfer and reuse from the source domain to the target domain.
另外,将第一人脸美丽预测模型的通用特征参数整体迁移至第二人脸美丽预测模型具体为:将第一特征共享层12、第一独立特征提取层13、第一分类层14各自的参数对应迁移至第二特征共享层22、第二独立特征提取层23、第二分类层25;使整个第二人脸美丽预测模型均接受迁移学习的参数,整体模型实现优化。In addition, the overall migration of the general feature parameters of the first face beauty prediction model to the second face beauty prediction model is specifically as follows: each of the first feature sharing layer 12, the first independent feature extraction layer 13, and the first classification layer 14 The parameters are correspondingly migrated to the second feature sharing layer 22, the second independent feature extraction layer 23, and the second classification layer 25; the entire second face beauty prediction model accepts the parameters of the migration learning, and the overall model is optimized.
进一步,特征融合层24融合几何特征、纹理特征和独立特征的融合方式为求和,计算方式为F fusion=[F CNN,G,H],式中,F fusion为融合特征,F CNN为独立特征,G为几何特征,H为纹理特征。 Further, the fusion method of the feature fusion layer 24 to fuse geometric features, texture features and independent features is summation, and the calculation method is F fusion = [F CNN ,G,H], where F fusion is the fusion feature and F CNN is the independent feature Features, G is geometric feature, H is texture feature.
上述人脸美丽预测方法对多个辅任务进行相似性度量,筛选出与主任务相似性高的辅任务,并基于此构建预训练网络模型,将预训练网络模型的参数迁移至主任务人脸美丽识别的网络模型中,使得主任务的网络模型得到优化,避免因无关辅任务训练导致的无用参数引起 的负迁移,从而能大大减少训练量,提升分类识别效率,提高分类识别精度。The above face beauty prediction method measures the similarity of multiple auxiliary tasks, selects the auxiliary tasks with high similarity to the main task, builds a pre-training network model based on this, and transfers the parameters of the pre-trained network model to the main task face In the beautiful recognition network model, the network model of the main task is optimized to avoid the negative transfer caused by useless parameters caused by the training of irrelevant auxiliary tasks, which can greatly reduce the amount of training, improve the efficiency of classification and recognition, and improve the accuracy of classification and recognition.
参照图6,本发明的另一个实施例,基于对抗迁移学习的人脸美丽预测装置,应用上述人脸美丽预测方法,包括:Referring to FIG. 6, another embodiment of the present invention, an apparatus for predicting face beauty based on anti-transfer learning, applying the above-mentioned face beauty prediction method includes:
相似性度量模块100,用于度量N个辅任务与主任务间的相似性,得到A个相似性最高的辅任务,其中主任务为人脸美丽预测任务,辅任务为识别人脸美丽因素的任务,N>A;The similarity measurement module 100 is used to measure the similarity between N auxiliary tasks and the main task to obtain A auxiliary tasks with the highest similarity, where the main task is the task of predicting the beauty of the face, and the auxiliary task is the task of identifying the factors of the beauty of the face , N>A;
第一模型建立模块210,用于建立对应A个相似度最高的辅任务的A个第一人脸美丽预测模型;The first model building module 210 is used to build A first face beauty prediction models corresponding to A auxiliary tasks with the highest similarity;
第二模型建立模块220,用于建立用于人脸美丽预测的第二人脸美丽预测模型;The second model building module 220 is used to build a second face beauty prediction model used for face beauty prediction;
参数迁移模块300,用于将A个所述第一人脸美丽预测模型作为源域和所述第二人脸美丽预测模型作为目标域通过对抗网络预训练以寻找源域相对目标域的通用特征参数,将所述通用特征参数迁移至所述第二人脸美丽预测模型;The parameter migration module 300 is configured to use A of the first face beauty prediction model as the source domain and the second face beauty prediction model as the target domain to find the common features of the source domain relative to the target domain through adversarial network pre-training Parameters, transferring the general feature parameters to the second face beauty prediction model;
测算模块400,用于输入待测人脸图像至再训练好的第二人脸美丽预测模型输出人脸美丽预测结果。The calculation module 400 is used to input the face image to be tested to the second face beauty prediction model that is retrained to output the face beauty prediction result.
在该实施例中,通过相似性度度量模块从多个识别人脸因素的辅任务中找到相关性最高的,并以此构建第一人脸美丽预测模型进行预训练;减少预训练的训练成本,减少存在不相关因素的辅任务对第一人脸美丽预测模型造成的偏差,避免带来负迁移。参数迁移模块300将预训练后形成的通用特征参数迁移至第二人脸美丽预测模型,并实现最终的人脸美丽预测,通过迁移学习减少算量,压缩训练时间,达到测算模块400利用更少的训练图像获得更精准的模型的效果。In this embodiment, the similarity measurement module is used to find the most relevant among the auxiliary tasks of identifying face factors, and to construct the first face beauty prediction model for pre-training; reducing the training cost of pre-training , To reduce the bias caused by the auxiliary tasks with irrelevant factors to the first face beauty prediction model, and avoid negative migration. The parameter migration module 300 migrates the general feature parameters formed after pre-training to the second face beauty prediction model, and realizes the final face beauty prediction, reduces the calculation amount through migration learning, compresses the training time, and reaches the calculation module 400 to use less The training images to obtain a more accurate model effect.
参照图6,进一步,相似性度量模块100包括:Referring to FIG. 6, further, the similarity measurement module 100 includes:
特征表达获取模块110,用于对N个辅任务与主任务分别构建全监督的特定网络并进行训练得到每个任务的特征表达E s(I); The feature expression acquisition module 110 is used to construct a fully-supervised specific network for the N auxiliary tasks and the main task respectively and perform training to obtain the feature expression E s (I) of each task;
紧密度度量模块120,用于构建N个辅任务与主任务间的迁移网络,并度量N个辅任务与主任务间的任务紧密度,任务紧密度的计算方式为:
Figure PCTCN2020112528-appb-000005
式中I是输入,D是数据集,f t(I)是第t个输入I的真实值,L t是真实值与预测值之间的损失,E I∈D表示期望值;
The tightness measurement module 120 is used to construct a migration network between N auxiliary tasks and the main task, and measure the task tightness between the N auxiliary tasks and the main task. The calculation method of the task tightness is:
Figure PCTCN2020112528-appb-000005
Where I is the input, D is the data set, f t (I) is the true value of the t-th input I, L t is the loss between the true value and the predicted value, and E I∈D represents the expected value;
归一化处理模块130,用于通过层次分析法对迁移网络的损失归一化处理得到关联矩阵;The normalization processing module 130 is used to normalize the loss of the migration network through the analytic hierarchy process to obtain an incidence matrix;
最优化处理模块140,用于对关联矩阵作最优化处理以得到A个相似性最高的辅任务。The optimization processing module 140 is used for optimizing the incidence matrix to obtain A auxiliary tasks with the highest similarity.
进一步,第一人脸美丽预测模型包括依次连接的用于预处理人脸图像的第一预处理层11,用于提取共享图像特征的第一特征共享层12,用于从共享图像特征中提取独立特征的第一独立特征提取层13,以及第一分类层14;第二人脸美丽预测模型包括依次连接的第二预处理层21,第二特征共享层22,第二独立特征提取层23,用于融合独立特征和对应人脸美丽预测任务的几何特征、纹理特征的特征融合层24,以及第二分类层25。Further, the first face beauty prediction model includes a first preprocessing layer 11 for preprocessing face images, a first feature sharing layer 12 for extracting shared image features, and a first feature sharing layer 12 for extracting features from the shared image. The first independent feature extraction layer 13 of independent features, and the first classification layer 14; the second face beauty prediction model includes a second preprocessing layer 21, a second feature sharing layer 22, and a second independent feature extraction layer 23 connected in sequence , A feature fusion layer 24 and a second classification layer 25 used to fuse independent features and geometric features and texture features of the corresponding facial beauty prediction task.
参照图7,进一步,参数迁移模块300包括:Referring to FIG. 7, further, the parameter migration module 300 includes:
第一提取模块310,用于提取对应输入第一人脸美丽预测模型的人脸图像的源特征f s=G(x s); The first extraction module 310 is configured to extract the source feature f s =G(x s ) of the face image corresponding to the input first face beauty prediction model;
第二提取模块320,用于提取对应输入第二人脸美丽预测模型的人脸图像的目标特征f t=G(x t); The second extraction module 320 is configured to extract the target feature f t =G(x t ) of the face image corresponding to the input second face beauty prediction model;
映射模块330,用于将源特征映射到目标特征空间,得到伪目标特征T(G(x s)); The mapping module 330 is used to map the source feature to the target feature space to obtain the pseudo target feature T(G(x s ));
区分模块340,用于区分目标特征和伪目标特征的来源并通过损失函数计算误差;The distinguishing module 340 is used to distinguish the source of the target feature and the pseudo-target feature and calculate the error through the loss function;
优化模块350,用于利用正则化项r(G(x s),T(G(x s)))测量源特征和伪目标特征之间的距离后结合所述误差优化源特征至目标特征空间的映射; The optimization module 350 is configured to use the regularization term r(G(x s ), T(G(x s ))) to measure the distance between the source feature and the pseudo target feature, and then combine the error to optimize the source feature to the target feature space The mapping;
参数获取模块360,用于当源域与目标域两者域适应时,获取通用特征参数;The parameter acquisition module 360 is used to acquire general characteristic parameters when both the source domain and the target domain are adapted;
迁移子模块370,用于将所述通用特征参数迁移至所述第二人脸美丽预测模型。The migration sub-module 370 is configured to migrate the general feature parameters to the second face beauty prediction model.
进一步,参数迁移模块300将第一特征共享层12、第一独立特征提取层13、第一分类层14各自的参数对应迁移至第二特征共享层22、第二独立特征提取层23、第二分类层25。Further, the parameter migration module 300 correspondingly migrates the respective parameters of the first feature sharing layer 12, the first independent feature extraction layer 13, and the first classification layer 14 to the second feature sharing layer 22, the second independent feature extraction layer 23, and the second feature sharing layer. Classification layer 25.
进一步,特征融合层24对几何特征、纹理特征和独立特征的融合的计算方式为F fusion=[F CNN,G,H],式中,F fusion为融合特征,F CNN为独立特征,G为几何特征,H为纹理特征。 Further, the calculation method of the feature fusion layer 24 for the fusion of geometric features, texture features and independent features is F fusion =[F CNN ,G,H], where F fusion is the fusion feature, F CNN is the independent feature, and G is Geometric feature, H is texture feature.
上述人脸美丽预测装置对多个辅任务进行相似性度量,筛选出与主任务相似性高的辅任务,并基于此构建基于对抗网络的参数迁移网络模型,将参数迁移网络模型生成的通用特征参数迁移至主任务人脸美丽识别的网络模型中,使得主任务的网络模型得到优化,避免因无关辅任务训练导致的无用参数引起的负迁移,从而能大大减少训练量,提升分类识别效率,提高分类识别精度。The above-mentioned facial beauty prediction device measures the similarity of multiple auxiliary tasks, selects the auxiliary tasks with high similarity to the main task, and builds a parameter migration network model based on the confrontation network based on this, and transfers the parameters to the general features generated by the network model The parameters are migrated to the main task face beauty recognition network model, so that the main task network model is optimized, and the negative transfer caused by useless parameters caused by the training of irrelevant auxiliary tasks can be avoided, which can greatly reduce the amount of training and improve the efficiency of classification and recognition. Improve classification and recognition accuracy.
本发明的另一个实施例,提供了一种存储介质,存储有可执行指令,可执行指令能使与该存储介质连接的处理器按照上述的基于对抗迁移学习的人脸美丽预测方法对人脸图像进 行处理,得到人脸美丽识别结果。Another embodiment of the present invention provides a storage medium that stores executable instructions, and the executable instructions enable a processor connected to the storage medium to perform facial beauty prediction methods based on the above-mentioned anti-migration learning-based face beauty prediction method. The image is processed to obtain the beautiful face recognition result.
以上,只是本发明的较佳实施例而已,本发明并不局限于上述实施方式,只要其以相同的手段达到本发明的技术效果,都应属于本发明的保护范围。The above are only preferred embodiments of the present invention. The present invention is not limited to the above-mentioned embodiments. As long as they achieve the technical effects of the present invention by the same means, they should fall within the protection scope of the present invention.

Claims (10)

  1. 基于对抗迁移学习的人脸美丽预测方法,其特征在于,包括以下步骤:The face beauty prediction method based on anti-migration learning is characterized in that it includes the following steps:
    度量N个辅任务与主任务间的相似性,得到A个相似性最高的辅任务,其中主任务为人脸美丽预测任务,辅任务为识别人脸美丽因素的任务,N>A;Measure the similarity between N auxiliary tasks and the main task, and obtain A auxiliary tasks with the highest similarity. The main task is the task of predicting the beauty of the face, and the auxiliary task is the task of identifying the beauty factors of the face, N>A;
    建立对应A个相似度最高的辅任务的A个第一人脸美丽预测模型以及建立用于人脸美丽预测的第二人脸美丽预测模型;Establish A first face beauty prediction model corresponding to A secondary tasks with the highest similarity, and establish a second face beauty prediction model for face beauty prediction;
    将A个所述第一人脸美丽预测模型作为源域和所述第二人脸美丽预测模型作为目标域通过对抗网络预训练以寻找源域相对目标域的通用特征参数,将所述通用特征参数迁移至所述第二人脸美丽预测模型;Taking A of the first face beauty prediction model as the source domain and the second face beauty prediction model as the target domain, through adversarial network pre-training to find the common feature parameters of the source domain relative to the target domain, the common feature The parameters are transferred to the second face beauty prediction model;
    输入待测人脸图像至训练好的第二人脸美丽预测模型输出人脸美丽预测结果。Input the face image to be tested to the trained second face beauty prediction model to output the face beauty prediction result.
  2. 根据权利要求1所述的基于对抗迁移学习的人脸美丽预测方法,其特征在于,所述度量N个辅任务与主任务间的相似性,得到A个相似性最高的辅任务包括以下步骤:The face beauty prediction method based on adversarial migration learning according to claim 1, wherein the measuring the similarity between the N auxiliary tasks and the main task, and obtaining A auxiliary tasks with the highest similarity comprises the following steps:
    对N个辅任务与主任务分别构建全监督的特定网络并进行训练得到每个任务的特征表达E s(I); Construct a fully-supervised specific network for the N auxiliary tasks and the main task respectively and perform training to obtain the characteristic expression E s (I) of each task;
    构建N个辅任务与主任务间的迁移网络,并度量N个辅任务与主任务间的任务紧密度,所述任务紧密度的计算方式为:
    Figure PCTCN2020112528-appb-100001
    式中I是输入,D是数据集,f t(I)是第t个输入I的真实值,L t是真实值与预测值之间的损失,E I∈D表示期望值;
    Construct a migration network between N auxiliary tasks and the main task, and measure the task tightness between the N auxiliary tasks and the main task, and the task tightness is calculated as follows:
    Figure PCTCN2020112528-appb-100001
    Where I is the input, D is the data set, f t (I) is the true value of the t-th input I, L t is the loss between the true value and the predicted value, and E I∈D represents the expected value;
    通过层次分析法对迁移网络的损失归一化处理得到关联矩阵;Use the analytic hierarchy process to normalize the loss of the migration network to obtain an incidence matrix;
    对关联矩阵作最优化处理以得到A个相似性最高的辅任务。Optimize the incidence matrix to get A auxiliary tasks with the highest similarity.
  3. 根据权利要求1或2所述的基于对抗迁移学习的人脸美丽预测方法,其特征在于,所述第一人脸美丽预测模型包括依次连接的用于预处理人脸图像的第一预处理层,用于提取共享图像特征的第一特征共享层,用于从共享图像特征中提取独立特征的第一独立特征提取层,以及第一分类层;所述第二人脸美丽预测模型包括依次连接的第二预处理层,第二特征共享层,第二独立特征提取层,用于融合独立特征和对应人脸美丽预测任务的几何特征、纹理特征的特征融合层,以及第二分类层。The face beauty prediction method based on adversarial migration learning according to claim 1 or 2, wherein the first face beauty prediction model includes a first preprocessing layer for preprocessing face images connected in sequence , A first feature sharing layer for extracting shared image features, a first independent feature extraction layer for extracting independent features from shared image features, and a first classification layer; the second face beauty prediction model includes sequential connections The second preprocessing layer, the second feature sharing layer, and the second independent feature extraction layer are used to fuse independent features with geometric features and texture features corresponding to the facial beauty prediction task, and the second classification layer.
  4. 根据权利要求3所述的基于对抗迁移学习的人脸美丽预测方法,其特征在于,所述将A 个所述第一人脸美丽预测模型作为源域和所述第二人脸美丽预测模型作为目标域通过对抗网络预训练以寻找源域相对目标域的通用特征参数,将所述通用特征参数迁移至所述第二人脸美丽预测模型具体包括以下步骤:The method for predicting face beauty based on anti-migration learning according to claim 3, wherein the A first face beauty prediction models are used as the source domain and the second face beauty prediction model is used as the source domain. The target domain searches for the common feature parameters of the source domain relative to the target domain through adversarial network pre-training, and migrating the common feature parameters to the second face beauty prediction model specifically includes the following steps:
    提取步骤:提取对应输入第一人脸美丽预测模型的人脸图像的源特征f s=G(x s)和对应输入第二人脸美丽预测模型的人脸图像的目标特征f t=G(x t); Extraction step: extract the source feature f s =G(x s ) corresponding to the face image input to the first face beauty prediction model and the target feature f t =G( x t );
    映射步骤:将源特征映射到目标特征空间,得到伪目标特征T(G(x s)); Mapping step: map the source feature to the target feature space to obtain the pseudo target feature T(G(x s ));
    区分步骤:区分目标特征和伪目标特征的来源并通过损失函数计算误差;Distinguish step: distinguish the source of the target feature and the pseudo-target feature and calculate the error through the loss function;
    优化步骤:利用正则化项r(G(x s),T(G(x s)))测量源特征和伪目标特征之间的距离后结合所述误差优化源特征至目标特征空间的映射; Optimization step: use the regularization term r(G(x s ), T(G(x s ))) to measure the distance between the source feature and the pseudo-target feature, and then combine the error to optimize the mapping of the source feature to the target feature space;
    重复映射步骤、区分步骤和优化步骤直至源域与目标域两者域适应以得到通用特征参数;Repeat the mapping step, the distinguishing step and the optimization step until the source domain and the target domain are adapted to obtain the common feature parameters;
    迁移步骤:将所述通用特征参数迁移至所述第二人脸美丽预测模型。Migration step: Migrate the general feature parameters to the second face beauty prediction model.
  5. 根据权利要求4所述的基于对抗迁移学习的人脸美丽预测方法,其特征在于,所述特征融合层融合几何特征、纹理特征和独立特征的融合方式为求和,计算方式为F fusion=[F CNN,G,H],式中,F fusion为融合特征,F CNN为独立特征,G为几何特征,H为纹理特征。 The face beauty prediction method based on anti-migration learning according to claim 4, wherein the fusion method of the feature fusion layer fusing geometric features, texture features, and independent features is summation, and the calculation method is F fusion =[ F CNN , G, H], where F fusion is a fusion feature, F CNN is an independent feature, G is a geometric feature, and H is a texture feature.
  6. 基于对抗迁移学习的人脸美丽预测装置,其特征在于,包括:The facial beauty prediction device based on anti-migration learning is characterized in that it includes:
    相似性度量模块,用于度量N个辅任务与主任务间的相似性,得到A个相似性最高的辅任务,其中主任务为人脸美丽预测任务,辅任务为识别人脸美丽因素的任务,N>A;The similarity measurement module is used to measure the similarity between N auxiliary tasks and the main task to obtain A auxiliary tasks with the highest similarity. The main task is the task of predicting the beauty of the face, and the auxiliary task is the task of identifying the beauty factors of the face. N>A;
    第一模型建立模块,用于建立对应A个相似度最高的辅任务的A个第一人脸美丽预测模型;The first model establishment module is used to establish A first face beauty prediction model corresponding to A auxiliary tasks with the highest similarity;
    第二模型建立模块,用于建立用于人脸美丽预测的第二人脸美丽预测模型;The second model building module is used to build a second face beauty prediction model for face beauty prediction;
    参数迁移模块,用于将A个所述第一人脸美丽预测模型作为源域和所述第二人脸美丽预测模型作为目标域通过对抗网络预训练以寻找源域相对目标域的通用特征参数,将所述通用特征参数迁移至所述第二人脸美丽预测模型;The parameter migration module is configured to use A of the first face beauty prediction model as the source domain and the second face beauty prediction model as the target domain to find the common feature parameters of the source domain relative to the target domain through pre-training of the confrontation network , Migrating the general feature parameters to the second face beauty prediction model;
    测算模块,用于输入待测人脸图像至再训练好的第二人脸美丽预测模型输出人脸美丽预测结果。The calculation module is used to input the face image to be tested to the second face beauty prediction model that is retrained to output the face beauty prediction result.
  7. 根据权利要求6所述的基于对抗迁移学习的人脸美丽预测装置,其特征在于,所述相似 性度量模块包括:The device for predicting the beauty of faces based on adversarial transfer learning according to claim 6, wherein the similarity measurement module comprises:
    特征表达获取模块,用于对N个辅任务与主任务分别构建全监督的特定网络并进行训练得到每个任务的特征表达E s(I); The feature expression acquisition module is used to construct a fully-supervised specific network for the N auxiliary tasks and the main task respectively and perform training to obtain the feature expression E s (I) of each task;
    紧密度度量模块,用于构建N个辅任务与主任务间的迁移网络,并度量N个辅任务与主任务间的任务紧密度,所述任务紧密度的计算方式为:
    Figure PCTCN2020112528-appb-100002
    式中I是输入,D是数据集,f t(I)是第t个输入I的真实值,L t是真实值与预测值之间的损失,E I∈D表示期望值;
    The tightness measurement module is used to construct a migration network between N auxiliary tasks and the main task, and measure the task tightness between the N auxiliary tasks and the main task, and the task tightness is calculated as follows:
    Figure PCTCN2020112528-appb-100002
    Where I is the input, D is the data set, f t (I) is the true value of the t-th input I, L t is the loss between the true value and the predicted value, and E I∈D represents the expected value;
    归一化处理模块,用于通过层次分析法对迁移网络的损失归一化处理得到关联矩阵;The normalization processing module is used to normalize the loss of the migration network through the analytic hierarchy process to obtain the correlation matrix;
    最优化处理模块,用于对关联矩阵作最优化处理以得到A个相似性最高的辅任务。The optimization processing module is used to optimize the incidence matrix to obtain A auxiliary tasks with the highest similarity.
  8. 根据权利要求6或7所述的基于对抗迁移学习的人脸美丽预测装置,其特征在于,所述第一人脸美丽预测模型包括依次连接的用于预处理人脸图像的第一预处理层,用于提取共享图像特征的第一特征共享层,用于从共享图像特征中提取独立特征的第一独立特征提取层,以及第一分类层;所述第二人脸美丽预测模型包括依次连接的第二预处理层,第二特征共享层,第二独立特征提取层,用于融合独立特征和对应人脸美丽预测任务的几何特征、纹理特征的特征融合层,以及第二分类层。The device for predicting face beauty based on adversarial transfer learning according to claim 6 or 7, wherein the first face beauty prediction model comprises a first preprocessing layer for preprocessing face images connected in sequence , A first feature sharing layer for extracting shared image features, a first independent feature extraction layer for extracting independent features from shared image features, and a first classification layer; the second face beauty prediction model includes sequential connections The second preprocessing layer, the second feature sharing layer, and the second independent feature extraction layer are used to fuse independent features with geometric features and texture features corresponding to the facial beauty prediction task, and the second classification layer.
  9. 根据权利要求8所述的基于对抗迁移学习的人脸美丽预测装置,其特征在于,所述参数迁移模块包括:The device for predicting the beauty of a face based on adversarial transfer learning according to claim 8, wherein the parameter transfer module comprises:
    第一提取模块,用于提取对应输入第一人脸美丽预测模型的人脸图像的源特征f s=G(x s); The first extraction module is used to extract the source feature f s =G(x s ) of the face image corresponding to the input first face beauty prediction model;
    第二提取模块,用于提取对应输入第二人脸美丽预测模型的人脸图像的目标特征f t=G(x t); The second extraction module is used to extract the target feature f t =G(x t ) of the face image corresponding to the input to the second face beauty prediction model;
    映射模块,用于将源特征映射到目标特征空间,得到伪目标特征T(G(x s)); The mapping module is used to map the source feature to the target feature space to obtain the pseudo target feature T(G(x s ));
    区分模块,用于区分目标特征和伪目标特征的来源并通过损失函数计算误差;The distinguishing module is used to distinguish the source of the target feature and the pseudo-target feature and calculate the error through the loss function;
    优化模块,用于利用正则化项r(G(x s),T(G(x s)))测量源特征和伪目标特征之间的距离后结合所述误差优化源特征至目标特征空间的映射; The optimization module is used to use the regularization term r(G(x s ), T(G(x s ))) to measure the distance between the source feature and the pseudo-target feature, and then combine the error to optimize the source feature to the target feature space Mapping
    参数获取模块,用于当源域与目标域两者域适应时,获取通用特征参数;The parameter acquisition module is used to acquire general characteristic parameters when both the source domain and the target domain are adapted;
    迁移模块,用于将所述通用特征参数迁移至所述第二人脸美丽预测模型。The migration module is used to migrate the general feature parameters to the second face beauty prediction model.
  10. 根据权利要求9所述的基于对抗迁移学习的人脸美丽预测装置,其特征在于,所述特征融合层对几何特征、纹理特征和独立特征的融合的计算方式为F fusion=[F CNN,G,H],式中,F fusion为融合特征,F CNN为独立特征,G为几何特征,H为纹理特征。 The face beauty prediction device based on anti-migration learning according to claim 9, wherein the calculation method of the fusion of geometric features, texture features, and independent features of the feature fusion layer is F fusion =[F CNN ,G ,H], where F fusion is a fusion feature, F CNN is an independent feature, G is a geometric feature, and H is a texture feature.
PCT/CN2020/112528 2019-09-20 2020-08-31 Adversarial transfer learning-based face beauty prediction method and device WO2021052159A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910893810.7A CN110705406B (en) 2019-09-20 2019-09-20 Face beauty prediction method and device based on transfer learning resistance
CN201910893810.7 2019-09-20

Publications (1)

Publication Number Publication Date
WO2021052159A1 true WO2021052159A1 (en) 2021-03-25

Family

ID=69195631

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/112528 WO2021052159A1 (en) 2019-09-20 2020-08-31 Adversarial transfer learning-based face beauty prediction method and device

Country Status (2)

Country Link
CN (1) CN110705406B (en)
WO (1) WO2021052159A1 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113128113A (en) * 2021-04-14 2021-07-16 国网上海市电力公司 Poor information building load prediction method based on deep learning and transfer learning
CN114444374A (en) * 2021-11-29 2022-05-06 河南工业大学 Multi-source to multi-target domain self-adaption method based on similarity measurement
CN114898424A (en) * 2022-04-01 2022-08-12 中南大学 Lightweight human face aesthetic prediction method based on dual label distribution
CN115879008A (en) * 2023-03-02 2023-03-31 中国空气动力研究与发展中心计算空气动力研究所 Data fusion model training method, device, equipment and storage medium
WO2023236594A1 (en) * 2022-06-09 2023-12-14 五邑大学 Facial beauty prediction method and apparatus, and electronic device and storage medium

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110705406B (en) * 2019-09-20 2022-11-15 五邑大学 Face beauty prediction method and device based on transfer learning resistance
CN111382846B (en) * 2020-05-28 2020-09-01 支付宝(杭州)信息技术有限公司 Method and device for training neural network model based on transfer learning
CN111784596A (en) * 2020-06-12 2020-10-16 北京理工大学 General endoscope image enhancement method and device based on generation of antagonistic neural network
CN111832435A (en) * 2020-06-24 2020-10-27 五邑大学 Beauty prediction method and device based on migration and weak supervision and storage medium
CN111914908B (en) * 2020-07-14 2023-10-24 浙江大华技术股份有限公司 Image recognition model training method, image recognition method and related equipment
CN112069916B (en) * 2020-08-14 2024-02-20 五邑大学 Face beauty prediction method, device and system and readable storage medium
CN112069946B (en) * 2020-08-25 2024-02-20 五邑大学 Face beauty prediction method, device, system and storage medium
CN113450267B (en) * 2021-05-14 2022-08-19 桂林电子科技大学 Transfer learning method capable of rapidly acquiring multiple natural degradation image restoration models
CN114548382B (en) * 2022-04-25 2022-07-15 腾讯科技(深圳)有限公司 Migration training method, device, equipment, storage medium and program product

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6826300B2 (en) * 2001-05-31 2004-11-30 George Mason University Feature based classification
CN108959522A (en) * 2018-04-26 2018-12-07 浙江工业大学 The migration search method of network is generated based on semi-supervised confrontation
CN110119689A (en) * 2019-04-18 2019-08-13 五邑大学 A kind of face beauty prediction technique based on multitask transfer learning
CN110705406A (en) * 2019-09-20 2020-01-17 五邑大学 Face beauty prediction method and device based on transfer learning resistance

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10719742B2 (en) * 2018-02-15 2020-07-21 Adobe Inc. Image composites using a generative adversarial neural network
CN109523018B (en) * 2019-01-08 2022-10-18 重庆邮电大学 Image classification method based on deep migration learning
CN109948648B (en) * 2019-01-31 2023-04-07 中山大学 Multi-target domain adaptive migration method and system based on meta-countermeasure learning
CN110084121A (en) * 2019-03-27 2019-08-02 南京邮电大学 Implementation method based on the human face expression migration for composing normalized circulation production confrontation network
CN110210486B (en) * 2019-05-15 2021-01-01 西安电子科技大学 Sketch annotation information-based generation countermeasure transfer learning method

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6826300B2 (en) * 2001-05-31 2004-11-30 George Mason University Feature based classification
CN108959522A (en) * 2018-04-26 2018-12-07 浙江工业大学 The migration search method of network is generated based on semi-supervised confrontation
CN110119689A (en) * 2019-04-18 2019-08-13 五邑大学 A kind of face beauty prediction technique based on multitask transfer learning
CN110705406A (en) * 2019-09-20 2020-01-17 五邑大学 Face beauty prediction method and device based on transfer learning resistance

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
甘俊英 等 (GAN, JUNYING ET AL.): "基于双激活层深度卷积特征的人脸美丽预测研究 (Research of Facial Beauty Prediction Based on Deep Convolutional Features Using Double Activation Layer)", 电子学报 (ACTA ELECTRONICA SINICA), vol. 47, no. 3, 31 March 2019 (2019-03-31), XP055772289, ISSN: 0372-2112, DOI: 10.3969/j.issn.0372-2112.2019.03.017 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113128113A (en) * 2021-04-14 2021-07-16 国网上海市电力公司 Poor information building load prediction method based on deep learning and transfer learning
CN113128113B (en) * 2021-04-14 2024-04-12 国网上海市电力公司 Lean information building load prediction method based on deep learning and transfer learning
CN114444374A (en) * 2021-11-29 2022-05-06 河南工业大学 Multi-source to multi-target domain self-adaption method based on similarity measurement
CN114898424A (en) * 2022-04-01 2022-08-12 中南大学 Lightweight human face aesthetic prediction method based on dual label distribution
CN114898424B (en) * 2022-04-01 2024-04-26 中南大学 Lightweight face aesthetic prediction method based on dual label distribution
WO2023236594A1 (en) * 2022-06-09 2023-12-14 五邑大学 Facial beauty prediction method and apparatus, and electronic device and storage medium
CN115879008A (en) * 2023-03-02 2023-03-31 中国空气动力研究与发展中心计算空气动力研究所 Data fusion model training method, device, equipment and storage medium
CN115879008B (en) * 2023-03-02 2023-05-26 中国空气动力研究与发展中心计算空气动力研究所 Data fusion model training method, device, equipment and storage medium

Also Published As

Publication number Publication date
CN110705406B (en) 2022-11-15
CN110705406A (en) 2020-01-17

Similar Documents

Publication Publication Date Title
WO2021052159A1 (en) Adversarial transfer learning-based face beauty prediction method and device
WO2021083241A1 (en) Facial image quality evaluation method, feature extraction model training method, image processing system, computer readable medium, and wireless communications terminal
CN108171209B (en) Face age estimation method for metric learning based on convolutional neural network
US11809485B2 (en) Method for retrieving footprint images
WO2020228446A1 (en) Model training method and apparatus, and terminal and storage medium
CN111126482B (en) Remote sensing image automatic classification method based on multi-classifier cascade model
WO2020182121A1 (en) Expression recognition method and related device
WO2021052160A1 (en) Face beauty prediction method based on multi-task migration and device
CN109145717B (en) Face recognition method for online learning
WO2017166137A1 (en) Method for multi-task deep learning-based aesthetic quality assessment on natural image
CN106570522B (en) Object recognition model establishing method and object recognition method
WO2022257487A1 (en) Method and apparatus for training depth estimation model, and electronic device and storage medium
WO2022057309A1 (en) Lung feature recognition method and apparatus, computer device, and storage medium
CN111401219B (en) Palm key point detection method and device
CN111127360A (en) Gray level image transfer learning method based on automatic encoder
CN112101262A (en) Multi-feature fusion sign language recognition method and network model
CN114677339A (en) Power transmission line bolt out-of-stock defect detection method introducing attention mechanism
CN113221695A (en) Method for training skin color recognition model, method for recognizing skin color and related device
WO2023231753A1 (en) Neural network training method, data processing method, and device
CN116977271A (en) Defect detection method, model training method, device and electronic equipment
CN114445875A (en) Deep learning-based identity recognition and face comparison system and training method
JP2020064364A (en) Learning device, image generating device, learning method, and learning program
CN114463628A (en) Deep learning remote sensing image ship target identification method based on threshold value constraint
JP2003344270A (en) System for evaluating deterioration of steel surface using self structurizing character map
CN114639132A (en) Feature extraction model processing method, device and equipment in face recognition scene

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20864493

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20864493

Country of ref document: EP

Kind code of ref document: A1

122 Ep: pct application non-entry in european phase

Ref document number: 20864493

Country of ref document: EP

Kind code of ref document: A1