CN106203395B - Face attribute recognition method based on multitask deep learning - Google Patents

Face attribute recognition method based on multitask deep learning Download PDF

Info

Publication number
CN106203395B
CN106203395B CN201610591877.1A CN201610591877A CN106203395B CN 106203395 B CN106203395 B CN 106203395B CN 201610591877 A CN201610591877 A CN 201610591877A CN 106203395 B CN106203395 B CN 106203395B
Authority
CN
China
Prior art keywords
face
image
training
neural network
key point
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201610591877.1A
Other languages
Chinese (zh)
Other versions
CN106203395A (en
Inventor
严严
陈日伟
王菡子
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xiamen University
Original Assignee
Xiamen University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xiamen University filed Critical Xiamen University
Priority to CN201610591877.1A priority Critical patent/CN106203395B/en
Publication of CN106203395A publication Critical patent/CN106203395A/en
Application granted granted Critical
Publication of CN106203395B publication Critical patent/CN106203395B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • G06V40/171Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Abstract

A face attribute identification method based on multitask deep learning relates to face attribute identification in computer vision. Preparing an image dataset; carrying out face detection on each image in the image data set one by one; performing face key point detection on all detected faces; aligning each face to a standard face image according to a face alignment method for the detected face key points to form a face image training set; calculating an average face image in a training set; constructing a multitask deep convolutional neural network, and carrying out network parameter training after subtracting an average face image from each face image in a face image training set to obtain a convolutional neural network model; carrying out face detection and face key point detection on a test image to be recognized, and aligning a face in the image to a standard face image according to a face key point; and subtracting the average face image from the standard face image, and putting the standard face image into the constructed convolutional neural network model to perform feed-forward operation, so as to obtain the face image.

Description

Face attribute recognition method based on multitask deep learning
Technical Field
The invention relates to face attribute recognition in computer vision, in particular to a face attribute recognition method based on multitask deep learning.
Background
The image-based face attribute recognition method is a process of judging the face attribute in an image by using a pattern recognition technology according to a given input image. The attributes of the human face contained in the human face image mainly include: age, gender, expression, race, whether to wear glasses, whether to make up, etc. The human face attribute recognition is automatically carried out by utilizing the computer, so that the human-computer interaction performance can be effectively improved, and the method has very important practical application value. The process of face attribute recognition comprises the following steps: the method comprises the steps of face detection technology, face image preprocessing technology, face feature extraction, training of a face attribute classifier and the like. The performance of the face feature extraction and the face attribute classifier directly affects the final face attribute recognition performance.
At present, the face attribute recognition technology is mainly completed by two steps: extracting human face features and training a human face attribute classifier. The face feature extraction technology is divided into two main categories according to different feature acquisition modes: manual design features and automatic learning features. The performance of the classifier is directly affected by the quality of the human face features. The manual design features are mainly as follows: SIFT feature (D.G.Lowe.visual image features from scale-innovative keys [ J ]. International Journal of Computer Vision,2004,60(2): 91-110), LBP feature (T.Ahonen, A.Hadid, M.Pietikain.face description with local binding patterns: Application to face registration [ J ]. IEEE Transactions on Pattern Analysis and machine Analysis, IE2006, 28(12): 2037-. However, these manually designed features are mostly based on expert experience design, and it is difficult to extract features that are effective for various tasks. Second, manually designed feature extraction is decoupled from classifier design, resulting in selected features that are not best suited for a particular classifier.
Deep learning has recently become one of the research hotspots in the field of computer vision. Unlike the traditional method of decomposing the recognition task into feature extraction and classifier training, the deep learning organically combines the feature extraction and the classifier training, directly takes the original data as input, and simultaneously performs feature extraction learning and classifier training learning. The deep learning method for statistical learning by putting feature learning and classifier training in the same frame effectively avoids the significance gap between feature extraction and a target task classifier, facilitates the feature extraction and the target task classifier, and overcomes the trouble of manually designing features. The depth model formed by the multilayer neural network has the characteristics of automatically acquiring characteristics from low level to high level, from simple to complex, and from general to special. For example: in a typical image classification network, edge information is often extracted from the previous layers of the network, angle information is extracted from the middle layer, and contour information, target information and the like are extracted from the later layers. The lower the hierarchy, the simpler and more universal the extracted features, and gradually extracting the features related to the target task.
Training a deep learning model often requires a large amount of label data to avoid over-fitting the learning model with a small amount of training data. However, it is very time-consuming and labor-consuming to acquire a large amount of tag data, and it is a key problem worth solving to explore the problem of insufficient data by acquiring characteristics of different features layer by using a deep network model.
Disclosure of Invention
The invention aims to provide a face attribute recognition method based on multitask deep learning.
The invention comprises the following steps:
A. an image data set is prepared that contains a large number of faces and corresponding face attribute labels.
B. And carrying out face detection on each image in the image data set one by one to obtain the position of the face in each image.
C. And performing face key point detection on all the detected faces to acquire the positions of the face key points in each image.
D. And aligning each face to a standard face image according to a face alignment method for the detected face key points to form a face image training set.
E. And calculating an average human face image in a training set by using the human face image training set.
F. And constructing a multi-task deep convolutional neural network, and carrying out network parameter training after subtracting the average face image from each face image in the face image training set to obtain a convolutional neural network model.
G. And respectively carrying out face detection and face key point detection on the test image to be recognized, and aligning the face in the image to the standard face image according to the face key point.
H. And subtracting the average face image from the standard face image, and putting the standard face image into the constructed convolutional neural network model for feedforward operation to obtain a plurality of attribute recognition results of the face.
In the step a, the prepared image data set may adopt image data including human faces with better diversity acquired in a complex scene, and provide corresponding K personal face attribute labels, where K is a number of learned tasks and a natural number; the convolutional neural network structure based on multi-task learning is adopted, and each face image does not need to have all face attribute label data at the same time, so that the existing face database can be fully utilized for combination to form a large-scale image data set.
In step B, the face detection for each image in the image data set one by one may adopt a common face detection method to obtain a position of a face in each image, and the common face detection method may adopt an OpenCV-self-contained face detection method.
In step C, the face key point detection may adopt a common face key point detection method to obtain the position of the face key point in each image, and the common face key point detection method may adopt a Dlib-carried face key point detection method.
In step D, the face alignment method is affine transformation based on a two-dimensional image, and specifically includes the following steps:
D1. and fitting by using a least square method according to the matching relation between the face key points and the standard face key points to obtain an optimal transformation matrix. Assuming d coordinates of the key points of the standard face as
Figure BDA0001059722060000031
WhereinThe coordinates of the ith standard face key point are obtained; d is the number of key points of the face and d is a natural number; and the coordinates of the key points of the detected face are { (x)1;y1),(x2;y2),...,(xd;yd) In which (x)i,yi) Is the ith individual face key point coordinate of the detected face and extends the coordinate to src { (x)1;y1;1),(x2;y2;1),...,(xd;yd(ii) a 1) And (6) optimizing by a least square method. The specific calculation formula is as follows:
Figure BDA0001059722060000033
wherein
Figure BDA0001059722060000034
Representing an affine transformation matrix; a. the0Is an optimal transformation matrix;
D2. transformation matrix A obtained by optimization0And aligning all the face images to the standard face images, and cutting the face images into images with uniform sizes.
In step E, the specific method for calculating the average face image in the training set for the face image training set may be:
E1. the mean image is an arithmetic mean image M obtained by calculating each channel (comprising 3 channels of RGB) of a face image training set, wherein the calculation mode of each channel and M is as follows:
Figure BDA0001059722060000035
wherein the content of the first and second substances, and
Figure BDA0001059722060000038
the R channel, the G channel and the B channel of the nth human face image are respectively arranged; n is the total image number of the face image training set, and N is a natural number. MR,MGAnd MBArithmetic mean images of an R channel, a G channel and a B channel respectively;
E2. the arithmetic mean value image obtained by the RGB three channels is combined into an average face image, and the calculation formula is as follows:
M=[MR,MG,MB];
E3. dividing the face image training set into training data and verification data according to the ratio of 9: 1.
In step F, the specific method for constructing the multitask deep convolutional neural network may be:
F1. training data are randomly disordered, each batch of samples with the size of m are set, and data are divided, wherein m is a natural number. In the process of training the deep convolutional neural network model, learning the weight parameters of the neural network by using a batch gradient updating method;
F2. a structure of a convolutional neural network is designed, which includes convolutional layers, downsampling layers, and full-connect layers. Each convolution layer and the full-connection layer adopt nonlinear rectification activation functions; in the face attribute recognition of the multi-task deep learning, the whole network structure is divided into a sharing layer and a unique layer, the sharing layer is shared by all tasks, and the tasks participate in the training of parameters of the sharing layer together. The unique layer is occupied by each task independently, parameter learning is carried out by using independent data of each task, and the number of the shared layers is assumed to be S, wherein S is a natural number; the number of the unique layers is U, wherein U is a natural number; setting different numbers of S and U according to different multi-task combinations;
F3. setting hyper-parameters such as the number of convolution layer filters and the number of characteristic graphs, the size of the filters, the size of kernels in down-sampling layers, the learning rate of each layer, the initial weight value and the like required in a convolution neural network structure;
F4. in the training process of the convolutional neural network, training skills such as impulse and abandon are adopted for accelerating the training of the convolutional neural network;
F5. judging whether to stop training according to the performance of the trained network model parameters on the verification data;
F6. and extracting the trained network model parameters W.
The method utilizes the deep convolutional neural network to simultaneously learn the feature extraction and the attribute identification of the face image, so that the learned features are more favorable for the identification of a classifier, and the feature extraction and the classifier training are not required to be respectively carried out. For multi-task face attribute recognition of K tasks, all training data do not need to have K label attributes at the same time, and face data with one or more attributes can be used for training network parameters, and different attribute tasks can benefit accordingly.
Unlike a traditional deep convolutional neural network which only has a single label and a single network output, the convolutional neural network based on the multitask deep learning has a plurality of outputs. The objective function of the network training is a combination of a plurality of Softmax loss functions and an L2 loss function. Assume that there are K tasks that need to be learned together. Then for the face attribute recognition of the ith classification task, the loss function is defined as follows:
Figure BDA0001059722060000041
wherein
Figure BDA0001059722060000042
Figure BDA0001059722060000043
Representing the probability value calculated by the Softmax loss function for each attribute class;
Figure BDA0001059722060000044
a value representing a fully connected classification output in the target class; ciThe number of categories of the ith task is represented, and i is a natural number.
For the face attribute recognition of the jth regression task, the loss function is defined as follows:
wherein y isnIn order to be the true tag value,
Figure BDA0001059722060000052
is the predicted value of the regressor.
In the network training, the cost loss functions of all tasks are combined, and a total optimization objective function is formed as follows:
Figure BDA0001059722060000053
wherein alpha iskThe penalty function representing the kth task is weighted with the total penalty function. By default αkThe values of (a) are all 1, indicating that the respective tasks are equally important.
Compared with the prior art, the method has the advantages of reducing the requirement of data quantity required by training network model parameters, reducing the risk of overfitting training data, reducing the recognition time which is spread in a single task, and effectively improving the accuracy of face attribute recognition.
Drawings
Fig. 1 is a schematic diagram of attributes of smile, gender and attraction of a female face.
Fig. 2 is a schematic diagram of attributes of smile, gender and attraction of a male face.
Detailed Description
The method of the present invention is described in detail below with reference to the accompanying drawings and examples.
The invention comprises the following steps:
s1, preparing an image data set which comprises a large number of human faces and corresponding human face attribute labels. The face attribute database used in this example is an image data set in the CelebrayA database, which contains over 20 million face images and 40 face attributes. Three representative face attribute tasks are employed (K — 3): and explaining the attribute of the human face gender, the attribute of the human face smile and the attribute of the human face attraction force. Schematic representation of three attributes as shown in FIGS. 1 and 2, with labels set to y, respectively1,y2,y3
And S2, carrying out face detection on each image in the image data set one by one to obtain the position of the face in each image. The step can adopt any one of the existing human face detection methods to carry out human face detection. The face detection method of the OpenCV is adopted in the embodiment, and the face detection method has the advantage of being capable of rapidly detecting the face.
And S3, detecting key points of the human face on all the detected human faces to obtain the position of the key points of the human face in each image. The step can adopt any one of the existing face key point detection methods to detect. In this embodiment, 68 personal face key points can be obtained by using a Dlib-carried face key point detection method.
S4, aligning each face to a standard face image according to a face alignment method for the detected face key points to form a face image training set, which specifically comprises the following steps:
(1) and fitting by using a least square method according to the matching relation between the face key points and the standard face key points to obtain an optimal transformation matrix. Assuming that the 68 coordinates of the key points of the standard face are
Figure BDA0001059722060000067
Wherein
Figure BDA0001059722060000068
The coordinates of the ith standard face key point are obtained; d is the number of key points of the face and d is a natural number; and the coordinates of the key points of the detected face are { (x)1;y1),(x2;y2),...,(x68;y68) In which (x)i,yi) Is the ith individual face key point coordinate of the detected face and extends the coordinate to src { (x)1;y1;1),(x2;y2;1),...,(x68;y68(ii) a 1) And (6) optimizing by a least square method. The specific calculation formula is as follows,
Figure BDA0001059722060000061
wherein
Figure BDA0001059722060000062
Representing an affine transformation matrix; a. the0Is the optimal transformation matrix.
(2) Transformation matrix A obtained by optimization0All face images are aligned to a standard face image and cut into images of uniform size 128 x 128.
S5, calculating an average face image in a training set for the face image training set, wherein the method specifically comprises the following steps:
(1) the mean image is an arithmetic mean image M calculated for each channel (including 3 channels of RGB) of the face image training set, wherein the calculation for each channel and M is as follows,
Figure BDA0001059722060000063
wherein
Figure BDA0001059722060000065
And
Figure BDA0001059722060000066
the R channel, the G channel and the B channel of the nth human face image are respectively arranged; n is the total image number of the face image training set, and N is a natural number. MR,MGAnd MBArithmetic mean images for the R, G and B channels, respectively.
(2) The arithmetic mean value images obtained by RGB three channels are combined into an average face image, and the calculation formula is M ═ MR,MG,MB]。
(3) Dividing the face image training set into training data and verification data according to the ratio of 9: 1.
S6, constructing a multi-task deep convolutional neural network, and training network parameters after subtracting an average face image from each face image in a face image training set to obtain a convolutional neural network model, wherein the method specifically comprises the following steps:
(1) training data are randomly scrambled, each batch is set to be m-128 samples, and data are divided. In the process of training the deep convolutional neural network model, a batch gradient updating method is used for learning the weight parameters of the neural network.
(2) A structure of a convolutional neural network is designed, which includes convolutional layers, downsampling layers, and full-connect layers. Each convolutional layer and the full link layer adopt nonlinear rectification activation functions. In the face attribute recognition of multi-task deep learning, the method integratesThe body network structure is divided into two parts of a shared layer and a unique layer. The sharing layer is shared by all tasks, and a plurality of tasks participate in the training of the sharing layer parameters together. The unique layer is occupied by each task independently, and parameter learning is carried out by using data of each task independently. Assuming that the number of sharing layers is S-10; the number of unique layers is U-2. The loss functions used are all Softmax functions, alphakAre all set to 1.
(3) The number of convolutional layer filters and the number of feature maps required in the convolutional neural network structure, the size of the filters, the size of kernels in downsampling layers, the learning rate of each layer, the initial value of weight and other hyper-parameters are set, and the network structure of the deep convolutional neural network is shown in table 1.
TABLE 1
Network layer name Type (B) Input size Output size Filter size/step size
Conv1_1 Convolutional layer 128*128*3 128*128*32 3*3/1
Conv1_2 Convolutional layer 128*128*32 128*128*64 3*3/1
Pool1 Downsampling layer 128*128*64 64*64*64 2*2/2
Conv2_1 Convolutional layer 64*64*64 64*64*96 3*3/1
Conv2_2 Convolutional layer 64*64*96 64*64*128 3*3/1
Pool2 Downsampling layer 64*64*128 32*32*128 2*2/2
Conv3_1 Convolutional layer 32*32*128 32*32*128 3*3/1
Conv3_2 Convolutional layer 32*32*128 32*32*192 3*3/1
Pool3 Downsampling layer 32*32*192 16*16*192 2*2/2
Conv4_1 Convolutional layer 16*16*192 16*16*256 3*3/1
Conv4_2 Convolutional layer 16*16*256 16*16*256 3*3/1
Pool4 Downsampling layer 16*16*256 8*8*256 2*2/2
Conv5_1 Convolutional layer 8*8*256 8*8*320 3*3/1
Conv5_2 Convolutional layer 8*8*320 8*8*320 3*3/1
Pool5 Downsampling layer 8*8*320 4*4*320 2*2/2
Dropout1 Dropout layer 4*4*320 4*4*320
Fc1 Full connection layer 4*4*320 1*1*128
Fc2 Full connection layer 1*1*128 1*1*Ci
(4) In the training process of the convolutional neural network, training skills such as impulse and abandon are adopted for accelerating the training of the convolutional neural network.
(5) And judging whether to stop training the trained network model parameters according to the performance of the trained network model parameters on the verification data.
(6) And extracting the trained network model parameters W.
S7, for any given image to be subjected to face attribute recognition, changing the image to be tested into a color image with the size of 128 multiplied by 128 by using the same data preprocessing method in the steps S1-S4, subtracting a mean value image in training data, inputting the color image into a trained deep convolutional neural network, and finally obtaining recognition results of 3 different face attributes.
And S8, performing operation of step S7 on each face image of the CelebrayA test data, and comparing the attribute identification precision with the prediction time. The precision of the single task network and the multitask network on the face CelebrayA test data and the time comparison result are shown in the table 2, and it can be seen from the table that under the same network structure, the accuracy of attribute identification can be improved by the multitask deep learning attribute identification method, and meanwhile, the average prediction time of each task can be greatly reduced by utilizing the multitask attribute identification.
TABLE 2
Method of producing a composite material T1 T2 T3 T1+T2 T1+T3 T2+T3 T1+T2+T3
Face gender (T1) 0.9670 N/A N/A 0.9580 0.9690 N/A 0.9680
Face smile (T2) N/A 0.9230 N/A 0.9210 N/A 0.9350 0.9360
Face attraction (T3) N/A N/A 0.8060 N/A 0.8130 0.8160 0.8220
Total predicted time (seconds) 0.0230 0.0230 0.0230 0.0270 0.0270 0.0270 0.0290
Average predicted time (seconds) 0.0230 0.0230 0.0230 0.0135 0.0135 0.0135 0.0097
The invention aims at different face attribute recognition tasks, can learn the shared network weight together, and train the unique network weight independently, thereby greatly reducing the requirement of the whole training data. The invention effectively improves the performance of face attribute recognition.

Claims (7)

1. The face attribute recognition method based on the multitask deep learning is characterized by comprising the following steps of:
A. preparing an image data set comprising a plurality of faces and corresponding face attribute labels;
B. carrying out face detection on each image in the image data set one by one to obtain the position of a face in each image;
C. performing face key point detection on all detected faces to acquire the position of the face key point in each image;
D. aligning each face to a standard face image according to a face alignment method for the detected face key points to form a face image training set;
E. calculating an average face image in a training set by a face image training set, wherein the specific method comprises the following steps:
E1. the mean image is each channel of a face image training set, comprises 3 channels of RGB, and is an arithmetic mean image M obtained through calculation, wherein the calculation mode of each channel and M is as follows:
Figure FDA0002218656420000011
wherein the content of the first and second substances,and
Figure FDA0002218656420000013
the R channel, the G channel and the B channel of the nth human face image are respectively arranged; n is the total image number of the face image training set, and N is a natural number; mR,MGAnd MBArithmetic mean images of an R channel, a G channel and a B channel respectively;
E2. the arithmetic mean value image obtained by the RGB three channels is combined into an average face image, and the calculation formula is as follows:
M=[MR,MG,MB];
E3. dividing a face image training set into training data and verification data according to the ratio of 9: 1;
F. constructing a multitask deep convolutional neural network, and carrying out network parameter training after subtracting an average face image from each face image in a face image training set to obtain a convolutional neural network model;
the specific method for constructing the multitask deep convolution neural network comprises the following steps:
F1. randomly disordering training data, setting the size of each batch of samples to be m, and dividing the data, wherein m is a natural number; in the process of training the deep convolutional neural network model, learning the weight parameters of the neural network by using a batch gradient updating method;
F2. designing a structure of a convolutional neural network, wherein the structure comprises a convolutional layer, a downsampling layer and a full-connection layer; each convolution layer and the full-connection layer adopt nonlinear rectification activation functions; in the face attribute recognition of multi-task deep learning, the whole network structure is divided into a sharing layer and a unique layer, the sharing layer is shared by all tasks, and a plurality of tasks participate in the training of parameters of the sharing layer together; the unique layer is occupied by each task independently, parameter learning is carried out by using independent data of each task, and the number of the shared layers is assumed to be S, wherein S is a natural number; the number of the unique layers is U, wherein U is a natural number; setting different numbers of S and U according to different multi-task combinations;
F3. setting the number of convolutional layer filters and the number of characteristic graphs required in a convolutional neural network structure, the size of the filters, the size of kernels in downsampling layers, the learning rate of each layer and a weight initial value;
F4. in the training process of the convolutional neural network, training skills such as impulse and abandon are adopted for accelerating the training of the convolutional neural network;
F5. judging whether to stop training according to the performance of the trained network model parameters on the verification data;
F6. extracting a trained network model parameter W;
G. respectively carrying out face detection and face key point detection on a test image to be recognized, and aligning a face in the image to a standard face image according to the face key point;
H. and subtracting the average face image from the standard face image, and putting the standard face image into the constructed convolutional neural network model for feedforward operation to obtain a plurality of attribute recognition results of the face.
2. The method for identifying human face attributes based on multitask deep learning as claimed in claim 1, characterized in that in step A, said prepared image data set adopts image data containing human faces collected under the scene, and provides corresponding K personal face attribute labels.
3. The method according to claim 1, wherein in step B, the face detection is performed on each image in the image data set one by using a common face detection method to obtain the position of the face in each image.
4. The method for identifying the human face attribute based on the multitask deep learning as claimed in claim 3, wherein the commonly used human face detection method adopts an OpenCV self-contained human face detection method.
5. The method for identifying human face attributes based on multitask deep learning as claimed in claim 1, wherein in step C, said human face key point detection adopts a commonly used human face key point detection method to obtain the position of the human face key point in each image.
6. The method for identifying face attributes based on multitask deep learning as claimed in claim 5, characterized in that said commonly used face key point detection method adopts a Dlib-carried face key point detection method.
7. The method for identifying the attributes of the human face based on the multitask deep learning as claimed in claim 1, wherein in the step D, the human face alignment method is affine transformation based on a two-dimensional image, and specifically comprises the following steps:
D1. fitting by using a least square method according to the matching relation between the face key points and the standard face key points to obtain an optimal transformation matrix; assuming d coordinates of the key points of the standard face asWherein
Figure FDA0002218656420000031
The coordinates of the ith standard face key point are obtained; d is the number of key points of the face and d is a natural number; and the coordinates of the key points of the detected face are { (x)1;y1),(x2;y2),...,(xd;yd) In which (x)i,yi) Is the ith individual face key point coordinate of the detected face and extends the coordinate to src { (x)1;y1;1),(x2;y2;1),...,(xd;yd(ii) a 1) Optimizing by a least square method; the specific calculation formula is as follows:
Figure FDA0002218656420000032
whereinRepresenting an affine transformation matrix; a. the0Is an optimal transformation matrix;
D2. transformation matrix A obtained by optimization0And aligning all the face images to the standard face images, and cutting the face images into images with uniform sizes.
CN201610591877.1A 2016-07-26 2016-07-26 Face attribute recognition method based on multitask deep learning Active CN106203395B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610591877.1A CN106203395B (en) 2016-07-26 2016-07-26 Face attribute recognition method based on multitask deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610591877.1A CN106203395B (en) 2016-07-26 2016-07-26 Face attribute recognition method based on multitask deep learning

Publications (2)

Publication Number Publication Date
CN106203395A CN106203395A (en) 2016-12-07
CN106203395B true CN106203395B (en) 2020-01-14

Family

ID=57495030

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610591877.1A Active CN106203395B (en) 2016-07-26 2016-07-26 Face attribute recognition method based on multitask deep learning

Country Status (1)

Country Link
CN (1) CN106203395B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3913545A3 (en) * 2020-12-14 2022-03-16 Beijing Baidu Netcom Science And Technology Co., Ltd. Method and apparatus for updating parameter of multi-task model, and electronic device

Families Citing this family (65)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106650653B (en) * 2016-12-14 2020-09-15 广东顺德中山大学卡内基梅隆大学国际联合研究院 Construction method of human face recognition and age synthesis combined model based on deep learning
CN106815566B (en) * 2016-12-29 2021-04-16 天津中科智能识别产业技术研究院有限公司 Face retrieval method based on multitask convolutional neural network
CN106874840B (en) * 2016-12-30 2019-10-22 东软集团股份有限公司 Vehicle information recognition method and device
CN107066941A (en) * 2017-03-01 2017-08-18 桂林电子科技大学 A kind of face identification method and system
CN107145857B (en) * 2017-04-29 2021-05-04 深圳市深网视界科技有限公司 Face attribute recognition method and device and model establishment method
CN107301381A (en) * 2017-06-01 2017-10-27 西安电子科技大学昆山创新研究院 Recognition Method of Radar Emitters based on deep learning and multi-task learning strategy
CN109033921A (en) * 2017-06-08 2018-12-18 北京君正集成电路股份有限公司 A kind of training method and device of identification model
CN109359499A (en) * 2017-07-26 2019-02-19 虹软科技股份有限公司 A kind of method and apparatus for face classifier
CN109299487B (en) * 2017-07-25 2023-01-06 展讯通信(上海)有限公司 Neural network system, accelerator, modeling method and device, medium and system
CN107545263B (en) * 2017-08-02 2020-12-15 清华大学 Object detection method and device
CN107437081A (en) * 2017-08-07 2017-12-05 北京中星微电子有限公司 Face identification method, device and storage medium based on depth volume neutral net
CN107704813B (en) * 2017-09-19 2020-11-17 北京一维大成科技有限公司 Face living body identification method and system
CN107832667A (en) * 2017-10-11 2018-03-23 哈尔滨理工大学 A kind of face identification method based on deep learning
CN107844781A (en) * 2017-11-28 2018-03-27 腾讯科技(深圳)有限公司 Face character recognition methods and device, electronic equipment and storage medium
CN107844782A (en) * 2017-11-29 2018-03-27 济南浪潮高新科技投资发展有限公司 A kind of face identification method based on the serial depth network of multitask
CN107766850B (en) * 2017-11-30 2020-12-29 电子科技大学 Face recognition method based on combination of face attribute information
CN107944416A (en) * 2017-12-06 2018-04-20 成都睿码科技有限责任公司 A kind of method that true man's verification is carried out by video
CN107977456B (en) * 2017-12-15 2018-10-30 清华大学 A kind of multi-source big data analysis method based on multitask depth network
CN107895160A (en) * 2017-12-21 2018-04-10 曙光信息产业(北京)有限公司 Human face detection and tracing device and method
CN108363948A (en) * 2017-12-29 2018-08-03 武汉烽火众智数字技术有限责任公司 A kind of face information structural method for video investigation
CN108596011A (en) * 2017-12-29 2018-09-28 中国电子科技集团公司信息科学研究院 A kind of face character recognition methods and device based on combined depth network
CN108428238B (en) * 2018-03-02 2022-02-15 南开大学 Multi-type task general detection method based on deep network
CN108510061B (en) * 2018-03-19 2022-03-29 华南理工大学 Method for synthesizing face by multiple monitoring videos based on condition generation countermeasure network
CN108596094B (en) * 2018-04-24 2021-02-05 杭州数为科技有限公司 Character style detection system, method, terminal and medium
CN110263603B (en) * 2018-05-14 2021-08-06 桂林远望智能通信科技有限公司 Face recognition method and device based on central loss and residual error visual simulation network
CN108764207B (en) * 2018-06-07 2021-10-19 厦门大学 Face expression recognition method based on multitask convolutional neural network
CN109325398B (en) * 2018-06-30 2020-10-09 东南大学 Human face attribute analysis method based on transfer learning
CN108960167B (en) * 2018-07-11 2023-08-18 腾讯科技(深圳)有限公司 Hairstyle identification method, device, computer readable storage medium and computer equipment
CN110737446B (en) * 2018-07-20 2021-10-12 杭州海康威视数字技术股份有限公司 Method and device for updating parameters
CN109214281A (en) * 2018-07-30 2019-01-15 苏州神指微电子有限公司 A kind of CNN hardware accelerator for AI chip recognition of face
CN109190514B (en) * 2018-08-14 2021-10-01 电子科技大学 Face attribute recognition method and system based on bidirectional long-short term memory network
CN109447259A (en) * 2018-09-21 2019-03-08 北京字节跳动网络技术有限公司 Multitasking and multitasking model training method, device and hardware device
CN109359575B (en) 2018-09-30 2022-05-10 腾讯科技(深圳)有限公司 Face detection method, service processing method, device, terminal and medium
CN109359688A (en) * 2018-10-19 2019-02-19 厦门理工学院 A kind of design method of the outer origin output compromise filter of premium class
CN109711252A (en) * 2018-11-16 2019-05-03 天津大学 A kind of face identification method of more ethnic groups
CN109558837B (en) * 2018-11-28 2024-03-22 北京达佳互联信息技术有限公司 Face key point detection method, device and storage medium
CN109727071A (en) * 2018-12-28 2019-05-07 中国科学院半导体研究所 Method and system for advertisement recommendation
CN111382642A (en) * 2018-12-29 2020-07-07 北京市商汤科技开发有限公司 Face attribute recognition method and device, electronic equipment and storage medium
CN110069994B (en) * 2019-03-18 2021-03-23 中国科学院自动化研究所 Face attribute recognition system and method based on face multiple regions
CN110046554B (en) * 2019-03-26 2022-07-12 青岛小鸟看看科技有限公司 Face alignment method and camera
CN110163269A (en) * 2019-05-09 2019-08-23 北京迈格威科技有限公司 Model generating method, device and computer equipment based on deep learning
CN110163151B (en) * 2019-05-23 2022-07-12 北京迈格威科技有限公司 Training method and device of face model, computer equipment and storage medium
CN110489951B (en) * 2019-07-08 2021-06-11 招联消费金融有限公司 Risk identification method and device, computer equipment and storage medium
CN110263768A (en) * 2019-07-19 2019-09-20 深圳市科葩信息技术有限公司 A kind of face identification method based on depth residual error network
CN110443189B (en) * 2019-07-31 2021-08-03 厦门大学 Face attribute identification method based on multitask multi-label learning convolutional neural network
CN110414489A (en) * 2019-08-21 2019-11-05 五邑大学 A kind of face beauty prediction technique based on multi-task learning
CN110633669B (en) * 2019-09-12 2024-03-26 华北电力大学(保定) Mobile terminal face attribute identification method based on deep learning in home environment
CN112825119A (en) * 2019-11-20 2021-05-21 北京眼神智能科技有限公司 Face attribute judgment method and device, computer readable storage medium and equipment
CN111507263B (en) * 2020-04-17 2022-08-05 电子科技大学 Face multi-attribute recognition method based on multi-source data
CN111626115A (en) * 2020-04-20 2020-09-04 北京市西城区培智中心学校 Face attribute identification method and device
CN111598000A (en) * 2020-05-18 2020-08-28 中移(杭州)信息技术有限公司 Face recognition method, device, server and readable storage medium based on multiple tasks
CN111753770A (en) * 2020-06-29 2020-10-09 北京百度网讯科技有限公司 Person attribute identification method and device, electronic equipment and storage medium
CN112200008A (en) * 2020-09-15 2021-01-08 青岛邃智信息科技有限公司 Face attribute recognition method in community monitoring scene
CN112149556A (en) * 2020-09-22 2020-12-29 南京航空航天大学 Face attribute recognition method based on deep mutual learning and knowledge transfer
WO2022061726A1 (en) * 2020-09-25 2022-03-31 Intel Corporation Method and system of multiple facial attributes recognition using highly efficient neural networks
CN112287765A (en) * 2020-09-30 2021-01-29 新大陆数字技术股份有限公司 Face living body detection method, device and equipment and readable storage medium
CN112488003A (en) * 2020-12-03 2021-03-12 深圳市捷顺科技实业股份有限公司 Face detection method, model creation method, device, equipment and medium
WO2022116163A1 (en) * 2020-12-04 2022-06-09 深圳市优必选科技股份有限公司 Portrait segmentation method, robot, and storage medium
CN112801138B (en) * 2021-01-05 2024-04-09 北京交通大学 Multi-person gesture estimation method based on human body topological structure alignment
CN112949382A (en) * 2021-01-22 2021-06-11 深圳市商汤科技有限公司 Camera movement detection method and device, and electronic device
CN112507978B (en) * 2021-01-29 2021-05-28 长沙海信智能系统研究院有限公司 Person attribute identification method, device, equipment and medium
CN113239727A (en) * 2021-04-03 2021-08-10 国家计算机网络与信息安全管理中心 Person detection and identification method
CN113743243A (en) * 2021-08-13 2021-12-03 厦门大学 Face beautifying method based on deep learning
CN113642541B (en) * 2021-10-14 2022-02-08 环球数科集团有限公司 Face attribute recognition system based on deep learning
CN117079337B (en) * 2023-10-17 2024-02-06 成都信息工程大学 High-precision face attribute feature recognition device and method

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6038337A (en) * 1996-03-29 2000-03-14 Nec Research Institute, Inc. Method and apparatus for object recognition
CN101950415A (en) * 2010-09-14 2011-01-19 武汉大学 Shape semantic model constraint-based face super-resolution processing method
CN104636755A (en) * 2015-01-31 2015-05-20 华南理工大学 Face beauty evaluation method based on deep learning

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6038337A (en) * 1996-03-29 2000-03-14 Nec Research Institute, Inc. Method and apparatus for object recognition
CN101950415A (en) * 2010-09-14 2011-01-19 武汉大学 Shape semantic model constraint-based face super-resolution processing method
CN104636755A (en) * 2015-01-31 2015-05-20 华南理工大学 Face beauty evaluation method based on deep learning

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
《多任务学习及卷积神经网络在人脸识别中的应用》;邵蔚元 等;《计算机工程与应用》;20160701;第52卷(第13期);第2-4节,图2 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3913545A3 (en) * 2020-12-14 2022-03-16 Beijing Baidu Netcom Science And Technology Co., Ltd. Method and apparatus for updating parameter of multi-task model, and electronic device

Also Published As

Publication number Publication date
CN106203395A (en) 2016-12-07

Similar Documents

Publication Publication Date Title
CN106203395B (en) Face attribute recognition method based on multitask deep learning
CN108615010B (en) Facial expression recognition method based on parallel convolution neural network feature map fusion
CN106203533B (en) Deep learning face verification method based on combined training
CN108182441B (en) Parallel multichannel convolutional neural network, construction method and image feature extraction method
CN106650806B (en) A kind of cooperating type depth net model methodology for pedestrian detection
CN103106265B (en) Similar image sorting technique and system
CN107016405A (en) A kind of insect image classification method based on classification prediction convolutional neural networks
CN107977671A (en) A kind of tongue picture sorting technique based on multitask convolutional neural networks
CN107506722A (en) One kind is based on depth sparse convolution neutral net face emotion identification method
CN106780466A (en) A kind of cervical cell image-recognizing method based on convolutional neural networks
CN107122375A (en) The recognition methods of image subject based on characteristics of image
CN109165674A (en) A kind of certificate photo classification method based on multi-tag depth convolutional network
CN106203356B (en) A kind of face identification method based on convolutional network feature extraction
CN104915643A (en) Deep-learning-based pedestrian re-identification method
CN106126585B (en) The unmanned plane image search method combined based on quality grading with perceived hash characteristics
CN104463209A (en) Method for recognizing digital code on PCB based on BP neural network
CN104268593A (en) Multiple-sparse-representation face recognition method for solving small sample size problem
CN104134071A (en) Deformable part model object detection method based on color description
CN105005765A (en) Facial expression identification method based on Gabor wavelet and gray-level co-occurrence matrix
CN109190643A (en) Based on the recognition methods of convolutional neural networks Chinese medicine and electronic equipment
CN109815967A (en) CNN ship seakeeping system and method based on Fusion Features
CN110516537B (en) Face age estimation method based on self-learning
CN105138975B (en) A kind of area of skin color of human body dividing method based on degree of depth conviction network
CN101169830A (en) Human face portrait automatic generation method based on embedded type hidden markov model and selective integration
CN114038037B (en) Expression label correction and identification method based on separable residual error attention network

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant