CN106203395B - Face attribute recognition method based on multitask deep learning - Google Patents
Face attribute recognition method based on multitask deep learning Download PDFInfo
- Publication number
- CN106203395B CN106203395B CN201610591877.1A CN201610591877A CN106203395B CN 106203395 B CN106203395 B CN 106203395B CN 201610591877 A CN201610591877 A CN 201610591877A CN 106203395 B CN106203395 B CN 106203395B
- Authority
- CN
- China
- Prior art keywords
- face
- image
- training
- neural network
- key point
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
- G06V40/171—Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
Abstract
A face attribute identification method based on multitask deep learning relates to face attribute identification in computer vision. Preparing an image dataset; carrying out face detection on each image in the image data set one by one; performing face key point detection on all detected faces; aligning each face to a standard face image according to a face alignment method for the detected face key points to form a face image training set; calculating an average face image in a training set; constructing a multitask deep convolutional neural network, and carrying out network parameter training after subtracting an average face image from each face image in a face image training set to obtain a convolutional neural network model; carrying out face detection and face key point detection on a test image to be recognized, and aligning a face in the image to a standard face image according to a face key point; and subtracting the average face image from the standard face image, and putting the standard face image into the constructed convolutional neural network model to perform feed-forward operation, so as to obtain the face image.
Description
Technical Field
The invention relates to face attribute recognition in computer vision, in particular to a face attribute recognition method based on multitask deep learning.
Background
The image-based face attribute recognition method is a process of judging the face attribute in an image by using a pattern recognition technology according to a given input image. The attributes of the human face contained in the human face image mainly include: age, gender, expression, race, whether to wear glasses, whether to make up, etc. The human face attribute recognition is automatically carried out by utilizing the computer, so that the human-computer interaction performance can be effectively improved, and the method has very important practical application value. The process of face attribute recognition comprises the following steps: the method comprises the steps of face detection technology, face image preprocessing technology, face feature extraction, training of a face attribute classifier and the like. The performance of the face feature extraction and the face attribute classifier directly affects the final face attribute recognition performance.
At present, the face attribute recognition technology is mainly completed by two steps: extracting human face features and training a human face attribute classifier. The face feature extraction technology is divided into two main categories according to different feature acquisition modes: manual design features and automatic learning features. The performance of the classifier is directly affected by the quality of the human face features. The manual design features are mainly as follows: SIFT feature (D.G.Lowe.visual image features from scale-innovative keys [ J ]. International Journal of Computer Vision,2004,60(2): 91-110), LBP feature (T.Ahonen, A.Hadid, M.Pietikain.face description with local binding patterns: Application to face registration [ J ]. IEEE Transactions on Pattern Analysis and machine Analysis, IE2006, 28(12): 2037-. However, these manually designed features are mostly based on expert experience design, and it is difficult to extract features that are effective for various tasks. Second, manually designed feature extraction is decoupled from classifier design, resulting in selected features that are not best suited for a particular classifier.
Deep learning has recently become one of the research hotspots in the field of computer vision. Unlike the traditional method of decomposing the recognition task into feature extraction and classifier training, the deep learning organically combines the feature extraction and the classifier training, directly takes the original data as input, and simultaneously performs feature extraction learning and classifier training learning. The deep learning method for statistical learning by putting feature learning and classifier training in the same frame effectively avoids the significance gap between feature extraction and a target task classifier, facilitates the feature extraction and the target task classifier, and overcomes the trouble of manually designing features. The depth model formed by the multilayer neural network has the characteristics of automatically acquiring characteristics from low level to high level, from simple to complex, and from general to special. For example: in a typical image classification network, edge information is often extracted from the previous layers of the network, angle information is extracted from the middle layer, and contour information, target information and the like are extracted from the later layers. The lower the hierarchy, the simpler and more universal the extracted features, and gradually extracting the features related to the target task.
Training a deep learning model often requires a large amount of label data to avoid over-fitting the learning model with a small amount of training data. However, it is very time-consuming and labor-consuming to acquire a large amount of tag data, and it is a key problem worth solving to explore the problem of insufficient data by acquiring characteristics of different features layer by using a deep network model.
Disclosure of Invention
The invention aims to provide a face attribute recognition method based on multitask deep learning.
The invention comprises the following steps:
A. an image data set is prepared that contains a large number of faces and corresponding face attribute labels.
B. And carrying out face detection on each image in the image data set one by one to obtain the position of the face in each image.
C. And performing face key point detection on all the detected faces to acquire the positions of the face key points in each image.
D. And aligning each face to a standard face image according to a face alignment method for the detected face key points to form a face image training set.
E. And calculating an average human face image in a training set by using the human face image training set.
F. And constructing a multi-task deep convolutional neural network, and carrying out network parameter training after subtracting the average face image from each face image in the face image training set to obtain a convolutional neural network model.
G. And respectively carrying out face detection and face key point detection on the test image to be recognized, and aligning the face in the image to the standard face image according to the face key point.
H. And subtracting the average face image from the standard face image, and putting the standard face image into the constructed convolutional neural network model for feedforward operation to obtain a plurality of attribute recognition results of the face.
In the step a, the prepared image data set may adopt image data including human faces with better diversity acquired in a complex scene, and provide corresponding K personal face attribute labels, where K is a number of learned tasks and a natural number; the convolutional neural network structure based on multi-task learning is adopted, and each face image does not need to have all face attribute label data at the same time, so that the existing face database can be fully utilized for combination to form a large-scale image data set.
In step B, the face detection for each image in the image data set one by one may adopt a common face detection method to obtain a position of a face in each image, and the common face detection method may adopt an OpenCV-self-contained face detection method.
In step C, the face key point detection may adopt a common face key point detection method to obtain the position of the face key point in each image, and the common face key point detection method may adopt a Dlib-carried face key point detection method.
In step D, the face alignment method is affine transformation based on a two-dimensional image, and specifically includes the following steps:
D1. and fitting by using a least square method according to the matching relation between the face key points and the standard face key points to obtain an optimal transformation matrix. Assuming d coordinates of the key points of the standard face asWhereinThe coordinates of the ith standard face key point are obtained; d is the number of key points of the face and d is a natural number; and the coordinates of the key points of the detected face are { (x)1;y1),(x2;y2),...,(xd;yd) In which (x)i,yi) Is the ith individual face key point coordinate of the detected face and extends the coordinate to src { (x)1;y1;1),(x2;y2;1),...,(xd;yd(ii) a 1) And (6) optimizing by a least square method. The specific calculation formula is as follows:
D2. transformation matrix A obtained by optimization0And aligning all the face images to the standard face images, and cutting the face images into images with uniform sizes.
In step E, the specific method for calculating the average face image in the training set for the face image training set may be:
E1. the mean image is an arithmetic mean image M obtained by calculating each channel (comprising 3 channels of RGB) of a face image training set, wherein the calculation mode of each channel and M is as follows:
wherein the content of the first and second substances, andthe R channel, the G channel and the B channel of the nth human face image are respectively arranged; n is the total image number of the face image training set, and N is a natural number. MR,MGAnd MBArithmetic mean images of an R channel, a G channel and a B channel respectively;
E2. the arithmetic mean value image obtained by the RGB three channels is combined into an average face image, and the calculation formula is as follows:
M=[MR,MG,MB];
E3. dividing the face image training set into training data and verification data according to the ratio of 9: 1.
In step F, the specific method for constructing the multitask deep convolutional neural network may be:
F1. training data are randomly disordered, each batch of samples with the size of m are set, and data are divided, wherein m is a natural number. In the process of training the deep convolutional neural network model, learning the weight parameters of the neural network by using a batch gradient updating method;
F2. a structure of a convolutional neural network is designed, which includes convolutional layers, downsampling layers, and full-connect layers. Each convolution layer and the full-connection layer adopt nonlinear rectification activation functions; in the face attribute recognition of the multi-task deep learning, the whole network structure is divided into a sharing layer and a unique layer, the sharing layer is shared by all tasks, and the tasks participate in the training of parameters of the sharing layer together. The unique layer is occupied by each task independently, parameter learning is carried out by using independent data of each task, and the number of the shared layers is assumed to be S, wherein S is a natural number; the number of the unique layers is U, wherein U is a natural number; setting different numbers of S and U according to different multi-task combinations;
F3. setting hyper-parameters such as the number of convolution layer filters and the number of characteristic graphs, the size of the filters, the size of kernels in down-sampling layers, the learning rate of each layer, the initial weight value and the like required in a convolution neural network structure;
F4. in the training process of the convolutional neural network, training skills such as impulse and abandon are adopted for accelerating the training of the convolutional neural network;
F5. judging whether to stop training according to the performance of the trained network model parameters on the verification data;
F6. and extracting the trained network model parameters W.
The method utilizes the deep convolutional neural network to simultaneously learn the feature extraction and the attribute identification of the face image, so that the learned features are more favorable for the identification of a classifier, and the feature extraction and the classifier training are not required to be respectively carried out. For multi-task face attribute recognition of K tasks, all training data do not need to have K label attributes at the same time, and face data with one or more attributes can be used for training network parameters, and different attribute tasks can benefit accordingly.
Unlike a traditional deep convolutional neural network which only has a single label and a single network output, the convolutional neural network based on the multitask deep learning has a plurality of outputs. The objective function of the network training is a combination of a plurality of Softmax loss functions and an L2 loss function. Assume that there are K tasks that need to be learned together. Then for the face attribute recognition of the ith classification task, the loss function is defined as follows:
wherein Representing the probability value calculated by the Softmax loss function for each attribute class;a value representing a fully connected classification output in the target class; ciThe number of categories of the ith task is represented, and i is a natural number.
For the face attribute recognition of the jth regression task, the loss function is defined as follows:
In the network training, the cost loss functions of all tasks are combined, and a total optimization objective function is formed as follows:
wherein alpha iskThe penalty function representing the kth task is weighted with the total penalty function. By default αkThe values of (a) are all 1, indicating that the respective tasks are equally important.
Compared with the prior art, the method has the advantages of reducing the requirement of data quantity required by training network model parameters, reducing the risk of overfitting training data, reducing the recognition time which is spread in a single task, and effectively improving the accuracy of face attribute recognition.
Drawings
Fig. 1 is a schematic diagram of attributes of smile, gender and attraction of a female face.
Fig. 2 is a schematic diagram of attributes of smile, gender and attraction of a male face.
Detailed Description
The method of the present invention is described in detail below with reference to the accompanying drawings and examples.
The invention comprises the following steps:
s1, preparing an image data set which comprises a large number of human faces and corresponding human face attribute labels. The face attribute database used in this example is an image data set in the CelebrayA database, which contains over 20 million face images and 40 face attributes. Three representative face attribute tasks are employed (K — 3): and explaining the attribute of the human face gender, the attribute of the human face smile and the attribute of the human face attraction force. Schematic representation of three attributes as shown in FIGS. 1 and 2, with labels set to y, respectively1,y2,y3。
And S2, carrying out face detection on each image in the image data set one by one to obtain the position of the face in each image. The step can adopt any one of the existing human face detection methods to carry out human face detection. The face detection method of the OpenCV is adopted in the embodiment, and the face detection method has the advantage of being capable of rapidly detecting the face.
And S3, detecting key points of the human face on all the detected human faces to obtain the position of the key points of the human face in each image. The step can adopt any one of the existing face key point detection methods to detect. In this embodiment, 68 personal face key points can be obtained by using a Dlib-carried face key point detection method.
S4, aligning each face to a standard face image according to a face alignment method for the detected face key points to form a face image training set, which specifically comprises the following steps:
(1) and fitting by using a least square method according to the matching relation between the face key points and the standard face key points to obtain an optimal transformation matrix. Assuming that the 68 coordinates of the key points of the standard face areWhereinThe coordinates of the ith standard face key point are obtained; d is the number of key points of the face and d is a natural number; and the coordinates of the key points of the detected face are { (x)1;y1),(x2;y2),...,(x68;y68) In which (x)i,yi) Is the ith individual face key point coordinate of the detected face and extends the coordinate to src { (x)1;y1;1),(x2;y2;1),...,(x68;y68(ii) a 1) And (6) optimizing by a least square method. The specific calculation formula is as follows,
(2) Transformation matrix A obtained by optimization0All face images are aligned to a standard face image and cut into images of uniform size 128 x 128.
S5, calculating an average face image in a training set for the face image training set, wherein the method specifically comprises the following steps:
(1) the mean image is an arithmetic mean image M calculated for each channel (including 3 channels of RGB) of the face image training set, wherein the calculation for each channel and M is as follows,
wherein Andthe R channel, the G channel and the B channel of the nth human face image are respectively arranged; n is the total image number of the face image training set, and N is a natural number. MR,MGAnd MBArithmetic mean images for the R, G and B channels, respectively.
(2) The arithmetic mean value images obtained by RGB three channels are combined into an average face image, and the calculation formula is M ═ MR,MG,MB]。
(3) Dividing the face image training set into training data and verification data according to the ratio of 9: 1.
S6, constructing a multi-task deep convolutional neural network, and training network parameters after subtracting an average face image from each face image in a face image training set to obtain a convolutional neural network model, wherein the method specifically comprises the following steps:
(1) training data are randomly scrambled, each batch is set to be m-128 samples, and data are divided. In the process of training the deep convolutional neural network model, a batch gradient updating method is used for learning the weight parameters of the neural network.
(2) A structure of a convolutional neural network is designed, which includes convolutional layers, downsampling layers, and full-connect layers. Each convolutional layer and the full link layer adopt nonlinear rectification activation functions. In the face attribute recognition of multi-task deep learning, the method integratesThe body network structure is divided into two parts of a shared layer and a unique layer. The sharing layer is shared by all tasks, and a plurality of tasks participate in the training of the sharing layer parameters together. The unique layer is occupied by each task independently, and parameter learning is carried out by using data of each task independently. Assuming that the number of sharing layers is S-10; the number of unique layers is U-2. The loss functions used are all Softmax functions, alphakAre all set to 1.
(3) The number of convolutional layer filters and the number of feature maps required in the convolutional neural network structure, the size of the filters, the size of kernels in downsampling layers, the learning rate of each layer, the initial value of weight and other hyper-parameters are set, and the network structure of the deep convolutional neural network is shown in table 1.
TABLE 1
Network layer name | Type (B) | Input size | Output size | Filter size/step size |
Conv1_1 | Convolutional layer | 128*128*3 | 128*128*32 | 3*3/1 |
Conv1_2 | Convolutional layer | 128*128*32 | 128*128*64 | 3*3/1 |
Pool1 | Downsampling layer | 128*128*64 | 64*64*64 | 2*2/2 |
Conv2_1 | Convolutional layer | 64*64*64 | 64*64*96 | 3*3/1 |
Conv2_2 | Convolutional layer | 64*64*96 | 64*64*128 | 3*3/1 |
Pool2 | Downsampling layer | 64*64*128 | 32*32*128 | 2*2/2 |
Conv3_1 | Convolutional layer | 32*32*128 | 32*32*128 | 3*3/1 |
Conv3_2 | Convolutional layer | 32*32*128 | 32*32*192 | 3*3/1 |
Pool3 | Downsampling layer | 32*32*192 | 16*16*192 | 2*2/2 |
Conv4_1 | Convolutional layer | 16*16*192 | 16*16*256 | 3*3/1 |
Conv4_2 | Convolutional layer | 16*16*256 | 16*16*256 | 3*3/1 |
Pool4 | Downsampling layer | 16*16*256 | 8*8*256 | 2*2/2 |
Conv5_1 | Convolutional layer | 8*8*256 | 8*8*320 | 3*3/1 |
Conv5_2 | Convolutional layer | 8*8*320 | 8*8*320 | 3*3/1 |
Pool5 | Downsampling layer | 8*8*320 | 4*4*320 | 2*2/2 |
Dropout1 | Dropout layer | 4*4*320 | 4*4*320 | |
Fc1 | Full connection layer | 4*4*320 | 1*1*128 | |
Fc2 | Full connection layer | 1*1*128 | 1*1*Ci |
(4) In the training process of the convolutional neural network, training skills such as impulse and abandon are adopted for accelerating the training of the convolutional neural network.
(5) And judging whether to stop training the trained network model parameters according to the performance of the trained network model parameters on the verification data.
(6) And extracting the trained network model parameters W.
S7, for any given image to be subjected to face attribute recognition, changing the image to be tested into a color image with the size of 128 multiplied by 128 by using the same data preprocessing method in the steps S1-S4, subtracting a mean value image in training data, inputting the color image into a trained deep convolutional neural network, and finally obtaining recognition results of 3 different face attributes.
And S8, performing operation of step S7 on each face image of the CelebrayA test data, and comparing the attribute identification precision with the prediction time. The precision of the single task network and the multitask network on the face CelebrayA test data and the time comparison result are shown in the table 2, and it can be seen from the table that under the same network structure, the accuracy of attribute identification can be improved by the multitask deep learning attribute identification method, and meanwhile, the average prediction time of each task can be greatly reduced by utilizing the multitask attribute identification.
TABLE 2
Method of producing a composite material | T1 | T2 | T3 | T1+T2 | T1+T3 | T2+T3 | T1+T2+T3 |
Face gender (T1) | 0.9670 | N/A | N/A | 0.9580 | 0.9690 | N/A | 0.9680 |
Face smile (T2) | N/A | 0.9230 | N/A | 0.9210 | N/A | 0.9350 | 0.9360 |
Face attraction (T3) | N/A | N/A | 0.8060 | N/A | 0.8130 | 0.8160 | 0.8220 |
Total predicted time (seconds) | 0.0230 | 0.0230 | 0.0230 | 0.0270 | 0.0270 | 0.0270 | 0.0290 |
Average predicted time (seconds) | 0.0230 | 0.0230 | 0.0230 | 0.0135 | 0.0135 | 0.0135 | 0.0097 |
The invention aims at different face attribute recognition tasks, can learn the shared network weight together, and train the unique network weight independently, thereby greatly reducing the requirement of the whole training data. The invention effectively improves the performance of face attribute recognition.
Claims (7)
1. The face attribute recognition method based on the multitask deep learning is characterized by comprising the following steps of:
A. preparing an image data set comprising a plurality of faces and corresponding face attribute labels;
B. carrying out face detection on each image in the image data set one by one to obtain the position of a face in each image;
C. performing face key point detection on all detected faces to acquire the position of the face key point in each image;
D. aligning each face to a standard face image according to a face alignment method for the detected face key points to form a face image training set;
E. calculating an average face image in a training set by a face image training set, wherein the specific method comprises the following steps:
E1. the mean image is each channel of a face image training set, comprises 3 channels of RGB, and is an arithmetic mean image M obtained through calculation, wherein the calculation mode of each channel and M is as follows:
wherein the content of the first and second substances,andthe R channel, the G channel and the B channel of the nth human face image are respectively arranged; n is the total image number of the face image training set, and N is a natural number; mR,MGAnd MBArithmetic mean images of an R channel, a G channel and a B channel respectively;
E2. the arithmetic mean value image obtained by the RGB three channels is combined into an average face image, and the calculation formula is as follows:
M=[MR,MG,MB];
E3. dividing a face image training set into training data and verification data according to the ratio of 9: 1;
F. constructing a multitask deep convolutional neural network, and carrying out network parameter training after subtracting an average face image from each face image in a face image training set to obtain a convolutional neural network model;
the specific method for constructing the multitask deep convolution neural network comprises the following steps:
F1. randomly disordering training data, setting the size of each batch of samples to be m, and dividing the data, wherein m is a natural number; in the process of training the deep convolutional neural network model, learning the weight parameters of the neural network by using a batch gradient updating method;
F2. designing a structure of a convolutional neural network, wherein the structure comprises a convolutional layer, a downsampling layer and a full-connection layer; each convolution layer and the full-connection layer adopt nonlinear rectification activation functions; in the face attribute recognition of multi-task deep learning, the whole network structure is divided into a sharing layer and a unique layer, the sharing layer is shared by all tasks, and a plurality of tasks participate in the training of parameters of the sharing layer together; the unique layer is occupied by each task independently, parameter learning is carried out by using independent data of each task, and the number of the shared layers is assumed to be S, wherein S is a natural number; the number of the unique layers is U, wherein U is a natural number; setting different numbers of S and U according to different multi-task combinations;
F3. setting the number of convolutional layer filters and the number of characteristic graphs required in a convolutional neural network structure, the size of the filters, the size of kernels in downsampling layers, the learning rate of each layer and a weight initial value;
F4. in the training process of the convolutional neural network, training skills such as impulse and abandon are adopted for accelerating the training of the convolutional neural network;
F5. judging whether to stop training according to the performance of the trained network model parameters on the verification data;
F6. extracting a trained network model parameter W;
G. respectively carrying out face detection and face key point detection on a test image to be recognized, and aligning a face in the image to a standard face image according to the face key point;
H. and subtracting the average face image from the standard face image, and putting the standard face image into the constructed convolutional neural network model for feedforward operation to obtain a plurality of attribute recognition results of the face.
2. The method for identifying human face attributes based on multitask deep learning as claimed in claim 1, characterized in that in step A, said prepared image data set adopts image data containing human faces collected under the scene, and provides corresponding K personal face attribute labels.
3. The method according to claim 1, wherein in step B, the face detection is performed on each image in the image data set one by using a common face detection method to obtain the position of the face in each image.
4. The method for identifying the human face attribute based on the multitask deep learning as claimed in claim 3, wherein the commonly used human face detection method adopts an OpenCV self-contained human face detection method.
5. The method for identifying human face attributes based on multitask deep learning as claimed in claim 1, wherein in step C, said human face key point detection adopts a commonly used human face key point detection method to obtain the position of the human face key point in each image.
6. The method for identifying face attributes based on multitask deep learning as claimed in claim 5, characterized in that said commonly used face key point detection method adopts a Dlib-carried face key point detection method.
7. The method for identifying the attributes of the human face based on the multitask deep learning as claimed in claim 1, wherein in the step D, the human face alignment method is affine transformation based on a two-dimensional image, and specifically comprises the following steps:
D1. fitting by using a least square method according to the matching relation between the face key points and the standard face key points to obtain an optimal transformation matrix; assuming d coordinates of the key points of the standard face asWhereinThe coordinates of the ith standard face key point are obtained; d is the number of key points of the face and d is a natural number; and the coordinates of the key points of the detected face are { (x)1;y1),(x2;y2),...,(xd;yd) In which (x)i,yi) Is the ith individual face key point coordinate of the detected face and extends the coordinate to src { (x)1;y1;1),(x2;y2;1),...,(xd;yd(ii) a 1) Optimizing by a least square method; the specific calculation formula is as follows:
whereinRepresenting an affine transformation matrix; a. the0Is an optimal transformation matrix;
D2. transformation matrix A obtained by optimization0And aligning all the face images to the standard face images, and cutting the face images into images with uniform sizes.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610591877.1A CN106203395B (en) | 2016-07-26 | 2016-07-26 | Face attribute recognition method based on multitask deep learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610591877.1A CN106203395B (en) | 2016-07-26 | 2016-07-26 | Face attribute recognition method based on multitask deep learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN106203395A CN106203395A (en) | 2016-12-07 |
CN106203395B true CN106203395B (en) | 2020-01-14 |
Family
ID=57495030
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610591877.1A Active CN106203395B (en) | 2016-07-26 | 2016-07-26 | Face attribute recognition method based on multitask deep learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106203395B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3913545A3 (en) * | 2020-12-14 | 2022-03-16 | Beijing Baidu Netcom Science And Technology Co., Ltd. | Method and apparatus for updating parameter of multi-task model, and electronic device |
Families Citing this family (65)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106650653B (en) * | 2016-12-14 | 2020-09-15 | 广东顺德中山大学卡内基梅隆大学国际联合研究院 | Construction method of human face recognition and age synthesis combined model based on deep learning |
CN106815566B (en) * | 2016-12-29 | 2021-04-16 | 天津中科智能识别产业技术研究院有限公司 | Face retrieval method based on multitask convolutional neural network |
CN106874840B (en) * | 2016-12-30 | 2019-10-22 | 东软集团股份有限公司 | Vehicle information recognition method and device |
CN107066941A (en) * | 2017-03-01 | 2017-08-18 | 桂林电子科技大学 | A kind of face identification method and system |
CN107145857B (en) * | 2017-04-29 | 2021-05-04 | 深圳市深网视界科技有限公司 | Face attribute recognition method and device and model establishment method |
CN107301381A (en) * | 2017-06-01 | 2017-10-27 | 西安电子科技大学昆山创新研究院 | Recognition Method of Radar Emitters based on deep learning and multi-task learning strategy |
CN109033921A (en) * | 2017-06-08 | 2018-12-18 | 北京君正集成电路股份有限公司 | A kind of training method and device of identification model |
CN109359499A (en) * | 2017-07-26 | 2019-02-19 | 虹软科技股份有限公司 | A kind of method and apparatus for face classifier |
CN109299487B (en) * | 2017-07-25 | 2023-01-06 | 展讯通信(上海)有限公司 | Neural network system, accelerator, modeling method and device, medium and system |
CN107545263B (en) * | 2017-08-02 | 2020-12-15 | 清华大学 | Object detection method and device |
CN107437081A (en) * | 2017-08-07 | 2017-12-05 | 北京中星微电子有限公司 | Face identification method, device and storage medium based on depth volume neutral net |
CN107704813B (en) * | 2017-09-19 | 2020-11-17 | 北京一维大成科技有限公司 | Face living body identification method and system |
CN107832667A (en) * | 2017-10-11 | 2018-03-23 | 哈尔滨理工大学 | A kind of face identification method based on deep learning |
CN107844781A (en) * | 2017-11-28 | 2018-03-27 | 腾讯科技(深圳)有限公司 | Face character recognition methods and device, electronic equipment and storage medium |
CN107844782A (en) * | 2017-11-29 | 2018-03-27 | 济南浪潮高新科技投资发展有限公司 | A kind of face identification method based on the serial depth network of multitask |
CN107766850B (en) * | 2017-11-30 | 2020-12-29 | 电子科技大学 | Face recognition method based on combination of face attribute information |
CN107944416A (en) * | 2017-12-06 | 2018-04-20 | 成都睿码科技有限责任公司 | A kind of method that true man's verification is carried out by video |
CN107977456B (en) * | 2017-12-15 | 2018-10-30 | 清华大学 | A kind of multi-source big data analysis method based on multitask depth network |
CN107895160A (en) * | 2017-12-21 | 2018-04-10 | 曙光信息产业(北京)有限公司 | Human face detection and tracing device and method |
CN108363948A (en) * | 2017-12-29 | 2018-08-03 | 武汉烽火众智数字技术有限责任公司 | A kind of face information structural method for video investigation |
CN108596011A (en) * | 2017-12-29 | 2018-09-28 | 中国电子科技集团公司信息科学研究院 | A kind of face character recognition methods and device based on combined depth network |
CN108428238B (en) * | 2018-03-02 | 2022-02-15 | 南开大学 | Multi-type task general detection method based on deep network |
CN108510061B (en) * | 2018-03-19 | 2022-03-29 | 华南理工大学 | Method for synthesizing face by multiple monitoring videos based on condition generation countermeasure network |
CN108596094B (en) * | 2018-04-24 | 2021-02-05 | 杭州数为科技有限公司 | Character style detection system, method, terminal and medium |
CN110263603B (en) * | 2018-05-14 | 2021-08-06 | 桂林远望智能通信科技有限公司 | Face recognition method and device based on central loss and residual error visual simulation network |
CN108764207B (en) * | 2018-06-07 | 2021-10-19 | 厦门大学 | Face expression recognition method based on multitask convolutional neural network |
CN109325398B (en) * | 2018-06-30 | 2020-10-09 | 东南大学 | Human face attribute analysis method based on transfer learning |
CN108960167B (en) * | 2018-07-11 | 2023-08-18 | 腾讯科技(深圳)有限公司 | Hairstyle identification method, device, computer readable storage medium and computer equipment |
CN110737446B (en) * | 2018-07-20 | 2021-10-12 | 杭州海康威视数字技术股份有限公司 | Method and device for updating parameters |
CN109214281A (en) * | 2018-07-30 | 2019-01-15 | 苏州神指微电子有限公司 | A kind of CNN hardware accelerator for AI chip recognition of face |
CN109190514B (en) * | 2018-08-14 | 2021-10-01 | 电子科技大学 | Face attribute recognition method and system based on bidirectional long-short term memory network |
CN109447259A (en) * | 2018-09-21 | 2019-03-08 | 北京字节跳动网络技术有限公司 | Multitasking and multitasking model training method, device and hardware device |
CN109359575B (en) | 2018-09-30 | 2022-05-10 | 腾讯科技(深圳)有限公司 | Face detection method, service processing method, device, terminal and medium |
CN109359688A (en) * | 2018-10-19 | 2019-02-19 | 厦门理工学院 | A kind of design method of the outer origin output compromise filter of premium class |
CN109711252A (en) * | 2018-11-16 | 2019-05-03 | 天津大学 | A kind of face identification method of more ethnic groups |
CN109558837B (en) * | 2018-11-28 | 2024-03-22 | 北京达佳互联信息技术有限公司 | Face key point detection method, device and storage medium |
CN109727071A (en) * | 2018-12-28 | 2019-05-07 | 中国科学院半导体研究所 | Method and system for advertisement recommendation |
CN111382642A (en) * | 2018-12-29 | 2020-07-07 | 北京市商汤科技开发有限公司 | Face attribute recognition method and device, electronic equipment and storage medium |
CN110069994B (en) * | 2019-03-18 | 2021-03-23 | 中国科学院自动化研究所 | Face attribute recognition system and method based on face multiple regions |
CN110046554B (en) * | 2019-03-26 | 2022-07-12 | 青岛小鸟看看科技有限公司 | Face alignment method and camera |
CN110163269A (en) * | 2019-05-09 | 2019-08-23 | 北京迈格威科技有限公司 | Model generating method, device and computer equipment based on deep learning |
CN110163151B (en) * | 2019-05-23 | 2022-07-12 | 北京迈格威科技有限公司 | Training method and device of face model, computer equipment and storage medium |
CN110489951B (en) * | 2019-07-08 | 2021-06-11 | 招联消费金融有限公司 | Risk identification method and device, computer equipment and storage medium |
CN110263768A (en) * | 2019-07-19 | 2019-09-20 | 深圳市科葩信息技术有限公司 | A kind of face identification method based on depth residual error network |
CN110443189B (en) * | 2019-07-31 | 2021-08-03 | 厦门大学 | Face attribute identification method based on multitask multi-label learning convolutional neural network |
CN110414489A (en) * | 2019-08-21 | 2019-11-05 | 五邑大学 | A kind of face beauty prediction technique based on multi-task learning |
CN110633669B (en) * | 2019-09-12 | 2024-03-26 | 华北电力大学(保定) | Mobile terminal face attribute identification method based on deep learning in home environment |
CN112825119A (en) * | 2019-11-20 | 2021-05-21 | 北京眼神智能科技有限公司 | Face attribute judgment method and device, computer readable storage medium and equipment |
CN111507263B (en) * | 2020-04-17 | 2022-08-05 | 电子科技大学 | Face multi-attribute recognition method based on multi-source data |
CN111626115A (en) * | 2020-04-20 | 2020-09-04 | 北京市西城区培智中心学校 | Face attribute identification method and device |
CN111598000A (en) * | 2020-05-18 | 2020-08-28 | 中移(杭州)信息技术有限公司 | Face recognition method, device, server and readable storage medium based on multiple tasks |
CN111753770A (en) * | 2020-06-29 | 2020-10-09 | 北京百度网讯科技有限公司 | Person attribute identification method and device, electronic equipment and storage medium |
CN112200008A (en) * | 2020-09-15 | 2021-01-08 | 青岛邃智信息科技有限公司 | Face attribute recognition method in community monitoring scene |
CN112149556A (en) * | 2020-09-22 | 2020-12-29 | 南京航空航天大学 | Face attribute recognition method based on deep mutual learning and knowledge transfer |
WO2022061726A1 (en) * | 2020-09-25 | 2022-03-31 | Intel Corporation | Method and system of multiple facial attributes recognition using highly efficient neural networks |
CN112287765A (en) * | 2020-09-30 | 2021-01-29 | 新大陆数字技术股份有限公司 | Face living body detection method, device and equipment and readable storage medium |
CN112488003A (en) * | 2020-12-03 | 2021-03-12 | 深圳市捷顺科技实业股份有限公司 | Face detection method, model creation method, device, equipment and medium |
WO2022116163A1 (en) * | 2020-12-04 | 2022-06-09 | 深圳市优必选科技股份有限公司 | Portrait segmentation method, robot, and storage medium |
CN112801138B (en) * | 2021-01-05 | 2024-04-09 | 北京交通大学 | Multi-person gesture estimation method based on human body topological structure alignment |
CN112949382A (en) * | 2021-01-22 | 2021-06-11 | 深圳市商汤科技有限公司 | Camera movement detection method and device, and electronic device |
CN112507978B (en) * | 2021-01-29 | 2021-05-28 | 长沙海信智能系统研究院有限公司 | Person attribute identification method, device, equipment and medium |
CN113239727A (en) * | 2021-04-03 | 2021-08-10 | 国家计算机网络与信息安全管理中心 | Person detection and identification method |
CN113743243A (en) * | 2021-08-13 | 2021-12-03 | 厦门大学 | Face beautifying method based on deep learning |
CN113642541B (en) * | 2021-10-14 | 2022-02-08 | 环球数科集团有限公司 | Face attribute recognition system based on deep learning |
CN117079337B (en) * | 2023-10-17 | 2024-02-06 | 成都信息工程大学 | High-precision face attribute feature recognition device and method |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6038337A (en) * | 1996-03-29 | 2000-03-14 | Nec Research Institute, Inc. | Method and apparatus for object recognition |
CN101950415A (en) * | 2010-09-14 | 2011-01-19 | 武汉大学 | Shape semantic model constraint-based face super-resolution processing method |
CN104636755A (en) * | 2015-01-31 | 2015-05-20 | 华南理工大学 | Face beauty evaluation method based on deep learning |
-
2016
- 2016-07-26 CN CN201610591877.1A patent/CN106203395B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6038337A (en) * | 1996-03-29 | 2000-03-14 | Nec Research Institute, Inc. | Method and apparatus for object recognition |
CN101950415A (en) * | 2010-09-14 | 2011-01-19 | 武汉大学 | Shape semantic model constraint-based face super-resolution processing method |
CN104636755A (en) * | 2015-01-31 | 2015-05-20 | 华南理工大学 | Face beauty evaluation method based on deep learning |
Non-Patent Citations (1)
Title |
---|
《多任务学习及卷积神经网络在人脸识别中的应用》;邵蔚元 等;《计算机工程与应用》;20160701;第52卷(第13期);第2-4节,图2 * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3913545A3 (en) * | 2020-12-14 | 2022-03-16 | Beijing Baidu Netcom Science And Technology Co., Ltd. | Method and apparatus for updating parameter of multi-task model, and electronic device |
Also Published As
Publication number | Publication date |
---|---|
CN106203395A (en) | 2016-12-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106203395B (en) | Face attribute recognition method based on multitask deep learning | |
CN108615010B (en) | Facial expression recognition method based on parallel convolution neural network feature map fusion | |
CN106203533B (en) | Deep learning face verification method based on combined training | |
CN108182441B (en) | Parallel multichannel convolutional neural network, construction method and image feature extraction method | |
CN106650806B (en) | A kind of cooperating type depth net model methodology for pedestrian detection | |
CN103106265B (en) | Similar image sorting technique and system | |
CN107016405A (en) | A kind of insect image classification method based on classification prediction convolutional neural networks | |
CN107977671A (en) | A kind of tongue picture sorting technique based on multitask convolutional neural networks | |
CN107506722A (en) | One kind is based on depth sparse convolution neutral net face emotion identification method | |
CN106780466A (en) | A kind of cervical cell image-recognizing method based on convolutional neural networks | |
CN107122375A (en) | The recognition methods of image subject based on characteristics of image | |
CN109165674A (en) | A kind of certificate photo classification method based on multi-tag depth convolutional network | |
CN106203356B (en) | A kind of face identification method based on convolutional network feature extraction | |
CN104915643A (en) | Deep-learning-based pedestrian re-identification method | |
CN106126585B (en) | The unmanned plane image search method combined based on quality grading with perceived hash characteristics | |
CN104463209A (en) | Method for recognizing digital code on PCB based on BP neural network | |
CN104268593A (en) | Multiple-sparse-representation face recognition method for solving small sample size problem | |
CN104134071A (en) | Deformable part model object detection method based on color description | |
CN105005765A (en) | Facial expression identification method based on Gabor wavelet and gray-level co-occurrence matrix | |
CN109190643A (en) | Based on the recognition methods of convolutional neural networks Chinese medicine and electronic equipment | |
CN109815967A (en) | CNN ship seakeeping system and method based on Fusion Features | |
CN110516537B (en) | Face age estimation method based on self-learning | |
CN105138975B (en) | A kind of area of skin color of human body dividing method based on degree of depth conviction network | |
CN101169830A (en) | Human face portrait automatic generation method based on embedded type hidden markov model and selective integration | |
CN114038037B (en) | Expression label correction and identification method based on separable residual error attention network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |