CN110569724A - Face alignment method based on residual hourglass network - Google Patents
Face alignment method based on residual hourglass network Download PDFInfo
- Publication number
- CN110569724A CN110569724A CN201910716528.1A CN201910716528A CN110569724A CN 110569724 A CN110569724 A CN 110569724A CN 201910716528 A CN201910716528 A CN 201910716528A CN 110569724 A CN110569724 A CN 110569724A
- Authority
- CN
- China
- Prior art keywords
- network
- residual
- hourglass
- face
- hourglass network
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/46—Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
- G06V10/462—Salient features, e.g. scale invariant feature transforms [SIFT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Data Mining & Analysis (AREA)
- Human Computer Interaction (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- Mathematical Physics (AREA)
- Molecular Biology (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Image Analysis (AREA)
Abstract
the invention discloses a face alignment method based on a residual error hourglass network, which comprises the steps of firstly, additionally arranging convolution branches on the basis of a residual error module of a basic unit of the hourglass network to increase the receptive field of the network so as to better extract characteristic information under different scales, and simultaneously keeping high-resolution information; and then, by combining the characteristics of the hourglass network, the core size of the newly added convolution branches is adjusted along with the number of layers of the hourglass network, so that the relationship between the feature map resolution and the receptive field is well balanced, the network can extract more detailed information, and the structural information from local to whole is also kept. Finally, the hourglass network is stacked and is assisted by an intermediate supervision mechanism, so that normal updating of low-layer parameters is guaranteed, and the network is allowed to reevaluate the initial estimation and characteristics of the whole image. According to the invention, by stacking the new residual hourglass network, more effective information is extracted, the capability of extracting local detail information by the network is increased, and the accuracy of detecting key points of the face is improved.
Description
Technical Field
The invention belongs to the technical field of computer vision, and relates to a face alignment method based on a residual hourglass network, in particular to a face alignment method based on a novel residual hourglass network in the field of face recognition of digital images.
Background
The face alignment is also called face key point detection and is used for positioning key points such as eyes, a nose, a mouth, a face contour and the like of a face, the face alignment can help to provide accurate face shape information with specific semantics, and the face alignment plays a vital role in the fields of face recognition, face expression analysis, gender age identification, three-dimensional face modeling and the like. Due to the natural scene or the existence of facial expression, exaggerated head posture, difference of illumination conditions and partial shielding under an unconstrained condition, the problem of face alignment still faces huge challenges, so that an efficient and accurate face alignment algorithm can better meet the requirements of people.
in recent years, with the wide application of the deep learning framework in the field of face alignment, human research on face problems has been rapidly progressing. The advantage of deep learning is that the feature extraction is particularly prominent, and the deep network structure of deep learning can extract data features layer by layer, so that the extracted features are more obvious and easier to classify.
Deep learning is introduced into a face alignment algorithm DCNN (Deep Convolutional Network) [ document 1] at the earliest, and by designing a cascaded Convolutional neural Network with three levels, the problem of local optimization caused by initial improper is solved, and more accurate key point detection is obtained by means of strong feature extraction capability of the CNN. Compared with the early face alignment algorithm based on the optimization (ASM [ document 2], AAM [ document 3] - [ document 5], CLM [ document 6] - [ document 7]), the aim of face alignment is achieved by optimizing an error equation, the nonlinear optimization problem is relatively complex to solve, the solving cost is increased due to overhigh middle dimension, the face alignment algorithm [ document 8] - [ document 14] based on the cascade shape regression continuously approaches to a standard shape by gradually estimating the shape increment according to the initialized shape, the requirement on the initialized shape is higher, the extraction of the features is more complicated, and the deep learning method is obviously simpler and more efficient.
with the recognition and use of the convolutional neural network, an hourglass network [ document 15] has been proposed to replace the CNN extraction features, and for the correlation task of face alignment, the parts of the face do not have the best recognition accuracy on the same feature map. For example, eyes may be easily recognized on the feature map of the layer 3, and mouths are more easily recognized on the layer 5, so that compared with a traditional convolutional neural network, only the last layer of convolution is used as a target feature, the hourglass network combines the features of the network at the lower level and the higher level by using a unique structure from bottom to top, so that the finally obtained feature is more effective, and therefore, the maximum utilization of image feature information is achieved. The novel residual error-based hourglass network increases convolution branches on the basis of the hourglass network, correspondingly adjusts the scale of a convolution kernel along with the number of layers of the hourglass network, increases the receptive field, balances the relationship between the feature map resolution and the receptive field, enables the network to extract more detailed information, maintains the structural information from local parts to the whole part, and increases the capability of the network to extract more effective characteristics.
[ document 1] Sun Y, Wang X, Tang X. deep conditional Network case for Facial Point Detection [ C ]// Computer Vision and Pattern recognition. IEEE,2013:3476-3483.
[ document 2] dyes T F, Taylor C J, Cooper D H, Graham J. active shape models-the guiding and the application. computer vision and image understating, 1995,61(1):38-59.
[ document 3] Sauer P, Cootes T F, Taylor C J. accurate regression processes for active application models// Proceedings of the British Machine Vision conference. Dundee, Scotland,2011:681-685.
[ document 3] Cootes T F, Edwards G J, Taylor C J.active appearance models.IEEE Transactions on Pattern Analysis and Machine understanding, 2001,23(6):581-585.
[ document 5] Asthana A, Zafeiriou S, Cheng S, cationic M.Robust reconstructed response map matching with constrained local models// IEEE Conference on computer Vision and Pattern recognition. Portland, USA,2013:3444-3451.
[ document 6] exists in Cristinacce D, Cootes T.feature detection and tracking with constrained local models// Proceedings of the British Machine Vision conference. Edinburgh, UK,2006:929-938.
[ document 7] Asthana A, Zafeiriou S, CHENG Shi-yang, cationic M.Inclusion Face Alignment in the wild.// IEEE Conference on Computer Vision and Pattern recognition. Columbus, USA 2014:1859-1867.
[ document 8] Xiong Xue-han, Torre F D L.Supervised device method and its applications to face alignment// IEEE Conference on Computer Vision and Pattern recognition. Portland, USA,2013:532-539.
[ document 9] Cao Xu-dong, Wei Yi-chen, Wen Fang, Sun Jian.face alignment by application program regression. International Journal of Computer Vision,2014,107(2):177-190.
[ document 10] Burgos-Artizzu X P, Perona P, Dollar P.Robust face and interference unit administration// IEEE International Conference on computer Vison.Sydney, Australia,2013:1513-1520.
[ document 11] Ren Shao-q, Cao Xu-dong, Wei Yi-chen, Sun Jian.face alignment at 3000fps via regressing local binding facilities// IEEE Conference on computer Vision and Pattern recognition. Columbus, USA 2014:1685-1692.
[ document 12] Dollar P, Welinder P, Perona P.Cascaded position regression.// IEEE Conference on Computer Vision and Pattern recognition. san Francisco, USA,2010:1078-1085.
[ document 13] Tzimiropoulos G, cationic M.Gauss-Newton deformed part models for face alignment in-the-wire// IEEE Conference on Computer Vision and Pattern recognition. Columbus, USA,2014:1851-1858.
[ document 14] Smith B M, Brandt J, Lin Z, Zhang L. Nonparametric context modifying of local impedance for position-and expression-robust surface and local impedance localization.// IEEE Conference on Computer Vision and PatternRecognition. Columbus, USA,2014:1741-1748.
document 15A. Newell, K.Yang, and J.Deng.Stacked hourglass networks for human position estimation. in European Conference on Computer Vision, pages 483-499 Springer,2016.1,2,3
Disclosure of Invention
The invention aims to provide a novel residual hourglass network-based face alignment method, which increases the capability of extracting effective features of a network and improves the accuracy of face alignment.
the technical scheme adopted by the invention is as follows: a face alignment method based on a residual hourglass network is characterized by comprising the following steps:
Step 1: constructing a novel residual hourglass network;
The novel residual error hourglass network comprises a novel residual error module and a residual error module; the novel residual error module is characterized in that a convolution branch is newly added on the basis of the residual error module, the kernel scale of the newly added convolution branch is k, and k changes along with the current layer number hg _ level of the hourglass network; the output of the residual error module is h (x), the output of the novel residual error module is h' (x), then:
h(x)=f(x)+x;
h′(x)=f(x)+gk(x)+x;
k=hg_level*2+1;
Where x is the input of the residual block and f (x) is the value obtained by passing x throughto output of triple layer convolution operation, gk(x) The output of the novel residual error module is obtained under the condition that the x is the k in the novel convolution branch kernel scale, hg _ level is the current layer of the hourglass network, hg _ levels is the total number of layers of the hourglass network, and H (x) is the final output of the hourglass network;
step 2: obtaining an estimated face key point thermodynamic diagram (H) from an input picture through a stacked novel residual hourglass network1,H2...,HNin which Hirepresenting an estimated human face key point thermodynamic diagram obtained by the ith novel residual hourglass network, wherein i is more than or equal to 1 and less than or equal to N, and N is the number of stacked novel residual hourglass networks;
And step 3: generating a real face key point thermodynamic diagram by combining a two-dimensional Gaussian function with real face key points of an input picture
and 4, step 4: predicted human face key point thermodynamic diagram H of each stage of novel residual hourglass networkiAnd real face key point thermodynamic diagramPass to obtain LiThen the loss { L ] obtained from the whole network stage1,L2...,LNTaking an average value to obtain a final L;
and 5: training a network to obtain a training model, carrying out the training model on an input picture to obtain a predicted human face key point thermodynamic diagram H, and converting the thermodynamic diagram H into predicted human face key point coordinates P;
Step 6: and drawing the key points of the human face on an original drawing.
The invention discloses a novel residual hourglass network-based face alignment method, which is a relatively simple face alignment method with good robustness. The description of the network on the whole and local associated information of the key points is enhanced by using a novel residual hourglass network model, the completeness of effective feature extraction is enhanced, the problem of gradient disappearance (variation) along with the deepening of the network depth can be effectively avoided by combining a stacked novel residual hourglass network of an intermediate supervision mechanism, the features can be processed in the context of the network while the normal updating of bottom layer parameters is ensured, the features are allowed to be reevaluated, the fault tolerance rate of the network is enhanced, and the algorithm effect is improved.
Drawings
FIG. 1: an overall framework diagram of an embodiment of the invention;
FIG. 2: a network flow diagram of an embodiment of the invention;
FIG. 3: the novel residual error hourglass network model schematic diagram of the embodiment of the invention;
FIG. 4: in the embodiment of the invention, an accumulated error distribution map (CED) of a residual hourglass (NRHG) on a 300W data set is formed by an hourglass network (HG) and the Novel Residual Hourglass (NRHG);
FIG. 5: in the embodiment of the invention, an accumulated error distribution map (CED) of an hourglass network (HG) and a Novel Residual Hourglass (NRHG) on an IBUG data set is provided;
FIG. 6: in the embodiment of the invention, an accumulated error distribution map (CED) of a residual hourglass (NRHG) on a COFW data set is formed by an hourglass network (HG) and the Novel Residual Hourglass (NRHG);
Detailed Description
in order to facilitate understanding and implementation of the present invention for those of ordinary skill in the art, the present invention is further described in detail with reference to the accompanying drawings and examples, which are described herein for the purpose of illustration and explanation only and are not intended to limit the present invention.
Referring to fig. 1 and fig. 2, the novel residual hourglass network human face alignment method provided by the invention comprises the following steps:
step 1: constructing a novel residual hourglass network;
Fig. 3 shows a schematic diagram of a novel residual hourglass network structure, the basic constituent units of the novel residual hourglass network are a novel residual module and a residual module, the novel residual module is a convolution branch newly added on the basis of the residual module, the size of the kernel of the newly added convolution branch is k, k changes along with the number of layers hg _ level of the hourglass network, the output of the residual module is h (x), and the output of the self-adaptive residual module is h' (x), and then:
h(x)=f(x)+x
h′(x)=f(x)+gk(x)+x
k=hg_level*2+1
where x is the input to the residual block, f (x) is the output of x via a triple convolution operation, gk(x) For the output obtained by x in the novel residual error module under the condition that the scale of the novel convolution branch kernel is k, hg _ level is the layer where the hourglass network is located currently, hg _ levels is the total number of layers of the hourglass network (designed to be 4 here), and H (x) is the final output of the hourglass network.
Step 2: obtaining an estimated face key point thermodynamic diagram (H) from an input picture through a stacked novel residual hourglass network1,H2...,HNIn which HiRepresenting an estimated human face key point thermodynamic diagram obtained by the ith novel residual hourglass network, wherein i is more than or equal to 1 and less than or equal to N, and N is the number of stacked novel residual hourglass networks;
and step 3: generating a real face key point thermodynamic diagram by combining a two-dimensional Gaussian function with real face key points of an input picture
hm(w,h,i)0<i<M=f(xi,yi)0<xi<w,0<yi<h
Where f (x, y) is a two-dimensional Gaussian function and ksize is the size of the Gaussian function, corresponding to amplitude, sigma2is the variance of a Gaussian function, (x)0,y0) Is the coordinate of the central point, here the coordinate of the key point of the real face, hm (w, h, i)0<i<MIs the thermodynamic diagram of the ith key point, w and h are the width and the height of the thermodynamic diagram respectively, and M is offtotal number of key points. Finally generated real face key point thermodynamic diagramThat is, hm (w, h, i) generated by all the key points0<i<MIn combination with (1)
and 4, step 4: the predictive face key point thermodynamic diagram H of each stage of the networki(i.e., output thermodynamic diagram of ith hourglass network) and real face key point thermodynamic diagramsPass to obtain LiThen the loss { L ] obtained from the whole network stage1,L2...,LNTaking an average value to obtain a final L;
In the formula hjandrespectively the estimated thermodynamic diagram and the real thermodynamic diagram of the j-th personal face key point of the current stage, wherein M is the total number of the key points of the face, and n is the total number of the network stages (namely the number of the stacked novel residual hourglass networks)
And 5: training a network to obtain a training model, carrying out the training model on an input picture to obtain a predicted human face key point thermodynamic diagram H, and converting the thermodynamic diagram H into predicted human face key point coordinates P;
The predicted face key point thermodynamic diagram H is composed of all key point thermodynamic diagrams hm (w, H, i)0<i<MThe deeper the color in the thermodynamic diagram, the larger the two-dimensional Gaussian function value corresponding to the position with the larger color, namely, the closer to the real key point, and the thermodynamic diagram hm (w, h, i) is formed0<i<MConverting into a one-bit vector, taking the maximum value, and calculating the position coordinate of the original thermodynamic diagram pixel where the value is locatedAnd (4) predicted face key point coordinates P.
Step 6: and drawing the key points of the human face on an original drawing.
The invention also provides a performance evaluation method of the face alignment method based on the residual hourglass network, which is used for evaluating the predicted face key point coordinate P and the real face key point coordinateCarrying out error comparison;
The quality of the face alignment algorithm is generally measured by Normalized Mean Error NME (Normalized Mean Error), which is defined as follows:
wherein M is the number of key points of the face, piAndI is the predicted coordinate and the real coordinate of the key point of the individual face, d is the normalization factor, and the embodiment adopts two normalization modes, one is Inter-pupil normalization, and the other is Inter-ocular normalization.
Compared with the face alignment method (HG) using the traditional hourglass network (network structure and training parameters are the same, and only the novel residual hourglass network is replaced by the common hourglass network), the face alignment accuracy is improved by the algorithm. The data sets used in the experiments were 300W, IBUG, COFW. 3148 images in the 300W data set training set, and 554 images in the test set; the IBUG data set training set is the same as the IBUG data set training set 300W, and the test set comprises 135 images; the COFW data set training set has 1345 image test sets with 507 images.
TABLE 1 comparison of the results of the two different normalization methods on 300W, IBUG and COFW data sets for NRHG and HG of this example (% omitted)
The data in table 1 are Normalized Mean Error (NME) under different test sets, and from the data comparison it can be seen that the new residual hourglass Network (NRHG) compares to the conventional hourglass network (HG) for both 300W and IBUG and COFW data sets, the average error obtained in two different normalization modes is reduced to different degrees, the reduction range on the difficult data set IBUG is large, this shows that the novel residual hourglass Network (NRHG) proposed in this embodiment is more effective in feature extraction of human face images than the conventional hourglass network (HG), can be more helpful for locating key points of human faces, therefore, the accuracy of the algorithm for face alignment is improved to a certain degree, and meanwhile, the NRHG provided by the embodiment has advantages over HG in positioning of key points of the face in complex scenes such as care and shielding.
referring to fig. 4, in the embodiment of the present invention, a cumulative error distribution map (CED) of a common hourglass network (HG) and a Novel Residual Hourglass (NRHG) on a 300W data set is provided;
Referring to fig. 5, in the embodiment of the present invention, a cumulative error distribution map (CED) of a common hourglass network (HG) and a Novel Residual Hourglass (NRHG) on an IBUG data set is provided;
Referring to fig. 6, in the embodiment of the present invention, a cumulative error distribution map (CED) of the common hourglass network (HG) and the Novel Residual Hourglass (NRHG) on the COFW data set is provided;
Fig. 4, fig. 5, and fig. 6 visually reflect that the face alignment effect of the Novel Residual Hourglass (NRHG) proposed by the present invention is improved to a certain extent in two different normalization manners of three different data sets compared with the conventional hourglass network (HG).
Table 2 comparison of experimental results of this example algorithm (NRHG) and other face alignment algorithms on the COFW dataset (% omitted)
Table 3 comparison of experimental results of this example algorithm (NRHG) and other face alignment algorithms on 300W dataset (% omitted)
As is clear from table 2, the experimental results of the novel residual hourglass network algorithm (NRHG) proposed in this embodiment are better than those of the above face alignment algorithms, except that the Normalized Mean Error (NME) is smaller than those of the above face alignment algorithms, FR (Feature Rate) is also significantly reduced, and the data results show the superiority of the algorithm proposed in this embodiment in the COFW data set.
Table 3 shows the comparison of the experimental results of the novel residual hourglass network face alignment algorithm (NRHG) proposed in this embodiment and other face alignment algorithms under three indexes (public subset, challenge subset, full set) on a 300W data set. The data in the table shows that the algorithm provided by the embodiment is superior to the face alignment algorithms, especially the improvement range of the experimental effect is larger compared with the traditional cascade regression RCPR, ESR, LBF and other algorithms, and it can be reflected to a certain extent that the algorithm provided by the embodiment has a significant advantage in face alignment compared with the traditional cascade regression algorithm, although the experimental result of the algorithm provided by the embodiment is not improved to a different extent compared with the face alignment algorithms, some problems still exist in the challenge subset, the novel residual hourglass network face alignment algorithm provided by the embodiment focuses on processing the face image features, although the image characterization has a certain help in the complex situations of caring and shielding, the algorithm for solving the problems of pose, illumination, shielding and the like in comparison RAR, TSR and the like also has a large improvement space, this is also the place that needs to be learned and improved later, but overall, the algorithm of the embodiment has certain advantages.
it should be understood that parts of the specification not set forth in detail are well within the prior art. The present invention may be replaced or modified by one of ordinary skill in the art without departing from the scope of the present invention as defined by the appended claims.
Claims (5)
1. A face alignment method based on a residual hourglass network is characterized by comprising the following steps:
Step 1: constructing a novel residual hourglass network;
The novel residual error hourglass network comprises a novel residual error module and a residual error module; the novel residual error module is characterized in that a convolution branch is newly added on the basis of the residual error module, the kernel scale of the newly added convolution branch is k, and k changes along with the current layer number hg _ level of the hourglass network; the output of the residual error module is h (x), the output of the novel residual error module is h' (x), then:
h(x)=f(x)+x;
h′(x)=f(x)+gk(x)+x;
k=hg_level*2+1;
Where x is the input to the residual block, f (x) is the output of x via a triple convolution operation, gk(x) The output of the novel residual error module is obtained under the condition that the x is the k in the novel convolution branch kernel scale, hg _ level is the current layer of the hourglass network, hg _ levels is the total number of layers of the hourglass network, and H (x) is the final output of the hourglass network;
Step 2: obtaining an estimated face key point thermodynamic diagram (H) from an input picture through a stacked novel residual hourglass network1,H2...,HNIn which HiRepresenting an estimated face key point thermodynamic diagram obtained by the ith novel residual hourglass network, wherein i is more than or equal to 1 and less than or equal to N, and N is a stacked novel residual hourglass networkThe number of the cells;
And step 3: generating a real face key point thermodynamic diagram by combining a two-dimensional Gaussian function with real face key points of an input picture
and 4, step 4: predicted human face key point thermodynamic diagram H of each stage of novel residual hourglass networkiAnd real face key point thermodynamic diagramPass to obtain LiThen the loss { L ] obtained from the whole network stage1,L2...,LNtaking an average value to obtain a final L;
And 5: training a network to obtain a training model, carrying out the training model on an input picture to obtain a predicted human face key point thermodynamic diagram H, and converting the thermodynamic diagram H into predicted human face key point coordinates P;
step 6: and drawing the key points of the human face on an original drawing.
2. The residual hourglass network-based face alignment method according to claim 1, wherein in step 3, a calculation formula for generating a face key point thermodynamic diagram by using a two-dimensional Gaussian function is as follows:
hm(w,h,i)0<i<M=f(xi,yi)0<xi<w,0<yi<h;
where f (x, y) is a two-dimensional Gaussian function, ksize is the size of the Gaussian function, and sigma2Is the variance of a Gaussian function, (x)0,y0) Is the coordinate of the central point, here the coordinate of the key point of the real face, hm (w, h, i)0<i<Mthe thermodynamic diagram of the ith key point is shown, w and h are respectively the width and the height of the thermodynamic diagram, and M is the total number of the key points; finally generated real face key point thermodynamic diagramThat is, hm (w, h, i) generated by all the key points0<i<MCombinations of (a) and (b).
3. The residual hourglass network-based face alignment method according to claim 1, wherein in step 4, the loss function L of each stageiAnd the final loss function L is expressed as:
In the formula, hjandRespectively an estimated thermodynamic diagram and a real thermodynamic diagram of the j-th personal face key point in the current stage, wherein M is the total number of the personal face key points, and n is the total number of the network stages, namely the number of the stacked novel residual hourglass networks.
4. The residual hourglass network-based face alignment method according to claim 2, wherein the specific implementation of step 5 comprises the following steps:
Step 5.1: will thermodynamic diagram hm (w, h, i)0<i<Mconverting into a bit vector, and taking the maximum value;
Step 5.2: calculating thermodynamic diagram hm (w, h, i)0<i<MAnd the position coordinate of the thermodynamic diagram pixel where the medium maximum value is located is the predicted face key point coordinate P.
5. A method for evaluating the performance of a face alignment method based on a residual hourglass network is characterized by comprising the following steps: for the predicted face key point coordinate P and the real face key point coordinateCarrying out error comparison;
The expression for the performance evaluation is:
Wherein M is the number of key points of the face, piAndi predicted coordinates and real coordinates of the face key points, respectively, and d is a normalization factor.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910716528.1A CN110569724B (en) | 2019-08-05 | 2019-08-05 | Face alignment method based on residual hourglass network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910716528.1A CN110569724B (en) | 2019-08-05 | 2019-08-05 | Face alignment method based on residual hourglass network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110569724A true CN110569724A (en) | 2019-12-13 |
CN110569724B CN110569724B (en) | 2021-06-04 |
Family
ID=68774545
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910716528.1A Active CN110569724B (en) | 2019-08-05 | 2019-08-05 | Face alignment method based on residual hourglass network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110569724B (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111402228A (en) * | 2020-03-13 | 2020-07-10 | 腾讯科技(深圳)有限公司 | Image detection method, device and computer readable storage medium |
CN111523484A (en) * | 2020-04-24 | 2020-08-11 | 北京嘀嘀无限科技发展有限公司 | Face key point detection method and device, electronic equipment and storage medium |
CN112417991A (en) * | 2020-11-02 | 2021-02-26 | 武汉大学 | Double-attention face alignment method based on hourglass capsule network |
CN112699847A (en) * | 2021-01-15 | 2021-04-23 | 苏州大学 | Face characteristic point detection method based on deep learning |
CN113610115A (en) * | 2021-07-14 | 2021-11-05 | 广州敏视数码科技有限公司 | Efficient face alignment method based on gray level image |
Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080247611A1 (en) * | 2007-04-04 | 2008-10-09 | Sony Corporation | Apparatus and method for face recognition and computer program |
CN101499128A (en) * | 2008-01-30 | 2009-08-05 | 中国科学院自动化研究所 | Three-dimensional human face action detecting and tracing method based on video stream |
CN106951875A (en) * | 2017-03-24 | 2017-07-14 | 深圳市唯特视科技有限公司 | The method that a kind of human body attitude estimation and face based on binary system convolution are alignd |
CN108090470A (en) * | 2018-01-10 | 2018-05-29 | 浙江大华技术股份有限公司 | A kind of face alignment method and device |
US20180263568A1 (en) * | 2017-03-09 | 2018-09-20 | The Board Of Trustees Of The Leland Stanford Junior University | Systems and Methods for Clinical Image Classification |
CN108764133A (en) * | 2018-05-25 | 2018-11-06 | 北京旷视科技有限公司 | Image-recognizing method, apparatus and system |
CN108898556A (en) * | 2018-05-24 | 2018-11-27 | 麒麟合盛网络技术股份有限公司 | A kind of image processing method and device of three-dimensional face |
CN109033946A (en) * | 2018-06-08 | 2018-12-18 | 东南大学 | Merge the estimation method of human posture of directional diagram |
CN109241910A (en) * | 2018-09-07 | 2019-01-18 | 高新兴科技集团股份有限公司 | A kind of face key independent positioning method returned based on the cascade of depth multiple features fusion |
CN109299669A (en) * | 2018-08-30 | 2019-02-01 | 清华大学 | Video human face critical point detection method and device based on double intelligent bodies |
CN109299659A (en) * | 2018-08-21 | 2019-02-01 | 中国农业大学 | A kind of human posture recognition method and system based on RGB camera and deep learning |
CN109657595A (en) * | 2018-12-12 | 2019-04-19 | 中山大学 | Based on the key feature Region Matching face identification method for stacking hourglass network |
-
2019
- 2019-08-05 CN CN201910716528.1A patent/CN110569724B/en active Active
Patent Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080247611A1 (en) * | 2007-04-04 | 2008-10-09 | Sony Corporation | Apparatus and method for face recognition and computer program |
CN101499128A (en) * | 2008-01-30 | 2009-08-05 | 中国科学院自动化研究所 | Three-dimensional human face action detecting and tracing method based on video stream |
US20180263568A1 (en) * | 2017-03-09 | 2018-09-20 | The Board Of Trustees Of The Leland Stanford Junior University | Systems and Methods for Clinical Image Classification |
CN106951875A (en) * | 2017-03-24 | 2017-07-14 | 深圳市唯特视科技有限公司 | The method that a kind of human body attitude estimation and face based on binary system convolution are alignd |
CN108090470A (en) * | 2018-01-10 | 2018-05-29 | 浙江大华技术股份有限公司 | A kind of face alignment method and device |
CN108898556A (en) * | 2018-05-24 | 2018-11-27 | 麒麟合盛网络技术股份有限公司 | A kind of image processing method and device of three-dimensional face |
CN108764133A (en) * | 2018-05-25 | 2018-11-06 | 北京旷视科技有限公司 | Image-recognizing method, apparatus and system |
CN109033946A (en) * | 2018-06-08 | 2018-12-18 | 东南大学 | Merge the estimation method of human posture of directional diagram |
CN109299659A (en) * | 2018-08-21 | 2019-02-01 | 中国农业大学 | A kind of human posture recognition method and system based on RGB camera and deep learning |
CN109299669A (en) * | 2018-08-30 | 2019-02-01 | 清华大学 | Video human face critical point detection method and device based on double intelligent bodies |
CN109241910A (en) * | 2018-09-07 | 2019-01-18 | 高新兴科技集团股份有限公司 | A kind of face key independent positioning method returned based on the cascade of depth multiple features fusion |
CN109657595A (en) * | 2018-12-12 | 2019-04-19 | 中山大学 | Based on the key feature Region Matching face identification method for stacking hourglass network |
Non-Patent Citations (2)
Title |
---|
OSHIN MISRA 等: "An Approach to Face Detection and Alignment Using Hough Transformation with Convolution Neural Network", 《ICACCA》 * |
王浩 等: "基于堆叠深度卷积沙漏网络的步态识别", 《计算机工程与应用》 * |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111402228A (en) * | 2020-03-13 | 2020-07-10 | 腾讯科技(深圳)有限公司 | Image detection method, device and computer readable storage medium |
CN111402228B (en) * | 2020-03-13 | 2021-05-07 | 腾讯科技(深圳)有限公司 | Image detection method, device and computer readable storage medium |
CN111523484A (en) * | 2020-04-24 | 2020-08-11 | 北京嘀嘀无限科技发展有限公司 | Face key point detection method and device, electronic equipment and storage medium |
CN112417991A (en) * | 2020-11-02 | 2021-02-26 | 武汉大学 | Double-attention face alignment method based on hourglass capsule network |
CN112417991B (en) * | 2020-11-02 | 2022-04-29 | 武汉大学 | Double-attention face alignment method based on hourglass capsule network |
CN112699847A (en) * | 2021-01-15 | 2021-04-23 | 苏州大学 | Face characteristic point detection method based on deep learning |
CN113610115A (en) * | 2021-07-14 | 2021-11-05 | 广州敏视数码科技有限公司 | Efficient face alignment method based on gray level image |
CN113610115B (en) * | 2021-07-14 | 2024-04-12 | 广州敏视数码科技有限公司 | Efficient face alignment method based on gray level image |
Also Published As
Publication number | Publication date |
---|---|
CN110569724B (en) | 2021-06-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110569724B (en) | Face alignment method based on residual hourglass network | |
CN109472198B (en) | Gesture robust video smiling face recognition method | |
CN110348330B (en) | Face pose virtual view generation method based on VAE-ACGAN | |
WO2020186886A1 (en) | Method and device for generating face recognition model | |
WO2020108362A1 (en) | Body posture detection method, apparatus and device, and storage medium | |
Li et al. | A comprehensive survey on 3D face recognition methods | |
CN109949255A (en) | Image rebuilding method and equipment | |
WO2017088432A1 (en) | Image recognition method and device | |
Wu et al. | MTCNN and FACENET based access control system for face detection and recognition | |
Zhang et al. | Content-adaptive sketch portrait generation by decompositional representation learning | |
WO2016023264A1 (en) | Fingerprint identification method and fingerprint identification device | |
CN111178208A (en) | Pedestrian detection method, device and medium based on deep learning | |
Zhou et al. | Pose-robust face recognition with Huffman-LBP enhanced by divide-and-rule strategy | |
CN112232184A (en) | Multi-angle face recognition method based on deep learning and space conversion network | |
CN111832405A (en) | Face recognition method based on HOG and depth residual error network | |
Yu et al. | Improvement of face recognition algorithm based on neural network | |
Hao et al. | Finger vein recognition based on multi-task learning | |
CN115376159A (en) | Cross-appearance pedestrian re-recognition method based on multi-mode information | |
Jiang et al. | Face recognition method based on sparse representation and feature fusion | |
Wan et al. | Palmprint recognition system for mobile device based on circle loss | |
CN114743234A (en) | Efficient face mask recognition method based on deep learning | |
CN110969101A (en) | Face detection and tracking method based on HOG and feature descriptor | |
CN108090460B (en) | Weber multidirectional descriptor-based facial expression recognition feature extraction method | |
Yan et al. | A parameter-free framework for general supervised subspace learning | |
Su et al. | A multiattribute sparse coding approach for action recognition from a single unknown viewpoint |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |