CN110569724B - Face alignment method based on residual hourglass network - Google Patents

Face alignment method based on residual hourglass network Download PDF

Info

Publication number
CN110569724B
CN110569724B CN201910716528.1A CN201910716528A CN110569724B CN 110569724 B CN110569724 B CN 110569724B CN 201910716528 A CN201910716528 A CN 201910716528A CN 110569724 B CN110569724 B CN 110569724B
Authority
CN
China
Prior art keywords
network
residual
hourglass
thermodynamic diagram
hourglass network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910716528.1A
Other languages
Chinese (zh)
Other versions
CN110569724A (en
Inventor
邵雄凯
阳邹
高榕
刘建舟
王春枝
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hubei University of Technology
Original Assignee
Hubei University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hubei University of Technology filed Critical Hubei University of Technology
Priority to CN201910716528.1A priority Critical patent/CN110569724B/en
Publication of CN110569724A publication Critical patent/CN110569724A/en
Application granted granted Critical
Publication of CN110569724B publication Critical patent/CN110569724B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Data Mining & Analysis (AREA)
  • Human Computer Interaction (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a face alignment method based on a residual error hourglass network, which comprises the steps of firstly, additionally arranging convolution branches on the basis of a residual error module of a basic unit of the hourglass network to increase the receptive field of the network so as to better extract characteristic information under different scales, and simultaneously keeping high-resolution information; and then, by combining the characteristics of the hourglass network, the core size of the newly added convolution branches is adjusted along with the number of layers of the hourglass network, so that the relationship between the feature map resolution and the receptive field is well balanced, the network can extract more detailed information, and the structural information from local to whole is also kept. Finally, the hourglass network is stacked and is assisted by an intermediate supervision mechanism, so that normal updating of low-layer parameters is guaranteed, and the network is allowed to reevaluate the initial estimation and characteristics of the whole image. According to the invention, by stacking the new residual hourglass network, more effective information is extracted, the capability of extracting local detail information by the network is increased, and the accuracy of detecting key points of the face is improved.

Description

Face alignment method based on residual hourglass network
Technical Field
The invention belongs to the technical field of computer vision, and relates to a face alignment method based on a residual hourglass network, in particular to a face alignment method based on a novel residual hourglass network in the field of face recognition of digital images.
Background
The face alignment is also called face key point detection and is used for positioning key points such as eyes, a nose, a mouth, a face contour and the like of a face, the face alignment can help to provide accurate face shape information with specific semantics, and the face alignment plays a vital role in the fields of face recognition, face expression analysis, gender age identification, three-dimensional face modeling and the like. Due to the natural scene or the existence of facial expression, exaggerated head posture, difference of illumination conditions and partial shielding under an unconstrained condition, the problem of face alignment still faces huge challenges, so that an efficient and accurate face alignment algorithm can better meet the requirements of people.
In recent years, with the wide application of the deep learning framework in the field of face alignment, human research on face problems has been rapidly progressing. The advantage of deep learning is that the feature extraction is particularly prominent, and the deep network structure of deep learning can extract data features layer by layer, so that the extracted features are more obvious and easier to classify.
At the earliest, deep learning is introduced into a face alignment algorithm DCNN (deep Convolutional network) [ document 1], and by designing a cascaded Convolutional neural network with three levels, the problem of local optimization caused by initial improper is solved, and more accurate key point detection is obtained by means of strong feature extraction capability of CNN. Compared with the early face alignment algorithm based on the optimization (ASM [ document 2], AAM [ document 3] - [ document 5], CLM [ document 6] - [ document 7]), the aim of face alignment is achieved by optimizing an error equation, the nonlinear optimization problem is relatively complex to solve, the solving cost is increased due to overhigh middle dimension, the face alignment algorithm [ document 8] - [ document 14] based on the cascade shape regression continuously approaches to a standard shape by gradually estimating the shape increment according to the initialized shape, the requirement on the initialized shape is higher, the extraction of the features is more complicated, and the deep learning method is obviously simpler and more efficient.
With the recognition and use of the convolutional neural network, an hourglass network [ document 15] has been proposed to replace the CNN extraction features, and for the correlation task of face alignment, the parts of the face do not have the best recognition accuracy on the same feature map. For example, eyes may be easily recognized on the feature map of the layer 3, and mouths are more easily recognized on the layer 5, so that compared with a traditional convolutional neural network, only the last layer of convolution is used as a target feature, the hourglass network combines the features of the network at the lower level and the higher level by using a unique structure from bottom to top, so that the finally obtained feature is more effective, and therefore, the maximum utilization of image feature information is achieved. The novel residual error-based hourglass network increases convolution branches on the basis of the hourglass network, correspondingly adjusts the scale of a convolution kernel along with the number of layers of the hourglass network, increases the receptive field, balances the relationship between the feature map resolution and the receptive field, enables the network to extract more detailed information, maintains the structural information from local parts to the whole part, and increases the capability of the network to extract more effective characteristics.
[ document 1] Sun Y, Wang X, Tang X. deep conditional Network case for Facial Point Detection [ C ]// Computer Vision and Pattern recognition. IEEE,2013: 3476-.
[ document 2] dyes T F, Taylor C J, Cooper D H, Graham J. active shape models-the guiding and the application. computer vision and image understating, 1995,61(1):38-59.
[ document 3] Sauer P, Cootes T F, Taylor C J. Accurate regression processes for active application models// Proceedings of the British Machine Vision conference. Dunde, Scotland,2011: 681-.
[ document 3] Cootes T F, Edwards G J, Taylor C J.active application models.IEEE Transactions on Pattern Analysis and Machine understanding, 2001,23(6):581-585.
[ document 5] Asthana A, Zafeiriou S, Cheng S, cationic M.Robust distributed reactive map matching with constrained local models// IEEE Conference on Computer Vision and Pattern recognition. Portland, USA,2013: 3444-.
[ document 6] cristiniace D, Cootes T.feature detection and tracking with constrained local models// Proceedings of the British Machine Vision conference. Edinburgh, UK,2006: 929-.
[ document 7] Asthana A, Zafeiriou S, CHENG Shi-yang, cationic M.Inclusion Face Alignment in the wild.// IEEE Conference on Computer Vision and Pattern recognition. Columbus, USA 2014: 1859-.
[ document 8] Xiong Xue-han, Torre F D L.Supervised device method and its applications to face alignment// IEEE Conference on Computer Vision and Pattern registration. Portland, USA,2013: 532-.
[ document 9] Cao Xu-dong, Wei Yi-chen, Wen Fang, Sun Jian.face alignment by application program regression. International Journal of Computer Vision,2014,107(2): 177-.
[ document 10] Burgos-Artizzu X P, Perona P, Dollar P.Robust face and evaluation unit oclusion// IEEE International Conference on Computer Vison.Sydney, Australia,2013: 1513-.
[ document 11] Ren Shao-q, Cao Xu-dong, Wei Yi-chen, Sun Jian.face alignment at 3000fps via regressing local binding facilities// IEEE Conference on Computer Vision and Pattern recognition. Columbus, USA 2014: 1685-.
[ document 12] Dollar P, Welinder P, Perona P.Cascaded position regression.// IEEE Conference on Computer Vision and Pattern recognition. san Francisco, USA,2010: 1078-.
[ document 13] Tzimiopoulos G, cationic M.Gauss-newton defined part models for face alignment in-the-wire.// IEEE Conference on Computer Vision and Pattern recognition. Columbus, USA,2014: 1851-.
[ document 14] Smith B M, Brandt J, Lin Z, Zhang L. Nonparametric context modifying of local impedance for position-and expression-robust surface localization.// IEEE Conference on Computer Vision and Pattern recognition. Columbus, USA,2014: 1741-.
Document 15A. Newell, K.Yang, and J.Deng.Stacked hourglass networks for human position estimation. in European Conference on Computer Vision, pages 483-499 Springer,2016.1,2,3
Disclosure of Invention
The invention aims to provide a novel residual hourglass network-based face alignment method, which increases the capability of extracting effective features of a network and improves the accuracy of face alignment.
The technical scheme adopted by the invention is as follows: a face alignment method based on a residual hourglass network is characterized by comprising the following steps:
step 1: constructing a novel residual hourglass network;
the novel residual error hourglass network comprises a novel residual error module and a residual error module; the novel residual error module is characterized in that a convolution branch is newly added on the basis of the residual error module, the kernel scale of the newly added convolution branch is k, and k changes along with the current layer number hg _ level of the hourglass network; the output of the residual error module is h (x), the output of the novel residual error module is h' (x), and then:
h(x)=f(x)+x;
h′(x)=f(x)+gk(x)+x;
k=hg_level*2+1;
Figure BDA0002155616930000041
where x is the input to the residual block, f (x) is the output of x via a triple convolution operation, gk(x) The output of the novel residual error module is obtained under the condition that the x is the k in the novel convolution branch kernel scale, hg _ level is the current layer of the hourglass network, hg _ levels is the total number of layers of the hourglass network, and H (x) is the final output of the hourglass network;
step 2: obtaining an estimated face key point thermodynamic diagram (H) from an input picture through a stacked novel residual hourglass network1,H2...,HNIn which HiObtained by representing ith novel residual hourglass networkEstimating a face key point thermodynamic diagram, wherein i is more than or equal to 1 and less than or equal to N, and N is the number of stacked novel residual hourglass networks;
and step 3: generating a real face key point thermodynamic diagram by combining a two-dimensional Gaussian function with real face key points of an input picture
Figure BDA0002155616930000042
And 4, step 4: predicted human face key point thermodynamic diagram H of each stage of novel residual hourglass networkiAnd real face key point thermodynamic diagram
Figure BDA0002155616930000043
Pass to obtain LiThen the loss { L ] obtained from the whole network stage1,L2...,LNTaking an average value to obtain a final L;
and 5: training a network to obtain a training model, carrying out the training model on an input picture to obtain a predicted human face key point thermodynamic diagram H, and converting the thermodynamic diagram H into predicted human face key point coordinates P;
step 6: and drawing the key points of the human face on an original drawing.
The invention discloses a novel residual hourglass network-based face alignment method, which is a relatively simple face alignment method with good robustness. The description of the network on the whole and local associated information of the key points is enhanced by using a novel residual hourglass network model, the completeness of effective feature extraction is enhanced, the problem of gradient disappearance (variation) along with the deepening of the network depth can be effectively avoided by combining a stacked novel residual hourglass network of an intermediate supervision mechanism, the features can be processed in the context of the network while the normal updating of bottom layer parameters is ensured, the features are allowed to be reevaluated, the fault tolerance rate of the network is enhanced, and the algorithm effect is improved.
Drawings
FIG. 1: an overall framework diagram of an embodiment of the invention;
FIG. 2: a network flow diagram of an embodiment of the invention;
FIG. 3: the novel residual error hourglass network model schematic diagram of the embodiment of the invention;
FIG. 4: in the embodiment of the invention, an accumulated error distribution map (CED) of a residual hourglass (NRHG) on a 300W data set is formed by an hourglass network (HG) and the Novel Residual Hourglass (NRHG);
FIG. 5: in the embodiment of the invention, an accumulated error distribution map (CED) of an hourglass network (HG) and a Novel Residual Hourglass (NRHG) on an IBUG data set is provided;
FIG. 6: in the embodiment of the invention, an accumulated error distribution map (CED) of a residual hourglass (NRHG) on a COFW data set is formed by an hourglass network (HG) and the Novel Residual Hourglass (NRHG);
Detailed Description
In order to facilitate understanding and implementation of the present invention for those of ordinary skill in the art, the present invention is further described in detail with reference to the accompanying drawings and examples, which are described herein for the purpose of illustration and explanation only and are not intended to limit the present invention.
Referring to fig. 1 and fig. 2, the novel residual hourglass network human face alignment method provided by the invention comprises the following steps:
step 1: constructing a novel residual hourglass network;
fig. 3 shows a schematic diagram of a novel residual hourglass network structure, the basic constituent units of the novel residual hourglass network are a novel residual module and a residual module, the novel residual module is formed by adding a convolution branch on the basis of the residual module, the size of the kernel of the newly added convolution branch is k, k changes along with the number of layers hg _ level of the hourglass network, the output of the residual module is h (x), the output of the self-adaptive residual module is h' (x), and then:
h(x)=f(x)+x
h′(x)=f(x)+gk(x)+x
k=hg_level*2+1
Figure BDA0002155616930000061
where x is the input to the residual blockF (x) is the output of x through a three-layer convolution operation, gk(x) For the output obtained by x in the novel residual error module under the condition that the scale of the novel convolution branch kernel is k, hg _ level is the layer where the hourglass network is located currently, hg _ levels is the total number of layers (designed to be 4) of the hourglass network, and H (x) is the final output of the hourglass network.
Step 2: obtaining an estimated face key point thermodynamic diagram (H) from an input picture through a stacked novel residual hourglass network1,H2...,HNIn which HiRepresenting an estimated human face key point thermodynamic diagram obtained by the ith novel residual hourglass network, wherein i is more than or equal to 1 and less than or equal to N, and N is the number of stacked novel residual hourglass networks;
and step 3: generating a real face key point thermodynamic diagram by combining a two-dimensional Gaussian function with real face key points of an input picture
Figure BDA0002155616930000062
Figure BDA0002155616930000063
hm(w,h,i)0<i<M=f(xi,yi)0<xi<w,0<yi<h
Where f (x, y) is a two-dimensional Gaussian function and ksize is the size of the Gaussian function, corresponding to amplitude, sigma2Is the variance of a Gaussian function, (x)0,y0) Is the coordinate of the central point, here the coordinate of the key point of the real face, hm (w, h, i)0<i<MAnd w and h are the width and the height of the thermodynamic diagram respectively, and M is the total number of the key points. Finally generated real face key point thermodynamic diagram
Figure BDA0002155616930000064
That is, hm (w, h, i) generated by all the key points0<i<MIn combination with (1)
And 4, step 4: the predictive face key point thermodynamic diagram H of each stage of the networki(i.e., output thermodynamic diagram of ith hourglass network) and real face key point thermodynamic diagrams
Figure BDA0002155616930000065
Pass to obtain LiThen the loss { L ] obtained from the whole network stage1,L2...,LNTaking an average value to obtain a final L;
Figure BDA0002155616930000071
Figure BDA0002155616930000072
in the formula hjAnd
Figure BDA0002155616930000073
respectively the estimated thermodynamic diagram and the real thermodynamic diagram of the j-th personal face key point of the current stage, wherein M is the total number of the key points of the face, and n is the total number of the network stages (namely the number of the stacked novel residual hourglass networks)
And 5: training a network to obtain a training model, carrying out the training model on an input picture to obtain a predicted human face key point thermodynamic diagram H, and converting the thermodynamic diagram H into predicted human face key point coordinates P;
the predicted face key point thermodynamic diagram H is composed of all key point thermodynamic diagrams hm (w, H, i)0<i<MThe deeper the color in the thermodynamic diagram, the larger the two-dimensional Gaussian function value corresponding to the position with the larger color, namely, the closer to the real key point, and the thermodynamic diagram hm (w, h, i) is formed0<i<MAnd converting the value into a one-bit vector, taking the maximum value, and calculating the position coordinate of the original thermodynamic diagram pixel where the value is located, namely the predicted face key point coordinate P.
Step 6: and drawing the key points of the human face on an original drawing.
The invention also provides a performance evaluation method of the face alignment method based on the residual hourglass network, which is used for evaluating the predicted face key point coordinate P and the real face key point coordinate
Figure BDA0002155616930000074
Carrying out error comparison;
the normalized Mean error nme (normalized Mean error) is generally used to measure the quality of the face alignment algorithm, and is defined as follows:
Figure BDA0002155616930000075
wherein M is the number of key points of the face, piAnd
Figure BDA0002155616930000076
i is the predicted coordinate and the real coordinate of the key point of the individual face, d is the normalization factor, and the embodiment adopts two normalization modes, one is Inter-pupil normalization, and the other is Inter-ocular normalization.
Compared with the face alignment method (HG) using the traditional hourglass network (network structure and training parameters are the same, and only the novel residual hourglass network is replaced by the common hourglass network), the face alignment accuracy is improved by the algorithm. The data sets used in the experiments were 300W, IBUG, COFW. 3148 images in the 300W data set training set, and 554 images in the test set; the IBUG data set training set is the same as the IBUG data set training set 300W, and the test set comprises 135 images; the COFW data set training set has 1345 image test sets with 507 images.
TABLE 1 comparison of the results of the two different normalization methods on 300W, IBUG and COFW data sets for NRHG and HG of this example (% omitted)
Figure BDA0002155616930000081
The data in table 1 are Normalized Mean Error (NME) under different test sets, and from the data comparison it can be seen that the new residual hourglass Network (NRHG) compares to the conventional hourglass network (HG) for both 300W and IBUG and COFW data sets, the average error obtained in two different normalization modes is reduced to different degrees, the reduction range on the difficult data set IBUG is large, this shows that the novel residual hourglass Network (NRHG) proposed in this embodiment is more effective in feature extraction of human face images than the conventional hourglass network (HG), can be more helpful for locating key points of human faces, therefore, the accuracy of the algorithm for face alignment is improved to a certain degree, and meanwhile, the NRHG provided by the embodiment has advantages over HG in positioning of key points of the face in complex scenes such as care and shielding.
Referring to fig. 4, in the embodiment of the present invention, a cumulative error distribution map (CED) of a common hourglass network (HG) and a Novel Residual Hourglass (NRHG) on a 300W data set is provided;
referring to fig. 5, in the embodiment of the present invention, a cumulative error distribution map (CED) of a common hourglass network (HG) and a Novel Residual Hourglass (NRHG) on an IBUG data set is provided;
referring to fig. 6, in the embodiment of the present invention, a cumulative error distribution map (CED) of the common hourglass network (HG) and the Novel Residual Hourglass (NRHG) on the COFW data set is provided;
fig. 4, fig. 5, and fig. 6 visually reflect that the face alignment effect of the Novel Residual Hourglass (NRHG) proposed by the present invention is improved to a certain extent in two different normalization manners of three different data sets compared with the conventional hourglass network (HG).
Table 2 comparison of experimental results of this example algorithm (NRHG) and other face alignment algorithms on the COFW dataset (% omitted)
Figure BDA0002155616930000091
Table 3 comparison of experimental results of this example algorithm (NRHG) and other face alignment algorithms on 300W dataset (% omitted)
Figure BDA0002155616930000092
As is clear from table 2, the experimental results of the novel residual hourglass network algorithm (NRHG) proposed in this embodiment are better than those of the face alignment algorithms, except that the Normalized Mean Error (NME) is smaller than those of the face alignment algorithms, fr (feature rate) is also significantly reduced, and the data results show the superiority of the algorithm proposed in this embodiment in the COFW data set.
Table 3 shows the comparison of the experimental results of the novel residual hourglass network face alignment algorithm (NRHG) proposed in this embodiment and other face alignment algorithms under three indexes (public subset, challenge subset, full set) on a 300W data set. The data in the table shows that the algorithm provided by the embodiment is superior to the face alignment algorithms, especially the improvement range of the experimental effect is larger compared with the traditional cascade regression RCPR, ESR, LBF and other algorithms, and it can be reflected to a certain extent that the algorithm provided by the embodiment has a significant advantage in face alignment compared with the traditional cascade regression algorithm, although the experimental result of the algorithm provided by the embodiment is not improved to a different extent compared with the face alignment algorithms, some problems still exist in the challenge subset, the novel residual hourglass network face alignment algorithm provided by the embodiment focuses on processing the face image features, although the image characterization has a certain help in the complex situations of caring and shielding, the algorithm for solving the problems of pose, illumination, shielding and the like in comparison RAR, TSR and the like also has a large improvement space, this is also the place that needs to be learned and improved later, but overall, the algorithm of the embodiment has certain advantages.
It should be understood that parts of the specification not set forth in detail are well within the prior art. The present invention may be replaced or modified by one of ordinary skill in the art without departing from the scope of the present invention as defined by the appended claims.

Claims (4)

1. A face alignment method based on a residual hourglass network is characterized by comprising the following steps:
step 1: constructing a novel residual hourglass network;
the novel residual error hourglass network comprises a novel residual error module and a residual error module; the novel residual error module is characterized in that a convolution branch is newly added on the basis of the residual error module, the kernel scale of the newly added convolution branch is k, and k changes along with the current layer number hg _ level of the hourglass network; the output of the residual error module is h (x), the output of the novel residual error module is h' (x), and then:
h(x)=f(x)+x;
h′(x)=f(x)+gk(x)+x;
k=hg_level*2+1;
Figure FDA0002915799690000011
where x is the input to the residual block, f (x) is the output of x via a triple convolution operation, gk(x) The output of the novel residual error module is obtained under the condition that the x is the k in the novel convolution branch kernel scale, hg _ level is the current layer of the hourglass network, hg _ levels is the total number of layers of the hourglass network, and H (x) is the final output of the hourglass network;
step 2: obtaining an estimated face key point thermodynamic diagram (H) from an input picture through a stacked novel residual hourglass network1,H2...,HNIn which HiRepresenting an estimated human face key point thermodynamic diagram obtained by the ith novel residual hourglass network, wherein i is more than or equal to 1 and less than or equal to N, and N is the number of stacked novel residual hourglass networks;
and step 3: generating a real face key point thermodynamic diagram by combining a two-dimensional Gaussian function with real face key points of an input picture
Figure FDA0002915799690000012
And 4, step 4: predicted human face key point thermodynamic diagram H of each stage of novel residual hourglass networkiAnd real face key point thermodynamic diagram
Figure FDA0002915799690000013
Pass to obtain LiThen the loss { L ] obtained from the whole network stage1,L2...,LNTaking an average value to obtain a final L;
and 5: training a network to obtain a training model, carrying out the training model on an input picture to obtain a predicted human face key point thermodynamic diagram H, and converting the thermodynamic diagram H into predicted human face key point coordinates P;
step 6: and drawing the key points of the human face on an original drawing.
2. The residual hourglass network-based face alignment method according to claim 1, wherein in step 3, a calculation formula for generating a face key point thermodynamic diagram by using a two-dimensional Gaussian function is as follows:
Figure FDA0002915799690000021
hm(w,h,i)0<i<M=f(xi,yi)0<xi<w,0<yi<h
where f (x, y) is a two-dimensional Gaussian function, ksize is the size of the Gaussian function, and sigma2Is the variance of a Gaussian function, (x)0,y0) Is the coordinate of the central point, here the coordinate of the key point of the real face, hm (w, h, i)0<i<MThe thermodynamic diagram of the ith key point is shown, w and h are respectively the width and the height of the thermodynamic diagram, and M is the total number of the key points; finally generated real face key point thermodynamic diagram
Figure FDA0002915799690000025
That is, hm (w, h, i) generated by all the key points0<i<MCombinations of (a) and (b).
3. The residual hourglass network-based face alignment method according to claim 1, wherein in step 4, the loss function L of each stageiAnd the final loss function L is expressed as:
Figure FDA0002915799690000022
Figure FDA0002915799690000023
in the formula, hjAnd
Figure FDA0002915799690000024
respectively an estimated thermodynamic diagram and a real thermodynamic diagram of the j-th personal face key point in the current stage, wherein M is the total number of the personal face key points, and n is the total number of the network stages, namely the number of the stacked novel residual hourglass networks.
4. The residual hourglass network-based face alignment method according to claim 2, wherein the specific implementation of step 5 comprises the following steps:
step 5.1: will thermodynamic diagram hm (w, h, i)0<i<MConverting into a bit vector, and taking the maximum value;
step 5.2: calculating thermodynamic diagram hm (w, h, i)0<i<MAnd the position coordinate of the thermodynamic diagram pixel where the medium maximum value is located is the predicted face key point coordinate P.
CN201910716528.1A 2019-08-05 2019-08-05 Face alignment method based on residual hourglass network Active CN110569724B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910716528.1A CN110569724B (en) 2019-08-05 2019-08-05 Face alignment method based on residual hourglass network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910716528.1A CN110569724B (en) 2019-08-05 2019-08-05 Face alignment method based on residual hourglass network

Publications (2)

Publication Number Publication Date
CN110569724A CN110569724A (en) 2019-12-13
CN110569724B true CN110569724B (en) 2021-06-04

Family

ID=68774545

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910716528.1A Active CN110569724B (en) 2019-08-05 2019-08-05 Face alignment method based on residual hourglass network

Country Status (1)

Country Link
CN (1) CN110569724B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111402228B (en) * 2020-03-13 2021-05-07 腾讯科技(深圳)有限公司 Image detection method, device and computer readable storage medium
CN111523484B (en) * 2020-04-24 2021-08-27 北京嘀嘀无限科技发展有限公司 Face key point detection method and device, electronic equipment and storage medium
CN112417991B (en) * 2020-11-02 2022-04-29 武汉大学 Double-attention face alignment method based on hourglass capsule network
CN112699847B (en) * 2021-01-15 2021-12-07 苏州大学 Face characteristic point detection method based on deep learning
CN113610115B (en) * 2021-07-14 2024-04-12 广州敏视数码科技有限公司 Efficient face alignment method based on gray level image

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101499128A (en) * 2008-01-30 2009-08-05 中国科学院自动化研究所 Three-dimensional human face action detecting and tracing method based on video stream
CN106951875A (en) * 2017-03-24 2017-07-14 深圳市唯特视科技有限公司 The method that a kind of human body attitude estimation and face based on binary system convolution are alignd
CN108090470A (en) * 2018-01-10 2018-05-29 浙江大华技术股份有限公司 A kind of face alignment method and device
CN108898556A (en) * 2018-05-24 2018-11-27 麒麟合盛网络技术股份有限公司 A kind of image processing method and device of three-dimensional face
CN109033946A (en) * 2018-06-08 2018-12-18 东南大学 Merge the estimation method of human posture of directional diagram
CN109241910A (en) * 2018-09-07 2019-01-18 高新兴科技集团股份有限公司 A kind of face key independent positioning method returned based on the cascade of depth multiple features fusion
CN109299669A (en) * 2018-08-30 2019-02-01 清华大学 Video human face critical point detection method and device based on double intelligent bodies
CN109299659A (en) * 2018-08-21 2019-02-01 中国农业大学 A kind of human posture recognition method and system based on RGB camera and deep learning
CN109657595A (en) * 2018-12-12 2019-04-19 中山大学 Based on the key feature Region Matching face identification method for stacking hourglass network

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4389956B2 (en) * 2007-04-04 2009-12-24 ソニー株式会社 Face recognition device, face recognition method, and computer program
WO2018165620A1 (en) * 2017-03-09 2018-09-13 The Board Of Trustees Of The Leland Stanford Junior University Systems and methods for clinical image classification
CN108764133B (en) * 2018-05-25 2020-10-20 北京旷视科技有限公司 Image recognition method, device and system

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101499128A (en) * 2008-01-30 2009-08-05 中国科学院自动化研究所 Three-dimensional human face action detecting and tracing method based on video stream
CN106951875A (en) * 2017-03-24 2017-07-14 深圳市唯特视科技有限公司 The method that a kind of human body attitude estimation and face based on binary system convolution are alignd
CN108090470A (en) * 2018-01-10 2018-05-29 浙江大华技术股份有限公司 A kind of face alignment method and device
CN108898556A (en) * 2018-05-24 2018-11-27 麒麟合盛网络技术股份有限公司 A kind of image processing method and device of three-dimensional face
CN109033946A (en) * 2018-06-08 2018-12-18 东南大学 Merge the estimation method of human posture of directional diagram
CN109299659A (en) * 2018-08-21 2019-02-01 中国农业大学 A kind of human posture recognition method and system based on RGB camera and deep learning
CN109299669A (en) * 2018-08-30 2019-02-01 清华大学 Video human face critical point detection method and device based on double intelligent bodies
CN109241910A (en) * 2018-09-07 2019-01-18 高新兴科技集团股份有限公司 A kind of face key independent positioning method returned based on the cascade of depth multiple features fusion
CN109657595A (en) * 2018-12-12 2019-04-19 中山大学 Based on the key feature Region Matching face identification method for stacking hourglass network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
An Approach to Face Detection and Alignment Using Hough Transformation with Convolution Neural Network;Oshin Misra 等;《ICACCA》;20161121;第1-5页 *
基于堆叠深度卷积沙漏网络的步态识别;王浩 等;《计算机工程与应用》;20180919;第127-133页 *

Also Published As

Publication number Publication date
CN110569724A (en) 2019-12-13

Similar Documents

Publication Publication Date Title
CN110569724B (en) Face alignment method based on residual hourglass network
CN109472198B (en) Gesture robust video smiling face recognition method
WO2020186886A1 (en) Method and device for generating face recognition model
CN110348330B (en) Face pose virtual view generation method based on VAE-ACGAN
WO2020108362A1 (en) Body posture detection method, apparatus and device, and storage medium
CN101964064B (en) Human face comparison method
Zhang et al. Content-adaptive sketch portrait generation by decompositional representation learning
CN109949255A (en) Image rebuilding method and equipment
WO2016023264A1 (en) Fingerprint identification method and fingerprint identification device
CN108446672B (en) Face alignment method based on shape estimation of coarse face to fine face
Zhou et al. Pose-robust face recognition with Huffman-LBP enhanced by divide-and-rule strategy
CN107066969A (en) A kind of face identification method
CN110909778A (en) Image semantic feature matching method based on geometric consistency
CN110610138A (en) Facial emotion analysis method based on convolutional neural network
Hao et al. Finger vein recognition based on multi-task learning
CN115376159A (en) Cross-appearance pedestrian re-recognition method based on multi-mode information
Wan et al. Palmprint recognition system for mobile device based on circle loss
Jiang et al. Face recognition method based on sparse representation and feature fusion
CN110969101A (en) Face detection and tracking method based on HOG and feature descriptor
CN114743234A (en) Efficient face mask recognition method based on deep learning
Karungaru et al. Automatic human faces morphing using genetic algorithms based control points selection
Su et al. A multiattribute sparse coding approach for action recognition from a single unknown viewpoint
Deepa et al. Age estimation in facial images using histogram equalization
Kumar et al. Face hallucination using OLPP and kernel ridge regression
Hong et al. Effective Face Inpainting by Conditional Generative Adversarial Network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant