CN109671108B - Single multi-view face image attitude estimation method capable of rotating randomly in plane - Google Patents
Single multi-view face image attitude estimation method capable of rotating randomly in plane Download PDFInfo
- Publication number
- CN109671108B CN109671108B CN201811550656.5A CN201811550656A CN109671108B CN 109671108 B CN109671108 B CN 109671108B CN 201811550656 A CN201811550656 A CN 201811550656A CN 109671108 B CN109671108 B CN 109671108B
- Authority
- CN
- China
- Prior art keywords
- dimensional
- human face
- face image
- angle
- gamma
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
- 238000000034 method Methods 0.000 title claims abstract description 45
- 238000004364 calculation method Methods 0.000 claims abstract description 12
- 230000006870 function Effects 0.000 claims description 41
- 239000011159 matrix material Substances 0.000 claims description 25
- 238000013519 translation Methods 0.000 claims description 9
- 238000010845 search algorithm Methods 0.000 claims description 6
- 239000000126 substance Substances 0.000 claims description 6
- 230000004888 barrier function Effects 0.000 claims description 3
- 238000011156 evaluation Methods 0.000 claims description 3
- 230000009467 reduction Effects 0.000 claims description 3
- 210000001747 pupil Anatomy 0.000 claims description 2
- 230000001815 facial effect Effects 0.000 claims 4
- 230000036544 posture Effects 0.000 description 10
- 230000008569 process Effects 0.000 description 4
- 238000010586 diagram Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 230000003190 augmentative effect Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 210000000887 face Anatomy 0.000 description 1
- 210000003128 head Anatomy 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
- 238000012549 training Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/30—Determination of transform parameters for the alignment of images, i.e. image registration
- G06T7/33—Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
- G06T7/344—Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods involving models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
- G06T2207/30201—Face
Abstract
The invention discloses a single multi-view face image posture estimation method capable of rotating randomly in a plane, which comprises the following steps of 1, establishing a pre-estimated value set of a gamma angle according to a face image to be detected, wherein the gamma angle represents a deflection angle around a Z axis, 2, traversing the pre-estimated value set of the gamma angle, calculating α and β angles corresponding to the pre-estimated values of the gamma angle by adopting a calculation method of solving α and β angles meeting an objective function under the condition of the specified gamma angle, α represents the deflection angle around the X axis, β represents the deflection angle around the Y axis, and 3, forming an alternative set by values of the objective function corresponding to the pre-estimated values of the gamma angle, selecting a minimum objective function value from the alternative set, and taking the gamma angle pre-estimated value, α and β angles corresponding to the minimum objective function value as a face posture of the face image to be detected.
Description
Technical Field
The invention relates to the technical field of face recognition, in particular to a face gesture recognition model and a face gesture recognition method.
Background
In the face identification, face images at various angles need to be collected in advance, and then the face image to be detected is aligned with the face image collected in advance, however, in practical application, the face image to be detected does not always face the camera, and deflection occurs in a three-dimensional space in most cases, so that the face pose (the deflection angle of the face around each coordinate axis in the three-dimensional space) in the face image to be detected needs to be identified, and the face alignment can be performed.
In addition, the head rotation direction and the eye fixation position can be obtained through posture estimation, the basis of human-computer interaction and visual monitoring in a multi-view environment is provided, the human face posture is corrected through the posture estimation, and the accuracy of multi-view human face recognition and analysis can be improved.
The face image is recorded with face information in a two-dimensional space (XY plane), and therefore, it is relatively easy to estimate the face pose γ (the angle of deflection around the Z axis perpendicular to the XY plane) on the XY plane, and generally the angle between the line between the two eyes and the X axis (horizontal direction).
In addition, for a frontal face, the rotation angle γ in the XY plane can be calculated by the inclination angle of the line connecting the centers of both eyes, but for a face that is severely affected by perspective and is deflected laterally, even if there is no rotation in the plane, the line connecting the centers of both eyes also has an inclination angle, as shown in fig. 1, and therefore, the rotation angle γ cannot be simply adopted as a method of calculating the inclination angle of the line connecting the centers of both eyes.
The general human face three-dimensional sparse model is a parameterized 3D model, is originally used for the application of human face coding based on the model, and has a small number of matching points and triangular surfaces, so that the modeling of the general human face three-dimensional sparse model only needs little computing time, and is widely applied to video animation and transmission, and fig. 2 is a standard wire frame structure of the general human face three-dimensional sparse model. Multi-view face pose fast estimation method on single image suitable for arbitrary rotation in plane
Disclosure of Invention
Aiming at the defects of the prior art, the invention aims to provide a single multi-view face image posture estimation method capable of rotating freely in a plane, solves the technical problem that the target face posture is detected by relying on prior conditions in the prior art, can realize target face posture estimation only by a single image to be detected, and does not need to be trained or learned in advance.
In order to solve the technical problems, the invention provides the following technical scheme: a single multi-view face image posture estimation method capable of rotating randomly in a plane comprises the following steps:
step 1: establishing a pre-evaluation value set of a gamma angle according to a face image to be detected, wherein the gamma angle represents a deflection angle around a Z axis;
and 3, combining the values of the objective function corresponding to the gamma angle estimated values into an alternative set, selecting a minimum objective function value from the alternative set, and taking the gamma angle estimated values, α angles and β angles corresponding to the minimum objective function value as the human face posture estimation parameters of the human face image to be detected.
Preferably, in step 1, a search algorithm is used to establish a set of estimated values of γ angles, and the method includes the following steps:
step 101: detecting an included angle theta between a central line of two eyes and a horizontal line on a face image to be detected;
step 103: search is performed in search steps: γ ═ θ ± t; wherein, t is the current searching times,carrying out positive and negative searching by using a searching step length t in each searching; storing the obtained estimated value of the gamma angle into an estimated value set of the gamma angle every time the search is carried out;
step 104: and after the search is finished, establishing an estimated value set of the gamma angle.
Preferably, in step 2, α angles and β angles corresponding to the estimated values of the gamma angles are calculated in parallel according to the estimated values of the gamma angles.
Preferably, the method for calculating α and β angles under the specified gamma angle in step 2 comprises the following steps:
step 201: establishing general human face three-dimensional sparse modelThe general human face three-dimensional sparse modelThe human face three-dimensional sparse model is a general human face three-dimensional sparse model which does not deflect on an X, Y, Z axis;
step 202: establishing general human face three-dimensional sparse modelThree-dimensional coordinate matrix V of upper designated point3D:
wherein the content of the first and second substances,three-dimensional sparse model representing general human faceThe (i) th designated point of (c),
step 203: establishing general human face three-dimensional sparse model of human face image to be detectedTwo-dimensional coordinate matrix V of matching points corresponding to the upper designated point2D: the matching points do not include reference points;
wherein the content of the first and second substances,it represents the i-th matching point and,n is the number of face matching points on the face image to be detected;
step 204, taking the vector X as an independent variable (s, α), and initializing values of s, α and β, wherein s represents a scaling coefficient, α represents a deflection angle around an X axis, and β represents a deflection angle around a Y axis;
step 205: universal human face three-dimensional sparse model rotating around Z axis at specified gamma angle
206-rotating the Utility model around the X-axis at an angle of αFace three-dimensional sparse modelAnd rotating the general human face three-dimensional sparse model around the Y axis at an angle of βUniversal human face three-dimensional sparse model after rotation by scaling factor sZooming is carried out;
step 207: the zoomed general human face three-dimensional sparse modelOrthographic projection is carried out on an XY plane, and a two-dimensional projection model is obtained
Step 208: translation two-dimensional projection modelModeling two-dimensional projectionThe reference point on the face image is superposed with the reference point on the face image to be detected; the reference point is contained in the matching point;
step 209: from two-dimensional projection modelsEstablishing a target function by the sum of squares of the distances between the upper designated point and the corresponding point on the face image to be detected;
step 2010, updating the vector X (s, α) by using a search algorithm, and repeating the steps 206 to 208 each time the vector X is updated until the optimal solution of the objective function is searched;
step 2011: and storing the optimal solution of the objective function into the alternative set.
Preferably, the designated points include a subnasal point, a corner point of eyes, a nose tip point, and a mouth corner point, and the reference point is a subnasal point among the matching points.
Preferably, the objective function minF (X) is constructed using an internal penalty function method as follows:
minF(X)=min[f(X)+r/s];
wherein r is a barrier factor, r > 0; s is a scaling factor;
(X) representing a two-dimensional projection modelThe sum of squares of the distances between the upper designated point and the corresponding point on the face image to be detected,
P3Dthree-dimensional sparse model representing general human faceThe forward projection matrix of (a);
T2Drepresenting two-dimensional projection modelsThe translation matrix of (a) is, indicating a reference point on the face image to be detected,representing two-dimensional projection modelsThe above reference point.
Preferably, in step 2010, the optimal solution of the objective function minf (x) is calculated by using a modified Newton method, and the method comprises the following steps:
step 111: let the current iteration number be k, k being an integer greater than or equal to zero, and the current gradient vectorXkRepresenting a current vector;
step 113: constructing a Newton direction: calculating the current function F (X)k) Current direction derivative P ofkAccording to the following formula:
step 114: one-dimensional search is carried out, and the current iteration step length t is calculated by using a golden section methodkTo update the current vector XkThe updated vector isXk+1,Xk+1=Xk+tkPkAnd calculating F (X)k+1);
Step 115: let k equal k +1, rk+1=crkC is the obstacle factor reduction coefficient, c is 0.1, and step 116 is entered;
step 116: judge g | |kIf the | is less than or equal to the established limit, the given limit is set; if yes, go to step 117; if not, go back to step 113;
step 117: judgment of rkWhether the/s is less than or equal to true or not; if yes, stopping iteration and using current vector XkOutputting as an optimal solution; if not, go back to step 113.
Compared with the prior art, the invention has the following advantages:
1. the method directly carries out attitude estimation on the target face on a single image without knowing the camera parameters of the image, taking other attitude pictures of the target face as references and needing no prior training or learning, and is simple to realize.
2. The method separates the estimated values of the gamma angle and the α and β angles, solves the problems of complex independent variable, large calculated amount and sensitivity to initial iteration values of a target function caused by simultaneous solving of rotation angles in three directions in the prior art, calculates the α and β angles corresponding to each gamma angle estimated value in a parallel mode after estimating the gamma angle, and greatly improves the operation speed.
3. According to the invention, based on the included angle theta between the central line of the two eyes and the horizontal line, the optimal human face plane external deflection angles α and β in a certain range around the included angle theta are searched to obtain the in-plane rotation angle gamma of the multi-view human face, so that the in-plane rotation angle calculation errors caused by large parallax and human face deflection are avoided.
4. The invention utilizes the universal human face three-dimensional sparse model after rotation, scaling, projection and translationα which is overlapped with the face image to be detected, thereby the two-dimensional image can not be directly detected,β corner, converting to solve general human face three-dimensional sparse modelThe human face postures around the X axis and the Y axis on the two-dimensional image are detected (under the condition of specifying the gamma angle).
5. The model coincidence process and the solving process of the objective function have synchronicity: the vector X needs to be continuously updated in the process of solving the optimal solution of the objective functionkEvery time vector X is updatedkThen the updated s, α and β pairs of universal human face three-dimensional sparse models are usedAnd rotating and zooming are carried out, and then projection and coincidence are carried out, so that the solving process of the optimal solution can be visually represented, and the coincidence degree of the model and the face image to be detected can be visually represented.
6. To ensure that s, α, β is calculated under the condition that s > 0, an internal penalty function is used to construct the augmented objective function of f (X).
7. The target function of the invention is calculated based on the distance from point to point in a two-dimensional plane, and the calculation in a three-dimensional space is not needed, so that the dimension is reduced, and the calculation speed can be improved.
8. The designated points are infranasal points, binocular corner points, nasal cusp points and mouth corner points, the characteristics of the points are prominent and easy to position, and only 8 designated points are adopted, so that the calculation speed can be greatly improved.
9. In the method for estimating the deflection angle of the human face around the X, Y axis, the translation amount of the model is determined by a method of overlapping the subnasal points, so that independent variables in an objective function are reduced to s and α, the model is constrained to rotate only by taking the subnasal point as the center by increasing the alignment point of the nasal tip, the complexity of an algorithm and the dependency on an initial point are reduced, and the α estimation can be completed only by small iteration times.
Drawings
FIG. 1 is a schematic illustration of the line connecting the centers of the eyes;
FIG. 2 is a standard wireframe structure diagram of a general human face three-dimensional sparse model;
FIG. 3 is a flow chart for calculating α, β angles at a specified γ angle;
fig. 4 is a test effect diagram of the method for estimating the pose of a single multi-view face image arbitrarily rotated in a plane in the present embodiment.
Detailed Description
The present invention will be described in further detail with reference to the accompanying drawings and preferred embodiments.
A single multi-view face image posture estimation method capable of rotating randomly in a plane comprises the following steps:
step 1: establishing a pre-evaluation value set of a gamma angle according to a face image to be detected, wherein the gamma angle represents a deflection angle around a Z axis;
and 3, combining the values of the objective function corresponding to the gamma angle estimated values into an alternative set, selecting a minimum objective function value from the alternative set, and taking the gamma angle estimated values, α and β angles corresponding to the minimum objective function value as estimation parameters of the human face pose of the human face image to be detected.
In the present embodiment, in step 2, α angles and β angles corresponding to the estimated values of each γ angle are calculated in parallel according to the estimated values of each γ angle.
In this embodiment, the search algorithm is used to establish the estimated value set of the γ angle in step 1, and the following steps are performed:
step 101: detecting an included angle theta between a central line of two eyes and a horizontal line on a face image to be detected;
step 103: search is performed in search steps: γ ═ θ ± t; wherein, t is the current searching times,carrying out positive and negative searching by using a searching step length t in each searching; storing the obtained estimated value of the gamma angle into an estimated value set of the gamma angle every time the search is carried out;
step 104: and after the search is finished, establishing an estimated value set of the gamma angle.
In this embodiment, as shown in fig. 3, the method for calculating α and β angles under the condition of specifying the γ angle in step 2 includes the following steps:
step 201: establishing general human face three-dimensional sparse modelThe general human face three-dimensional sparse modelThe human face three-dimensional sparse model is a general human face three-dimensional sparse model which does not deflect on an X, Y, Z axis; the general face three-dimensional sparse model can adopt a code-3 and sparse face grid model and the like;
step 202: establishing general human face three-dimensional sparse modelThree-dimensional coordinate matrix V of upper designated point3D:
wherein the content of the first and second substances,three-dimensional sparse model representing general human faceTo (1) aThe number of the i designated points is,
step 203: establishing general human face three-dimensional sparse model of human face image to be detectedTwo-dimensional coordinate matrix V of matching points corresponding to the upper designated point2D: the matching points do not include reference points;
wherein the content of the first and second substances,it represents the i-th matching point and,n is the number of face matching points on the face image to be detected;
step 204, taking the vector X as an independent variable (s, α), and initializing values of s, α and β, wherein s represents a scaling coefficient, α represents a deflection angle around an X axis, and β represents a deflection angle around a Y axis;
step 205: universal human face three-dimensional sparse model rotating around Z axis at specified gamma angle
206, rotating the general human face three-dimensional sparse model around the X axis at an angle of αAnd rotating the general human face three-dimensional sparse model around the Y axis at an angle of βUniversal human face three-dimensional sparse model after rotation by scaling factor sZooming is carried out;
step 207: the zoomed general human face three-dimensional sparse modelOrthographic projection is carried out on an XY plane, and a two-dimensional projection model is obtained
Step 208: translation two-dimensional projection modelModeling two-dimensional projectionThe reference point on the face image is superposed with the reference point on the face image to be detected; the reference point is contained in the matching point;
step 209: from two-dimensional projection modelsEstablishing a target function by the sum of squares of the distances between the upper designated point and the corresponding point on the face image to be detected;
step 2010, updating the vector X (s, α) by using a search algorithm, and repeating the steps 206 to 208 each time the vector X is updated until the optimal solution of the objective function is searched;
step 2011: and storing the optimal solution of the objective function into the alternative set.
In this specific embodiment, the designated points include a subnasal point, a corner point of eyes, a nose tip point, and a mouth corner point, and the reference point is a subnasal point in the matching points.
In this embodiment, the objective function minF (X) is constructed using an interior penalty function method as follows:
minF(X)=min[f(X)+r/s];
wherein r is a barrier factor, r > 0; s is a scaling factor;
(X) representing a two-dimensional projection modelThe sum of squares of the distances between the upper designated point and the corresponding point on the face image to be detected,
P3Dthree-dimensional sparse model representing general human faceThe forward projection matrix of (a);
T2Drepresenting two-dimensional projection modelsThe translation matrix of (a) is, indicating a reference point on the face image to be detected,representing two-dimensional projection modelsThe above reference point.
Of course, min f (X) can also be directly used as the objective function, but convergence is slow, and a negative solution of s occurs.
In this embodiment, s is an initial value s in step 2040For pupil distance and two-dimensional projection model on human face image to be detectedRatio of interpupillary distance above, initial value α of α, β0、β0Are all 0 degrees; initial value r of r0=100。
In this embodiment, in step 2010, a modified Newton method is used to calculate an optimal solution of an objective function minf (x), and the method includes the following steps:
step 111: let the current iteration number be k, k being an integer greater than or equal to zero, and the current gradient vectorXkRepresenting a current vector;
step 113: constructing a Newton direction: calculating the current function F (X)k) Current direction derivative P ofkAccording to the following formula:
step (ii) of114: one-dimensional search is carried out, and the current iteration step length t is calculated by using a golden section methodkTo update the current vector XkThe updated vector is Xk+1,Xk+1=Xk+tkPkAnd calculating F (X)k+1);
Step 115: let k equal k +1, rk+1=crkC is the obstacle factor reduction coefficient, c is 0.1, and step 116 is entered;
step 116: judge g | |kIf the | is less than or equal to the established limit, the given limit is set; if yes, go to step 117; if not, go back to step 113;
step 117: judgment of rkWhether the/s is less than or equal to true or not; if yes, stopping iteration and using current vector XkOutputting as an optimal solution; if not, go back to step 113.
In the present embodiment, f (x) is expanded and simplified as:
wherein, c0To c12Are all constants and are calculated as follows:
Experiments on virtual 3D faces and real face images show that the average estimation error of the invention for α angles is 6.5 degrees, the average estimation error of β angles is 4.6 degrees, the calculation time of the deflection estimation of the face around the X, Y axis of a specified in-plane rotation angle is less than 1ms, a parallel calculation mode of an independent thread is created for the deflection estimation calculation of each axis around the X, Y axis, the average calculation time of the whole algorithm is less than 5ms, and partial test effects are shown in figure 4.
Claims (9)
1. A single multi-view face image posture estimation method capable of rotating randomly in a plane is characterized in that: the method comprises the following steps:
step 1: establishing a pre-evaluation value set of a gamma angle according to a face image to be detected, wherein the gamma angle represents a deflection angle around a Z axis;
step 2, traversing the estimated value set of the gamma angles, and calculating α and β angles corresponding to the estimated values of the gamma angles by adopting a calculation method of solving α and β angles meeting an objective function under the condition of the specified gamma angles, wherein α represents a deflection angle around an X axis, and β represents a deflection angle around a Y axis;
step 3, combining the values of the objective functions corresponding to the gamma angle pre-estimated values into an alternative set, selecting a minimum objective function value from the alternative set, and taking the gamma angle pre-estimated values, α and β angles corresponding to the minimum objective function value as the human face posture estimation parameters of the human face image to be detected;
in the step 1, a search algorithm is adopted to establish an estimated value set of a gamma angle, and the method comprises the following steps:
step 101: detecting an included angle theta between a central line of two eyes and a horizontal line on a face image to be detected;
step 103: search is performed in search steps: γ ═ θ ± t; wherein, t is the current searching times,carrying out positive and negative searching by using a searching step length t in each searching; storing the obtained estimated value of the gamma angle into an estimated value set of the gamma angle every time the search is carried out;
step 104: and after the search is finished, establishing an estimated value set of the gamma angle.
2. The method for estimating pose of single multi-view face image with arbitrary rotation in plane according to claim 1, wherein the method for calculating α and β angles under the designated γ angle in step 2 comprises the following steps:
step 201: establishing general human face three-dimensional sparse modelThe general human face three-dimensional sparse modelThe human face three-dimensional sparse model is a general human face three-dimensional sparse model which does not deflect on an X, Y, Z axis;
step 202: establishing general human face three-dimensional sparse modelThree-dimensional coordinate matrix V of upper designated point3D:
wherein the content of the first and second substances,three-dimensional sparse model representing general human faceThe (i) th designated point of (c),
step 203: establishing three-dimensional sparse model of face to be detected and general face on face imageTwo-dimensional coordinate matrix V of matching points corresponding to the upper designated point2D: the matching points do not include reference points;
wherein the content of the first and second substances,it represents the i-th matching point and,n is the number of face matching points on the face image to be detected;
step 204, taking the vector X as an independent variable (s, α), and initializing values of s, α and β, wherein s represents a scaling coefficient, α represents a deflection angle around an X axis, and β represents a deflection angle around a Y axis;
step 205: universal human face three-dimensional sparse model rotating around Z axis at specified gamma angle
206, rotating the general human face three-dimensional sparse model around the X axis at an angle of αAnd rotating the general human face three-dimensional sparse model around the Y axis at an angle of βUniversal human face three-dimensional sparse model after rotation by scaling factor sZooming is carried out;
step 207: the zoomed general human face three-dimensional sparse modelOrthographic projection is carried out on an XY plane, and a two-dimensional projection model is obtained
Step 208: translation two-dimensional projection modelModeling two-dimensional projectionThe reference point on the face image is superposed with the reference point on the face image to be detected; the reference point is contained in the matching point;
step 209: from two-dimensional projection modelsEstablishing a target function by the sum of squares of the distances between the upper designated point and the corresponding point on the face image to be detected;
step 2010, updating the vector X (s, α) by using a search algorithm, and repeating the steps 206 to 208 each time the vector X is updated until the optimal solution of the objective function is searched;
step 2011: and storing the optimal solution of the objective function into the alternative set.
3. The method for estimating the pose of a single multi-view face image rotated arbitrarily in a plane according to claim 2, wherein: the designated points comprise subnasal points, binocular eye corner points, nasal cusp points and mouth corner points, and the reference points are the subnasal points.
4. The method for estimating the pose of a single multi-view face image rotated arbitrarily in a plane according to claim 2, wherein: the objective function min f (X) is as follows:
P3Dthree-dimensional sparse model representing general human faceThe forward projection matrix of (a);
5. The method for estimating the pose of a single multi-view face image rotated arbitrarily in a plane according to claim 2, wherein: the objective function min F (X) is constructed using the internal penalty function method as follows:
min F(X)=min[f(X)+r/s];
wherein r is a barrier factor, r > 0; s is a scaling factor;
(X) representing a two-dimensional projection modelThe sum of squares of the distances between the upper designated point and the corresponding point on the face image to be detected,
P3Dthree-dimensional sparse model representing general human faceThe forward projection matrix of (a);
6. The method for estimating the pose of a single multi-view facial image rotated arbitrarily in a plane according to claim 5, wherein: initial value s of s in step 2040For pupil distance and two-dimensional projection model on human face image to be detectedRatio of interpupillary distance above α, βInitial value α0、β0Are all 0 degrees; initial value r of r0=100。
7. The method for estimating the pose of a single multi-view facial image rotated arbitrarily in a plane according to claim 5, wherein: in step 2010, an optimal solution of an objective function minF (X) is calculated by adopting a modified Newton method, and the method comprises the following steps:
step 111: let the current iteration number be k, k being an integer greater than or equal to zero, and the current gradient vectorXkRepresenting a current vector;
step 113: constructing a Newton direction: calculating the current function F (X)k) Current direction derivative P ofkAccording to the following formula:
step 114: one-dimensional search is carried out, and the current iteration step length t is calculated by using a golden section methodkTo update the current vector XkThe updated vector is Xk+1,Xk+1=Xk+tkPkAnd calculating F (X)k+1);
Step 115: let k equal k +1, rk+1=crkC is the obstacle factor reduction coefficient, c is 0.1, and step 116 is entered;
step 116: judge g | |kWhether | is less than or equal toGiven a termination limit; if yes, go to step 117; if not, go back to step 113;
step 117: judgment of rkWhether the/s is less than or equal to true or not; if yes, stopping iteration and using current vector XkOutputting as an optimal solution; if not, go back to step 113.
8. The method for estimating the pose of a single multi-view facial image rotated arbitrarily in a plane according to claim 5, wherein: (x) expanded and simplified as:
wherein, c0To c12Are all constants and are calculated as follows:
9. The method for estimating the pose of a single multi-view facial image rotated arbitrarily in a plane according to claim 1, wherein in step 2, α angles and β angles corresponding to each gamma angle estimated value are calculated in parallel according to each gamma angle estimated value.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811550656.5A CN109671108B (en) | 2018-12-18 | 2018-12-18 | Single multi-view face image attitude estimation method capable of rotating randomly in plane |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811550656.5A CN109671108B (en) | 2018-12-18 | 2018-12-18 | Single multi-view face image attitude estimation method capable of rotating randomly in plane |
Publications (2)
Publication Number | Publication Date |
---|---|
CN109671108A CN109671108A (en) | 2019-04-23 |
CN109671108B true CN109671108B (en) | 2020-07-28 |
Family
ID=66143987
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201811550656.5A Expired - Fee Related CN109671108B (en) | 2018-12-18 | 2018-12-18 | Single multi-view face image attitude estimation method capable of rotating randomly in plane |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109671108B (en) |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6937745B2 (en) * | 2001-12-31 | 2005-08-30 | Microsoft Corporation | Machine vision system and method for estimating and tracking facial pose |
CN101561710B (en) * | 2009-05-19 | 2011-02-09 | 重庆大学 | Man-machine interaction method based on estimation of human face posture |
CN102054291A (en) * | 2009-11-04 | 2011-05-11 | 厦门市美亚柏科信息股份有限公司 | Method and device for reconstructing three-dimensional face based on single face image |
US9437011B2 (en) * | 2012-06-11 | 2016-09-06 | Samsung Electronics Co., Ltd. | Method and apparatus for estimating a pose of a head for a person |
CN104036546B (en) * | 2014-06-30 | 2017-01-11 | 清华大学 | Method for carrying out face three-dimensional reconstruction at any viewing angle on basis of self-adaptive deformable model |
CN105678241B (en) * | 2015-12-30 | 2019-02-26 | 四川川大智胜软件股份有限公司 | A kind of cascade two dimensional image face pose estimation |
-
2018
- 2018-12-18 CN CN201811550656.5A patent/CN109671108B/en not_active Expired - Fee Related
Also Published As
Publication number | Publication date |
---|---|
CN109671108A (en) | 2019-04-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109166149B (en) | Positioning and three-dimensional line frame structure reconstruction method and system integrating binocular camera and IMU | |
Ye et al. | Accurate 3d pose estimation from a single depth image | |
CN111325823A (en) | Method, device and equipment for acquiring face texture image and storage medium | |
US7027618B2 (en) | Head motion estimation from four feature points | |
JP6483832B2 (en) | Method and system for scanning an object using an RGB-D sensor | |
WO2021063271A1 (en) | Human body model reconstruction method and reconstruction system, and storage medium | |
Gu et al. | MedUCC: Medium-driven underwater camera calibration for refractive 3-D reconstruction | |
Su et al. | Cross-validated locally polynomial modeling for 2-D/3-D gaze tracking with head-worn devices | |
WO2019136588A1 (en) | Cloud computing-based calibration method, device, electronic device, and computer program product | |
CN113095274B (en) | Sight estimation method, system, device and storage medium | |
CN109671108B (en) | Single multi-view face image attitude estimation method capable of rotating randomly in plane | |
Wang et al. | Automatic non-rigid registration of 3d dynamic data for facial expression synthesis and transfer | |
CN111915739A (en) | Real-time three-dimensional panoramic information interactive information system | |
KR20200057572A (en) | Hand recognition augmented reality-intraction apparatus and method | |
Zhang et al. | Stereo vision information system using median theorem and attitude compensation with nonlinear differential equations | |
Barros et al. | Real-time monocular 6-dof head pose estimation from salient 2d points | |
Marinescu et al. | A versatile 3d face reconstruction from multiple images for face shape classification | |
Li et al. | Detail 3D face reconstruction based on 3DMM and displacement map | |
CN110097644B (en) | Expression migration method, device and system based on mixed reality and processor | |
CN112836544A (en) | Novel sitting posture detection method | |
Kendrick et al. | An online tool for the annotation of 3d models | |
CN113536931A (en) | Hand posture estimation method and device | |
CN112381952A (en) | Face contour point cloud model reconstruction method and device based on multiple cameras | |
CN111612912A (en) | Rapid three-dimensional reconstruction and optimization method based on Kinect2 camera face contour point cloud model | |
CN113554745B (en) | Three-dimensional face reconstruction method based on image |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
TR01 | Transfer of patent right | ||
TR01 | Transfer of patent right |
Effective date of registration: 20210707 Address after: Room 821, K3 / F, Fuxing Huiyu Shuian international, Sanjiao Road, Wuchang District, Wuhan City, Hubei Province, 430061 Patentee after: Tuoerte Intelligent Technology (Wuhan) Co.,Ltd. Address before: No. 69 lijiatuo Chongqing District of Banan City Road 400054 red Patentee before: Chongqing University of Technology |
|
CF01 | Termination of patent right due to non-payment of annual fee | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20200728 |