CN102402691A - Method for tracking gestures and actions of human face - Google Patents

Method for tracking gestures and actions of human face Download PDF

Info

Publication number
CN102402691A
CN102402691A CN2010102780635A CN201010278063A CN102402691A CN 102402691 A CN102402691 A CN 102402691A CN 2010102780635 A CN2010102780635 A CN 2010102780635A CN 201010278063 A CN201010278063 A CN 201010278063A CN 102402691 A CN102402691 A CN 102402691A
Authority
CN
China
Prior art keywords
face
human face
people
frame image
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN2010102780635A
Other languages
Chinese (zh)
Inventor
王阳生
周明才
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Automation of Chinese Academy of Science
Original Assignee
Institute of Automation of Chinese Academy of Science
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Automation of Chinese Academy of Science filed Critical Institute of Automation of Chinese Academy of Science
Priority to CN2010102780635A priority Critical patent/CN102402691A/en
Publication of CN102402691A publication Critical patent/CN102402691A/en
Pending legal-status Critical Current

Links

Images

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses a method for tracking gestures and actions of a human face, which comprises steps as follows: a step S1 includes that frame-by-frame images are extracted from a video streaming, human face detection is carried out for a first frame of image of an input video or when tracking is failed, and a human face surrounding frame is obtained, a step S2 includes that after convergent iteration of a previous frame of image, more remarkable feature points of textural features of a human face area of the previous frame of image match with corresponding feather points found in a current frame of image during normal tracking, and matching results of the feather points are obtained, a step S3 includes that the shape of an active appearance model is initialized according to the human face surrounding frame or the feature point matching results, and an initial value of the shape of a human face in the current frame of image is obtained, and a step S4 includes that the active appearance model is fit by a reversal synthesis algorithm, so that human face three-dimensional gestures and face action parameters are obtained. By the aid of the method, online tracking can be completed full-automatically in real time under the condition of common illumination.

Description

A kind of method that human face posture and action are followed the tracks of
Technical field
The present invention relates to Flame Image Process and computer vision field, particularly based on the human face posture and the motion tracking method of image.
Background technology
The estimation of people's face in image or the video being carried out attitude and face action has important use with tracking and is worth in fields such as man-machine interaction, virtual reality, intelligent monitoring, colourful attitude recognition of face and Expression Recognition.Such as; In expression animation system based on video; Can drive a visual human or cartoon character according to true man's head pose that from video, extracts and face action and do similar expression action, be with a wide range of applications in industries such as interactive digital amusement, animation making.
Different according to the employed information of track algorithm, can face tracking method be divided into based on characteristic method (Feature Based) and based on the method (Appearance Based) of outward appearance.Usually select some easy tracking based on the method for characteristic, and to the characteristics of image of robust relatively such as illumination, attitude, expression, semantic point etc. is arranged like color, edge, angle point or some.Usually do not need training data because local feature point matees, thereby usually illumination variation and people's face texture variations are compared robust based on the method for characteristic.Yet a shortcoming of these class methods is that tracking results is accurate inadequately, and with shake.Attempt whole people's face display model and input picture are mated based on the method for outward appearance, thereby realize face tracking.Method based on outward appearance generally requires current frame image whole people's face texture and reconstructed image to be complementary.Compare with method, owing to utilized the texture information of whole human face region, can follow the tracks of more accurately usually, stable and not shake based on the method for outward appearance based on characteristic.But this method is relatively more responsive for the original shape position, is absorbed in local minimum easily.
In method, the method based on the method for three-dimensional deformation model and two and three dimensions deformable faceform's active appearance models is arranged typically based on outward appearance.Because two and three dimensions deformable faceform's active appearance models algorithm has very big advantage than the three-dimensional deformation model method on match speed, thereby is a kind of practical method.
Two and three dimensions deformable faceform's active appearance models (2D+3D AAM) algorithm is a kind of two-dimentional active appearance models algorithm that adds the 3 d human face mesh model constraint in essence, thereby its core still is two-dimentional active appearance models algorithm.For further improving the performance of two-dimentional active appearance models algorithm; People improve two-dimentional active appearance models algorithm from all angles, are one of very important directions in many improvement directions to the improvement of active appearance models texture expression aspect.Everybody recognizes and uses original half-tone information to express as texture, is difficult to satisfy the demand of practical application.In practical application, the sample that we can collect is limited after all, can not contain into many variable factors such as all imaging circumstances, ethnic group, the colour of skin, attitude, expression, age, picture quality.Even if be encompassed in the inside, the ability to express of the texture model that the use gray feature trains out also is limited.
(Active Appearance Model, AAM) fitting algorithm is based on the iterative optimization method that gradient descends to active appearance models.Provide one preferably initial value be the successful prerequisite of iterative optimization method.In video tracking; When the rapid speed of user movement; If simply use result behind the previous frame image iteration convergence as the initial value of people's face shape in the current frame image; At this moment the method that descends based on gradient is easy to be absorbed in local minimum, but not correct global optimum causes following the tracks of and interrupts or failure.To this problem, common disposal route is in the process of following the tracks of, to merge particle filter algorithm.Yet a very big problem of these class methods is exactly that calculated amount is too big, is difficult to requirement of real time.
Summary of the invention
To above-mentioned prior art problems; The demand of balance various aspects of performance; Consider simultaneously in the practical application requirement of computing velocity the objective of the invention is the people's face to proper motion in the video, realize full-automatic, in real time, robust ground carries out people's face 3 d pose and face action tracking; For this reason, a kind of full automatic method that human face posture and action are followed the tracks of is provided.
To achieve these goals, the present invention is following with the technical scheme steps that the action method of following the tracks of proposes to human face posture:
Step S1: from video flowing, extract by two field picture, carry out people's face during to input video first two field picture or tracking failure and detect, obtain people's face and surround frame;
Step S2: behind previous frame image iteration convergence; During normal the tracking; Compare the notable attribute point for some textural characteristics of human face region in the previous frame image, find these corresponding with it unique points to mate, obtain the matching result of these unique points at current frame image;
Step S3: according to people's face encirclement frame or Feature Points Matching result the shape of active appearance models is carried out initialization, obtain the people's face shape initial value in the current frame image;
Step S4: use the counter-rotating composition algorithm that active appearance models is carried out match, obtain people's face 3 d pose and face action parameter.
The said method that human face posture and action are followed the tracks of is used based on the method for adaptive boosting and is carried out the detection of people's face, obtains position and the size of people's face in image.
Said unique point is to calculate the value of each gloomy matrix determinant in pixel sea earlier, chooses the relatively large pixel of extra large gloomy matrix determinant then as unique point.
Said previous frame image and said current frame image are used multi-resolution representation respectively, in order to improve Feature Points Matching speed.
Said Feature Points Matching is based on the multiresolution framework of image, the method that adopts piece coupling again current frame image find with the previous frame image in the point that is complementary of unique point.
Said active appearance models uses the multiband texture that merges gray scale and marginal information to express, in order to improve the active appearance models algorithm to the adaptive faculty of illumination variation and to not meeting the generalization ability of people's face.
It is said that based on the Feature Points Matching result shape of active appearance models to be carried out initialized treatment step following:
Step S31: in the previous frame image, on the basis, people's face location, select unique point;
Step S32: calculated characteristics point is in the barycentric coordinates that belong to separately in the triangular plate;
Step S33: carry out Feature Points Matching at current frame image;
Step S34: the unique point according to coupling is estimated people's face shape.
Said active appearance models comprises two-dimension human face shape, two-dimension human face texture model, 3-d deformable faceform and imaging model; Said imaging model adopts complete perspective projection, so that obtain human face posture parameter accurately.
Advantage of the present invention is:
1. can realize full-automatic human face posture and motion tracking.At first two field picture or follow-up when losing; The method that can adopt people's face to detect detects people's face and the face encirclement frame of choosing comes the initialization active appearance models; And under normal situation of following the tracks of, utilize Feature Points Matching to obtain the initial value of people's face shape of current frame image; Therefore whole tracing process can automatically be accomplished, and does not need manual intervention.
2. tracking velocity is fast, for the computing machine of Pentium 2.8G, under 320 * 240 resolution, can carry out real-time online and follow the tracks of.
3. track algorithm has good extensive performance, can under common illumination condition, carry out robust tracking to unseen people.
Description of drawings
Fig. 1 is the process flow diagram of full-automatic human face posture of the present invention and motion tracking.
Fig. 2 a to Fig. 2 c is the shape initialization procedure synoptic diagram based on Feature Points Matching of the present invention.
Fig. 3 a to Fig. 3 c is the average face synoptic diagram that multiband texture of the present invention is expressed.
Fig. 4 is a complete perspective projection model synoptic diagram of the present invention.
Embodiment
To combine accompanying drawing that the present invention is specified below, and be to be noted that described embodiment only is intended to be convenient to understanding of the present invention, and it is not played any qualification effect.
Referring to Fig. 1 the automatic watch mutual affection analysis method based on motion tracking is shown, implements according to following steps:
1. people's face detects
The purpose that people's face detects is to detect the position of people's face in image automatically, and the present invention utilizes adaptive boosting (Adaboost) algorithm to carry out people's face and detects automatically.Adaptive boosting is a kind of statistical learning algorithm commonly used, successfully has been applied to detection of people's face and the classification of people's face.Adaptive boosting is provided with weights for each training sample; And revise the weights of sample iteratively; To suitably be reduced by the weights of the sample of correct classification, and will suitably be improved, so just can focus on the sample of difficult classification by the weights of the sample of mis-classification.
People's face only detects and when first two field picture or follow-up tracking failure, just carries out.At this moment choose people's face that face detection algorithm obtains surrounds frame and comes active appearance models (AAM) is carried out initialization.Because people's face detection block has only size and positional information, thus the x in can only active appearance models global affine transformation parameter, y direction translation parameters and whole zooming parameter carry out initialization, and other parameters then are set to zero.
2. based on the shape initialization of Feature Points Matching
In normal tracing process, the present invention utilizes Feature Points Matching to obtain the initial value of the people's face shape in the current frame image, can effectively improve the capturing ability to rapid movement.Basic thought based on the shape initialization algorithm of Feature Points Matching is: behind previous frame image iteration convergence; Choose some relatively notable attribute points at human face region; At current frame image these unique points are mated then, obtain the matching result of these unique points; Again according to the unique point of these couplings to estimating the form parameter of people's face in the current frame image.The concrete steps of algorithm are following:
1) on basis, previous frame image people face location, select unique point:
In the present invention, unique point is made up of two parts: a part is the more preassigned semantic point that has, and in the present invention, we have selected 30 semantic points, comprise some angle points and point on the eyes eyebrow nose face; Another part is the comparison notable attribute point of choosing according to the value of each gloomy matrix determinant in pixel sea, and the computing formula of extra large gloomy matrix is:
D ( u , v ) = I xx ( u , v ) I xy ( u , v ) I xy ( u , v ) I yy ( u , v ) ,
I wherein Xx(u v) is pixel (u, square (or the weighted sum of squares of field point to play a smoothing effect) on every side of the x direction gradient of v) locating, I Yy(u is v) with I Xx(u, v) similar, I Xy(u v) is that (u v) locates x and y direction gradient product (or the weighted product sum of field point) on every side to pixel.
Last selected unique point is shown in Fig. 2 a.
2) calculated characteristics point is established x=(x in the barycentric coordinates that belong to separately in the triangular plate; Y) be with
Figure BSA00000264766400052
be in the triangle on summit a bit; Wherein subscript r, s, t are used to identify leg-of-mutton three summits, and then this unique point can be expressed as the weighted mean on Atria summit:
x = α x r 0 + β x s 0 + γ x t 0 ,
Its weight, β and γ are exactly these barycentric coordinates in triangle, and computing formula is:
α = x s 0 y t 0 - y s 0 x t 0 - xy t 0 + yx t 0 - yx s 0 + xy s 0 x s 0 y t 0 - x r 0 y t 0 - x s 0 y r 0 - y s 0 x t 0 + y r 0 x t 0 + y s 0 x r 0 ,
β = xy t 0 - x r 0 y t 0 - xy r 0 - yx t 0 + y r 0 x t 0 + yx r 0 x s 0 y t 0 - x r 0 y t 0 - x s 0 y r 0 - y s 0 x t 0 + y r 0 x t 0 + y s 0 x r 0 , γ = 1 - α - β ,
After selecting unique point, calculate each unique point x iBarycentric coordinates in its place triangular mesh are (c I1, c I2, c I3), and the sequence number of an Atria summit, record place in shape can be used when calculating in the current frame image people's face shape initial value in the back, i=1, and 2,3.....M, M are unique point quantity.
3) carry out Feature Points Matching at current frame image
The present invention adopts the method for piece template matches to carry out the coupling of unique point.Concrete grammar is in the previous frame image, is the center with the unique point, takes out a pocket T, on current frame image, is that a bigger area I is selected at the center with this unique point also then, and area I should be able to inclusion region T.Subregion T in the previous frame image is slided in the current frame image area I, calculate the normalized correlation coefficient of lap.With the maximum position of normalized correlation coefficient as with the previous frame image on the position that is complementary of unique point.In the present invention, the size of regional T is taken as 9 * 9 (empirical values).The two frame peak excursions that the size of area I is followed the tracks of are as required confirmed.In order to accelerate matching speed, can use by thick to smart multiresolution framework.
What Fig. 2 b showed is the example as a result of a Feature Points Matching.Can find out that from this figure most of unique point all correct match has arrived, have only the fraction unique point not match.Fail the reason of correct match have multiple, such as since the variation of attitude cause local grain generation marked change even be blocked.Fig. 2 c is the original shape of the active appearance models that obtains according to Feature Points Matching.Can find out from this figure,, thereby obtain more satisfactory active appearance models original shape in view of the above owing to Feature Points Matching gets better.
4) estimate people's face shape according to the unique point of coupling
Then we can utilize the unique point of these couplings and information such as barycentric coordinates that the front obtains to roughly estimate people's face shape of current frame image to suppose in current frame image, to have M unique point
Figure BSA00000264766400061
that matches.This can realize through minimizing following formula:
p 0 = min p Σ i = 1 M - w i ρ ( | | Σ j = 1 3 c ij W ( x ij ; p ) - z i | | 2 , r ) ,
W wherein iNormalized correlation coefficient when being coupling,
Figure BSA00000264766400063
To be i unique point be mapped to when the given form parameter p coordinate position on the current frame image, Be i the corresponding average face key point coordinate in shape in triangle 3 summits, unique point place, ρ (, r) be a robust error function.We can use Gauss-newton (Gauss-Newton) algorithm to carry out iterative optimum shape parameter p 0ρ (, definition r) is following:
&rho; ( &delta; , r ) = 3 ( r 2 - &delta; 2 ) 4 r 3 &delta; < r 0 otherwise ,
Wherein r is for putting the letter radius.
3. human face posture and action parameter extract
The present invention uses two and three dimensions deformable faceform's active appearance models algorithm to obtain people's face 3 d pose and face action parameter.Two and three dimensions deformable faceform's active appearance models algorithm can be regarded the two-dimentional active appearance models algorithm that has added the 3 d human face mesh model constraint in essence as.
3.1 merging the multiband texture of gray scale and marginal information expresses.
The present invention proposes a kind of multiband texture that merges gray scale and marginal information and express, to improve the active appearance models algorithm to the adaptive faculty of illumination variation and to not meeting the generalization ability of people's face.Its basic thought is that first wave filter with 3 quadratures carries out filtering to image, then filtered 2 width of cloth gradient images is done a square processing, to remove directional information, at last three width of cloth images is done (Sigmoid) normalization processing of S shape.The multiband texture that specifically can adopt following steps to obtain fusion gray scale and marginal information is expressed:
Figure BSA00000264766400072
Wherein the sigmoid function definition is as follows:
f ( m ) = m m + m &OverBar; ,
Wherein is the average of m in the whole area-of-interest; The effect of this function is to approach 0 value being transformed to much smaller than the input value of average
Figure BSA00000264766400075
, approaches 1 value and be transformed to the input value much larger than average
Figure BSA00000264766400076
.When this function is applied to the normalization edge image, can strengthen real edge image effectively, suppress possible noise simultaneously.
When calculating the multiband texture expression of merging gray scale and marginal information in the above, we have done square operation to the gradient of x direction and y direction.This operation has the benefit of two aspects:
1) effectively strengthens possible edge;
2) directional information of removal edge gradient, a preserving edge strength information.When not knowing that background is bright or darker than people face, it is necessary removing gradient direction information.In addition; Here be
Figure BSA00000264766400081
and
Figure BSA00000264766400082
to be done Sigmoid normalization respectively handle, calculated amount is less relatively.
There is any need to prove in addition; Here during compute gradient, need earlier input picture is deformed on the average shape, obtain the irrelevant facial image of shape; And then on the irrelevant facial image of shape compute gradient, with the gradient that guarantees to calculate in the image irrelevant to rotation direction.
Fig. 3 a to Fig. 3 c has shown the average face image of this multiband texture model.Wherein Fig. 3 a is the average face image of gray scale wave band, and Fig. 3 b and Fig. 3 c are respectively the average face images of x and y direction gradient wave band.As can be seen from the figure, the information of these three wave bands has very strong complementarity, and Fig. 3 a characterizes overall intensity and distributes, and Fig. 3 b characterizes x direction coboundary and distributes, and Fig. 3 c characterizes y direction coboundary and distributes.Simultaneously owing to used square operation and the operation of Sigmoid function normalization, the essential marginal texture information of people's face has been contained in the multiband texture expression of merging gray scale and marginal information, filtering nonessential interfere information.
3.2 three-dimensional face model and imaging model.
In the present invention, used deformable three-dimensional face model (Candide-3) is a three-dimensional model that is called Kan Daide-3 (Candide-3).This model not only can suitably be out of shape according to different people's face shapes approaching personalized human face, and can produce corresponding deformation according to face action, is very suitable for human face action and describes and follow the tracks of.
The shape g of 3-d deformable face wire frame model can be with each apex coordinate P of three-dimensional model i=(x i, y i, z i) TThe vector that is formed by connecting is represented:
g=(x 1,y 1,z 1,x 2,y 2,z 2,…,x n,y n,z n) T
Wherein n is a grid vertex quantity, (x i, y i, z i) TBe i grid vertex P iThree-dimensional coordinate, T representing matrix transposition.The three-dimensional coordinate here is the coordinate figure in model self local coordinate system, for people's face shape of propertyization one by one, can use following shape to generate:
g = g &OverBar; + S&sigma; + A&alpha; ,
Wherein
Figure BSA00000264766400084
is the average face shape; S and A are respectively change in shape and action transformation matrices; The all corresponding a kind of independently changing pattern of each row of matrix, σ and α are respectively change in shape and action variation factor vector.Therefore, S σ has described grid model to the different variation of people's face on global shape, like whole fat or thin of people's face, distance between two and the position of eyes eyebrow nose mouth in people's face etc.A α describes the variation of the mesh shape that face action causes, as opens one's mouth, and lifts eyebrow etc.In addition, we suppose that it is separate that change in shape changes with action.
The present invention has adopted complete perspective projection model to describe imaging process, and is as shown in Figure 4.Each coordinate system is the cartesian coordinate system that satisfies the right-hand rule among the figure.
If the Candide-3 faceform is last 1 P o=(X o, Y o, Z o) T, coordinate is P under camera coordinates system after rotating R, translation T rigid body translation c=(X c, Y c, Z c) T:
X c Y c Z c = R X o Y o Z o + T ,
Point P cThrough complete perspective projection 1 P to the plane of delineation i, establishing this some coordinate under camera coordinates system is P i=(X i, Y i, Z i) T, then the relation between these two coordinates is:
X i Y i Z i = f &CenterDot; X c / Z c f &CenterDot; Y c / Z c f ,
Wherein f is the digital focal length of camera, can calculate according to the resolution of entire image and the visual angle size of camera.
Some P on the plane of delineation iCoordinate under image coordinate system can obtain through computes:
u i v i = X i + W / 2 - Y i + H / 2 ,
Wherein W and H are respectively the width and the height of image.Because y axle positive dirction typically refers to downwards under image coordinate system, so following formula is to Y iDone individual inversion operation.In addition since the initial point of image coordinate system usually in the upper left corner of image, so x, the y coordinate adds respectively and equals image half-breadth and half high skew.
3.3 active appearance models algorithm based on the two and three dimensions deformable faceform of Kan Daide-3 three-dimensional face model.
The proposition of initial two and three dimensions active appearance models algorithm is in order to improve the performance of two-dimentional active appearance models algorithm; Force a constraint promptly for two-dimentional active appearance models, make that the two-dimensional shapes of utilizing two-dimentional global affine transformation parameter and form parameter to generate is legal.Here, legal implication is to have deformation parameter in effective three-dimensional rotation and translation parameters and the three-dimensional face model, and the feasible plane of delineation gained two-dimensional shapes that projects to is consistent with the shape that the two-dimensional shapes model generates.On the mathematics, this constraint can be expressed as:
minE m=min||s′(p)-P(Q(g′(σ,α)))|| 2
G (σ wherein; α) expression Kan Daide-3 people face shape; Q (x) expression is done the three-dimensional rigid body conversion that comprises rotation matrix R and translation vector T to the three-dimensional vertices vector x; P representes that each three-dimensional vertices among the three-dimensional vertices vector Q (x) is projected to the plane of delineation obtains its pixel coordinate under image coordinate system, the two-dimensional shapes that s (p) expression generates according to the two-dimensional shapes parameter p, the part of subscript ' whole vector of expression.Because the summit is not one to one on the key point of two-dimensional shapes and the three-dimensional face grid, have only the sub-fraction ability corresponding, subscript ' be exactly this part on the expression ability correspondence.
Above-mentioned bound term is added in the objective function of original two dimensional active appearance models, promptly obtain so-called two and three dimensions deformable faceform's active appearance models, that is:
| | A 0 + &Sigma; i = 1 m &lambda; i A i - I ( W ( p ) ) | | + w m | | s &prime; ( p ) - P ( Q ( g &prime; ( &sigma; , &alpha; ) ) ) | | 2 ,
A wherein 0Be average face outward appearance, A iFor with preceding n people's face outward appearance base vector, λ iBe apparent coefficient, W (p) is a coordinate transform function, w mBe the weight of 3D shape constraint portions, w in the present invention m=0.1.
Optimization for two and three dimensions deformable faceform's active appearance models is found the solution, and still can use the fast algorithm in the two-dimentional active appearance models.The first step, the span (A in the subspace i) The optimization aim function:
| | A 0 - I ( W ( p ) ) | | span ( A i ) &perp; 2 + w m &Sigma; i F i 2 ( p ; P ; &sigma; ; &alpha; ) ,
F wherein I(p; P; σ; α) the site error of i obligatory point of expression on x, y direction.Here succinct in order to write, with the synthetic expression formula of the site error on x, the y both direction.The method that second step was found the solution apparent parameter lambda is with two-dimentional active appearance models.
In order to represent that conveniently we are linked to be a vectorial q=(p with all unknown numbers; P; σ; α), then can accomplish through following two step iteration the optimization of following formula:
The first step: calculating parameter renewal amount:
&Delta;q = - H 3 D - 1 [ &Delta; p SD 0 + w m &CenterDot; &Sigma; i ( &PartialD; F &PartialD; q ) T &CenterDot; Fi ( q ) ] ,
Δ p wherein SDBe the steepest decline figure of two dimension part, H 3DBe the extra large gloomy matrix of two and three dimensions active appearance models, F iBe each bound term, the concrete definition as follows:
&Delta; p SD = &Sigma; x &Element; s 0 SD ( x ) &CenterDot; ( A ( x ) - I ( W ( x ; p ) ) ) ,
H 3 D = H 2 D 0 0 0 + w m &CenterDot; &Sigma; i ( &PartialD; F i &PartialD; q ) T ( &PartialD; F i &PartialD; q ) ,
H 2 D = &Sigma; x &Element; s 0 SD T ( x ) &CenterDot; SD ( x ) ,
SD ( x ) = &dtri; A 0 ( x ) &PartialD; N &PartialD; q &dtri; A 0 ( x ) &PartialD; W &PartialD; p span ( A i ) &perp; ,
Second step: parameter update:
Utilize counter-rotating synthetic method undated parameter p, use addition rule undated parameter P, σ and α.
Need to prove, in above-mentioned arthmetic statement,, ignored to F in order to describe and to be concise in expression i(p; P; σ; Necessity correction that need do when α) carrying out the single order Taylor expansion, i.e. F i(p; P; σ; Increment problem about p α) need be come from the mapping of the increment on the average shape.
In superincumbent two and three dimensions deformable faceform's the active appearance models fitting algorithm, can try to achieve human face posture parameter, personalized form parameter σ and face action parameter alpha simultaneously.Under the tracking environmental of reality, confirm personalized form parameter σ in the tracking initiation stage usually, remain unchanged at follow-up tracking phase, only if just the tracking failure needs calculating once more in the time of need reinitializing grid model.In follow-up tracing process, only need tracks facial action parameter α.This can bring the benefit of two aspects:
1) reduced the number of parameters that to find the solution when following the tracks of, and then improved tracking velocity, again the necessary operation parameter can be provided simultaneously.
2) can be to the form generation of the two-dimentional active appearance models stronger constraint strength of fixing personalized shape coefficient.
Top description is to be used to realize the present invention and embodiment, and therefore, scope of the present invention should not described by this and limit.It should be appreciated by those skilled in the art,, all belong to claim of the present invention and come restricted portion in any modification that does not depart from the scope of the present invention or local replacement.

Claims (8)

1. one kind to human face posture and the action method of following the tracks of, and comprises step:
Step S1: from video flowing, extract by two field picture, carry out people's face during to input video first two field picture or tracking failure and detect, obtain people's face and surround frame;
Step S2: behind previous frame image iteration convergence; During normal the tracking; Compare the notable attribute point for some textural characteristics of human face region in the previous frame image, find these corresponding with it unique points to mate, obtain the matching result of these unique points at current frame image;
Step S3: according to people's face encirclement frame or Feature Points Matching result the shape of active appearance models is carried out initialization, obtain the people's face shape initial value in the current frame image;
Step S4: use the counter-rotating composition algorithm that active appearance models is carried out match, obtain people's face 3 d pose and face action parameter.
2. according to the said method that human face posture and action are followed the tracks of of claim 1, it is characterized in that also comprising step: use based on the method for adaptive boosting and carry out the detection of people's face, obtain position and the size of people's face in image.
3. according to the said method that human face posture and action are followed the tracks of of claim 1; It is characterized in that: said unique point is to calculate the value of each gloomy matrix determinant in pixel sea earlier, chooses the relatively large pixel of extra large gloomy matrix determinant then as unique point.
4. according to the said method that human face posture and action are followed the tracks of of claim 1, it is characterized in that: said previous frame image and said current frame image are used multi-resolution representation respectively, in order to improve Feature Points Matching speed.
5. according to the said method that human face posture and action are followed the tracks of of claim 1; It is characterized in that; Said Feature Points Matching is based on the multiresolution framework of image, the method that adopts piece coupling again current frame image find with the previous frame image in the point that is complementary of unique point.
6. according to the said method that human face posture and action are followed the tracks of of claim 1; It is characterized in that: said active appearance models uses the multiband texture that merges gray scale and marginal information to express, in order to improve the active appearance models algorithm to the adaptive faculty of illumination variation and to not meeting the generalization ability of people's face.
7. according to claim 1 is said human face posture and the action method of following the tracks of is characterized in that said based on the Feature Points Matching result shape of active appearance models to be carried out initialized treatment step following:
Step S31: in the previous frame image, on the basis, people's face location, select unique point;
Step S32: calculated characteristics point is in the barycentric coordinates that belong to separately in the triangular plate;
Step S33: carry out Feature Points Matching at current frame image;
Step S34: the unique point according to coupling is estimated people's face shape.
8. according to the said method that human face posture and action are followed the tracks of of claim 1, it is characterized in that: said active appearance models comprises two-dimension human face shape, two-dimension human face texture model, 3-d deformable faceform and imaging model; Said imaging model adopts complete perspective projection, so that obtain human face posture parameter accurately.
CN2010102780635A 2010-09-08 2010-09-08 Method for tracking gestures and actions of human face Pending CN102402691A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN2010102780635A CN102402691A (en) 2010-09-08 2010-09-08 Method for tracking gestures and actions of human face

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN2010102780635A CN102402691A (en) 2010-09-08 2010-09-08 Method for tracking gestures and actions of human face

Publications (1)

Publication Number Publication Date
CN102402691A true CN102402691A (en) 2012-04-04

Family

ID=45884881

Family Applications (1)

Application Number Title Priority Date Filing Date
CN2010102780635A Pending CN102402691A (en) 2010-09-08 2010-09-08 Method for tracking gestures and actions of human face

Country Status (1)

Country Link
CN (1) CN102402691A (en)

Cited By (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102750527A (en) * 2012-06-26 2012-10-24 浙江捷尚视觉科技有限公司 Long-time stable human face detection and tracking method in bank scene and long-time stable human face detection and tracking device in bank scene
CN102867174A (en) * 2012-08-30 2013-01-09 中国科学技术大学 Method and device for positioning human face features
CN102867173A (en) * 2012-08-28 2013-01-09 华南理工大学 Human face recognition method and system thereof
CN102880866A (en) * 2012-09-29 2013-01-16 宁波大学 Method for extracting face features
CN103400119A (en) * 2013-07-31 2013-11-20 南京融图创斯信息科技有限公司 Face recognition technology-based mixed reality spectacle interactive display method
CN103870824A (en) * 2014-03-28 2014-06-18 海信集团有限公司 Method and device for capturing face in face detecting and tracking process
CN104392447A (en) * 2014-11-28 2015-03-04 西南科技大学 Image matching method based on gray scale gradient
CN105447462A (en) * 2015-11-20 2016-03-30 小米科技有限责任公司 Facial pose estimation method and device
CN105989326A (en) * 2015-01-29 2016-10-05 北京三星通信技术研究有限公司 Method and device for determining three-dimensional position information of human eyes
CN106228113A (en) * 2016-07-12 2016-12-14 电子科技大学 Human face characteristic point quick alignment method based on AAM
CN106295511A (en) * 2016-07-26 2017-01-04 北京小米移动软件有限公司 Face tracking method and device
CN106462738A (en) * 2014-05-20 2017-02-22 埃西勒国际通用光学公司 Method for constructing a model of the face of a person, method and device for posture analysis using such a model
CN106778474A (en) * 2016-11-14 2017-05-31 深圳奥比中光科技有限公司 3D human body recognition methods and equipment
CN107316029A (en) * 2017-07-03 2017-11-03 腾讯科技(深圳)有限公司 A kind of live body verification method and equipment
CN107633526A (en) * 2017-09-04 2018-01-26 腾讯科技(深圳)有限公司 A kind of image trace point acquisition methods and equipment, storage medium
CN107992825A (en) * 2017-12-01 2018-05-04 青岛海尔智能家电科技有限公司 A kind of method and system of the recognition of face based on augmented reality
CN108229246A (en) * 2016-12-14 2018-06-29 上海交通大学 Real-time three-dimensional human face posture method for tracing based on vehicle computing machine platform
CN108345821A (en) * 2017-01-24 2018-07-31 成都理想境界科技有限公司 Face tracking method and apparatus
CN108510520A (en) * 2018-02-13 2018-09-07 视辰信息科技(上海)有限公司 A kind of image processing method, device and AR equipment
CN108875506A (en) * 2017-11-17 2018-11-23 北京旷视科技有限公司 Face shape point-tracking method, device and system and storage medium
CN109711304A (en) * 2013-05-21 2019-05-03 深圳市腾讯计算机系统有限公司 A kind of man face characteristic point positioning method and device
WO2019097285A1 (en) * 2017-08-31 2019-05-23 Banuba Limited Computer-implemented methods and computer systems for real-time detection of human's emotions from visual recordings
CN109859322A (en) * 2019-01-22 2019-06-07 广西大学 A kind of spectrum posture moving method based on deformation pattern
CN109874021A (en) * 2017-12-04 2019-06-11 腾讯科技(深圳)有限公司 Living broadcast interactive method, apparatus and system
CN112464918A (en) * 2021-01-27 2021-03-09 昆山恒巨电子有限公司 Body-building action correcting method and device, computer equipment and storage medium
CN114863506A (en) * 2022-03-18 2022-08-05 珠海优特电力科技股份有限公司 Method, device and system for verifying access permission and identity authentication terminal
WO2023087891A1 (en) * 2021-11-18 2023-05-25 中兴通讯股份有限公司 Real-time facial image driving method and apparatus, electronic device, and storage medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1811793A (en) * 2006-03-02 2006-08-02 复旦大学 Automatic positioning method for characteristic point of human faces
CN1866271A (en) * 2006-06-13 2006-11-22 北京中星微电子有限公司 AAM-based head pose real-time estimating method and system

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1811793A (en) * 2006-03-02 2006-08-02 复旦大学 Automatic positioning method for characteristic point of human faces
CN1866271A (en) * 2006-06-13 2006-11-22 北京中星微电子有限公司 AAM-based head pose real-time estimating method and system

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
周明才: "表情互动中的若干关键问题研究", 《HTTP://LIB.IA.AC.CN/DLIB/LIST.ASP?LANG=GB&TYPE=&DOCGROUPID=10&DOCID=10071》 *

Cited By (48)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102750527A (en) * 2012-06-26 2012-10-24 浙江捷尚视觉科技有限公司 Long-time stable human face detection and tracking method in bank scene and long-time stable human face detection and tracking device in bank scene
CN102750527B (en) * 2012-06-26 2015-08-19 浙江捷尚视觉科技股份有限公司 The medium-term and long-term stable persona face detection method of a kind of bank scene and device
CN102867173A (en) * 2012-08-28 2013-01-09 华南理工大学 Human face recognition method and system thereof
CN102867173B (en) * 2012-08-28 2015-01-28 华南理工大学 Human face recognition method and system thereof
CN102867174B (en) * 2012-08-30 2016-01-20 中国科学技术大学 A kind of human face characteristic positioning method and device
CN102867174A (en) * 2012-08-30 2013-01-09 中国科学技术大学 Method and device for positioning human face features
CN102880866A (en) * 2012-09-29 2013-01-16 宁波大学 Method for extracting face features
CN102880866B (en) * 2012-09-29 2014-12-17 宁波大学 Method for extracting face features
CN109711304A (en) * 2013-05-21 2019-05-03 深圳市腾讯计算机系统有限公司 A kind of man face characteristic point positioning method and device
CN109711304B (en) * 2013-05-21 2022-06-14 深圳市腾讯计算机系统有限公司 Face feature point positioning method and device
CN103400119A (en) * 2013-07-31 2013-11-20 南京融图创斯信息科技有限公司 Face recognition technology-based mixed reality spectacle interactive display method
CN103400119B (en) * 2013-07-31 2017-02-15 徐坚 Face recognition technology-based mixed reality spectacle interactive display method
CN103870824A (en) * 2014-03-28 2014-06-18 海信集团有限公司 Method and device for capturing face in face detecting and tracking process
CN103870824B (en) * 2014-03-28 2017-10-20 海信集团有限公司 A kind of face method for catching and device during Face datection tracking
CN106462738B (en) * 2014-05-20 2020-10-09 依视路国际公司 Method for constructing a model of a person's face, method and apparatus for analyzing a pose using such a model
US10380411B2 (en) 2014-05-20 2019-08-13 Essilor International Method for constructing a model of the face of a person, method and device for posture analysis using such a model
CN106462738A (en) * 2014-05-20 2017-02-22 埃西勒国际通用光学公司 Method for constructing a model of the face of a person, method and device for posture analysis using such a model
CN104392447A (en) * 2014-11-28 2015-03-04 西南科技大学 Image matching method based on gray scale gradient
CN104392447B (en) * 2014-11-28 2017-10-13 西南科技大学 A kind of image matching method based on shade of gray
CN105989326A (en) * 2015-01-29 2016-10-05 北京三星通信技术研究有限公司 Method and device for determining three-dimensional position information of human eyes
CN105989326B (en) * 2015-01-29 2020-03-03 北京三星通信技术研究有限公司 Method and device for determining three-dimensional position information of human eyes
CN105447462A (en) * 2015-11-20 2016-03-30 小米科技有限责任公司 Facial pose estimation method and device
CN105447462B (en) * 2015-11-20 2018-11-20 小米科技有限责任公司 Face pose estimation and device
CN106228113A (en) * 2016-07-12 2016-12-14 电子科技大学 Human face characteristic point quick alignment method based on AAM
CN106295511A (en) * 2016-07-26 2017-01-04 北京小米移动软件有限公司 Face tracking method and device
CN106295511B (en) * 2016-07-26 2019-05-21 北京小米移动软件有限公司 Face tracking method and device
CN106778474A (en) * 2016-11-14 2017-05-31 深圳奥比中光科技有限公司 3D human body recognition methods and equipment
CN108229246A (en) * 2016-12-14 2018-06-29 上海交通大学 Real-time three-dimensional human face posture method for tracing based on vehicle computing machine platform
CN108345821A (en) * 2017-01-24 2018-07-31 成都理想境界科技有限公司 Face tracking method and apparatus
CN108345821B (en) * 2017-01-24 2022-03-08 成都理想境界科技有限公司 Face tracking method and device
CN107316029B (en) * 2017-07-03 2018-11-23 腾讯科技(深圳)有限公司 A kind of living body verification method and equipment
CN107316029A (en) * 2017-07-03 2017-11-03 腾讯科技(深圳)有限公司 A kind of live body verification method and equipment
WO2019097285A1 (en) * 2017-08-31 2019-05-23 Banuba Limited Computer-implemented methods and computer systems for real-time detection of human's emotions from visual recordings
WO2019042419A1 (en) * 2017-09-04 2019-03-07 腾讯科技(深圳)有限公司 Image tracking point acquisition method and device, and storage medium
US11164323B2 (en) * 2017-09-04 2021-11-02 Tencent Technology (Shenzhen) Company Limited Method for obtaining image tracking points and device and storage medium thereof
CN107633526A (en) * 2017-09-04 2018-01-26 腾讯科技(深圳)有限公司 A kind of image trace point acquisition methods and equipment, storage medium
CN108875506A (en) * 2017-11-17 2018-11-23 北京旷视科技有限公司 Face shape point-tracking method, device and system and storage medium
CN108875506B (en) * 2017-11-17 2022-01-07 北京旷视科技有限公司 Face shape point tracking method, device and system and storage medium
CN107992825A (en) * 2017-12-01 2018-05-04 青岛海尔智能家电科技有限公司 A kind of method and system of the recognition of face based on augmented reality
CN109874021A (en) * 2017-12-04 2019-06-11 腾讯科技(深圳)有限公司 Living broadcast interactive method, apparatus and system
WO2019157922A1 (en) * 2018-02-13 2019-08-22 视辰信息科技(上海)有限公司 Image processing method and device and ar apparatus
CN108510520B (en) * 2018-02-13 2019-03-08 视辰信息科技(上海)有限公司 A kind of image processing method, device and AR equipment
CN108510520A (en) * 2018-02-13 2018-09-07 视辰信息科技(上海)有限公司 A kind of image processing method, device and AR equipment
CN109859322A (en) * 2019-01-22 2019-06-07 广西大学 A kind of spectrum posture moving method based on deformation pattern
CN109859322B (en) * 2019-01-22 2022-12-06 广西大学 Spectral attitude migration method based on deformation graph
CN112464918A (en) * 2021-01-27 2021-03-09 昆山恒巨电子有限公司 Body-building action correcting method and device, computer equipment and storage medium
WO2023087891A1 (en) * 2021-11-18 2023-05-25 中兴通讯股份有限公司 Real-time facial image driving method and apparatus, electronic device, and storage medium
CN114863506A (en) * 2022-03-18 2022-08-05 珠海优特电力科技股份有限公司 Method, device and system for verifying access permission and identity authentication terminal

Similar Documents

Publication Publication Date Title
CN102402691A (en) Method for tracking gestures and actions of human face
Zuffi et al. Lions and tigers and bears: Capturing non-rigid, 3d, articulated shape from images
Dai et al. A 3d morphable model of craniofacial shape and texture variation
CN100416612C (en) Video flow based three-dimensional dynamic human face expression model construction method
Burl et al. A probabilistic approach to object recognition using local photometry and global geometry
CN101968846B (en) Face tracking method
US20220358770A1 (en) Scene reconstruction in three-dimensions from two-dimensional images
CN101499128B (en) Three-dimensional human face action detecting and tracing method based on video stream
US9298257B2 (en) Apparatus and method for controlling avatar using expression control point
CN102999942B (en) Three-dimensional face reconstruction method
Blake et al. Active contours: the application of techniques from graphics, vision, control theory and statistics to visual tracking of shapes in motion
Balan et al. Detailed human shape and pose from images
CN101964064B (en) Human face comparison method
Agudo et al. Simultaneous pose and non-rigid shape with particle dynamics
Bascle et al. Stereo matching, reconstruction and refinement of 3D curves using deformable contours
CN104036546A (en) Method for carrying out face three-dimensional reconstruction at any viewing angle on basis of self-adaptive deformable model
CN102376100A (en) Single-photo-based human face animating method
CN102654903A (en) Face comparison method
CN103733226A (en) Fast articulated motion tracking
Pan et al. Sketch-based skeleton-driven 2D animation and motion capture
CN105893984A (en) Face projection method for facial makeup based on face features
CN103714556A (en) Moving target tracking method based on pyramid appearance model
Bao et al. High-quality face capture using anatomical muscles
Chen et al. Single and sparse view 3d reconstruction by learning shape priors
Fayad et al. Non-rigid Structure from Motion using Quadratic Deformation Models.

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C02 Deemed withdrawal of patent application after publication (patent law 2001)
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20120404