CN102402691A - Method for tracking gestures and actions of human face - Google Patents
Method for tracking gestures and actions of human face Download PDFInfo
- Publication number
- CN102402691A CN102402691A CN2010102780635A CN201010278063A CN102402691A CN 102402691 A CN102402691 A CN 102402691A CN 2010102780635 A CN2010102780635 A CN 2010102780635A CN 201010278063 A CN201010278063 A CN 201010278063A CN 102402691 A CN102402691 A CN 102402691A
- Authority
- CN
- China
- Prior art keywords
- face
- human face
- people
- frame image
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 51
- 230000009471 action Effects 0.000 title claims abstract description 30
- 238000001514 detection method Methods 0.000 claims abstract description 6
- 239000011159 matrix material Substances 0.000 claims description 10
- 230000008878 coupling Effects 0.000 claims description 9
- 238000010168 coupling process Methods 0.000 claims description 9
- 238000005859 coupling reaction Methods 0.000 claims description 9
- 230000003044 adaptive effect Effects 0.000 claims description 8
- 238000003384 imaging method Methods 0.000 claims description 7
- 230000000295 complement effect Effects 0.000 claims description 4
- 239000000284 extract Substances 0.000 claims description 4
- 239000000203 mixture Substances 0.000 claims description 2
- 238000005286 illumination Methods 0.000 abstract description 4
- 210000003746 feather Anatomy 0.000 abstract 2
- 230000015572 biosynthetic process Effects 0.000 abstract 1
- 238000003786 synthesis reaction Methods 0.000 abstract 1
- 230000006870 function Effects 0.000 description 8
- 230000008569 process Effects 0.000 description 7
- 238000005457 optimization Methods 0.000 description 5
- 238000013519 translation Methods 0.000 description 5
- 230000008901 benefit Effects 0.000 description 4
- 230000008859 change Effects 0.000 description 4
- 238000010586 diagram Methods 0.000 description 4
- 238000010606 normalization Methods 0.000 description 4
- 230000000694 effects Effects 0.000 description 3
- 210000004709 eyebrow Anatomy 0.000 description 3
- 230000001815 facial effect Effects 0.000 description 3
- 230000009466 transformation Effects 0.000 description 3
- PXFBZOLANLWPMH-UHFFFAOYSA-N 16-Epiaffinine Natural products C1C(C2=CC=CC=C2N2)=C2C(=O)CC2C(=CC)CN(C)C1C2CO PXFBZOLANLWPMH-UHFFFAOYSA-N 0.000 description 2
- 238000013459 approach Methods 0.000 description 2
- 239000004744 fabric Substances 0.000 description 2
- 238000001914 filtration Methods 0.000 description 2
- 210000002837 heart atrium Anatomy 0.000 description 2
- 230000006872 improvement Effects 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- 238000012549 training Methods 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 244000188472 Ilex paraguariensis Species 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 238000009499 grossing Methods 0.000 description 1
- 210000003128 head Anatomy 0.000 description 1
- 230000000977 initiatory effect Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 239000002245 particle Substances 0.000 description 1
- 238000012797 qualification Methods 0.000 description 1
- 238000010189 synthetic method Methods 0.000 description 1
- 230000017105 transposition Effects 0.000 description 1
Images
Landscapes
- Image Analysis (AREA)
Abstract
The invention discloses a method for tracking gestures and actions of a human face, which comprises steps as follows: a step S1 includes that frame-by-frame images are extracted from a video streaming, human face detection is carried out for a first frame of image of an input video or when tracking is failed, and a human face surrounding frame is obtained, a step S2 includes that after convergent iteration of a previous frame of image, more remarkable feature points of textural features of a human face area of the previous frame of image match with corresponding feather points found in a current frame of image during normal tracking, and matching results of the feather points are obtained, a step S3 includes that the shape of an active appearance model is initialized according to the human face surrounding frame or the feature point matching results, and an initial value of the shape of a human face in the current frame of image is obtained, and a step S4 includes that the active appearance model is fit by a reversal synthesis algorithm, so that human face three-dimensional gestures and face action parameters are obtained. By the aid of the method, online tracking can be completed full-automatically in real time under the condition of common illumination.
Description
Technical field
The present invention relates to Flame Image Process and computer vision field, particularly based on the human face posture and the motion tracking method of image.
Background technology
The estimation of people's face in image or the video being carried out attitude and face action has important use with tracking and is worth in fields such as man-machine interaction, virtual reality, intelligent monitoring, colourful attitude recognition of face and Expression Recognition.Such as; In expression animation system based on video; Can drive a visual human or cartoon character according to true man's head pose that from video, extracts and face action and do similar expression action, be with a wide range of applications in industries such as interactive digital amusement, animation making.
Different according to the employed information of track algorithm, can face tracking method be divided into based on characteristic method (Feature Based) and based on the method (Appearance Based) of outward appearance.Usually select some easy tracking based on the method for characteristic, and to the characteristics of image of robust relatively such as illumination, attitude, expression, semantic point etc. is arranged like color, edge, angle point or some.Usually do not need training data because local feature point matees, thereby usually illumination variation and people's face texture variations are compared robust based on the method for characteristic.Yet a shortcoming of these class methods is that tracking results is accurate inadequately, and with shake.Attempt whole people's face display model and input picture are mated based on the method for outward appearance, thereby realize face tracking.Method based on outward appearance generally requires current frame image whole people's face texture and reconstructed image to be complementary.Compare with method, owing to utilized the texture information of whole human face region, can follow the tracks of more accurately usually, stable and not shake based on the method for outward appearance based on characteristic.But this method is relatively more responsive for the original shape position, is absorbed in local minimum easily.
In method, the method based on the method for three-dimensional deformation model and two and three dimensions deformable faceform's active appearance models is arranged typically based on outward appearance.Because two and three dimensions deformable faceform's active appearance models algorithm has very big advantage than the three-dimensional deformation model method on match speed, thereby is a kind of practical method.
Two and three dimensions deformable faceform's active appearance models (2D+3D AAM) algorithm is a kind of two-dimentional active appearance models algorithm that adds the 3 d human face mesh model constraint in essence, thereby its core still is two-dimentional active appearance models algorithm.For further improving the performance of two-dimentional active appearance models algorithm; People improve two-dimentional active appearance models algorithm from all angles, are one of very important directions in many improvement directions to the improvement of active appearance models texture expression aspect.Everybody recognizes and uses original half-tone information to express as texture, is difficult to satisfy the demand of practical application.In practical application, the sample that we can collect is limited after all, can not contain into many variable factors such as all imaging circumstances, ethnic group, the colour of skin, attitude, expression, age, picture quality.Even if be encompassed in the inside, the ability to express of the texture model that the use gray feature trains out also is limited.
(Active Appearance Model, AAM) fitting algorithm is based on the iterative optimization method that gradient descends to active appearance models.Provide one preferably initial value be the successful prerequisite of iterative optimization method.In video tracking; When the rapid speed of user movement; If simply use result behind the previous frame image iteration convergence as the initial value of people's face shape in the current frame image; At this moment the method that descends based on gradient is easy to be absorbed in local minimum, but not correct global optimum causes following the tracks of and interrupts or failure.To this problem, common disposal route is in the process of following the tracks of, to merge particle filter algorithm.Yet a very big problem of these class methods is exactly that calculated amount is too big, is difficult to requirement of real time.
Summary of the invention
To above-mentioned prior art problems; The demand of balance various aspects of performance; Consider simultaneously in the practical application requirement of computing velocity the objective of the invention is the people's face to proper motion in the video, realize full-automatic, in real time, robust ground carries out people's face 3 d pose and face action tracking; For this reason, a kind of full automatic method that human face posture and action are followed the tracks of is provided.
To achieve these goals, the present invention is following with the technical scheme steps that the action method of following the tracks of proposes to human face posture:
Step S1: from video flowing, extract by two field picture, carry out people's face during to input video first two field picture or tracking failure and detect, obtain people's face and surround frame;
Step S2: behind previous frame image iteration convergence; During normal the tracking; Compare the notable attribute point for some textural characteristics of human face region in the previous frame image, find these corresponding with it unique points to mate, obtain the matching result of these unique points at current frame image;
Step S3: according to people's face encirclement frame or Feature Points Matching result the shape of active appearance models is carried out initialization, obtain the people's face shape initial value in the current frame image;
Step S4: use the counter-rotating composition algorithm that active appearance models is carried out match, obtain people's face 3 d pose and face action parameter.
The said method that human face posture and action are followed the tracks of is used based on the method for adaptive boosting and is carried out the detection of people's face, obtains position and the size of people's face in image.
Said unique point is to calculate the value of each gloomy matrix determinant in pixel sea earlier, chooses the relatively large pixel of extra large gloomy matrix determinant then as unique point.
Said previous frame image and said current frame image are used multi-resolution representation respectively, in order to improve Feature Points Matching speed.
Said Feature Points Matching is based on the multiresolution framework of image, the method that adopts piece coupling again current frame image find with the previous frame image in the point that is complementary of unique point.
Said active appearance models uses the multiband texture that merges gray scale and marginal information to express, in order to improve the active appearance models algorithm to the adaptive faculty of illumination variation and to not meeting the generalization ability of people's face.
It is said that based on the Feature Points Matching result shape of active appearance models to be carried out initialized treatment step following:
Step S31: in the previous frame image, on the basis, people's face location, select unique point;
Step S32: calculated characteristics point is in the barycentric coordinates that belong to separately in the triangular plate;
Step S33: carry out Feature Points Matching at current frame image;
Step S34: the unique point according to coupling is estimated people's face shape.
Said active appearance models comprises two-dimension human face shape, two-dimension human face texture model, 3-d deformable faceform and imaging model; Said imaging model adopts complete perspective projection, so that obtain human face posture parameter accurately.
Advantage of the present invention is:
1. can realize full-automatic human face posture and motion tracking.At first two field picture or follow-up when losing; The method that can adopt people's face to detect detects people's face and the face encirclement frame of choosing comes the initialization active appearance models; And under normal situation of following the tracks of, utilize Feature Points Matching to obtain the initial value of people's face shape of current frame image; Therefore whole tracing process can automatically be accomplished, and does not need manual intervention.
2. tracking velocity is fast, for the computing machine of Pentium 2.8G, under 320 * 240 resolution, can carry out real-time online and follow the tracks of.
3. track algorithm has good extensive performance, can under common illumination condition, carry out robust tracking to unseen people.
Description of drawings
Fig. 1 is the process flow diagram of full-automatic human face posture of the present invention and motion tracking.
Fig. 2 a to Fig. 2 c is the shape initialization procedure synoptic diagram based on Feature Points Matching of the present invention.
Fig. 3 a to Fig. 3 c is the average face synoptic diagram that multiband texture of the present invention is expressed.
Fig. 4 is a complete perspective projection model synoptic diagram of the present invention.
Embodiment
To combine accompanying drawing that the present invention is specified below, and be to be noted that described embodiment only is intended to be convenient to understanding of the present invention, and it is not played any qualification effect.
Referring to Fig. 1 the automatic watch mutual affection analysis method based on motion tracking is shown, implements according to following steps:
1. people's face detects
The purpose that people's face detects is to detect the position of people's face in image automatically, and the present invention utilizes adaptive boosting (Adaboost) algorithm to carry out people's face and detects automatically.Adaptive boosting is a kind of statistical learning algorithm commonly used, successfully has been applied to detection of people's face and the classification of people's face.Adaptive boosting is provided with weights for each training sample; And revise the weights of sample iteratively; To suitably be reduced by the weights of the sample of correct classification, and will suitably be improved, so just can focus on the sample of difficult classification by the weights of the sample of mis-classification.
People's face only detects and when first two field picture or follow-up tracking failure, just carries out.At this moment choose people's face that face detection algorithm obtains surrounds frame and comes active appearance models (AAM) is carried out initialization.Because people's face detection block has only size and positional information, thus the x in can only active appearance models global affine transformation parameter, y direction translation parameters and whole zooming parameter carry out initialization, and other parameters then are set to zero.
2. based on the shape initialization of Feature Points Matching
In normal tracing process, the present invention utilizes Feature Points Matching to obtain the initial value of the people's face shape in the current frame image, can effectively improve the capturing ability to rapid movement.Basic thought based on the shape initialization algorithm of Feature Points Matching is: behind previous frame image iteration convergence; Choose some relatively notable attribute points at human face region; At current frame image these unique points are mated then, obtain the matching result of these unique points; Again according to the unique point of these couplings to estimating the form parameter of people's face in the current frame image.The concrete steps of algorithm are following:
1) on basis, previous frame image people face location, select unique point:
In the present invention, unique point is made up of two parts: a part is the more preassigned semantic point that has, and in the present invention, we have selected 30 semantic points, comprise some angle points and point on the eyes eyebrow nose face; Another part is the comparison notable attribute point of choosing according to the value of each gloomy matrix determinant in pixel sea, and the computing formula of extra large gloomy matrix is:
I wherein
Xx(u v) is pixel (u, square (or the weighted sum of squares of field point to play a smoothing effect) on every side of the x direction gradient of v) locating, I
Yy(u is v) with I
Xx(u, v) similar, I
Xy(u v) is that (u v) locates x and y direction gradient product (or the weighted product sum of field point) on every side to pixel.
Last selected unique point is shown in Fig. 2 a.
2) calculated characteristics point is established x=(x in the barycentric coordinates that belong to separately in the triangular plate; Y) be with
be in the triangle on summit a bit; Wherein subscript r, s, t are used to identify leg-of-mutton three summits, and then this unique point can be expressed as the weighted mean on Atria summit:
Its weight, β and γ are exactly these barycentric coordinates in triangle, and computing formula is:
After selecting unique point, calculate each unique point x
iBarycentric coordinates in its place triangular mesh are (c
I1, c
I2, c
I3), and the sequence number of an Atria summit, record place in shape can be used when calculating in the current frame image people's face shape initial value in the back, i=1, and 2,3.....M, M are unique point quantity.
3) carry out Feature Points Matching at current frame image
The present invention adopts the method for piece template matches to carry out the coupling of unique point.Concrete grammar is in the previous frame image, is the center with the unique point, takes out a pocket T, on current frame image, is that a bigger area I is selected at the center with this unique point also then, and area I should be able to inclusion region T.Subregion T in the previous frame image is slided in the current frame image area I, calculate the normalized correlation coefficient of lap.With the maximum position of normalized correlation coefficient as with the previous frame image on the position that is complementary of unique point.In the present invention, the size of regional T is taken as 9 * 9 (empirical values).The two frame peak excursions that the size of area I is followed the tracks of are as required confirmed.In order to accelerate matching speed, can use by thick to smart multiresolution framework.
What Fig. 2 b showed is the example as a result of a Feature Points Matching.Can find out that from this figure most of unique point all correct match has arrived, have only the fraction unique point not match.Fail the reason of correct match have multiple, such as since the variation of attitude cause local grain generation marked change even be blocked.Fig. 2 c is the original shape of the active appearance models that obtains according to Feature Points Matching.Can find out from this figure,, thereby obtain more satisfactory active appearance models original shape in view of the above owing to Feature Points Matching gets better.
4) estimate people's face shape according to the unique point of coupling
Then we can utilize the unique point of these couplings and information such as barycentric coordinates that the front obtains to roughly estimate people's face shape of current frame image to suppose in current frame image, to have M unique point
that matches.This can realize through minimizing following formula:
W wherein
iNormalized correlation coefficient when being coupling,
To be i unique point be mapped to when the given form parameter p coordinate position on the current frame image,
Be i the corresponding average face key point coordinate in shape in triangle 3 summits, unique point place, ρ (, r) be a robust error function.We can use Gauss-newton (Gauss-Newton) algorithm to carry out iterative optimum shape parameter p
0ρ (, definition r) is following:
Wherein r is for putting the letter radius.
3. human face posture and action parameter extract
The present invention uses two and three dimensions deformable faceform's active appearance models algorithm to obtain people's face 3 d pose and face action parameter.Two and three dimensions deformable faceform's active appearance models algorithm can be regarded the two-dimentional active appearance models algorithm that has added the 3 d human face mesh model constraint in essence as.
3.1 merging the multiband texture of gray scale and marginal information expresses.
The present invention proposes a kind of multiband texture that merges gray scale and marginal information and express, to improve the active appearance models algorithm to the adaptive faculty of illumination variation and to not meeting the generalization ability of people's face.Its basic thought is that first wave filter with 3 quadratures carries out filtering to image, then filtered 2 width of cloth gradient images is done a square processing, to remove directional information, at last three width of cloth images is done (Sigmoid) normalization processing of S shape.The multiband texture that specifically can adopt following steps to obtain fusion gray scale and marginal information is expressed:
Wherein the sigmoid function definition is as follows:
Wherein
is the average of m in the whole area-of-interest; The effect of this function is to approach 0 value being transformed to much smaller than the input value of average
, approaches 1 value and be transformed to the input value much larger than average
.When this function is applied to the normalization edge image, can strengthen real edge image effectively, suppress possible noise simultaneously.
When calculating the multiband texture expression of merging gray scale and marginal information in the above, we have done square operation to the gradient of x direction and y direction.This operation has the benefit of two aspects:
1) effectively strengthens possible edge;
2) directional information of removal edge gradient, a preserving edge strength information.When not knowing that background is bright or darker than people face, it is necessary removing gradient direction information.In addition; Here be
and
to be done Sigmoid normalization respectively handle, calculated amount is less relatively.
There is any need to prove in addition; Here during compute gradient, need earlier input picture is deformed on the average shape, obtain the irrelevant facial image of shape; And then on the irrelevant facial image of shape compute gradient, with the gradient that guarantees to calculate in the image irrelevant to rotation direction.
Fig. 3 a to Fig. 3 c has shown the average face image of this multiband texture model.Wherein Fig. 3 a is the average face image of gray scale wave band, and Fig. 3 b and Fig. 3 c are respectively the average face images of x and y direction gradient wave band.As can be seen from the figure, the information of these three wave bands has very strong complementarity, and Fig. 3 a characterizes overall intensity and distributes, and Fig. 3 b characterizes x direction coboundary and distributes, and Fig. 3 c characterizes y direction coboundary and distributes.Simultaneously owing to used square operation and the operation of Sigmoid function normalization, the essential marginal texture information of people's face has been contained in the multiband texture expression of merging gray scale and marginal information, filtering nonessential interfere information.
3.2 three-dimensional face model and imaging model.
In the present invention, used deformable three-dimensional face model (Candide-3) is a three-dimensional model that is called Kan Daide-3 (Candide-3).This model not only can suitably be out of shape according to different people's face shapes approaching personalized human face, and can produce corresponding deformation according to face action, is very suitable for human face action and describes and follow the tracks of.
The shape g of 3-d deformable face wire frame model can be with each apex coordinate P of three-dimensional model
i=(x
i, y
i, z
i)
TThe vector that is formed by connecting is represented:
g=(x
1,y
1,z
1,x
2,y
2,z
2,…,x
n,y
n,z
n)
T,
Wherein n is a grid vertex quantity, (x
i, y
i, z
i)
TBe i grid vertex P
iThree-dimensional coordinate, T representing matrix transposition.The three-dimensional coordinate here is the coordinate figure in model self local coordinate system, for people's face shape of propertyization one by one, can use following shape to generate:
Wherein
is the average face shape; S and A are respectively change in shape and action transformation matrices; The all corresponding a kind of independently changing pattern of each row of matrix, σ and α are respectively change in shape and action variation factor vector.Therefore, S σ has described grid model to the different variation of people's face on global shape, like whole fat or thin of people's face, distance between two and the position of eyes eyebrow nose mouth in people's face etc.A α describes the variation of the mesh shape that face action causes, as opens one's mouth, and lifts eyebrow etc.In addition, we suppose that it is separate that change in shape changes with action.
The present invention has adopted complete perspective projection model to describe imaging process, and is as shown in Figure 4.Each coordinate system is the cartesian coordinate system that satisfies the right-hand rule among the figure.
If the Candide-3 faceform is last 1 P
o=(X
o, Y
o, Z
o)
T, coordinate is P under camera coordinates system after rotating R, translation T rigid body translation
c=(X
c, Y
c, Z
c)
T:
Point P
cThrough complete perspective projection 1 P to the plane of delineation
i, establishing this some coordinate under camera coordinates system is P
i=(X
i, Y
i, Z
i)
T, then the relation between these two coordinates is:
Wherein f is the digital focal length of camera, can calculate according to the resolution of entire image and the visual angle size of camera.
Some P on the plane of delineation
iCoordinate under image coordinate system can obtain through computes:
Wherein W and H are respectively the width and the height of image.Because y axle positive dirction typically refers to downwards under image coordinate system, so following formula is to Y
iDone individual inversion operation.In addition since the initial point of image coordinate system usually in the upper left corner of image, so x, the y coordinate adds respectively and equals image half-breadth and half high skew.
3.3 active appearance models algorithm based on the two and three dimensions deformable faceform of Kan Daide-3 three-dimensional face model.
The proposition of initial two and three dimensions active appearance models algorithm is in order to improve the performance of two-dimentional active appearance models algorithm; Force a constraint promptly for two-dimentional active appearance models, make that the two-dimensional shapes of utilizing two-dimentional global affine transformation parameter and form parameter to generate is legal.Here, legal implication is to have deformation parameter in effective three-dimensional rotation and translation parameters and the three-dimensional face model, and the feasible plane of delineation gained two-dimensional shapes that projects to is consistent with the shape that the two-dimensional shapes model generates.On the mathematics, this constraint can be expressed as:
minE
m=min||s′(p)-P(Q(g′(σ,α)))||
2,
G (σ wherein; α) expression Kan Daide-3 people face shape; Q (x) expression is done the three-dimensional rigid body conversion that comprises rotation matrix R and translation vector T to the three-dimensional vertices vector x; P representes that each three-dimensional vertices among the three-dimensional vertices vector Q (x) is projected to the plane of delineation obtains its pixel coordinate under image coordinate system, the two-dimensional shapes that s (p) expression generates according to the two-dimensional shapes parameter p, the part of subscript ' whole vector of expression.Because the summit is not one to one on the key point of two-dimensional shapes and the three-dimensional face grid, have only the sub-fraction ability corresponding, subscript ' be exactly this part on the expression ability correspondence.
Above-mentioned bound term is added in the objective function of original two dimensional active appearance models, promptly obtain so-called two and three dimensions deformable faceform's active appearance models, that is:
A wherein
0Be average face outward appearance, A
iFor with preceding n people's face outward appearance base vector, λ
iBe apparent coefficient, W (p) is a coordinate transform function, w
mBe the weight of 3D shape constraint portions, w in the present invention
m=0.1.
Optimization for two and three dimensions deformable faceform's active appearance models is found the solution, and still can use the fast algorithm in the two-dimentional active appearance models.The first step, the span (A in the subspace
i)
⊥The optimization aim function:
F wherein
I(p; P; σ; α) the site error of i obligatory point of expression on x, y direction.Here succinct in order to write, with the synthetic expression formula of the site error on x, the y both direction.The method that second step was found the solution apparent parameter lambda is with two-dimentional active appearance models.
In order to represent that conveniently we are linked to be a vectorial q=(p with all unknown numbers; P; σ; α), then can accomplish through following two step iteration the optimization of following formula:
The first step: calculating parameter renewal amount:
Δ p wherein
SDBe the steepest decline figure of two dimension part, H
3DBe the extra large gloomy matrix of two and three dimensions active appearance models, F
iBe each bound term, the concrete definition as follows:
Second step: parameter update:
Utilize counter-rotating synthetic method undated parameter p, use addition rule undated parameter P, σ and α.
Need to prove, in above-mentioned arthmetic statement,, ignored to F in order to describe and to be concise in expression
i(p; P; σ; Necessity correction that need do when α) carrying out the single order Taylor expansion, i.e. F
i(p; P; σ; Increment problem about p α) need be come from the mapping of the increment on the average shape.
In superincumbent two and three dimensions deformable faceform's the active appearance models fitting algorithm, can try to achieve human face posture parameter, personalized form parameter σ and face action parameter alpha simultaneously.Under the tracking environmental of reality, confirm personalized form parameter σ in the tracking initiation stage usually, remain unchanged at follow-up tracking phase, only if just the tracking failure needs calculating once more in the time of need reinitializing grid model.In follow-up tracing process, only need tracks facial action parameter α.This can bring the benefit of two aspects:
1) reduced the number of parameters that to find the solution when following the tracks of, and then improved tracking velocity, again the necessary operation parameter can be provided simultaneously.
2) can be to the form generation of the two-dimentional active appearance models stronger constraint strength of fixing personalized shape coefficient.
Top description is to be used to realize the present invention and embodiment, and therefore, scope of the present invention should not described by this and limit.It should be appreciated by those skilled in the art,, all belong to claim of the present invention and come restricted portion in any modification that does not depart from the scope of the present invention or local replacement.
Claims (8)
1. one kind to human face posture and the action method of following the tracks of, and comprises step:
Step S1: from video flowing, extract by two field picture, carry out people's face during to input video first two field picture or tracking failure and detect, obtain people's face and surround frame;
Step S2: behind previous frame image iteration convergence; During normal the tracking; Compare the notable attribute point for some textural characteristics of human face region in the previous frame image, find these corresponding with it unique points to mate, obtain the matching result of these unique points at current frame image;
Step S3: according to people's face encirclement frame or Feature Points Matching result the shape of active appearance models is carried out initialization, obtain the people's face shape initial value in the current frame image;
Step S4: use the counter-rotating composition algorithm that active appearance models is carried out match, obtain people's face 3 d pose and face action parameter.
2. according to the said method that human face posture and action are followed the tracks of of claim 1, it is characterized in that also comprising step: use based on the method for adaptive boosting and carry out the detection of people's face, obtain position and the size of people's face in image.
3. according to the said method that human face posture and action are followed the tracks of of claim 1; It is characterized in that: said unique point is to calculate the value of each gloomy matrix determinant in pixel sea earlier, chooses the relatively large pixel of extra large gloomy matrix determinant then as unique point.
4. according to the said method that human face posture and action are followed the tracks of of claim 1, it is characterized in that: said previous frame image and said current frame image are used multi-resolution representation respectively, in order to improve Feature Points Matching speed.
5. according to the said method that human face posture and action are followed the tracks of of claim 1; It is characterized in that; Said Feature Points Matching is based on the multiresolution framework of image, the method that adopts piece coupling again current frame image find with the previous frame image in the point that is complementary of unique point.
6. according to the said method that human face posture and action are followed the tracks of of claim 1; It is characterized in that: said active appearance models uses the multiband texture that merges gray scale and marginal information to express, in order to improve the active appearance models algorithm to the adaptive faculty of illumination variation and to not meeting the generalization ability of people's face.
7. according to claim 1 is said human face posture and the action method of following the tracks of is characterized in that said based on the Feature Points Matching result shape of active appearance models to be carried out initialized treatment step following:
Step S31: in the previous frame image, on the basis, people's face location, select unique point;
Step S32: calculated characteristics point is in the barycentric coordinates that belong to separately in the triangular plate;
Step S33: carry out Feature Points Matching at current frame image;
Step S34: the unique point according to coupling is estimated people's face shape.
8. according to the said method that human face posture and action are followed the tracks of of claim 1, it is characterized in that: said active appearance models comprises two-dimension human face shape, two-dimension human face texture model, 3-d deformable faceform and imaging model; Said imaging model adopts complete perspective projection, so that obtain human face posture parameter accurately.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2010102780635A CN102402691A (en) | 2010-09-08 | 2010-09-08 | Method for tracking gestures and actions of human face |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2010102780635A CN102402691A (en) | 2010-09-08 | 2010-09-08 | Method for tracking gestures and actions of human face |
Publications (1)
Publication Number | Publication Date |
---|---|
CN102402691A true CN102402691A (en) | 2012-04-04 |
Family
ID=45884881
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN2010102780635A Pending CN102402691A (en) | 2010-09-08 | 2010-09-08 | Method for tracking gestures and actions of human face |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN102402691A (en) |
Cited By (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102750527A (en) * | 2012-06-26 | 2012-10-24 | 浙江捷尚视觉科技有限公司 | Long-time stable human face detection and tracking method in bank scene and long-time stable human face detection and tracking device in bank scene |
CN102867174A (en) * | 2012-08-30 | 2013-01-09 | 中国科学技术大学 | Method and device for positioning human face features |
CN102867173A (en) * | 2012-08-28 | 2013-01-09 | 华南理工大学 | Human face recognition method and system thereof |
CN102880866A (en) * | 2012-09-29 | 2013-01-16 | 宁波大学 | Method for extracting face features |
CN103400119A (en) * | 2013-07-31 | 2013-11-20 | 南京融图创斯信息科技有限公司 | Face recognition technology-based mixed reality spectacle interactive display method |
CN103870824A (en) * | 2014-03-28 | 2014-06-18 | 海信集团有限公司 | Method and device for capturing face in face detecting and tracking process |
CN104392447A (en) * | 2014-11-28 | 2015-03-04 | 西南科技大学 | Image matching method based on gray scale gradient |
CN105447462A (en) * | 2015-11-20 | 2016-03-30 | 小米科技有限责任公司 | Facial pose estimation method and device |
CN105989326A (en) * | 2015-01-29 | 2016-10-05 | 北京三星通信技术研究有限公司 | Method and device for determining three-dimensional position information of human eyes |
CN106228113A (en) * | 2016-07-12 | 2016-12-14 | 电子科技大学 | Human face characteristic point quick alignment method based on AAM |
CN106295511A (en) * | 2016-07-26 | 2017-01-04 | 北京小米移动软件有限公司 | Face tracking method and device |
CN106462738A (en) * | 2014-05-20 | 2017-02-22 | 埃西勒国际通用光学公司 | Method for constructing a model of the face of a person, method and device for posture analysis using such a model |
CN106778474A (en) * | 2016-11-14 | 2017-05-31 | 深圳奥比中光科技有限公司 | 3D human body recognition methods and equipment |
CN107316029A (en) * | 2017-07-03 | 2017-11-03 | 腾讯科技(深圳)有限公司 | A kind of live body verification method and equipment |
CN107633526A (en) * | 2017-09-04 | 2018-01-26 | 腾讯科技(深圳)有限公司 | A kind of image trace point acquisition methods and equipment, storage medium |
CN107992825A (en) * | 2017-12-01 | 2018-05-04 | 青岛海尔智能家电科技有限公司 | A kind of method and system of the recognition of face based on augmented reality |
CN108229246A (en) * | 2016-12-14 | 2018-06-29 | 上海交通大学 | Real-time three-dimensional human face posture method for tracing based on vehicle computing machine platform |
CN108345821A (en) * | 2017-01-24 | 2018-07-31 | 成都理想境界科技有限公司 | Face tracking method and apparatus |
CN108510520A (en) * | 2018-02-13 | 2018-09-07 | 视辰信息科技(上海)有限公司 | A kind of image processing method, device and AR equipment |
CN108875506A (en) * | 2017-11-17 | 2018-11-23 | 北京旷视科技有限公司 | Face shape point-tracking method, device and system and storage medium |
CN109711304A (en) * | 2013-05-21 | 2019-05-03 | 深圳市腾讯计算机系统有限公司 | A kind of man face characteristic point positioning method and device |
WO2019097285A1 (en) * | 2017-08-31 | 2019-05-23 | Banuba Limited | Computer-implemented methods and computer systems for real-time detection of human's emotions from visual recordings |
CN109859322A (en) * | 2019-01-22 | 2019-06-07 | 广西大学 | A kind of spectrum posture moving method based on deformation pattern |
CN109874021A (en) * | 2017-12-04 | 2019-06-11 | 腾讯科技(深圳)有限公司 | Living broadcast interactive method, apparatus and system |
CN112464918A (en) * | 2021-01-27 | 2021-03-09 | 昆山恒巨电子有限公司 | Body-building action correcting method and device, computer equipment and storage medium |
CN114863506A (en) * | 2022-03-18 | 2022-08-05 | 珠海优特电力科技股份有限公司 | Method, device and system for verifying access permission and identity authentication terminal |
WO2023087891A1 (en) * | 2021-11-18 | 2023-05-25 | 中兴通讯股份有限公司 | Real-time facial image driving method and apparatus, electronic device, and storage medium |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1811793A (en) * | 2006-03-02 | 2006-08-02 | 复旦大学 | Automatic positioning method for characteristic point of human faces |
CN1866271A (en) * | 2006-06-13 | 2006-11-22 | 北京中星微电子有限公司 | AAM-based head pose real-time estimating method and system |
-
2010
- 2010-09-08 CN CN2010102780635A patent/CN102402691A/en active Pending
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1811793A (en) * | 2006-03-02 | 2006-08-02 | 复旦大学 | Automatic positioning method for characteristic point of human faces |
CN1866271A (en) * | 2006-06-13 | 2006-11-22 | 北京中星微电子有限公司 | AAM-based head pose real-time estimating method and system |
Non-Patent Citations (1)
Title |
---|
周明才: "表情互动中的若干关键问题研究", 《HTTP://LIB.IA.AC.CN/DLIB/LIST.ASP?LANG=GB&TYPE=&DOCGROUPID=10&DOCID=10071》 * |
Cited By (48)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102750527A (en) * | 2012-06-26 | 2012-10-24 | 浙江捷尚视觉科技有限公司 | Long-time stable human face detection and tracking method in bank scene and long-time stable human face detection and tracking device in bank scene |
CN102750527B (en) * | 2012-06-26 | 2015-08-19 | 浙江捷尚视觉科技股份有限公司 | The medium-term and long-term stable persona face detection method of a kind of bank scene and device |
CN102867173A (en) * | 2012-08-28 | 2013-01-09 | 华南理工大学 | Human face recognition method and system thereof |
CN102867173B (en) * | 2012-08-28 | 2015-01-28 | 华南理工大学 | Human face recognition method and system thereof |
CN102867174B (en) * | 2012-08-30 | 2016-01-20 | 中国科学技术大学 | A kind of human face characteristic positioning method and device |
CN102867174A (en) * | 2012-08-30 | 2013-01-09 | 中国科学技术大学 | Method and device for positioning human face features |
CN102880866A (en) * | 2012-09-29 | 2013-01-16 | 宁波大学 | Method for extracting face features |
CN102880866B (en) * | 2012-09-29 | 2014-12-17 | 宁波大学 | Method for extracting face features |
CN109711304A (en) * | 2013-05-21 | 2019-05-03 | 深圳市腾讯计算机系统有限公司 | A kind of man face characteristic point positioning method and device |
CN109711304B (en) * | 2013-05-21 | 2022-06-14 | 深圳市腾讯计算机系统有限公司 | Face feature point positioning method and device |
CN103400119A (en) * | 2013-07-31 | 2013-11-20 | 南京融图创斯信息科技有限公司 | Face recognition technology-based mixed reality spectacle interactive display method |
CN103400119B (en) * | 2013-07-31 | 2017-02-15 | 徐坚 | Face recognition technology-based mixed reality spectacle interactive display method |
CN103870824A (en) * | 2014-03-28 | 2014-06-18 | 海信集团有限公司 | Method and device for capturing face in face detecting and tracking process |
CN103870824B (en) * | 2014-03-28 | 2017-10-20 | 海信集团有限公司 | A kind of face method for catching and device during Face datection tracking |
CN106462738B (en) * | 2014-05-20 | 2020-10-09 | 依视路国际公司 | Method for constructing a model of a person's face, method and apparatus for analyzing a pose using such a model |
US10380411B2 (en) | 2014-05-20 | 2019-08-13 | Essilor International | Method for constructing a model of the face of a person, method and device for posture analysis using such a model |
CN106462738A (en) * | 2014-05-20 | 2017-02-22 | 埃西勒国际通用光学公司 | Method for constructing a model of the face of a person, method and device for posture analysis using such a model |
CN104392447A (en) * | 2014-11-28 | 2015-03-04 | 西南科技大学 | Image matching method based on gray scale gradient |
CN104392447B (en) * | 2014-11-28 | 2017-10-13 | 西南科技大学 | A kind of image matching method based on shade of gray |
CN105989326A (en) * | 2015-01-29 | 2016-10-05 | 北京三星通信技术研究有限公司 | Method and device for determining three-dimensional position information of human eyes |
CN105989326B (en) * | 2015-01-29 | 2020-03-03 | 北京三星通信技术研究有限公司 | Method and device for determining three-dimensional position information of human eyes |
CN105447462A (en) * | 2015-11-20 | 2016-03-30 | 小米科技有限责任公司 | Facial pose estimation method and device |
CN105447462B (en) * | 2015-11-20 | 2018-11-20 | 小米科技有限责任公司 | Face pose estimation and device |
CN106228113A (en) * | 2016-07-12 | 2016-12-14 | 电子科技大学 | Human face characteristic point quick alignment method based on AAM |
CN106295511A (en) * | 2016-07-26 | 2017-01-04 | 北京小米移动软件有限公司 | Face tracking method and device |
CN106295511B (en) * | 2016-07-26 | 2019-05-21 | 北京小米移动软件有限公司 | Face tracking method and device |
CN106778474A (en) * | 2016-11-14 | 2017-05-31 | 深圳奥比中光科技有限公司 | 3D human body recognition methods and equipment |
CN108229246A (en) * | 2016-12-14 | 2018-06-29 | 上海交通大学 | Real-time three-dimensional human face posture method for tracing based on vehicle computing machine platform |
CN108345821A (en) * | 2017-01-24 | 2018-07-31 | 成都理想境界科技有限公司 | Face tracking method and apparatus |
CN108345821B (en) * | 2017-01-24 | 2022-03-08 | 成都理想境界科技有限公司 | Face tracking method and device |
CN107316029B (en) * | 2017-07-03 | 2018-11-23 | 腾讯科技(深圳)有限公司 | A kind of living body verification method and equipment |
CN107316029A (en) * | 2017-07-03 | 2017-11-03 | 腾讯科技(深圳)有限公司 | A kind of live body verification method and equipment |
WO2019097285A1 (en) * | 2017-08-31 | 2019-05-23 | Banuba Limited | Computer-implemented methods and computer systems for real-time detection of human's emotions from visual recordings |
WO2019042419A1 (en) * | 2017-09-04 | 2019-03-07 | 腾讯科技(深圳)有限公司 | Image tracking point acquisition method and device, and storage medium |
US11164323B2 (en) * | 2017-09-04 | 2021-11-02 | Tencent Technology (Shenzhen) Company Limited | Method for obtaining image tracking points and device and storage medium thereof |
CN107633526A (en) * | 2017-09-04 | 2018-01-26 | 腾讯科技(深圳)有限公司 | A kind of image trace point acquisition methods and equipment, storage medium |
CN108875506A (en) * | 2017-11-17 | 2018-11-23 | 北京旷视科技有限公司 | Face shape point-tracking method, device and system and storage medium |
CN108875506B (en) * | 2017-11-17 | 2022-01-07 | 北京旷视科技有限公司 | Face shape point tracking method, device and system and storage medium |
CN107992825A (en) * | 2017-12-01 | 2018-05-04 | 青岛海尔智能家电科技有限公司 | A kind of method and system of the recognition of face based on augmented reality |
CN109874021A (en) * | 2017-12-04 | 2019-06-11 | 腾讯科技(深圳)有限公司 | Living broadcast interactive method, apparatus and system |
WO2019157922A1 (en) * | 2018-02-13 | 2019-08-22 | 视辰信息科技(上海)有限公司 | Image processing method and device and ar apparatus |
CN108510520B (en) * | 2018-02-13 | 2019-03-08 | 视辰信息科技(上海)有限公司 | A kind of image processing method, device and AR equipment |
CN108510520A (en) * | 2018-02-13 | 2018-09-07 | 视辰信息科技(上海)有限公司 | A kind of image processing method, device and AR equipment |
CN109859322A (en) * | 2019-01-22 | 2019-06-07 | 广西大学 | A kind of spectrum posture moving method based on deformation pattern |
CN109859322B (en) * | 2019-01-22 | 2022-12-06 | 广西大学 | Spectral attitude migration method based on deformation graph |
CN112464918A (en) * | 2021-01-27 | 2021-03-09 | 昆山恒巨电子有限公司 | Body-building action correcting method and device, computer equipment and storage medium |
WO2023087891A1 (en) * | 2021-11-18 | 2023-05-25 | 中兴通讯股份有限公司 | Real-time facial image driving method and apparatus, electronic device, and storage medium |
CN114863506A (en) * | 2022-03-18 | 2022-08-05 | 珠海优特电力科技股份有限公司 | Method, device and system for verifying access permission and identity authentication terminal |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN102402691A (en) | Method for tracking gestures and actions of human face | |
Zuffi et al. | Lions and tigers and bears: Capturing non-rigid, 3d, articulated shape from images | |
Dai et al. | A 3d morphable model of craniofacial shape and texture variation | |
CN100416612C (en) | Video flow based three-dimensional dynamic human face expression model construction method | |
Burl et al. | A probabilistic approach to object recognition using local photometry and global geometry | |
CN101968846B (en) | Face tracking method | |
US20220358770A1 (en) | Scene reconstruction in three-dimensions from two-dimensional images | |
CN101499128B (en) | Three-dimensional human face action detecting and tracing method based on video stream | |
US9298257B2 (en) | Apparatus and method for controlling avatar using expression control point | |
CN102999942B (en) | Three-dimensional face reconstruction method | |
Blake et al. | Active contours: the application of techniques from graphics, vision, control theory and statistics to visual tracking of shapes in motion | |
Balan et al. | Detailed human shape and pose from images | |
CN101964064B (en) | Human face comparison method | |
Agudo et al. | Simultaneous pose and non-rigid shape with particle dynamics | |
Bascle et al. | Stereo matching, reconstruction and refinement of 3D curves using deformable contours | |
CN104036546A (en) | Method for carrying out face three-dimensional reconstruction at any viewing angle on basis of self-adaptive deformable model | |
CN102376100A (en) | Single-photo-based human face animating method | |
CN102654903A (en) | Face comparison method | |
CN103733226A (en) | Fast articulated motion tracking | |
Pan et al. | Sketch-based skeleton-driven 2D animation and motion capture | |
CN105893984A (en) | Face projection method for facial makeup based on face features | |
CN103714556A (en) | Moving target tracking method based on pyramid appearance model | |
Bao et al. | High-quality face capture using anatomical muscles | |
Chen et al. | Single and sparse view 3d reconstruction by learning shape priors | |
Fayad et al. | Non-rigid Structure from Motion using Quadratic Deformation Models. |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C02 | Deemed withdrawal of patent application after publication (patent law 2001) | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20120404 |