CN104376333A - Facial expression recognition method based on random forests - Google Patents
Facial expression recognition method based on random forests Download PDFInfo
- Publication number
- CN104376333A CN104376333A CN201410503590.XA CN201410503590A CN104376333A CN 104376333 A CN104376333 A CN 104376333A CN 201410503590 A CN201410503590 A CN 201410503590A CN 104376333 A CN104376333 A CN 104376333A
- Authority
- CN
- China
- Prior art keywords
- mrow
- msup
- facial expression
- math
- aam
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 230000008921 facial expression Effects 0.000 title claims abstract description 100
- 238000000034 method Methods 0.000 title claims abstract description 49
- 238000007637 random forest analysis Methods 0.000 title claims abstract description 43
- 238000006073 displacement reaction Methods 0.000 claims abstract description 30
- 238000012549 training Methods 0.000 claims abstract description 30
- 238000013145 classification model Methods 0.000 claims abstract description 14
- 238000004422 calculation algorithm Methods 0.000 claims description 19
- 239000013598 vector Substances 0.000 claims description 13
- 230000003287 optical effect Effects 0.000 claims description 11
- 238000003066 decision tree Methods 0.000 claims description 10
- 238000000513 principal component analysis Methods 0.000 claims description 9
- 238000012545 processing Methods 0.000 claims description 8
- 230000009467 reduction Effects 0.000 claims description 3
- 230000014509 gene expression Effects 0.000 abstract description 43
- 230000008569 process Effects 0.000 abstract description 16
- 238000000605 extraction Methods 0.000 abstract description 9
- 230000000694 effects Effects 0.000 abstract description 6
- 238000012706 support-vector machine Methods 0.000 abstract description 4
- 230000007935 neutral effect Effects 0.000 description 7
- 230000001815 facial effect Effects 0.000 description 6
- 238000004364 calculation method Methods 0.000 description 4
- 210000001097 facial muscle Anatomy 0.000 description 4
- 210000000056 organ Anatomy 0.000 description 3
- 238000011160 research Methods 0.000 description 3
- 206010063659 Aversion Diseases 0.000 description 2
- 241000288110 Fulica Species 0.000 description 2
- 230000009471 action Effects 0.000 description 2
- 230000006399 behavior Effects 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 230000007547 defect Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 210000004709 eyebrow Anatomy 0.000 description 2
- 238000005286 illumination Methods 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 238000002372 labelling Methods 0.000 description 2
- 210000003205 muscle Anatomy 0.000 description 2
- 238000013528 artificial neural network Methods 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 239000000470 constituent Substances 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 238000004377 microelectronic Methods 0.000 description 1
- 230000035936 sexual power Effects 0.000 description 1
- 108010047303 von Willebrand Factor Proteins 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/174—Facial expression recognition
- G06V40/176—Dynamic expression
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a facial expression recognition method based on random forests. The facial expression recognition method based on the random forests comprises the step of extraction of a displacement feature of an AAM, the step of extraction of AUs in a facial expression sequence, the step of training of a facial expression classification model and the step of facial expression recognition. According to the facial expression recognition method, the novel AAM displacement feature is provided to be used for training and learning the AUs, and finally facial expression recognition is carried out by depending on the AUs. Compared with other feature representations in identification of the same classification, the facial expression recognition method based on the random forests better describes expression information and changing process information contained in the expression sequence. The random forests are used for facial expression recognition for the first time, and the random forests in the method have a better classified recognition effect in the field compared with a frequently used support vector machine (SVM) method at present. For the aspect of CK and AU recognition of databases, the facial expression recognition method based on the random forests can achieve a perfect recognition effect.
Description
Technical Field
The invention relates to a facial expression recognition method, in particular to a facial expression recognition method based on a random forest.
Background
The human face expression is used as important non-language information, so that very rich information can be provided, and language communication is effectively supplemented. Many times, one's joy, anger, sadness and fun is directly reflected in facial expressions. The human face expression can be automatically identified by the computer, so that the computer can better interact with people, and humanized service is provided. Facial expression recognition can be widely applied to various fields such as medical treatment, human-computer interaction (HCI), video games, data-driven animation and the like. Facial expression recognition has attracted more and more research attention over the past twenty years, and a variety of advanced algorithms have been proposed[1]. Facial expressions have some of the following characteristics: (1) is delicate. This is reflected in, on the one hand, the very similarity between certain expressions; on the other hand, the facial expression contains abundant micro expression and is not easy to perceive and capture. (2) The facial expression is vanishing instantly. (3) The facial expression is complicated and changeable. From the perspective of organs, the facial expression is mainly reflected in the eyebrows, eyes, nose, mouth and other places, and each organ has multiple states such as open eyes, closed eyes, squinting and the like, and different states of different organs are combined to form different facial expressions. (4) Is easily influenced by factors such as posture, illumination and the like. Facial expressions are closely related to facial poses, and significant differences may occur for the same expression, viewed from the front, side, or other angles, respectively. In addition, the influence of different lighting on facial expressions is very obvious, for example, strong lighting makes the whole face image white, or darker light can cause the face in the image to be blurred, and the expressions are more difficult to distinguish, which are common problems and challenges in the field of biometric identification. Therefore, most of the facial expression recognition research still stays in the experimental exploration stage at present, and widely adopted facial expression databases are established under certain limiting conditions, particularly data collected under controlled environments such as illumination, angles and the like. Therefore, facial expression recognition remains a significant difficulty and challengeAnd (4) the problem of sexual performance.
The invention of the "facial expression recognition method and system, and the training method and system of expression classifier" of the invention of the treble national duty of the beijing zhongxing microelectronics ltd is patented and approved by the chinese national intellectual property office in 12.21.2009, and is published in 11.03.2010, with the publication numbers: CN 101877056a, this method adopts Hidden Markov Model (HMM), and the overall performance is poor. Secondly, in the training process of the system, the neutral expression and the weak and small expression in the facial expression sequence are adopted, the information contained in the weak and small expression is not as much as the information contained in the exaggerated expression, and therefore partial information can be lost. In addition, not only the fusion binary image but also the gray level image are used in the process of extracting the features, and the calculation process is complex.
The invention of 'facial expression recognition method based on feature point vector and texture deformation energy parameter' of the major university of aerospace, xu xia li, chenlijiang and wangxiang kan applies for patent and obtains approval from the intellectual property office of china 10 and 17 days 2012, and is published in 27.02 and 2013 with the publication numbers as follows: CN 102945361A; feature point positioning is respectively carried out on neutral expressions and terminal expressions (exaggerated expressions) of the facial expression sequence by using an AAM tool in OPENCV, 26 feature points are selected to form feature point vectors, and the feature point vectors are divided into Euclidean distances among the feature points and included angles of connecting lines. And establishing a feature block according to the feature points, calculating a texture deformation energy coefficient matrix, and obtaining a final texture deformation energy parameter by using PCA. The three are combined to be used as training data of the RBF neural network, and finally face expression recognition is achieved. Although the AAM feature includes features of both shape and texture, it is not directly related to the facial expression, and the effect of directly performing facial expression recognition using AAM as a feature is not ideal. And in the extraction process of the features, redundancy is removed from three aspects of the constituent features, resulting in several problems. On one hand, the process of removing redundancy inevitably removes a part of useful information, on the other hand, the redundant information is sometimes necessary in the experimental process, and the proper redundancy can be beneficial to the experimental result, such as the calculation of the HOG characteristic can be well embodied.
The invention of von willebrand and Pengzheng industry of northwest industry university, namely the facial expression recognition method based on facial action unit combination characteristics, applies for patent and is approved by the intellectual property office of China in 2012 12 and 21 and is published in 2013 in 04 and 24, and the publication numbers are as follows: CN 103065122A; the face database adopted in the patent is self-constructed, and the rationality and the effectiveness of the face database are questionable. Second, a single image is used in the algorithm rather than a sequence of images. In fact, the expression is a continuous change process, and the single image contains insufficient expression information. In addition, in the training process of the algorithm, manual work (10 researchers) is adopted for AU labeling, and the AU labeling needs experts in the aspect to ensure high accuracy, so that errors are inevitable. Although the patent states that it defines a final adopted threshold, its reliability is poor. Moreover, LBP does not directly link with AU, and the learning of AU combination is carried out by using LBP, and the reasonability is worth exploring.
The facial expression recognition can be applied to various fields, such as automatic smile capturing and photographing carried by a camera, natural interaction between an intelligent robot and a person, capturing of painful expressions of patients in a medical treatment process so as to timely rescue and the like. Although facial expression recognition has been studied for a long time and many solutions have been proposed, the results of practical application are not ideal and the recognition effect is still in the space of further improvement in the experimental research stage.
Disclosure of Invention
In view of the above disadvantages, the present invention aims to provide a new facial expression recognition algorithm, which performs 7 basic facial expression recognitions on the CK + database and obtains a higher recognition rate than the previous algorithms.
The invention aims at researching the facial expression of an image sequence. The core methods used are Random forest (Random forest) and Active Appearance Model (Active Appearance Model). The random forest is mainly composed of some decision trees, and the AAM is mainly used for describing shape and texture features.
The method comprises the steps of extracting AAM characteristic points of a first frame (neutral expression) from an expression sequence image (from the neutral expression to an exaggerated expression), and tracking through an LK algorithm to obtain the AAM characteristic points of a last frame (the exaggerated expression). And (3) taking the displacement of the AAM feature points as feature vectors, inputting Random forest (Random forest) for training, and identifying AUs contained in the expression. And the AUs are used as input and are transmitted to Random forest (Random forest) for new training to obtain an expression classification model, and finally, the model is used for facial expression recognition.
The specific technical scheme of the invention is as follows:
step 1 AAM Displacement feature extraction
AAM is based on the active Shape model ASM (active Shape model), originally proposed by Edwards, Cootes and Taylor. AAM includes two parts: a shape model and a texture model. The present invention uses shape models. The shape model is defined as the coordinates of n feature points:
s=(x1,y1,x2,y2,…,xn,yn)T
it can be changed into a linear model:
wherein p isiIs the shape parameter, siThe training data is obtained by Principal Component Analysis (PCA) (principal Component analysis) dimensionality reduction. However, for the purposes of calculation and tracking, the present invention does not use the second expression, but uses the first simple expression.
Step 1.1, selecting a first frame (neutral expression) of a facial expression image sequence, extracting AAM feature points of the first frame, and setting as:
s0=[x10,y10,x20,y20,...,xn0,yn0]
step 1.2, tracking the AAM characteristic points from the first frame to the last frame by using an LK tracking algorithm. The LK optical flow tracking algorithm employed by the present invention is an improved version of the classic LK algorithm proposed by j
Wherein,is the initial estimate at pyramid L level, which can be calculated from L-1 level:
gL=2(gL-1+dL-1)
the final optical flow can be expressed in the form:
wherein L ismIs the deepest level of the pyramid. Thus, dLThe solution can be solved as follows:
wherein, <math><mrow>
<msubsup>
<mi>η</mi>
<mi>k</mi>
<mi>L</mi>
</msubsup>
<mo>=</mo>
<msup>
<mrow>
<mo>(</mo>
<msup>
<mi>G</mi>
<mrow>
<mo>-</mo>
<mn>1</mn>
</mrow>
</msup>
<mo>)</mo>
</mrow>
<mi>L</mi>
</msup>
<msubsup>
<mi>b</mi>
<mi>k</mi>
<mi>L</mi>
</msubsup>
<mo>,</mo>
</mrow></math> while <math><mrow>
<msup>
<mrow>
<mo>(</mo>
<msup>
<mi>G</mi>
<mrow>
<mo>-</mo>
<mn>1</mn>
</mrow>
</msup>
<mo>)</mo>
</mrow>
<mi>L</mi>
</msup>
<mo>=</mo>
<munder>
<mi>Σ</mi>
<mrow>
<mi>x</mi>
<mo>∈</mo>
<mi>N</mi>
<mrow>
<mo>(</mo>
<mi>u</mi>
<mo>)</mo>
</mrow>
</mrow>
</munder>
<mo>▿</mo>
<msup>
<mi>I</mi>
<mi>L</mi>
</msup>
<msup>
<mrow>
<mo>(</mo>
<mo>▿</mo>
<msup>
<mi>I</mi>
<mi>L</mi>
</msup>
<mo>)</mo>
</mrow>
<mi>T</mi>
</msup>
<mo>,</mo>
</mrow></math> <math><mrow>
<msubsup>
<mi>b</mi>
<mi>k</mi>
<mi>L</mi>
</msubsup>
<mo>=</mo>
<munder>
<mi>Σ</mi>
<mrow>
<mi>x</mi>
<mo>∈</mo>
<mi>N</mi>
<mrow>
<mo>(</mo>
<mi>u</mi>
<mo>)</mo>
</mrow>
</mrow>
</munder>
<mo>▿</mo>
<msup>
<mi>I</mi>
<mi>L</mi>
</msup>
<mi>δ</mi>
<msubsup>
<mi>I</mi>
<mi>k</mi>
<mi>L</mi>
</msubsup>
<mo>,</mo>
</mrow></math> In both formulaeIL k(x)=IL(x)-JL(x+gL+vL)
Wherein v isLIs the optical flow vector of L layers, K is of each layer in the LK algorithmThe maximum number of iterations.
Through an LK optical flow tracking algorithm, obtaining AAM feature points of the last frame (exaggerated expression), and setting the AAM feature points at the moment as:
sn=[x1n,y1n,x2n,y2n,...,xnn,ynn]
step 1.3, according to the results obtained in step 1.1 and step 1.2, the displacement of the AAM characteristic point is calculated:
st=[x1n-x10,y1n-y10,x2n-x20,y2n-y20,...,xnn-xn0,ynn-yn0];
the obtained AAM displacement characteristic is obtained.
Step 2, AU extraction in facial expression sequence
Action units AU come from the facial behavior Coding system facs (facial Action Coding system), which was proposed by Paul Ekman and Wallace v.
And 2.1, taking the AAM displacement characteristics obtained in the step 1 as the input of the random forest.
Input AAM displacement characteristic data are firstly subjected to randomization processing to obtain K pieces of fractional data, then the K pieces of fractional data are transmitted into corresponding decision trees to obtain K results, and classification is carried out according to the following random forest definition to obtain a final classification result.
The definition of random forest is as follows:
where H (x) is the final fused classifier, K is the number of decision trees, hi(x) Represents a decision tree, Y is a class label, I (h)i(x) Y) indicates that x belongs to category Y.
Step 2.2 AAM feature point displacement reflects changes of facial expressions, such as eye closing, eye enlargement, mouth opening and the like, which are information represented by AU, and the two are consistent. Therefore, it is reasonable to input AAM displacement features into a random forest for training to obtain human face AU information in the sequence, and the feasibility is also verified by the final experimental result. Thus, the corresponding AUs contained in the sequence of facial expressions is finally obtained via step 2.1.
Step 3, training a facial expression classification model
Facial expressions are primarily combinations of facial muscle group movements, and AU is the movement of a block or group of facial muscles. Therefore, the combination of AUs represents the corresponding facial expression. The corresponding expression of the expression sequence can be obtained by analyzing the AU combination contained in the expression sequence.
And (3) taking AUs information of the facial expression sequence obtained in the step (2) as training data, and inputting the training data into the random forest again for training to obtain the facial expression classification model.
Step 4 facial expression recognition
The process of facial expression recognition is similar to the process of training a facial expression classification model. And (3) processing the facial expression to be classified according to the step 1 to obtain an AAM displacement characteristic vector, processing the AAM displacement characteristic vector in the step 2 to obtain AU information contained in the sequence, and finally inputting the AU information into the random forest facial expression classification model obtained in the step 3 to obtain a final facial expression recognition result.
The invention mainly aims at 7 basic facial expressions: anger, apprehension, aversion, fear, happiness, sadness and surprise.
The invention has the beneficial effects that: a new facial expression recognition algorithm is provided, which specifically comprises the following steps: (1) a new AAM displacement feature is proposed for training learning AUs and ultimately relying on AUs for facial expression recognition. It describes the expression information and change process information contained in an expression sequence better than other feature representations in the same kind of recognition. (2) The random forest is used for facial expression recognition for the first time, and compared with a Support Vector Machine (SVM) method which is commonly used in the past, the random forest has a better classification recognition effect in the field. (3) Aiming at the aspect of AU identification of CK + database, the automatic identification method provided by the invention can achieve perfect identification effect.
The method adopts a CK + database which is provided by field authority experts PatrickLucy and the like, is an international universal database researched in the aspect of face expression recognition at present, and also provides AUs, expression and other information of an image sequence in the database, so that the defects of the prior art III (publication number CN 103065122A) are avoided, and meanwhile, the method well avoids the defects of the prior art I (publication number CN 101877056A) and II (publication number CN 102945361A).
Drawings
Fig. 1 is a flowchart of a facial expression recognition method according to the present invention.
Fig. 2 is a diagram of the AAM displacement feature extraction process.
FIG. 3 is a diagram of a random forest training and recognition process.
Detailed Description
A facial expression recognition method based on random forests comprises the following steps:
step 1 AAM Displacement feature extraction
AAM is based on the active Shape model ASM (active Shape model), originally proposed by Edwards, Cootes and Taylor. AAM includes two parts: a shape model and a texture model. The present invention uses shape models. The shape model is defined as the coordinates of n feature points:
s=(x1,y1,x2,y2,…,xn,yn)T
it can be changed into a linear model:
wherein p isiIs the shape parameter, siThe training data is obtained by Principal Component Analysis (PCA) (principal Component analysis) dimensionality reduction. However, for the purposes of calculation and tracking, the present invention does not use the second expression, but uses the first simple expression. The detailed process is shown in fig. 2.
Step 1.1 selects the first frame (for example, neutral expression) of the facial expression image sequence, extracts the AAM feature points, and sets as:
s0=[x10,y10,x20,y20,...,xn0,yn0]
step 1.2 tracking the AAM feature points from the first frame (neutral expression) to the last frame (exaggerated expression) using the LK tracking algorithm. The LK optical flow tracking algorithm employed by the present invention is an improved version of the classic LK algorithm proposed by j
Wherein,is the initial estimate at pyramid L level, which can be calculated from L-1 level:
gL=2(gL-1+dL-1)
the final optical flow can be expressed in the form:
wherein L ismIs the deepest level of the pyramid. Thus, dLThe solution can be solved as follows:
wherein, <math><mrow>
<msubsup>
<mi>η</mi>
<mi>k</mi>
<mi>L</mi>
</msubsup>
<mo>=</mo>
<msup>
<mrow>
<mo>(</mo>
<msup>
<mi>G</mi>
<mrow>
<mo>-</mo>
<mn>1</mn>
</mrow>
</msup>
<mo>)</mo>
</mrow>
<mi>L</mi>
</msup>
<msubsup>
<mi>b</mi>
<mi>k</mi>
<mi>L</mi>
</msubsup>
<mo>,</mo>
</mrow></math> while <math><mrow>
<msup>
<mrow>
<mo>(</mo>
<msup>
<mi>G</mi>
<mrow>
<mo>-</mo>
<mn>1</mn>
</mrow>
</msup>
<mo>)</mo>
</mrow>
<mi>L</mi>
</msup>
<mo>=</mo>
<munder>
<mi>Σ</mi>
<mrow>
<mi>x</mi>
<mo>∈</mo>
<mi>N</mi>
<mrow>
<mo>(</mo>
<mi>u</mi>
<mo>)</mo>
</mrow>
</mrow>
</munder>
<mo>▿</mo>
<msup>
<mi>I</mi>
<mi>L</mi>
</msup>
<msup>
<mrow>
<mo>(</mo>
<mo>▿</mo>
<msup>
<mi>I</mi>
<mi>L</mi>
</msup>
<mo>)</mo>
</mrow>
<mi>T</mi>
</msup>
<mo>,</mo>
</mrow></math> <math><mrow>
<msubsup>
<mi>b</mi>
<mi>k</mi>
<mi>L</mi>
</msubsup>
<mo>=</mo>
<munder>
<mi>Σ</mi>
<mrow>
<mi>x</mi>
<mo>∈</mo>
<mi>N</mi>
<mrow>
<mo>(</mo>
<mi>u</mi>
<mo>)</mo>
</mrow>
</mrow>
</munder>
<mo>▿</mo>
<msup>
<mi>I</mi>
<mi>L</mi>
</msup>
<mi>δ</mi>
<msubsup>
<mi>I</mi>
<mi>k</mi>
<mi>L</mi>
</msubsup>
<mo>,</mo>
</mrow></math> In both formulaeIL k(x)=IL(x)-JL(x+gL+vL)
Wherein v isLIs the optical flow vector of the L layers, and K is the maximum number of iterations of each layer in the LK algorithm.
Through an LK optical flow tracking algorithm, obtaining AAM feature points of the last frame (exaggerated expression), and setting the AAM feature points at the moment as:
sn=[x1n,y1n,x2n,y2n,...,xnn,ynn]
step 1.3, according to the results obtained in step 1.1 and step 1.2, the displacement of the AAM characteristic point is calculated:
st=[x1n-x10,y1n-y10,x2n-x20,y2n-y20,...,xnn-xn0,ynn-yn0]。
i.e. the calculated AAM displacement characteristic, as shown in block 1 of fig. 1.
Step 2, AU extraction in facial expression sequence
Action units AU come from the facial behavior Coding system facs (facial Action Coding system), which was proposed by Paul Ekman and Wallace v. The core idea of AU is mainly to describe the movement of a muscle or a group of muscles of the face, such as the raising of eyebrows in AU 1. Since the probability of occurrence of the part AUs is very low, only 17 of them with a high probability of occurrence AUs: {1,2,4,5,6,7,9, 12,14,15,17,20,23,24,25,27,38 }. The extraction process of AU is shown in block 2 of fig. 1.
And 2.1, taking the AAM displacement characteristics obtained in the step 1 as the input of the random forest. The random forest is shown in figure 3;
as shown in fig. 3, the input AAM displacement feature data (i.e., D in the figure) is first randomized to obtain K pieces of data, and then the K pieces of data are transmitted to corresponding decision trees to obtain K results, and then classified according to the following random forest definitions to obtain the final classification result.
The definition of random forest is as follows:
where H (x) is the final fused classifier, K is the number of decision trees, hi(x) Represents a decision tree, Y is a class label, I (h)i(x) Y) indicates that x belongs to category Y.
Step 2.2 AAM feature point displacement reflects changes of facial expressions, such as eye closing, eye enlargement, mouth opening and the like, which are information represented by AU, and the two are consistent. Therefore, it is reasonable to input AAM displacement features into a random forest for training to obtain human face AU information in the sequence, and the feasibility is also verified by the final experimental result. Thus, the corresponding AUs contained in the sequence of facial expressions is finally obtained via step 2.1.
Step 3, training a facial expression classification model
Facial expressions are primarily combinations of facial muscle group movements, and AU is the movement of a block or group of facial muscles. Therefore, the combination of AUs represents the corresponding facial expression. The corresponding expression of the expression sequence can be obtained by analyzing the AU combination contained in the expression sequence.
As shown in block 3 in fig. 1, AUs information of the facial expression sequence obtained in step 2 is used as training data, and random forests are input again for training, so that a facial expression classification model can be obtained.
Step 4 facial expression recognition
The process of facial expression recognition is similar to the process of training a facial expression classification model. And (3) processing the facial expression to be classified according to the step 1 to obtain an AAM displacement characteristic vector, processing the AAM displacement characteristic vector in the step 2 to obtain AU information contained in the sequence, and finally inputting the AU information into the random forest facial expression classification model obtained in the step 3 to obtain a final facial expression recognition result.
The invention mainly aims at 7 basic facial expressions: anger, apprehension, aversion, fear, happiness, sadness and surprise.
Claims (5)
1. A facial expression recognition method based on random forests comprises the following steps:
step 1, extracting AAM displacement characteristics;
step 2, extracting AU in the facial expression sequence;
step 3, training a facial expression classification model;
and 4, recognizing the facial expression.
2. The random forest based facial expression recognition method of claim 1, wherein: the method for extracting the AAM displacement characteristics in the step 1 comprises the following steps:
using a shape model, the shape model is defined as the coordinates of n feature points:
s=(x1,y1,x2,y2,…,xn,yn)T
it can be changed into a linear model:
wherein p isiIs the shape parameter, siCarrying out Principal Component Analysis (PCA) dimensionality reduction on training data to obtain the training data;
step 1.1, selecting a first frame of a facial expression image sequence, extracting AAM feature points of the first frame, and setting as:
s0=[x10,y10,x20,y20,...,xn0,yn0]
step 1.2, tracking the AAM characteristic points from the first frame to the last frame by using an LK tracking algorithm; finding an image rate that minimizes the value of the error matching function
Wherein,is the initial estimate at pyramid L level, which can be calculated from L-1 level:
gL=2(gL-1+dL-1)
the final optical flow can be expressed in the form:
wherein L ismIs the deepest level of the pyramid, and thus, dLThe solution can be solved as follows:
wherein, <math>
<mrow>
<msubsup>
<mi>η</mi>
<mi>k</mi>
<mi>L</mi>
</msubsup>
<mo>=</mo>
<msup>
<mrow>
<mo>(</mo>
<msup>
<mi>G</mi>
<mrow>
<mo>-</mo>
<mn>1</mn>
</mrow>
</msup>
<mo>)</mo>
</mrow>
<mi>L</mi>
</msup>
<msubsup>
<mi>b</mi>
<mi>k</mi>
<mi>L</mi>
</msubsup>
<mo>,</mo>
</mrow>
</math> while <math>
<mrow>
<msup>
<mrow>
<mo>(</mo>
<msup>
<mi>G</mi>
<mrow>
<mo>-</mo>
<mn>1</mn>
</mrow>
</msup>
<mo>)</mo>
</mrow>
<mi>L</mi>
</msup>
<mo>=</mo>
<munder>
<mi>Σ</mi>
<mrow>
<mi>x</mi>
<mo>∈</mo>
<mi>N</mi>
<mrow>
<mo>(</mo>
<mi>u</mi>
<mo>)</mo>
</mrow>
</mrow>
</munder>
<mo>▿</mo>
<msup>
<mi>I</mi>
<mi>L</mi>
</msup>
<msup>
<mrow>
<mo>(</mo>
<mo>▿</mo>
<msup>
<mi>I</mi>
<mi>L</mi>
</msup>
<mo>)</mo>
</mrow>
<mi>T</mi>
</msup>
<mo>,</mo>
</mrow>
</math> <math>
<mrow>
<msubsup>
<mi>b</mi>
<mi>k</mi>
<mi>L</mi>
</msubsup>
<mo>=</mo>
<munder>
<mi>Σ</mi>
<mrow>
<mi>x</mi>
<mo>∈</mo>
<mi>N</mi>
<mrow>
<mo>(</mo>
<mi>u</mi>
<mo>)</mo>
</mrow>
</mrow>
</munder>
<mo>▿</mo>
<msup>
<mi>I</mi>
<mi>L</mi>
</msup>
<mi>δ</mi>
<msubsup>
<mi>I</mi>
<mi>k</mi>
<mi>L</mi>
</msubsup>
<mo>,</mo>
</mrow>
</math> In both formulaeIL k(x)=IL(x)-JL(x+gL+vL)
Wherein v isLIs the optical flow vector of L layers, K is the maximum iteration number of each layer in the LK algorithm;
obtaining AAM characteristic points of the last frame through an LK optical flow tracking algorithm, and setting the AAM characteristic points as follows:
sn=[x1n,y1n,x2n,y2n,...,xnn,ynn]
step 1.3, according to the results obtained in step 1.1 and step 1.2, the displacement of the AAM characteristic point is calculated:
st=[x1n-x10,y1n-y10,x2n-x20,y2n-y20,...,xnn-xn0,ynn-yn0]。
3. the random forest based facial expression recognition method of claim 1, wherein: the method for extracting AU in the human face expression sequence in the step 2 comprises the following steps:
step 2.1, the AAM displacement characteristics obtained in the step 1 are used as input of a random forest, input AAM displacement characteristic data are firstly subjected to randomization processing to obtain K sub-data, then the K sub-data are transmitted into a corresponding decision tree to obtain K results, and then classification is carried out according to the following random forest definition to obtain a final classification result;
the definition of random forest is as follows:
where H (x) is the final fused classifier, K is the number of decision trees, hi(x) Represents a decision tree, Y is a class label, I (h)i(x) Y) indicates that x belongs to category Y.
Step 2.2 finally obtains AUs corresponding to the sequence of facial expressions.
4. The random forest based facial expression recognition method of claim 1, wherein: the method for training the human face expression classification model in the step 3 comprises the following steps:
and (3) taking AUs information of the facial expression sequence obtained in the step (2) as training data, and inputting the training data into the random forest again for training to obtain the facial expression classification model.
5. The random forest based facial expression recognition method of claim 1, wherein: the method for recognizing the facial expressions in the step 4 comprises the following steps:
and (3) processing the facial expression to be classified according to the step 1 to obtain an AAM displacement characteristic vector, processing the AAM displacement characteristic vector in the step 2 to obtain AU information contained in the sequence, and finally inputting the AU information into the random forest facial expression classification model obtained in the step 3 to obtain a final facial expression recognition result.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410503590.XA CN104376333A (en) | 2014-09-25 | 2014-09-25 | Facial expression recognition method based on random forests |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410503590.XA CN104376333A (en) | 2014-09-25 | 2014-09-25 | Facial expression recognition method based on random forests |
Publications (1)
Publication Number | Publication Date |
---|---|
CN104376333A true CN104376333A (en) | 2015-02-25 |
Family
ID=52555229
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201410503590.XA Pending CN104376333A (en) | 2014-09-25 | 2014-09-25 | Facial expression recognition method based on random forests |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN104376333A (en) |
Cited By (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105528611A (en) * | 2015-06-24 | 2016-04-27 | 广州三瑞医疗器械有限公司 | Ache identification classifier training method and device |
CN106169072A (en) * | 2016-07-07 | 2016-11-30 | 中国科学院上海微系统与信息技术研究所 | A kind of face identification method based on Taylor expansion and system |
CN106384083A (en) * | 2016-08-31 | 2017-02-08 | 上海交通大学 | Automatic face expression identification and information recommendation method |
CN107066951A (en) * | 2017-03-15 | 2017-08-18 | 中国地质大学(武汉) | A kind of recognition methods of spontaneous expression of face and system |
CN107633207A (en) * | 2017-08-17 | 2018-01-26 | 平安科技(深圳)有限公司 | AU characteristic recognition methods, device and storage medium |
CN107704810A (en) * | 2017-09-14 | 2018-02-16 | 南京理工大学 | A kind of expression recognition method suitable for medical treatment and nursing |
CN107862292A (en) * | 2017-11-15 | 2018-03-30 | 平安科技(深圳)有限公司 | Personage's mood analysis method, device and storage medium |
CN107895154A (en) * | 2017-11-28 | 2018-04-10 | 中国地质大学(武汉) | The forming method and system of facial expression strength model |
CN109063679A (en) * | 2018-08-24 | 2018-12-21 | 广州多益网络股份有限公司 | A kind of human face expression detection method, device, equipment, system and medium |
CN109409273A (en) * | 2018-10-17 | 2019-03-01 | 中联云动力(北京)科技有限公司 | A kind of motion state detection appraisal procedure and system based on machine vision |
CN109711378A (en) * | 2019-01-02 | 2019-05-03 | 河北工业大学 | Human face expression automatic identifying method |
CN109753922A (en) * | 2018-12-29 | 2019-05-14 | 北京建筑大学 | Anthropomorphic robot expression recognition method based on dense convolutional neural networks |
CN109784123A (en) * | 2017-11-10 | 2019-05-21 | 浙江思考者科技有限公司 | The analysis and judgment method of real's expression shape change |
CN109961054A (en) * | 2019-03-29 | 2019-07-02 | 山东大学 | It is a kind of based on area-of-interest characteristic point movement anxiety, depression, angry facial expression recognition methods |
CN110175578A (en) * | 2019-05-29 | 2019-08-27 | 厦门大学 | Micro- expression recognition method based on depth forest applied to criminal investigation |
CN110532971A (en) * | 2019-09-02 | 2019-12-03 | 京东方科技集团股份有限公司 | Image procossing and device, training method and computer readable storage medium |
CN110992455A (en) * | 2019-12-08 | 2020-04-10 | 北京中科深智科技有限公司 | Real-time expression capturing method and system |
CN111353390A (en) * | 2020-01-17 | 2020-06-30 | 道和安邦(天津)安防科技有限公司 | Micro-expression recognition method based on deep learning |
CN111460952A (en) * | 2020-03-26 | 2020-07-28 | 心图熵动科技(苏州)有限责任公司 | Generation method, system and prediction system of schizophrenia face recognition rule |
CN111523367A (en) * | 2020-01-22 | 2020-08-11 | 湖北科技学院 | Intelligent facial expression recognition method and system based on facial attribute analysis |
CN112287745A (en) * | 2020-07-23 | 2021-01-29 | 南京航空航天大学 | Cognitive emotion recognition method based on geodesic distance and sample entropy |
CN112949522A (en) * | 2021-03-11 | 2021-06-11 | 重庆邮电大学 | Portrait data classification method based on support vector machine |
CN113221825A (en) * | 2021-05-31 | 2021-08-06 | 国网浙江省电力有限公司杭州供电公司 | Electric power safety control face recognition method based on machine learning |
CN113313020A (en) * | 2021-05-27 | 2021-08-27 | 成都威爱新经济技术研究院有限公司 | Unmarked facial expression capturing method and system based on virtual human |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102831447A (en) * | 2012-08-30 | 2012-12-19 | 北京理工大学 | Method for identifying multi-class facial expressions at high precision |
-
2014
- 2014-09-25 CN CN201410503590.XA patent/CN104376333A/en active Pending
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102831447A (en) * | 2012-08-30 | 2012-12-19 | 北京理工大学 | Method for identifying multi-class facial expressions at high precision |
Non-Patent Citations (1)
Title |
---|
陈雄: "基于序列特征的随机森林表情识别", 《中国优秀硕士学位论文全文数据库信息科技辑》 * |
Cited By (39)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105528611A (en) * | 2015-06-24 | 2016-04-27 | 广州三瑞医疗器械有限公司 | Ache identification classifier training method and device |
CN106169072B (en) * | 2016-07-07 | 2019-03-19 | 中国科学院上海微系统与信息技术研究所 | A kind of face identification method and system based on Taylor expansion |
CN106169072A (en) * | 2016-07-07 | 2016-11-30 | 中国科学院上海微系统与信息技术研究所 | A kind of face identification method based on Taylor expansion and system |
CN106384083A (en) * | 2016-08-31 | 2017-02-08 | 上海交通大学 | Automatic face expression identification and information recommendation method |
CN107066951A (en) * | 2017-03-15 | 2017-08-18 | 中国地质大学(武汉) | A kind of recognition methods of spontaneous expression of face and system |
CN107066951B (en) * | 2017-03-15 | 2020-01-14 | 中国地质大学(武汉) | Face spontaneous expression recognition method and system |
CN107633207A (en) * | 2017-08-17 | 2018-01-26 | 平安科技(深圳)有限公司 | AU characteristic recognition methods, device and storage medium |
CN107633207B (en) * | 2017-08-17 | 2018-10-12 | 平安科技(深圳)有限公司 | AU characteristic recognition methods, device and storage medium |
CN107704810A (en) * | 2017-09-14 | 2018-02-16 | 南京理工大学 | A kind of expression recognition method suitable for medical treatment and nursing |
CN109784123A (en) * | 2017-11-10 | 2019-05-21 | 浙江思考者科技有限公司 | The analysis and judgment method of real's expression shape change |
CN107862292B (en) * | 2017-11-15 | 2019-04-12 | 平安科技(深圳)有限公司 | Personage's mood analysis method, device and storage medium |
WO2019095571A1 (en) * | 2017-11-15 | 2019-05-23 | 平安科技(深圳)有限公司 | Human-figure emotion analysis method, apparatus, and storage medium |
CN107862292A (en) * | 2017-11-15 | 2018-03-30 | 平安科技(深圳)有限公司 | Personage's mood analysis method, device and storage medium |
CN107895154B (en) * | 2017-11-28 | 2020-08-25 | 中国地质大学(武汉) | Method and system for forming facial expression intensity calculation model |
CN107895154A (en) * | 2017-11-28 | 2018-04-10 | 中国地质大学(武汉) | The forming method and system of facial expression strength model |
CN109063679A (en) * | 2018-08-24 | 2018-12-21 | 广州多益网络股份有限公司 | A kind of human face expression detection method, device, equipment, system and medium |
CN109409273A (en) * | 2018-10-17 | 2019-03-01 | 中联云动力(北京)科技有限公司 | A kind of motion state detection appraisal procedure and system based on machine vision |
CN109753922A (en) * | 2018-12-29 | 2019-05-14 | 北京建筑大学 | Anthropomorphic robot expression recognition method based on dense convolutional neural networks |
CN109711378A (en) * | 2019-01-02 | 2019-05-03 | 河北工业大学 | Human face expression automatic identifying method |
CN109711378B (en) * | 2019-01-02 | 2020-12-22 | 河北工业大学 | Automatic facial expression recognition method |
CN109961054A (en) * | 2019-03-29 | 2019-07-02 | 山东大学 | It is a kind of based on area-of-interest characteristic point movement anxiety, depression, angry facial expression recognition methods |
CN110175578A (en) * | 2019-05-29 | 2019-08-27 | 厦门大学 | Micro- expression recognition method based on depth forest applied to criminal investigation |
CN110175578B (en) * | 2019-05-29 | 2021-06-22 | 厦门大学 | Deep forest-based micro expression identification method applied to criminal investigation |
US11961327B2 (en) | 2019-09-02 | 2024-04-16 | Boe Technology Group Co., Ltd. | Image processing method and device, classifier training method, and readable storage medium |
CN110532971B (en) * | 2019-09-02 | 2023-04-28 | 京东方科技集团股份有限公司 | Image processing apparatus, training method, and computer-readable storage medium |
CN110532971A (en) * | 2019-09-02 | 2019-12-03 | 京东方科技集团股份有限公司 | Image procossing and device, training method and computer readable storage medium |
CN110992455A (en) * | 2019-12-08 | 2020-04-10 | 北京中科深智科技有限公司 | Real-time expression capturing method and system |
CN111353390A (en) * | 2020-01-17 | 2020-06-30 | 道和安邦(天津)安防科技有限公司 | Micro-expression recognition method based on deep learning |
CN111523367B (en) * | 2020-01-22 | 2022-07-22 | 湖北科技学院 | Intelligent facial expression recognition method and system based on facial attribute analysis |
CN111523367A (en) * | 2020-01-22 | 2020-08-11 | 湖北科技学院 | Intelligent facial expression recognition method and system based on facial attribute analysis |
CN111460952B (en) * | 2020-03-26 | 2023-11-21 | 心图熵动科技(苏州)有限责任公司 | Method, system and prediction system for generating face recognition rule of schizophrenia |
CN111460952A (en) * | 2020-03-26 | 2020-07-28 | 心图熵动科技(苏州)有限责任公司 | Generation method, system and prediction system of schizophrenia face recognition rule |
CN112287745A (en) * | 2020-07-23 | 2021-01-29 | 南京航空航天大学 | Cognitive emotion recognition method based on geodesic distance and sample entropy |
CN112287745B (en) * | 2020-07-23 | 2024-05-03 | 南京航空航天大学 | Cognitive emotion recognition method based on geodesic distance and sample entropy |
CN112949522A (en) * | 2021-03-11 | 2021-06-11 | 重庆邮电大学 | Portrait data classification method based on support vector machine |
CN112949522B (en) * | 2021-03-11 | 2022-06-21 | 重庆邮电大学 | Portrait data classification method based on support vector machine |
CN113313020A (en) * | 2021-05-27 | 2021-08-27 | 成都威爱新经济技术研究院有限公司 | Unmarked facial expression capturing method and system based on virtual human |
CN113221825A (en) * | 2021-05-31 | 2021-08-06 | 国网浙江省电力有限公司杭州供电公司 | Electric power safety control face recognition method based on machine learning |
CN113221825B (en) * | 2021-05-31 | 2022-03-18 | 国网浙江省电力有限公司杭州供电公司 | Electric power safety control face recognition method based on machine learning |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN104376333A (en) | Facial expression recognition method based on random forests | |
Mohammadpour et al. | Facial emotion recognition using deep convolutional networks | |
Burkert et al. | Dexpression: Deep convolutional neural network for expression recognition | |
Fan et al. | A dynamic framework based on local Zernike moment and motion history image for facial expression recognition | |
Tian et al. | Recognizing action units for facial expression analysis | |
Jalal et al. | Human daily activity recognition with joints plus body features representation using Kinect sensor | |
CN106778796B (en) | Human body action recognition method and system based on hybrid cooperative training | |
Youssif et al. | Arabic sign language (arsl) recognition system using hmm | |
CN102254180B (en) | Geometrical feature-based human face aesthetics analyzing method | |
CN109815826A (en) | The generation method and device of face character model | |
Haq et al. | Boosting the face recognition performance of ensemble based LDA for pose, non-uniform illuminations, and low-resolution images | |
CN111680550B (en) | Emotion information identification method and device, storage medium and computer equipment | |
Youssef et al. | Auto-optimized multimodal expression recognition framework using 3D kinect data for ASD therapeutic aid | |
Gudipati et al. | Efficient facial expression recognition using adaboost and haar cascade classifiers | |
Ullah et al. | Emotion recognition from occluded facial images using deep ensemble model | |
Ahmad et al. | Facial expression recognition using lightweight deep learning modeling | |
CN107895154B (en) | Method and system for forming facial expression intensity calculation model | |
Uddin et al. | Human activity recognition using robust spatiotemporal features and convolutional neural network | |
Nayak et al. | Distribution-based dimensionality reduction applied to articulated motion recognition | |
Lv et al. | A spontaneous facial expression recognition method using head motion and AAM features | |
Manolova et al. | Facial expression classification using supervised descent method combined with PCA and SVM | |
Dembani et al. | UNSUPERVISED FACIAL EXPRESSION DETECTION USING GENETIC ALGORITHM. | |
Wang et al. | PAU-Net: Privileged action unit network for facial expression recognition | |
Li et al. | Hmm based eyebrow recognition | |
Ali et al. | Smile detection using data amalgamation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20150225 |