CN102831447A - Method for identifying multi-class facial expressions at high precision - Google Patents

Method for identifying multi-class facial expressions at high precision Download PDF

Info

Publication number
CN102831447A
CN102831447A CN2012103144354A CN201210314435A CN102831447A CN 102831447 A CN102831447 A CN 102831447A CN 2012103144354 A CN2012103144354 A CN 2012103144354A CN 201210314435 A CN201210314435 A CN 201210314435A CN 102831447 A CN102831447 A CN 102831447A
Authority
CN
China
Prior art keywords
prime
expression
image
node
haar
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN2012103144354A
Other languages
Chinese (zh)
Other versions
CN102831447B (en
Inventor
罗森林
谢尔曼
潘丽敏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Institute of Technology BIT
Original Assignee
Beijing Institute of Technology BIT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Institute of Technology BIT filed Critical Beijing Institute of Technology BIT
Priority to CN201210314435.4A priority Critical patent/CN102831447B/en
Publication of CN102831447A publication Critical patent/CN102831447A/en
Application granted granted Critical
Publication of CN102831447B publication Critical patent/CN102831447B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Image Analysis (AREA)

Abstract

The invention relates to a method for identifying multi-class facial expressions at a high precision based on Haar-like features, which belongs to the technical field of computer science and graphic image process. Firstly, the high-accuracy face detection is achieved by using the Haar-like features and a series-wound face detection classifier; further, the feature selection is carried out on the high-dimension Haar-like feature by using the Ada Boost. MH algorithm; and finally, the expression classifier training is carried out by using the Random Forest algorithm to complete the expression identification. Compared with the prior art, the method can reduce training and identifying time while increasing the multi-class expression identification rate, and can implement the parallelization conveniently to increase the identification rate and meet the requirement of real-time processing and mobile computing. The method can identify the static image and the dynamic image at a high precision, is not only applicable to the desktop computer but also to the mobile computing platforms, such as cellphone, tablet personal computer and the like.

Description

The high-precision recognition methods of multi-class facial expression
Technical field
The present invention relates to a kind of high-precision recognition methods of multi-class facial expression based on Haar-like features, belong to computer science and graph and image processing technical field.
Background technology
Facial expression is the important channel of Human communication, expression recognition(Facial expression recognition, FER)As a technology in man-machine interaction, just increasingly it is taken seriously.Diversified expression is generally divided into several basic class by people, and then solves the problems, such as identification using sorting technique.For example, Cohn-Kanade, JAFFE Facial expression database have recorded it is angry, detest, frightened, glad, sad, surprised 6 kinds of expressions, CAS-PEAL-R1 face expression databases have recorded smile, frown, it is surprised, dehisce, 5 kinds of expressions of closing one's eyes.
Human facial expression recognition needs to solve 2 basic problems:1. the high characteristic vector of representative strong, discrimination how is extracted to characterize different facial expressions;2. which kind of accuracy rate of use is high, fireballing recognition methods makes a distinction to different facial expressions.Existing human facial expression recognition technology is taken a broad view of, commonly used approach has:
1. in terms of feature extraction:
(1) Optical-flow Feature:Binaryzation or gray processing are carried out to sequence of video images, and then feature extraction is carried out to the light stream sports ground of the sequence, characteristic sequence is obtained.The problem of this method is in progress Expression Recognition application one is that sign extraction rate is not fast enough, and two be the accuracy of identification deficiency of discrimination model.
(2) Gabor characteristic:Gabor filter is divided into the passage of certain amount, and then carries out Two-Dimensional Gabor Wavelets conversion to the Facial Expression Image after standardization processing to extract the textural characteristics of Facial Expression Image by Gabor filter.The shortcoming of this method is that extraction rate is slower, is had difficulties in real-time application.
(3) expressive features moment characteristics:Each two field picture that battle array is directed in facial expression image sequence extracts the length of normalized face's key point displacement and particular geometric feature successively, and these data are constituted into a feature column vector;All feature column vectors in sequence one eigenmatrix of arrangement form in order, each eigenmatrix represents a facial expression image sequence.Identification of this method due to being related to face's key point, therefore extraction rate and precision all existing defects.
(4) the image local feature extracting method based on 2 D partial least square method:Sample graph is divided into several equal-sized sub-blocks by expression classification first, recycle the textural characteristics of each sub-block of LBP operator extractions, and Local textural feature matrix is constituted, using adaptive weighted mechanism, statistical nature extraction is carried out to Local textural feature matrix using 2 D partial least square method.The algorithm design comparison of this method is complicated, and extraction rate is not suitable for real-time disposition than relatively low.
(5) feature based on AVR and enhancing LBP:Wavelet decomposition is carried out to standard faces image, then extracts LBP features, enhancing variance-rate AVR characteristic values and the additional penalty factor are calculated afterwards, the characteristic value of some groups of different dimensions mutually distinguished with AVR values is finally extracted.This method needs to use the steps such as wavelet transformation, LBP feature extractions, penalty factor calculating, and extraction rate is relatively low, it is impossible to meet the demand handled in real time.
(6) face parameter attribute:The position of face first in identification facial zone, and then extract each organ according to the image information of face(Such as eye, nose, eyebrow, the corners of the mouth)Texture, profile parameters be used as characteristic vector.This method is related to the identification to face organ, therefore accuracy of identification and the representative aspect existing defects of feature.
In addition, earlier some study expressive features point motion features also including histogram, histogram of gradients, based on piecewise affine transformations etc..For the characteristic type that dimension is larger, dimension-reduction treatment is also often referred to, common Feature Dimension Reduction processing method has:Cluster linear discriminant analysis method, PCA etc..
2. in terms of differentiating method of expressing one's feelings:
(1) SVMs(SVM)Algorithm;SVMs (Support Vector Machine, SVM) be built upon that the VC of Statistical Learning Theory dimensions are theoretical and Structural risk minization basis on, according to limited sample information model complexity(I.e. to the study precision of specific training sample)And learning ability(The ability of arbitrary sample is recognized without error)Between seek optimal compromise, to obtain best Generalization Ability.SVM algorithm to kernel function, kernel functional parameter in training, it is necessary to be constantly adjusted to optimize, therefore training process is often more complicated, and this is the important deficiency in the algorithm use;In addition, SVM algorithm is a kind of two sorting algorithm, the identification for plurality of classes is, it is necessary to which further to algorithm improved.
(2) Canonical Correlation Analysis:This method borrows principal component analysis dimensionality reduction thought, principal component is extracted to two groups of variables respectively, and the degree of correlation between the principal component extracted from two groups of variables is reached maximum, and from orthogonal between each principal component of same group of internal extraction, the overall linear relationship of two groups of variables is described with the correlation for the principal component extracted respectively between two groups.This method is more accurate for the description of linear relationship, but precision when being measured for more complicated relation is not satisfactory, and this is the limitation of the algorithm in use.
(3) Histogram Matching:The input of this method is two groups of statistics with histogram amounts, is generally regarded as two groups of one-dimensional vectors, and then use the distance metric method of one-dimensional vector(Such as Euclidean distance, card side, histogram intersection, Bhattacharyya distances, land mobile distance), similarity measurement is carried out to statistics with histogram amount.But this method designs histogrammic statistic unit, requires stricter by the representative of statistic, if above-mentioned 2 points can not meet very well, recognition effect will be greatly affected.
(4) AdaBoost algorithms:The algorithm is a kind of iterative algorithm, and its core concept is that different graders (Weak Classifier) are trained for same training set, and then these weak classifier sets are got up, a stronger final classification device (strong classifier) is constituted.Its algorithm is realized by change data distribution in itself.The limitation one that this method is used is its training time, for the high dimensional data of larger data amount, and this method generally requires the plenty of time and is trained;Two be the selection of Weak Classifier, and optimal Weak Classifier can just be searched out by generally requiring to carry out largely test.
In summary, this application scenarios is recognized for a variety of expression high-precision high-speeds, existing feature extracting method existing characteristics representativeness is limited, precision and not high enough the deficiency of extraction rate;Meanwhile, existing expression differentiating method also having that accuracy of identification is undesirable, complexity is too high, recognizable expression categorical measure is limited, the low limitation of recognition speed.
The content of the invention
The purpose of the present invention is to recognize problems to solve a variety of facial expression high-precision high-speeds, proposes a kind of human facial expression recognition method based on Haar-like features.
The present invention design principle be:The Face datection of high accuracy is realized first by Haar-like features and series connection Face datection grader;And then Feature Selection is carried out to higher-dimension Haar-like features using AdaBoost.MH algorithms;It is final to carry out expression classifier training using random forests algorithm, to complete Expression Recognition.
The technical scheme is that be achieved by the steps of:
Step 1, in order to realize automatically extracting for facial zone image, first by multiple facial zone images(It is used as positive sample)With multiple non-face area images(It is used as negative sample)Off-line training is carried out, face recognition grader is obtained.Face recognition grader can be obtained by a variety of conventional training methods in the prior art.The present invention uses the AdaBoost Cascading Classifier training methods based on Haar-like features.
Step 2, on the basis of step 1, the off-line training of facial expression grader is carried out.Detailed process is as follows:
Step 2.1, expression mark is carried out to face-image training data first, specific method is:Collect the picture or video of various expression classifications to be identified(For expression video, extract key frame therein and be used as training image), training image collection A is formed, wherein the picture number included is m;Then use continuous integer numbering as each pictures or the class label of key frame, it is expression classification number to be identified to form expression class label collection Y={ 1,2 ..., k }, wherein k.
Step 2.2, facial zone data extraction, the face-image being cut out are carried out to every width training image after step 2.1 mark.
Facial zone data extract specific method be:The integrogram of each width image is calculated first.Described integrogram is identical with original image size, and the value of any point (x, y) is artwork corresponding points (x ', y ') and its all pixel value sums in upper left side thereon:
ii ( x , y ) = Σ x ′ ≤ x , y ′ ≤ y i ( x ′ , y ′ ) ,
Ii (x, y) represents the value of point (x, y) on integrogram in formula, and i (x ', y ') represents the pixel value of point (x ', y ') on original image.
Calculating is obtained after integrogram, the face recognition grader obtained according to step 1, and the Haar-like features of image in sliding window region are extracted using sliding window method, facial zone is quickly determined;Then the image cutting-out of facial zone is come out, the size zoomed to needed for Expression Recognition, keeps original expression mark, form training image collection B.
Step 2.3, in order to train expression classifier, each image in the training image collection B formed to step 2.2 carries out the second extraction of Haar-like features, and the specific method of Haar-like feature extractions is:The integrogram for the image that each width is cut out is calculated, calculating corresponding H according to each integrogram ties up Haar-like characteristic values(Wherein H is determined by the Haar-like characteristic types and the size of picture used).
The corresponding H dimensions Haar-like characteristic vectors of each image are denoted as a line, whole H of all m width images is tieed up Haar-like characteristic vectors and constitutes a m row, the eigenmatrix X of H row;
Step 2.4, using AdaBoost.MH algorithms, the Haar-like eigenmatrixes X obtained to step 2.3 carries out Feature Selection;Described AdaBoost.MH algorithms take turns computing by F, and screening is iterated to single feature Weak Classifier, and selection F dimension principal characters obtain a m row, the principal character matrix X ' of F row to complete Feature Selection from H dimension Haar-like characteristic value collections;
Weak Classifier used in the iteration screening, need to meet following condition:1. the input of Weak Classifier is one-dimensional characteristic value(A certain specific dimension i.e. in characteristic vector);2. for class label to be identified, the output of Weak Classifier is 1 or -1.
Using AdaBoost.MH algorithms carry out Feature Selection detailed process be:
Step 2.4.1, the corresponding weight of initialization each image, is denoted as D1(i, yi)=1/ (mk), yi∈ Y represent the expression class label of i-th of image, i=1 ... m;
Step 2.4.2, starts f wheel iteration (f=1 ... F):Eigenmatrix X each column data is carried out H computing, obtain r as the input of a Weak Classifier successivelyF, j
r f , j = Σ i = 1 m D f ( i , y i ) K [ l i ] h j ( x i , j , y i ) ,
Wherein, j=1 ... H, xI, jRepresent j-th of element in X the i-th row(J-th of characteristic value in characteristic vector corresponding to i.e. i-th training image);hj(xI, j, yi) represent with xI, jIt is used as the Weak Classifier of input, Df(i, yi) represent that f takes turns the weighted value of i-th of training image in iteration, K [ y i ] = + 1 y i ∈ Y = { 1,2 , . . . , k } - 1 y i ∉ Y = { 1,2 , . . . , k } ;
Terminate after H computing, H r for taking epicycle iteration to obtainF, jIn maximum, be denoted as rf, and by rfJth dimensional feature value x corresponding, using XjIt is used as the Weak Classifier h of inputj(xj, Y), the Weak Classifier h filtered out is taken turns as ff(xj, Y), while by xjNew feature space is added to as the feature dimensions filtered out;
Step 2.4.3, calculates the Weak Classifier h selected by step 2.4.2f(xj, Y) weight αf
α f = 1 2 ln ( 1 + r f 1 - r f ) ;
Step 2.4.4, calculates the weight D of each image in f+1 wheel iterationf+1
D f + 1 = D f ( i , y i ) exp ( - α f K [ y i ] h f ( x i , j , y i ) ) Z f , i = 1 . . . m .
Wherein, hf(xI, j,yi) represent f wheel iteration in filter out, the Weak Classifier of input, Z are used as using the jth dimensional feature value of i-th of imagefIt is normalization factor
Z f = Σ i = 1 m D f ( i , y i ) exp ( - α f K [ y i ] h f ( x i , j , y i ) ) ;
Step 2.4.5, the new weight that step 2.4.4 is obtained substitutes into step 2.4.2, according to step 2.4.2 to step 2.4.4 method iteration, until filtering out F dimension principal characters, and F row are extracted in eigenmatrix X, form a m row, the principal character matrix X ' of F row;
Step 3, the expression class label collection Y marked using the principal character matrix X ' obtained through step 2 and by step 2.1, training generation Expression Recognition grader, the process of training follows random forests algorithm, and specific method is:
Step 3.1, decision tree number T and node intrinsic dimensionality u in design requirement, generate T CART categorised decision tree.The record format of the root node of the decision tree is N (J), and the record format of intermediate node is N (V, J), and the record format of leafy node is (V, J, yt).Wherein, J represents node N disruptive features dimension, and V represents node N characteristic value, ytRepresent node N class label.
The generation method of every CART categorised decision tree is:
Step 3.1.1, carries out that grab sample can be put back to m times, and principal character matrix X ' a line is extracted every time, constitutes a new m row, the matrix X " of F row, the growth for this CART categorised decision tree;The corresponding training sample label of each row feature constitutes new expression class label collection Y " in X ";
Step 3.1.2, since root node, node split is carried out by node, is finally completed the growth of whole tree.Each the fission process of node is:
a)U row are randomly choosed from eigenmatrix X "
Figure BDA00002075520300061
As the training data needed for this node split, wherein
Figure BDA00002075520300062
Represent X " jth row(That is the jth dimension in F dimension Haar-like feature spaces);
b)The x " j selected information gain IG is calculated respectivelyj, obtain u IGj;
IG j = IG ( Y ′ ′ | x ‾ j ′ ′ ) = H ( Y ′ ′ ) - H ( Y ′ ′ | x ‾ j ′ ′ ) , - - - ( 1 )
Wherein, H (Y ") is expression class label collection Y " comentropy:
H ( Y ′ ′ ) = - Σ w = 1 k p ( y ′ ′ w = V w ) · log 2 p ( y ′ ′ w = V w ) ,
Vw represents the value of w-th of class label in Y ", namely Vw∈ { 1,2 ..., k };
Figure BDA00002075520300065
It is based on for expression class label collection Y "
Figure BDA00002075520300066
Conditional entropy:
H ( Y ′ ′ | x ‾ j ′ ′ ) = Σ s = 1 h p ( x ′ ′ j = V s ) · H ( Y ′ ′ | x ′ ′ j = V s )
= - Σ s = 1 h p ( x ′ ′ j = V s ) · Σ w = 1 q p ( y ′ ′ w = V w | s ) · log 2 p ( y ′ ′ w = V w | s )
VsRepresent x "jS kind values in middle all different values of element;Vw/sRepresent VsCorresponding expression class label;H (Y " | x "J, s=Vs) it is VsThe comentropy of corresponding expression class label set, x "jRepresent the element in X " jth row, h≤m, q≤k;
c)The u information gain value IG that comparison step b is obtainedj, and IG will be made in X "jThe maximum row of value are extracted, and are denoted as x "J, while the columns J that this is listed in X ' is recorded, tie up, used during for identification as the disruptive features of this node;
d)Count x "JIn all different characteristic values quantity c, then set up c respectively using different characteristic value as node characteristic value V child node to current node, and using the child node as the root node of subtree, the new subtree of generation, the growing method of subtree is:By x "JIn all values be equal to the element of this feature value where row vector in X " propose, constitute new eigenmatrix Xv, then by the corresponding institute's espressiove class label proposition of proposed row vector, constitute new expression class label set YvThen (X is usedv, Yv) (X ", Y ") is substituted, step a-d operation is recursively carried out, until meeting following condition for the moment, terminates the growth of this subtree:
1. this node can not continue division(That is XvLine number be less than 2, or each row in all characteristic values it is all equal), corresponding expression class label set YvMiddle frequency of occurrences highest label as the node class label ytPreserve;
2. the Y under this nodevIn expression class label it is all identical, unique expression class label is used as to the class label y of the nodetPreserve.
Step 3.2, all T CART categorised decision trees are preserved, final random forest Expression Recognition grader is formed.
Step 4, the random forest Expression Recognition grader obtained using step 3 off-line training, ONLINE RECOGNITION is carried out to still image to be measured or dynamic video;
1) recognition methods to still image is:
Step a, extracts the facial zone in still image to be identified;
Step b, on the basis of step a, the Haar-like feature extracting methods according to step 2.2, and the principal character matrix X ' that step 2.4.5 is obtained, the F dimension Haar-like features needed for identification are extracted, the characteristic vector of facial expression image to be identified is constituted, is denoted as x;Use xJRepresent characteristic vector x J dimensional feature values;
The characteristic vector x of facial expression image to be identified is identified respectively for T CART categorised decisions tree in step c, the random forest Expression Recognition grader obtained using step 3 off-line training, and the identification of every CART categorised decision tree is since root node, and detailed process is:
C.1. the disruptive features dimension J of current node is obtained from grader, the characteristic vector x of facial expression image to be identified J dimensional feature values x is readJ
C.2. searched in the child node of current node, select node characteristic value and xJMost close child node;
C.3. operation c.1-c.2 is recursively carried out repeatedly, until current node is leafy node, stops recurrence, and by the class label y of the leafy nodetExported as the recognition result of this CART categorised decision tree;
Step d, to the T output result y of T CART categorised decision tree in random forest Expression Recognition gradertCounted, will appear from frequency highest class label and exported as final recognition result.
2) recognition methods to dynamic video is:
Step e, is decoded to video file, is extracted per frame data, is obtained images to be recognized sequence;
Step f, on the basis of step e, carries out facial zone data extraction to each image in image sequence to be identified, obtains face-image sequence to be identified;
Step g, according to the Haar-like feature extracting methods described in step 2.2, the F dimension Haar-like features filtered out to every breadth portion image extracting step 2.4.5 in the face-image sequence to be identified obtained by step f;
Step h, on the basis of step g, every width face-image in face-image sequence to be identified is identified the random forest Expression Recognition grader obtained using step 3 off-line training, obtains classification sequence of expressing one's feelings;The identification process of the face-image is identical with step c, d;
Step i, the expression classification sequence obtained to step h carries out smooth, the burr judgement among removal recognition sequence, obtains final recognition result.
Beneficial effect
Compared to the method based on features such as the colour of skin, border, Gabor, wavelet transformations, the method for detecting human face that the present invention is used has the characteristics of recognition speed is fast, accuracy rate is high.
Compared with the methods such as Gabor characteristic, Wavelet Transform Feature, Optical-flow Feature, the technology that the present invention is used has higher accuracy rate and smaller calculating consumption, is applicable not only to desktop computer, is also applied for the mobile computing platforms such as mobile phone, tablet personal computer.
Compared with the machine learning methods such as AdaBoost, SVM and traditional Canonical Correlation Analysis, the method based on template matches and similarity measurement, the present invention realizes the final identification of expression using the method for " Feature Selection+random forest ", recognition accuracy with faster recognition speed and Geng Gao, and parallelization can be conveniently realized, with the demand for further improving recognition efficiency, meeting processing and mobile computing in real time.
Brief description of the drawings
Fig. 1 is human facial expression recognition method schematic of the invention;
Fig. 2 is the schematic diagram of embodiment septum reset image extraction method;
Fig. 3 is 10 class Haar-like features used in embodiment septum reset image extraction method;
The 5 class Haar-like features that Fig. 4 is used by embodiment septum reset Expression Recognition;
Fig. 5 is the schematic diagram of " Feature Selection+random forest " method in embodiment;
Fig. 6 is in embodiment, when being tested using CAS-PEAL-R1 expressions storehouse, performance test of the invention with traditional AdaBoost.MH algorithms, (a) figure is recognition accuracies of traditional AdaBoost.MH to all kinds of expressions, and (b) figure is recognition accuracy of " Feature Selection+random forest " method proposed by the invention to all kinds of expressions;
It is the overall accuracy performance that the present invention carries " Feature Selection+random forest " method and AdaBoost.MH when being tested using JAFFE expressions storehouse during Fig. 7 is embodiment.
Embodiment
In order to better illustrate objects and advantages of the present invention, the embodiment to the inventive method is described in further details with reference to the accompanying drawings and examples.
Respectively so that static images and dynamic video as input, are designed and disposed with 3 tests:(1) for CAS-PEAL-R1 express one's feelings storehouse static images test, (2) for JAFFE express one's feelings storehouse static images test, (3) be directed to dynamic video test.
CAS-PEAL-R1 is the face database that Inst. of Computing Techn. Academia Sinica builds, wherein front photograph of the face expression database comprising 379 people, and everyone records 5 kinds of expressions:Frown, smile, closing one's eyes, dehiscing, it is surprised.JAFFE describes the 6 classes expression of 10 Japanese womens:Glad, sad, dejected, angry, surprised, detest.
In order to describe algorithm performance curve, it is necessary to test different parameter combinations(That is u, T value)The performance impact that static images are recognized, therefore in test (1), (2), 400 random forest graders trained under different node intrinsic dimensionality u are tested respectively.During first round Feature Selection, it is step-length that F value, which is defined as 10000, u value with 10, and 4000 are risen to by 10;T value is defined as u2Upper round numbers, record every group(u、T)Overall accuracy under value.For a certain N-dimensional confusion matrix C, its overall accuracy P definition is:
p = Σ i = 1 N c ii Σ i = 1 N Σ j = 1 N c ij × 100 % . - - - ( 2 )
More objectively parser performance, in test (1), the overall accuracy of every bit is all obtained using 10 folding interior extrapolation methods on algorithm performance curve.For test (2), because expression storehouse picture sum is less, be not suitable for 10 folding cross-betas of deployment, thus using the method for more strict opener test.
For test (3), the video shot using camera is included on screen as input, by the recognition result of every two field picture and identification are time-consuming.
Above-mentioned 3 testing process will one by one be illustrated below, all tests are completed on same computer, and concrete configuration is:Intel double-cores CPU(Dominant frequency 1.8G), 1G internal memories, WindowsXP SP3 operating systems.
In above-mentioned 3 are tested, automatically extracting for facial zone image is carried out using identical face recognition grader.The specific training flow of face recognition grader using 10 class Haar-like features shown in Fig. 3 as shown in Fig. 2 using the AdaBoost Cascading Classifier training methods based on Haar-like features, be trained.
In addition, in above-mentioned 3 are tested, using identical Weak Classifier.The definition of Weak Classifier is:
h j ( x , y ) = 1 p j , y x j < p j , y &theta; j , y - 1 p j , y x j &GreaterEqual; p j , y &theta; j , y - - - ( 3 )
Wherein, xjRepresent the input of Weak Classifier, θJ, yRepresent the threshold value obtained after training, pJ, yIndicate the direction of the sign of inequality.
1. the static images test in storehouse of being expressed one's feelings for CAS-PEAL-R1
150 width are respectively selected as experimental data in 5 kinds of facial expression images for expressing one's feelings storehouse from CAS-PEAL-R1.In order to carry out 10 folding cross validations, for every group of F, T valued combinations, 750 width images are randomly divided into 10 groups, 10 training for taking turns expression classifier and identification test are carried out respectively.In each wheel test, using 1 group in 10 groups of images as test data, it is estimated for the accuracy rate to grader;Remaining 9 groups of data is as training data, the off-line training for facial expression grader.After the completion of 10 wheel tests, each image was tested once, and test result is collected, and generated confusion matrix, and calculate overall accuracy according to formula (2).
10 above-mentioned wheel test process are identical, and the idiographic flow for often taking turns test is:
Step 1, it is loaded into face recognition grader.
Step 2, the off-line training of facial expression grader is carried out.
Step 2.1, expression mark is carried out to face-image training data first.Because the expression in CAS-PEAL-R1 expressions storehouse is documented on filename, therefore the work of this step is directly to correspond to integer { 1,2,3,4,5 } according to the mark keyword on filename to dump among annotation repository.
Step 2.2, for the face-image being cut out, facial zone data extraction is carried out to every width training image after step 2.1 mark.Facial zone data extract specific method be:The integrogram of entire image, and then the face recognition grader obtained according to step 1 are calculated first, and the Haar-like features of image in sliding window region are extracted using sliding window method, facial zone is quickly determined(The increase coefficient of sliding window is 1.2);Finally the image cutting-out of facial zone is come out, the size zoomed to needed for Expression Recognition(In the present embodiment, the size used is 32 × 32pix), original expression mark is kept, training image collection is formed.
Step 2.3, in order to carry out Expression Recognition, the face-image being cut out to step 2.2 carries out the second extraction of Haar-like features.
Fig. 4 illustrates the 5 class Haar-like features that the present embodiment is used, and this feature has following three feature:
1. arithmetic speed is fast.Coordinate integrogram, the extraction of any size Haar-like features only need to perform the digital independent and plus and minus calculation of fixed number of times.Haar-like features comprising 2 rectangles need to only read 6 points from integrogram and carry out plus/minus computing, and the feature of 3 rectangles need to only read 8 points, and the feature of 4 rectangles need to only read 9 points.
2. distinction is strong.The dimension of Haar-like feature spaces is very high, by taking 5 category features that the present embodiment is used as an example, and the image of one 32 × 32, total dimension of 5 category features has exceeded 510,000, and particular number is as shown in table 1.
The quantity of the class Haar-like features of 1 one 32 × 32 images of table 5
Figure BDA00002075520300111
The dimension of pixel count of this dimension considerably beyond picture in itself, the also traditional characteristic such as significantly larger than Gabor, facial expression feature point feature, therefore with higher differentiation potentiality.
The specific method of extraction is:The integrogram for the face-image that each width is cut out is calculated, 510112 all dimension Haar-like characteristic values are calculated according to integrogram, Haar-like characteristic value collections are obtained.
Calculate the integrogram for the image that each width is cut out, 510112 corresponding dimension Haar-like characteristic values are calculated according to each integrogram, the corresponding 510112 dimension Haar-like characteristic vectors of each image are denoted as a line, whole Haar-like characteristic vectors of all 675 width images is constituted 675 rows, the eigenmatrix X of 510112 row.
In the following description, using yi∈ Y represent the expression class label of i-th of image, xiRepresent eigenmatrix X the i-th row(510112 dimension Haar-like characteristic vectors corresponding to i.e. i-th training image), xjRepresent X jth row(Jth dimension in i.e. 510112 dimension Haar-like feature spaces), xI, jRepresent j-th of element in X the i-th row(J-th of characteristic value in characteristic vector corresponding to i.e. i-th training image);
Step 2.4, using AdaBoost.MH algorithms, the Haar-like eigenmatrixes X obtained to step 2.3 carries out Feature Selection;Described AdaBoost.MH algorithms pass through 10000 wheel computings, screening is iterated to single feature Weak Classifier, selection 10000 ties up principal character to complete Feature Selection from H dimension Haar-like characteristic value collections, obtains 675 rows, the principal character matrix X ' of 10000 row;
Weak Classifier used in above-mentioned interative computation, need to meet following condition:1. the input of Weak Classifier is one-dimensional characteristic value(A certain specific dimension i.e. in characteristic vector);2. for class label l to be identified, the output of Weak Classifier is 1 or -1.
Using AdaBoost.MH algorithms carry out Feature Selection detailed process be:
Step 2.4.1, the corresponding weight of initialization each image, is denoted as D1(i, yi)=1/(675×5)。
Step 2.4.2, starts epicycle iteration(In illustrating below, the wheel number of iteration is represented using f), eigenmatrix X each column data is carried out 510112 computings, r is calculated according to the following formula as the input of a Weak Classifier successivelyF, jValue:
r f , j = &Sigma; i = 1 m D f ( i , y i ) K [ l i ] h j ( x i , j , y i ) ,
Wherein, j=1 ... 510112, xI, jRepresent j-th of element in X the i-th row;hj(xI, j, yi) represent with xI, jIt is used as the Weak Classifier of input, Df(i, yi) represent epicycle iteration(That is f takes turns iteration)In i-th of training image weighted value, K [ y i ] = + 1 y i &Element; Y = { 1,2 , . . . , k } - 1 y i &NotElement; Y = { 1,2 , . . . , k } ;
After above-mentioned 510112 computings terminate, compare 510112 r that epicycle is iterated to calculate outF, jValue, take its maximum, be denoted as rf, and then find and make rF, jReach maximum occurrences rf, using Weak Classifier h of the jth dimensional feature value as inputj(xj, Y), the Weak Classifier filtered out as epicycle(In order to which following description is convenient, h is denoted asf(xj, Y)), while the jth dimensional feature x that the Weak Classifier is usedjNew feature space is added to as the feature dimensions filtered out;
Step 2.4.3, calculates the Weak Classifier h selected by step 2.4.2f(xj, Y) weight αf
&alpha; f = 1 2 ln ( 1 + r f 1 - r f ) ;
Step 2.4.4, calculates the weight D of each image in next round iterationf+1
D f + 1 = D f ( i , y i ) exp ( - &alpha; f K [ y i ] h f ( x i , j , y i ) ) Z f , i = 1 . . . 675 .
Wherein, hf(xi, l) represent that the jth dimensional feature value extracted in f wheel iteration using in i-th of image is used as the Weak Classifier of input, ZfIt is normalization factor
Z f = &Sigma; il D f ( i , y i ) exp ( - &alpha; f K i [ y i ] h f ( x i , l ) ) i = 1 . . . 675 .
Wherein, hf(xI, j,yi) represent f wheel iteration in filter out, the Weak Classifier of input, Z are used as using the jth dimensional feature value of i-th of imagefIt is normalization factor
Z f = &Sigma; i = 1 m D f ( i , y i ) exp ( - &alpha; f K [ y i ] h f ( x i , j , y i ) ) ;
Step 2.4.5, the new weight that step 2.4.4 is obtained substitutes into step 2.4.2, is taken turns according to step 2.4.2 to step 2.4.4 method iteration 10000, so as to filter out 10000 dimension principal characters, forms new feature space, i.e.,:The characteristic series filtered out is extracted by wheel from eigenmatrix X, 675 rows, the principal character matrix X ' of 10000 row is formed;
Step 3, expression class label collection Y={ 1,2,3,4,5 } marked using the principal character matrix X ' obtained through step 2 and by step 2.1, training generation Expression Recognition grader, the process of training follows random forests algorithm, and specific method is:
Step 3.1, decision tree number T and node intrinsic dimensionality u in design requirement, generate T CART categorised decision tree.The record format of the root node of the decision tree is N (J), and the record format of intermediate node is N (V, J), and the record format of leafy node is (V, J, yt).Wherein, J represents node N disruptive features dimension, and V represents node N characteristic value, ytRepresent node N class label.
The generation method of every CART categorised decision tree is:
Step 3.1.1, carries out that grab sample can be put back to 675 times, and principal character matrix X ' a line is extracted every time, a 675 new rows, the matrix X " of 10000 row is constituted, dedicated for the growth of this CART categorised decision tree;X " in the corresponding training sample label of each row feature constitute new expression class label collection Y ";
Step 3.1.2, since root node, node split is carried out by node, is finally completed the growth of whole tree.Each the fission process of node is:
a)U row are randomly choosed from eigenmatrix X
Figure BDA00002075520300133
As the training data needed for this node split, wherein
Figure BDA00002075520300134
Represent X " jth row(That is the jth dimension in F dimension Haar-like feature spaces);
b)The x selected " is calculated respectivelyjInformation gain IGj, obtain u IGj;
IG j = IG ( Y &prime; &prime; | x &OverBar; j &prime; &prime; ) = H ( Y &prime; &prime; ) - H ( Y &prime; &prime; | x &OverBar; j &prime; &prime; ) , - - - ( 1 )
Wherein, H (Y ") is expression class label collection Y " comentropy:
H ( Y &prime; &prime; ) = - &Sigma; w = 1 k p ( y &prime; &prime; w = V w ) &CenterDot; log 2 p ( y &prime; &prime; w = V w ) ,
VwRepresent the value of w-th of class label in Y ", namely Vw∈ { 1,2 ..., k };
Figure BDA00002075520300141
It is based on for expression class label collection Y "Conditional entropy:
H ( Y &prime; &prime; | x &OverBar; j &prime; &prime; ) = &Sigma; s = 1 h p ( x &prime; &prime; j = V s ) &CenterDot; H ( Y &prime; &prime; | x &prime; &prime; j = V s )
= - &Sigma; s = 1 h p ( x &prime; &prime; j = V s ) &CenterDot; &Sigma; w = 1 q p ( y &prime; &prime; w = V w | s ) &CenterDot; log 2 p ( y &prime; &prime; w = V w | s )
VsRepresent x "jS kind values in middle all different values of element;Vw/sRepresent VsCorresponding expression class label;H(Y″|x″J, s=Vs) it is VsThe comentropy of corresponding expression class label set, x "jRepresent the element in X " jth row, h≤m, q≤k;
c)The u information gain value IG that comparison step b is obtainedj, and IG will be made in X "jThe maximum row of value are extracted, and are denoted as x "J, while the columns J that this is listed in X ' is recorded, tie up, used during for identification as the disruptive features of this node;
d)Count x "JIn all different characteristic values quantity c, then set up c respectively using different characteristic value as node characteristic value V child node to current node, and using the child node as the root node of subtree, the new subtree of generation, the growing method of subtree is:By x "JIn all values be equal to the element of this feature value where row vector in X " propose, constitute new eigenmatrix Xv, then by the corresponding institute's espressiove class label proposition of proposed row vector, constitute new expression class label set Yv;Then use (Xv, Yv) to substitute (X ", Y "), recursively carry out step a-d operation, until meeting following condition for the moment, terminate the growth of this subtree:
1. this node can not continue division(That is XvLine number be less than 2, or each row in all characteristic values it is all equal), corresponding expression class label set YvMiddle frequency of occurrences highest label as the node class label ytPreserve;
2. the Y under this nodevIn expression class label it is all identical, unique expression class label is used as to the class label y of the nodetPreserve.
Step 3.2, by the unified preservation of all T CART categorised decisions trees, final random forest Expression Recognition grader is formed.
Step 4, every width test pictures are carried out Expression Recognition by the Expression Recognition grader trained using step 3 respectively, and record recognition result and identification are time-consuming.The specific method of every width test pictures identification is:
Step a, extracts the facial zone in still image to be identified;
Step b, on the basis of step a, the Haar-like feature extracting methods according to step 2.2, and step 2.4.5 Feature Selection result, extract 10000 dimension Haar-like features needed for identification, constitute the characteristic vector of facial expression image to be identified, be denoted as x;Use xJRepresent characteristic vector x J dimensional feature values;
Step c, the characteristic vector x extracted using step b, the random forest Expression Recognition grader obtained using step 3 off-line training, the characteristic vector x of facial expression image to be identified is identified using T CART categorised decisions tree in grader respectively, the identification of every CART categorised decision tree is since root node, and detailed process is:
C.1. the disruptive features dimension J of current node is obtained from grader, the characteristic vector x of facial expression image to be identified J dimensional feature values x is readJ
C.2. searched in the child node of current node, select node value v and xJMost close child node childv
C.3. operation c.1-c.2 is recursively carried out repeatedly, until current node is leaf node, stops recurrence, and the class label of the leaf node is made into ytExported for the recognition result of this CART categorised decision tree;
Step d, to the T output result y of T CART categorised decision tree in random forest Expression Recognition gradertCounted, will appear from frequency highest class label and exported as final recognition result.
Step e, after the test of 10 wheels terminates, carries out collecting comparison, obtains confusion matrix, and according to formula (2), calculate overall accuracy to the recognition result of all 750 pictures.
2. the static images test in storehouse of being expressed one's feelings for JAFFE
JAFFE describes the 6 classes expression of 10 Japanese womens:Glad, sad, dejected, angry, surprised, detest.Because expression storehouse picture sum is less(216 width image altogether), be not suitable for 10 folding cross-betas of deployment, thus using the method for more strict opener test, 10 width are extracted as test set in expression per class, remaining 26 width is used as training set.
Idiographic flow is similar with test 1, and different part is:(1) expression classification number k value is 6, and (2) training image quantity m value is 156.
3. for the test of dynamic video
In order to test the present invention to the recognition performance of dynamic video, the video shot using camera is included on screen as input, by the recognition result of every two field picture and identification are time-consuming.The best expression classifier parameter of recognition performance in test 1 is chosen at, is set it to:T=60, u=3550(I.e. grader has 60 CART categorised decision trees;The progress of 3550 dimensions is randomly selected in each node growth of each tree from 10000 dimensional features).Concretely comprise the following steps:
Step 1:The video data obtained to USB camera, extracts per frame data, obtains images to be recognized sequence.
Step 2:On the basis of step 1, facial zone data extraction is carried out to each image in image sequence to be identified, face-image sequence to be identified is obtained.
Step 3:Haar-like features are tieed up to every breadth portion image zooming-out F in the face-image sequence to be identified obtained by step 2(In this example, F=10000).
Step 4:On the basis of step 3, the random forest expression classifier obtained using step 3 off-line training in test 1(T=60, u=3550)Every width face-image in face-image sequence to be identified is identified, classification sequence of expressing one's feelings is obtained, and by the recognition result of every frame and the time-consuming output of identification over the display.The identification process of every width face-image is identical with step 4 in test 1.
Test result
For test 1, in order to contrast institute's extracting method of the present invention and difference of the AdaBoost.MH methods in Expression Recognition accuracy rate, carried out using traditional AdaBoost.MH methods with 1 similar experiment of test, " Feature Selection+random forest " method carried with the present invention is compared.Two methods are as shown in Figure 6 to the recognition accuracy of all kinds of expressions.As can clearly see from the figure:AdaBoost.MH to the discrimination highest of eye closing, to dehiscing, the discrimination of surprised two class expression it is relatively low(It is not above 75%).Correspondingly, as u > 900, " Feature Selection+random forest " method has exceeded 90% to the recognition accuracy that 5 classes are expressed one's feelings.
In terms of recognition speed, table 3 have recorded the average of each link when in test 1 750 width pictures are carried out with Expression Recognition and take.It can be seen that, the identification of " Feature Selection+random forest " method is taken as 5.2ms, adds the time overhead of recognition of face, recognition speed can reach 27.62 frames/second.
The identification of table 3 " Feature Selection+random forest " method takes
Figure BDA00002075520300161
For test 2, equally carried out using traditional AdaBoost.MH methods with 2 similar experiments of test, " Feature Selection+random forest " method carried with the present invention is compared.Fig. 7 illustrates the performance of two methods overall accuracy, it is seen then that the accuracy rate of method of the invention is significantly higher than AdaBoost.MH methods.
In test 3, this method has very high recognition accuracy;Meanwhile, the Expression Recognition of single frames is time-consuming in 5ms or so.
Above-mentioned 3 tests test result indicates that, the present invention has that accuracy rate is high, fireballing feature.The 10 folding cross validation results on CAS-PEAL-R1 expressions storehouse show that overall recognition accuracy reaches 94.7%;In the opener test that JAFFE expresses one's feelings storehouse, 91.2% recognition accuracy also obtain;In terms of recognition speed, the average identification of every face is taken as 5.2ms, can meet the demand of Real time identification.

Claims (5)

1. the high-precision recognition methods of multi-class facial expression, it is characterised in that:Comprise the following steps:
Step 1, using multiple facial zone images as positive sample, multiple non-face area images as negative sample carry out off-line training, obtain face recognition grader;
Step 2, on the basis of step 1, the off-line training of facial expression grader is carried out;Detailed process is as follows:
Step 2.1, expression mark is carried out to face-image training data, specific method is:The picture of various expression classifications to be identified or the key frame of video are collected, training image collection A is formed, wherein the picture number included is m;Using continuous integer numbering as each pictures or the class label of key frame, it is expression classification number to be identified to form expression class label collection Y={ 1,2 ..., k }, wherein k;
Step 2.2, facial zone data extraction is carried out to every width training image after step 2.1 mark, the face-image being cut out forms training image collection B;
Step 2.3, in order to train expression classifier, each image in the training image collection B formed to step 2.2 carries out the second extraction of Haar-like features, and the specific method of Haar-like feature extractions is:The integrogram for the image that each width is cut out is calculated, calculating corresponding H according to each integrogram ties up Haar-like characteristic values;
The corresponding H dimensions Haar-like characteristic vectors of each image are denoted as a line, whole H of all m width images is tieed up Haar-like characteristic vectors and constitutes a m row, the eigenmatrix X of H row;
Step 2.4, using AdaBoost.MH algorithms, the Haar-like eigenmatrixes X obtained to step 2.3 carries out Feature Selection;Detailed process is:
Step 2.4.1, the corresponding weight of initialization each image, is denoted as D1(i, yi)=1/ (mk), yi∈ Y represent the expression class label of i-th of image, i=1 ... m;
Step 2.4.2, starts f wheel iteration, f=1 ... F:Eigenmatrix X each column data is carried out H computing, obtain r as the input of a Weak Classifier successivelyF, j
r f , j = &Sigma; i = 1 m D f ( i , y i ) K [ l i ] h j ( x i , j , y i ) ,
Wherein, j=1 ... H, xI, jRepresent j-th of element in X the i-th row;hj(xI, j, yi) represent with xI, jIt is used as the Weak Classifier of input, Df(i, yi) represent that f takes turns the weighted value of i-th of training image in iteration,
K [ y i ] = + 1 y i &Element; Y = { 1,2 , . . . , k } - 1 y i &NotElement; Y = { 1,2 , . . . , k } ;
Terminate after H computing, H r for taking epicycle iteration to obtainF, jIn maximum, be denoted as rf, and by rfIt is corresponding, using X jth dimensional feature value xj as input Weak Classifier hj(xj, Y), the Weak Classifier h filtered out is taken turns as ff(xj, Y), while by xjNew feature space is added to as the feature dimensions filtered out;
Step 2.4.3, calculates the Weak Classifier h selected by step 2.4.2f(xj, Y) weight αf
&alpha; f = 1 2 ln ( 1 + r f 1 - r f ) ;
Step 2.4.4, calculates the weight D of each image in f+1 wheel iterationf+1
D f + 1 = D f ( i , y i ) exp ( - &alpha; f K [ y i ] h f ( x i , j , y i ) ) Z f , i = 1 . . . m .
Wherein, hf(xI, j,yi) represent f wheel iteration in filter out, the Weak Classifier of input, Z are used as using the jth dimensional feature value of i-th of imagefIt is normalization factor
Z f = &Sigma; i = 1 m D f ( i , y i ) exp ( - &alpha; f K [ y i ] h f ( x i , j , y i ) ) ;
Step 2.4.5, the new weight that step 2.4.4 is obtained substitutes into step 2.4.2, according to step 2.4.2 to step 2.4.4 method iteration, until filtering out F dimension principal characters, and F row are extracted in eigenmatrix X, form a m row, the principal character matrix X ' of F row;
Step 3, the expression class label collection Y marked using the principal character matrix X ' obtained through step 2 and by step 2.1, training generation Expression Recognition grader, the process of training follows random forests algorithm, and specific method is:
Step 3.1, decision tree number T and node intrinsic dimensionality u in design requirement, generate T CART categorised decision tree;The record format of the root node of the decision tree is N (J), and the record format of intermediate node is N (V, J), and the record format of leafy node is (V, J, yt);Wherein, J represents node N disruptive features dimension, and V represents node N characteristic value, ytRepresent node N class label;
The generation method of every CART categorised decision tree is:
Step 3.1.1, carries out that grab sample can be put back to m times, and principal character matrix X ' a line is extracted every time, constitutes a new m row, the matrix X " of F row, the growth for this CART categorised decision tree;The corresponding training sample label of each row feature constitutes new expression class label collection Y " in X ";
Step 3.1.2, since root node, node split is carried out by node, is finally completed the growth of whole tree;Each the fission process of node is:
a)The random selection u row from eigenmatrix X "
Figure FDA00002075520200024
As the training data needed for this node split, wherein
Figure FDA00002075520200031
Represent X " jth row;
b)The x selected " j information gain IG is calculated respectivelyj, obtain u IGj;
IG j = IG ( Y &prime; &prime; | x &OverBar; j &prime; &prime; ) = H ( Y &prime; &prime; ) - H ( Y &prime; &prime; | x &OverBar; j &prime; &prime; ) ,
Wherein, H (Y ") is expression class label collection Y " comentropy:
H ( Y &prime; &prime; ) = - &Sigma; w = 1 k p ( y &prime; &prime; w = V w ) &CenterDot; log 2 p ( y &prime; &prime; w = V w ) ,
VwRepresent the value of w-th of class label in Y, Vw∈ { 1,2 ..., k };
It is based on for expression class label collection Y "
Figure FDA00002075520200035
Conditional entropy:
H ( Y &prime; &prime; | x &OverBar; j &prime; &prime; ) = &Sigma; s = 1 h p ( x &prime; &prime; j = V s ) &CenterDot; H ( Y &prime; &prime; | x &prime; &prime; j = V s )
= - &Sigma; s = 1 h p ( x &prime; &prime; j = V s ) &CenterDot; &Sigma; w = 1 q p ( y &prime; &prime; w = V w | s ) &CenterDot; log 2 p ( y &prime; &prime; w = V w | s )
VsRepresent x "jS kind values in middle all different values of element;Vw/sRepresent VsCorresponding expression class label;H (Y " | x "J, s=Vs) it is VsThe comentropy of corresponding expression class label set, x "jRepresent the element in X " jth row, h≤m, q≤k;
c)The u information gain value IG that comparison step b is obtainedj, and IG will be made in X "jThe maximum row of value are extracted, and are denoted as x "J, while the columns J that this is listed in X ' is recorded, tieed up as the disruptive features of this node;
d)Count x "JIn all different characteristic values quantity c, then set up c respectively using different characteristic value as node characteristic value V child node to current node, and using the child node as the root node of subtree, the new subtree of generation, the growing method of subtree is:By x "JIn all values be equal to the element of this feature value where row vector in X " propose, constitute new eigenmatrix Xv, then by the corresponding institute's espressiove class label proposition of proposed row vector, constitute new expression class label set Yv;Then (X is usedv, Yv) (X ", Y ") is substituted, step a-d operation is recursively carried out, until meeting following condition for the moment, terminates the growth of this subtree:
①XvLine number to be less than in 2, or each row all characteristic values all equal when causing this node can not to continue division, corresponding expression class label set YvMiddle frequency of occurrences highest label as the node class label ytPreserve;
2. the Y under this nodevIn expression class label it is all identical when, unique expression class label is used as to the class label y of the nodetPreserve;
Step 3.2, all T CART categorised decision trees are preserved, final random forest Expression Recognition grader is formed;
Step 4, the random forest Expression Recognition grader obtained using step 3 off-line training, ONLINE RECOGNITION is carried out to still image to be measured or dynamic video;
1) recognition methods to still image is:
Step a, extracts the facial zone in still image to be identified;
Step b, on the basis of step a, according to Haar-like feature extracting methods, and the principal character matrix X ' that step 2.4.5 is obtained, extracts the F dimension Haar-like features needed for identification, constitutes the characteristic vector of facial expression image to be identified, be denoted as x;Use xJRepresent characteristic vector x J dimensional feature values;
The characteristic vector x of facial expression image to be identified is identified respectively for T CART categorised decisions tree in step c, the random forest Expression Recognition grader obtained using step 3 off-line training, and the identification of every CART categorised decision tree is since root node, and detailed process is:
C.1. the disruptive features dimension J of current node is obtained from grader, the characteristic vector x of facial expression image to be identified J dimensional feature values x is readJ
C.2. searched in the child node of current node, select node characteristic value and xJMost close child node;
C.3. operation c.1-c.2 is recursively carried out repeatedly, until current node is leafy node, stops recurrence, and by the class label y of the leafy nodetExported as the recognition result of this CART categorised decision tree;
Step d, to the T output result y of T CART categorised decision tree in random forest Expression Recognition gradertCounted, will appear from frequency highest class label and exported as final recognition result;
2) recognition methods to dynamic video is:
Step e, is decoded to video file, is extracted per frame data, is obtained images to be recognized sequence;
Step f, on the basis of step e, carries out facial zone data extraction to each image in image sequence to be identified, obtains face-image sequence to be identified;
Step g, according to the Haar-like feature extracting methods described in step 2.2, the F dimension Haar-like features filtered out to every breadth portion image extracting step 2.4.5 in the face-image sequence to be identified obtained by step f;
Step h, on the basis of step g, every width face-image in face-image sequence to be identified is identified the random forest Expression Recognition grader obtained using step 3 off-line training, obtains classification sequence of expressing one's feelings;The identification process of the face-image is identical with step c, d;
Step i, the expression classification sequence obtained to step h carries out smooth, the burr judgement among removal recognition sequence, obtains final recognition result.
2. the high-precision recognition methods of multi-class facial expression according to claim 1, it is characterised in that:In Feature Selection method described in step 2.4, the input of the Weak Classifier used in interative computation is the one-dimensional characteristic value in characteristic vector, meanwhile, for expression class label y to be identified, the output of Weak Classifier is 1 or -1.
3. the high-precision recognition methods of multi-class facial expression according to claim 1, it is characterised in that:The specific method of facial zone data extraction is described in step 2.2:The integrogram of each width image is calculated first;Described integrogram is identical with original image size, and the value of any point (x, y) is artwork corresponding points (x ', y ') and its all pixel value sums in upper left side thereon:
ii ( x , y ) = &Sigma; x &prime; &le; x , y &prime; &le; y i ( x &prime; , y &prime; ) ,
Ii (x, y) represents point x, y on integrogram in formula) value, i (x ', y ') represents the pixel value of point (x ', y ') on original image;
Calculating is obtained after integrogram, the face recognition grader obtained according to step 1, and the Haar-like features of image in sliding window region are extracted using sliding window method, facial zone is quickly determined;Then the image cutting-out of facial zone is come out, the size zoomed to needed for Expression Recognition, keeps original expression mark.
4. the high-precision recognition methods of multi-class facial expression according to claim 1, it is characterised in that:The value of H described in step 2.3 is determined by the Haar-like characteristic types and dimension of picture used.
5. the high-precision recognition methods of multi-class facial expression according to claim 1, it is characterised in that:Face recognition grader described in step 1 is obtained using the AdaBoost Cascading Classifier training methods based on Haar-like features.
CN201210314435.4A 2012-08-30 2012-08-30 Method for identifying multi-class facial expressions at high precision Expired - Fee Related CN102831447B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201210314435.4A CN102831447B (en) 2012-08-30 2012-08-30 Method for identifying multi-class facial expressions at high precision

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201210314435.4A CN102831447B (en) 2012-08-30 2012-08-30 Method for identifying multi-class facial expressions at high precision

Publications (2)

Publication Number Publication Date
CN102831447A true CN102831447A (en) 2012-12-19
CN102831447B CN102831447B (en) 2015-01-21

Family

ID=47334573

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201210314435.4A Expired - Fee Related CN102831447B (en) 2012-08-30 2012-08-30 Method for identifying multi-class facial expressions at high precision

Country Status (1)

Country Link
CN (1) CN102831447B (en)

Cited By (39)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103248955A (en) * 2013-04-22 2013-08-14 深圳Tcl新技术有限公司 Identity recognition method and device based on intelligent remote control system
CN104284252A (en) * 2014-09-10 2015-01-14 康佳集团股份有限公司 Method for generating electronic photo album automatically
CN104376333A (en) * 2014-09-25 2015-02-25 电子科技大学 Facial expression recognition method based on random forests
WO2015078007A1 (en) * 2013-11-29 2015-06-04 徐勇 Quick human face alignment method
WO2015089949A1 (en) * 2013-12-19 2015-06-25 成都品果科技有限公司 Human face clustering method merging lbp and gabor features
CN104966088A (en) * 2015-06-02 2015-10-07 南昌航空大学 Fracture image recognition method based on Grouplet-variational relevance vector machine
CN105117740A (en) * 2015-08-21 2015-12-02 北京旷视科技有限公司 Font identification method and device
CN105139041A (en) * 2015-08-21 2015-12-09 北京旷视科技有限公司 Method and device for recognizing languages based on image
CN105303149A (en) * 2014-05-29 2016-02-03 腾讯科技(深圳)有限公司 Figure image display method and apparatus
CN105718937A (en) * 2014-12-03 2016-06-29 财团法人资讯工业策进会 Multi-class object classification method and system
CN106528586A (en) * 2016-05-13 2017-03-22 上海理工大学 Human behavior video identification method
CN106778677A (en) * 2016-12-30 2017-05-31 东北农业大学 Feature based selection and driver's fatigue state recognition method and device of facial multizone combining classifiers
CN106845531A (en) * 2016-12-30 2017-06-13 东北农业大学 The method and system of face fatigue state identification are carried out using the first yojan of relative covering
CN106874921A (en) * 2015-12-11 2017-06-20 清华大学 Image classification method and device
CN107341688A (en) * 2017-06-14 2017-11-10 北京万相融通科技股份有限公司 The acquisition method and system of a kind of customer experience
CN108288048A (en) * 2018-02-09 2018-07-17 中国矿业大学 Based on the facial emotions identification feature selection method for improving brainstorming optimization algorithm
CN108460364A (en) * 2018-03-27 2018-08-28 百度在线网络技术(北京)有限公司 Method and apparatus for generating information
CN108710820A (en) * 2018-03-30 2018-10-26 百度在线网络技术(北京)有限公司 Infantile state recognition methods, device and server based on recognition of face
CN108734214A (en) * 2018-05-21 2018-11-02 Oppo广东移动通信有限公司 Image-recognizing method and device, electronic equipment, storage medium
CN108982544A (en) * 2018-06-20 2018-12-11 青岛联合创智科技有限公司 A kind of printed circuit board flaw component detection method
CN109086657A (en) * 2018-06-08 2018-12-25 华南理工大学 A kind of ear detection method, system and model based on machine learning
CN109359532A (en) * 2018-09-12 2019-02-19 中国人民解放军国防科技大学 BGP face recognition method based on heuristic information
CN109359675A (en) * 2018-09-28 2019-02-19 腾讯科技(武汉)有限公司 Image processing method and equipment
CN109376717A (en) * 2018-12-14 2019-02-22 中科软科技股份有限公司 Personal identification method, device, electronic equipment and the storage medium of face comparison
WO2019042080A1 (en) * 2017-08-29 2019-03-07 Hu Man Ren Gong Zhi Neng Ke Ji (Shanghai) Limited Image data processing system and method
CN109784143A (en) * 2018-11-27 2019-05-21 中国电子科技集团公司第二十八研究所 A kind of micro- expression classification method based on optical flow method
CN109829959A (en) * 2018-12-25 2019-05-31 中国科学院自动化研究所 Expression edition method and device based on face parsing
CN109840485A (en) * 2019-01-23 2019-06-04 科大讯飞股份有限公司 A kind of micro- human facial feature extraction method, apparatus, equipment and readable storage medium storing program for executing
CN110175578A (en) * 2019-05-29 2019-08-27 厦门大学 Micro- expression recognition method based on depth forest applied to criminal investigation
CN110321845A (en) * 2019-07-04 2019-10-11 北京奇艺世纪科技有限公司 A kind of method, apparatus and electronic equipment for extracting expression packet from video
WO2019223513A1 (en) * 2018-05-21 2019-11-28 Oppo广东移动通信有限公司 Image recognition method, electronic device and storage medium
CN110532971A (en) * 2019-09-02 2019-12-03 京东方科技集团股份有限公司 Image procossing and device, training method and computer readable storage medium
CN112149596A (en) * 2020-09-29 2020-12-29 厦门理工学院 Abnormal behavior detection method, terminal device and storage medium
CN112492389A (en) * 2019-09-12 2021-03-12 上海哔哩哔哩科技有限公司 Video pushing method, video playing method, computer device and storage medium
CN113111789A (en) * 2021-04-15 2021-07-13 山东大学 Facial expression recognition method and system based on video stream
WO2021218415A1 (en) * 2020-04-30 2021-11-04 京东方科技集团股份有限公司 Expression recognition method and apparatus, electronic device, and storage medium
CN114004982A (en) * 2021-10-27 2022-02-01 中国科学院声学研究所 Acoustic Haar feature extraction method and system for underwater target recognition
CN114373214A (en) * 2022-01-14 2022-04-19 平安普惠企业管理有限公司 User psychological analysis method, device, equipment and storage medium based on micro expression
US11843878B2 (en) 2021-06-11 2023-12-12 Infineon Technologies Ag Sensor devices, electronic devices, method for performing object detection by a sensor device, and method for performing object detection by an electronic device

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
TAKESHI MITA,TOSHIMITSU KANEKO,OSAMU HORI: "《Joint Haar-like Features for Face Detection》", 《PROCEEDINGS OF THE TENTH IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV’05)》 *
谢尔曼, 罗森林, 潘丽敏: "《基于Haar 特征的Turbo-Boost表情识别算法》", 《计算机辅助设计与图形学学报》 *
马景义 等: "《拟自适应分类随机森林算法》", 《数理统计与管理》 *

Cited By (50)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103248955A (en) * 2013-04-22 2013-08-14 深圳Tcl新技术有限公司 Identity recognition method and device based on intelligent remote control system
WO2015078007A1 (en) * 2013-11-29 2015-06-04 徐勇 Quick human face alignment method
WO2015089949A1 (en) * 2013-12-19 2015-06-25 成都品果科技有限公司 Human face clustering method merging lbp and gabor features
CN105303149A (en) * 2014-05-29 2016-02-03 腾讯科技(深圳)有限公司 Figure image display method and apparatus
CN104284252A (en) * 2014-09-10 2015-01-14 康佳集团股份有限公司 Method for generating electronic photo album automatically
CN104376333A (en) * 2014-09-25 2015-02-25 电子科技大学 Facial expression recognition method based on random forests
CN105718937A (en) * 2014-12-03 2016-06-29 财团法人资讯工业策进会 Multi-class object classification method and system
CN105718937B (en) * 2014-12-03 2019-04-05 财团法人资讯工业策进会 Multi-class object classification method and system
CN104966088A (en) * 2015-06-02 2015-10-07 南昌航空大学 Fracture image recognition method based on Grouplet-variational relevance vector machine
CN104966088B (en) * 2015-06-02 2018-10-23 南昌航空大学 Based on small echo in groups-variation interconnection vector machine fracture surface image recognition methods
CN105117740A (en) * 2015-08-21 2015-12-02 北京旷视科技有限公司 Font identification method and device
CN105139041A (en) * 2015-08-21 2015-12-09 北京旷视科技有限公司 Method and device for recognizing languages based on image
CN106874921A (en) * 2015-12-11 2017-06-20 清华大学 Image classification method and device
CN106528586A (en) * 2016-05-13 2017-03-22 上海理工大学 Human behavior video identification method
CN106778677A (en) * 2016-12-30 2017-05-31 东北农业大学 Feature based selection and driver's fatigue state recognition method and device of facial multizone combining classifiers
CN106845531A (en) * 2016-12-30 2017-06-13 东北农业大学 The method and system of face fatigue state identification are carried out using the first yojan of relative covering
CN107341688A (en) * 2017-06-14 2017-11-10 北京万相融通科技股份有限公司 The acquisition method and system of a kind of customer experience
WO2019042080A1 (en) * 2017-08-29 2019-03-07 Hu Man Ren Gong Zhi Neng Ke Ji (Shanghai) Limited Image data processing system and method
CN108288048A (en) * 2018-02-09 2018-07-17 中国矿业大学 Based on the facial emotions identification feature selection method for improving brainstorming optimization algorithm
CN108288048B (en) * 2018-02-09 2021-11-23 中国矿业大学 Facial emotion recognition feature selection method based on improved brainstorming optimization algorithm
CN108460364A (en) * 2018-03-27 2018-08-28 百度在线网络技术(北京)有限公司 Method and apparatus for generating information
CN108460364B (en) * 2018-03-27 2022-03-11 百度在线网络技术(北京)有限公司 Method and apparatus for generating information
CN108710820A (en) * 2018-03-30 2018-10-26 百度在线网络技术(北京)有限公司 Infantile state recognition methods, device and server based on recognition of face
CN108734214A (en) * 2018-05-21 2018-11-02 Oppo广东移动通信有限公司 Image-recognizing method and device, electronic equipment, storage medium
WO2019223513A1 (en) * 2018-05-21 2019-11-28 Oppo广东移动通信有限公司 Image recognition method, electronic device and storage medium
CN109086657A (en) * 2018-06-08 2018-12-25 华南理工大学 A kind of ear detection method, system and model based on machine learning
CN108982544A (en) * 2018-06-20 2018-12-11 青岛联合创智科技有限公司 A kind of printed circuit board flaw component detection method
CN109359532A (en) * 2018-09-12 2019-02-19 中国人民解放军国防科技大学 BGP face recognition method based on heuristic information
CN109359675A (en) * 2018-09-28 2019-02-19 腾讯科技(武汉)有限公司 Image processing method and equipment
CN109359675B (en) * 2018-09-28 2022-08-12 腾讯科技(武汉)有限公司 Image processing method and apparatus
CN109784143A (en) * 2018-11-27 2019-05-21 中国电子科技集团公司第二十八研究所 A kind of micro- expression classification method based on optical flow method
CN109376717A (en) * 2018-12-14 2019-02-22 中科软科技股份有限公司 Personal identification method, device, electronic equipment and the storage medium of face comparison
CN109829959A (en) * 2018-12-25 2019-05-31 中国科学院自动化研究所 Expression edition method and device based on face parsing
CN109840485B (en) * 2019-01-23 2021-10-08 科大讯飞股份有限公司 Micro-expression feature extraction method, device, equipment and readable storage medium
CN109840485A (en) * 2019-01-23 2019-06-04 科大讯飞股份有限公司 A kind of micro- human facial feature extraction method, apparatus, equipment and readable storage medium storing program for executing
CN110175578A (en) * 2019-05-29 2019-08-27 厦门大学 Micro- expression recognition method based on depth forest applied to criminal investigation
CN110175578B (en) * 2019-05-29 2021-06-22 厦门大学 Deep forest-based micro expression identification method applied to criminal investigation
CN110321845A (en) * 2019-07-04 2019-10-11 北京奇艺世纪科技有限公司 A kind of method, apparatus and electronic equipment for extracting expression packet from video
CN110321845B (en) * 2019-07-04 2021-06-18 北京奇艺世纪科技有限公司 Method and device for extracting emotion packets from video and electronic equipment
US11961327B2 (en) 2019-09-02 2024-04-16 Boe Technology Group Co., Ltd. Image processing method and device, classifier training method, and readable storage medium
CN110532971A (en) * 2019-09-02 2019-12-03 京东方科技集团股份有限公司 Image procossing and device, training method and computer readable storage medium
CN112492389B (en) * 2019-09-12 2022-07-19 上海哔哩哔哩科技有限公司 Video pushing method, video playing method, computer device and storage medium
CN112492389A (en) * 2019-09-12 2021-03-12 上海哔哩哔哩科技有限公司 Video pushing method, video playing method, computer device and storage medium
WO2021218415A1 (en) * 2020-04-30 2021-11-04 京东方科技集团股份有限公司 Expression recognition method and apparatus, electronic device, and storage medium
CN112149596A (en) * 2020-09-29 2020-12-29 厦门理工学院 Abnormal behavior detection method, terminal device and storage medium
CN113111789B (en) * 2021-04-15 2022-12-20 山东大学 Facial expression recognition method and system based on video stream
CN113111789A (en) * 2021-04-15 2021-07-13 山东大学 Facial expression recognition method and system based on video stream
US11843878B2 (en) 2021-06-11 2023-12-12 Infineon Technologies Ag Sensor devices, electronic devices, method for performing object detection by a sensor device, and method for performing object detection by an electronic device
CN114004982A (en) * 2021-10-27 2022-02-01 中国科学院声学研究所 Acoustic Haar feature extraction method and system for underwater target recognition
CN114373214A (en) * 2022-01-14 2022-04-19 平安普惠企业管理有限公司 User psychological analysis method, device, equipment and storage medium based on micro expression

Also Published As

Publication number Publication date
CN102831447B (en) 2015-01-21

Similar Documents

Publication Publication Date Title
CN102831447B (en) Method for identifying multi-class facial expressions at high precision
Liu et al. The treasure beneath convolutional layers: Cross-convolutional-layer pooling for image classification
Bar et al. Classification of artistic styles using binarized features derived from a deep neural network
Dollár et al. Integral channel features.
EP2948877B1 (en) Content based image retrieval
Shotton et al. Semantic texton forests for image categorization and segmentation
CN102938065B (en) Face feature extraction method and face identification method based on large-scale image data
CN103605952B (en) Based on the Human bodys&#39; response method that Laplce&#39;s canonical group is sparse
CN104778457A (en) Video face identification algorithm on basis of multi-instance learning
CN101187986A (en) Face recognition method based on supervisory neighbour keeping inlaying and supporting vector machine
Folego et al. From impressionism to expressionism: Automatically identifying van Gogh's paintings
Chen et al. Face recognition algorithm based on VGG network model and SVM
Wang et al. S 3 D: Scalable pedestrian detection via score scale surface discrimination
CN106778714A (en) LDA face identification methods based on nonlinear characteristic and model combination
CN103942572A (en) Method and device for extracting facial expression features based on bidirectional compressed data space dimension reduction
Tian et al. Support vector machine with mixture of kernels for image classification
CN112560894A (en) Improved 3D convolutional network hyperspectral remote sensing image classification method and device
Sasankar et al. A study for Face Recognition using techniques PCA and KNN
CN113887509B (en) Rapid multi-modal video face recognition method based on image set
Lu et al. Real-time facial expression recognition based on pixel-pattern-based texture feature
Liu et al. Gabor feature representation method based on block statistics and its application to facial expression recognition
Zhao et al. Sign text detection in street view images using an integrated feature
Jiang et al. Face recognition by combining wavelet transform and K-nearest neighbor
Agarwal et al. HOG feature and vocabulary tree for content-based image retrieval
Li et al. Block-based bag of words for robust face recognition under variant conditions of facial expression, illumination, and partial occlusion

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20150121

Termination date: 20150830

EXPY Termination of patent right or utility model