CN109492529A - A kind of Multi resolution feature extraction and the facial expression recognizing method of global characteristics fusion - Google Patents

A kind of Multi resolution feature extraction and the facial expression recognizing method of global characteristics fusion Download PDF

Info

Publication number
CN109492529A
CN109492529A CN201811167972.4A CN201811167972A CN109492529A CN 109492529 A CN109492529 A CN 109492529A CN 201811167972 A CN201811167972 A CN 201811167972A CN 109492529 A CN109492529 A CN 109492529A
Authority
CN
China
Prior art keywords
training
data
test
set data
random
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201811167972.4A
Other languages
Chinese (zh)
Inventor
王海波
叶宾
李会军
张家铭
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China University of Mining and Technology CUMT
Chinese University of Hong Kong CUHK
Original Assignee
China University of Mining and Technology CUMT
Chinese University of Hong Kong CUHK
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China University of Mining and Technology CUMT, Chinese University of Hong Kong CUHK filed Critical China University of Mining and Technology CUMT
Priority to CN201811167972.4A priority Critical patent/CN109492529A/en
Publication of CN109492529A publication Critical patent/CN109492529A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/174Facial expression recognition
    • G06V40/175Static expression
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • G06V40/169Holistic features and representations, i.e. based on the facial image taken as a whole

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Evolutionary Computation (AREA)
  • Computational Linguistics (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses the facial expression recognizing methods that a kind of Multi resolution feature extraction and global characteristics merge, and select a face expression data collection as initial data, initial data is divided into training set data and test set data;The convolutional neural networks of Multi resolution feature extraction and global characteristics fusion are constructed using TensorFlow artificial intelligence learning system;Convolutional neural networks read training set data, model training is carried out after pre-processing to training set data, then read test collection data, affiliated expression classification is successively identified to each expression in test set data, after the identification for completing institute's espressiove, the Average Accuracy of calculating institute's espressiove and average F1-score index, are finally completed the process of facial expression recognition.Present invention recognition speed in the case where guaranteeing that recognition accuracy is high is fast, while adapting to a variety of light environments with stronger robustness, so as to effectively meet application request.

Description

A kind of Multi resolution feature extraction and the facial expression recognizing method of global characteristics fusion
Technical field
The present invention relates to a kind of facial expression recognizing method, specifically a kind of Multi resolution feature extraction and global characteristics fusion Facial expression recognizing method.
Background technique
For people in usual exchange, facial expression is a kind of wherein important mode, a kind of its carrier as information, Many information that can't be conveyed by words are contained, we are known as non-linguistic information.Since face can very directly in exchange It is observed, therefore expression can transmit emotion information in a manner of more intuitive.In recent years, sentiment analysis is gradually ground by numerous The concern for the person of studying carefully, and facial expression recognition is the important component of emotion analysis, has also obtained quick development, especially In intelligent human-machine interaction, computer is by acquiring the expression of people come the emotion information of reasoning people and carrying out imitation of emotion.
The step of being currently based on the facial expression recognition algorithm of traditional computer vision can be generally divided into two steps: feature mentions It takes and tagsort.Have the feature of the engineers of many maturations at present, such as: LBP feature, HOG feature, Haar feature, Gabor characteristic etc., what these were characterized in designing by artificial experimental summary, wherein the subjective factor comprising people is inside, The distribution that sufficiently can not objectively reflect feature in image brings uncertain factor to the identification of expression.Also have simultaneously The traditional characteristic sorting algorithm that many is had excellent performance, such as: PCA+LDA, SVM, Adaboost etc., deep learning upsurge it Before, these algorithms are since itself excellent performance has obtained everybody favor, and wherein SVM even more once becomes numerous classification problems Preferred model.But there is also deficiencies for traditional calculations visible sensation method: 1, artificial design rule carries out feature extraction, makes finally to extract Feature include a part of subjective factor, really cannot objectively reflect the true distribution of data;2, the feature haveing excellent performance mentions It takes method often computationally intensive, is difficult to realize identification in real time on limited hardware resource.
In view of the above problems, document (Zhang Guoliang, Zhao Zhu Jun, Du Jixiang, Wang Zhanni, Wang Tian view-based access control model Expression analysis The small-sized microcomputer system of interactive expression robot system research [J], 2017 (6): 1381-1386.) it proposes using more Size, multi-angle Gabor wavelet filter extract facial characteristics, obtain character representation more better than LBP, Gabor is small The principle of wave conversion is to carry out Fourier transformation using signal of the window function to each time slice, due to the items of its filter Characteristic is all similar with human visual system, this is also the major reason that it is excellent in computer vision problem;Liu Weifeng Deng (Liu Weifeng, Japanese plum Juan, LBP signature analysis [J] the computer engineering and application of Wang Yanjiang human face expression, 2011,47 (2): The method for 149-152.) first passing through integral projection orients the key point in the regions such as eyes, nose and mouth to come, then people Face be divided into several sub-regions, then calculate the LBP feature of each subregion for final classification, this method is in JAFFE Classification accuracy on data set is higher by 9 percentage points than traditional LBP method.Document (Andrew J Calder, A.Mike Burton,Paul Miller,Andrew W Young,Shigeru Akamatsu.A principal component Analysis of facial expressions [J] .Vision Research, 2001,41 (9): 1179-1208.) it uses PCA algorithm, the low-dimensional for obtaining face indicate, carry out expression classification with this, but its calculation amount is larger, and PCA cannot be very The detailed information of good reservation image, eventually leading to cannot be applied to well in practical problem, but this method is to face The information such as scale size, direction, position are insensitive, provide thinking for subsequent work;Brave wait of print (prints brave, Shi Jinyu, Liu Dan Facial expression recognition [J] the photoelectric project of flat based on Gabor wavelet, 2009,36 (5): 111-116.) extract face Gabor characteristic, then by PCA algorithm obtain Gabor characteristic low-dimensional indicate, finally using multi-classification algorithm to feature vector into Row classification, has biggish promotion in the discrimination on JAFFE data set compared with 2DPCA algorithm.Document (Wu T, Bartlett M S,Movellan J R.Facial expression recognition using gabor motion energy filters[C].San Francisco:2010IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 2010.) it is filtered by Gabor Method obtains corresponding eigenface, then SVM will be used to classify after eigenface dimensionality reduction;California, USA university develops one kind can With the system for automatically adjusting video playout speed according to the expression of people;In addition the Japanese laboratory ATR discloses a Japan Female face expression data collection JAFFE, while also proposed the facial expression recognition technology based on Gabor wavelet coding (Michael J.Lyons,Shigeru Akemastu,Miyuki Kamachi,Jiro Gyoba.Coding Facial Expressions with Gabor Wavelets[C],Nara:Third IEEE International Conference On Automatic Face and Gesture Recognition, 1998.) and a kind of automatic facial expression recognition system (Michael J.Lyons,Julien Budynek,ShigeruAkamatsu.Automatic Classification of Single Facial Images[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,1999,21(12):1357-1362.).But to there is Expression Recognition accurate for these existing recognition methods In the case that rate is high, recognition speed is slower;And in the faster situation of recognition speed, Expression Recognition accuracy rate is low.
Summary of the invention
In view of the above existing problems in the prior art, the present invention provides a kind of Multi resolution feature extractions and global characteristics to merge Facial expression recognizing method, in the case where guaranteeing that recognition accuracy is high, recognition speed is fast, while adapting to a variety of illumination rings Border has stronger robustness, so as to effectively meet application request.
To achieve the goals above, the technical solution adopted by the present invention is that: a kind of Multi resolution feature extraction and global characteristics The facial expression recognizing method of fusion, specific steps are as follows:
A, it selects a face expression data collection as initial data, initial data is divided into training set data and test set Training set data and test set data are converted into TFrecord lattice using TensorFlow artificial intelligence learning system by data Formula file is saved;
B, the volume of Multi resolution feature extraction and global characteristics fusion is constructed using TensorFlow artificial intelligence learning system Product neural network: Multi resolution feature extraction module is located to the inlet of convolutional neural networks, convolutional neural networks are by multiple volumes Lamination, multiple global pool layers, fused layer and classification layer composition, multiple convolutional layers mention the characteristic pattern of different depth respectively It takes, completes to pass to multiple global pool layers after extracting, multiple global pool layer combination fused layers merge pond by global characteristics It is sent into classification layer after changing operation, the expression classification in layer of classifying is divided into: it is angry, detest, fears, it is glad, it is sad, it is surprised, it is tranquil Totally seven kinds, corresponding English classification is Angry, Disgust, Fear, Happy, Sad, Surprise, Neutal;
C, convolutional neural networks read training set data, pre-process to training set data, and pretreatment includes turning at random Turn, random cropping, random brightness variation and random Gaussian, after the completion with the training set data to convolutional neural networks carry out Model training, after model training of every completion, convolutional neural networks read training set data and carry out facial expression recognition test, The accuracy rate for obtaining the facial expression recognition after the secondary model training, then proceedes to model training next time, in model training mistake Intermediate output model file is saved in journey;
D, training/test accuracy rate data and training/test loss data are exported after each model training, and with curve Form real-time update show, pass through observation training/test accuracy rate curve and training/test loss curve, dynamic adjustment mould Shape parameter;
E, when training/test accuracy rate no longer rises and stablizes, and training/test loss value no longer declines and stablizes, Terminate the model training of convolutional neural networks, and saves the model parameter of last time;
F, convolutional neural networks load the model parameter of last time, then read test collection data, to test set data In each expression successively identify belonging to expression classification, after the identification for completing institute's espressiove, calculate the average standard of institute's espressiove True rate and average F1-score index, are finally completed the process of facial expression recognition.
Further, the detailed process of step A are as follows: select FER2013 human face expression data set as initial data, picture number It is .csv format according to file, by Usage attribute, data set is divided into Training, PrivateTest and PublicTest tri- Point, wherein PrivateTest data set is as test set data using Training data set as training set data;Using TensorFlow artificial intelligence learning system by training set data and test set data be converted into TFrecord formatted file into Row saves, and the efficiency of data is read when this format can improve trained, convenient for the processing and preservation of image.
Further, in the step B pond layer by the characteristic pattern of different depth by global poolization operation after, be transformed to one Dimensional vector, then all vectors are spliced, finally it is sent into classification layer.
Further, the detailed process of the step C are as follows:
Convolutional neural networks are read using multithreadingTFrecordThe training set data of format, reduce data reading The time-consuming taken pre-processes training set data after data are converted to numpy array from tensor type, pretreatment packet Overturning at random, random cropping, random brightness variation and random Gaussian are included, wherein random overturn including flip horizontal, vertically Overturning and horizontal vertical overturning;Random cropping is to be cut to cut origin to original image with the random center of selection, then will give birth to At new images size be normalized to the input size of network;Random brightness transformation is the linear change of the pixel value about original image It changes, effect intuitively is brightness change;Random Gaussian is superposition random Gaussian signal on the basis of original image, is used for mould The electrical noise that imitative camera component generates;Model training, mould are carried out to convolutional neural networks with the training set data after the completion Type training is divided into stress model structure, optimization loss function and error back propagation;Loss function uses Softmax cross entropy letter Number, optimizes loss function using Adam algorithm;After model training of every completion, convolutional neural networks read training set Data carry out facial expression recognition test, obtain the accuracy rate of facial expression recognition after the secondary model training and save the secondary mould Type file;Model training next time is then proceeded to, updates the preservation secondary model file after completing the secondary model training.
Further, the step D detailed process are as follows: after each model training output training/test accuracy rate data with Training/test loss data, and real-time update is shown in graph form, passes through observation training/test accuracy rate curve and instruction White silk/test loss curve, dynamic adjustment model parameter, and the model parameter adjusted will be needed to be stored in individual configuration file In, take automatic being adjusted with the mode combined manually.
Compared with prior art, the present invention has the advantage that
1, the data set FER2013 that the present invention uses has authority in facial expression recognition problem, and has numerous Research institution and scholar conduct a research around the data set;
2, the present invention is the facial expression recognition new method based on convolutional neural networks, has stronger robustness, in light According in changeable environment also have stable recognition effect;
3, the data preprocessing method that uses of the present invention efficiently solves that sample class is uneven and lazy weight is asked Topic;
4, recognition accuracy of the present invention on FER2013 data set is high, and recognition speed is fast, solves other identification sides In the case that method has Expression Recognition accuracy rate height, recognition speed is slower;And in the faster situation of recognition speed, Expression Recognition The low problem of accuracy rate, so as to effectively meet application request.
Detailed description of the invention
Fig. 1 is training flow chart of the invention;
Fig. 2 is the result schematic diagram overturn at random in training data pretreatment of the present invention;
Wherein, the first behavior original image;Second row is followed successively by flip vertical, flip horizontal and vertical-horizontal overturning;
Fig. 3 is the result schematic diagram of random cropping in training data pretreatment of the present invention;
Wherein, the first behavior original image;Three kinds of images that second behavior random cropping obtains;
Fig. 4 is the result schematic diagram of random brightness variation in training data pretreatment of the present invention;
Wherein, the first behavior original image;Three kinds of images that second behavior random brightness changes;
Fig. 5 is the result schematic diagram of random Gaussian in training data pretreatment of the present invention;
Wherein, the first behavior original image;Three kinds of images that second behavior addition random Gaussian obtains;
Fig. 6 is pretreatment process figure of the present invention to training data;
Wherein, k is the probability for controlling final output enhancing image, and range is 0~1;
Fig. 7 is the schematic diagram of Multi resolution feature extraction module in the present invention;
Fig. 8 is the schematic diagram of global characteristics fused layer in the present invention;
Fig. 9 is pretreatment schematic diagram of the present invention to test set data;
Figure 10 is the present invention to the corresponding confusion matrix of test set data recognition result.
Specific embodiment
The present invention will be further described below.
As shown, specific steps of the invention are as follows:
A, select FER2013 human face expression data set as initial data, image data file is .csv format, is passed through Usage attribute, data set are divided into Training, PrivateTest and PublicTest three parts, wherein by training data Collection is used as training set data, and PrivateTest data set is as test set data;System is learnt using TensorFlow artificial intelligence Training set data and test set data are converted into TFrecord formatted file and saved by system, and this format can improve training When read data efficiency, convenient for the processing and preservation of image;Since the training of neural network relies on a large amount of data set, The number and diversity of training data largely affect the performance of neural network;
B, the volume of Multi resolution feature extraction and global characteristics fusion is constructed using TensorFlow artificial intelligence learning system Product neural network: Multi resolution feature extraction module is located to the inlet of convolutional neural networks, convolutional neural networks are by multiple volumes Lamination, multiple global pool layers, fused layer and classification layer composition, multiple convolutional layers mention the characteristic pattern of different depth respectively It takes, completes to pass to multiple global pool layers after extracting, multiple global pool layer combination fused layers merge pond by global characteristics Final classification layer is sent into after changing operation;(such as table 1) sets X as input picture, and A is the convolution kernel having a size of 1*1, k1, k2 3*3 With the convolution kernel of 5*5, Y is that the characteristic pattern splicing generated by upper and lower two-way obtains.Because it is contemplated that various sizes of feature, it should Multi resolution feature extraction module allows neural network relatively sufficiently using the information of original pixels, and the module is in network Inlet, the quantity of characteristic pattern or fewer can improve the property of network under the premise of unobvious increase parameter amount Energy.Two-way head and the tail have 1*1 convolution kernel up and down simultaneously, but purpose is different, the former output channel number is greater than input channel number, make It with being augmented features, prepares for subsequent convolution, because the feature that convolution kernel extracts is past when feature port number is less Toward be it is insufficient, generally improve this problem by expanding channel;The output channel number of the latter is less than input channel number, main Acting on is pressure channel, reduces calculation amount.Image data after extracting obtains after being handled respectively by multiple convolutional layers The characteristic pattern of different depth, neural network pass through global pool layer by the characteristic pattern of different depth after global poolization operation, It is transformed to one-dimensional vector, then all vectors are spliced, is sent into final classification layer.Global pool layer provides for network Invariance on feature locations replaces response in this position with the statistical property of part, and in Feature-scale invariance, Neural network is maintained using various sizes of convolution kernel, a kind of existing method be generated using various sizes of image it is corresponding The characteristic pattern of size, but the calculation amount of network can be greatly increased in this way, the method that the present invention uses is different deep using network On the one hand the characteristic pattern of degree improves the utilization rate of data, does not in addition also have since this method uses the generated data of network Obviously increase network query function amount;
C, convolutional neural networks are read using multithreadingTFrecordThe training set data of format reduces data reading The time-consuming taken, after data are converted to numpy array from tensor type, as shown in fig. 6, being located in advance to training set data Reason, pretreatment includes random overturning, random cropping, random brightness variation and random Gaussian, wherein random overturning includes water Flat overturning, flip vertical and horizontal vertical overturning;Random cropping is to be carried out with the random center of selection to cut origin to original image It cuts, then the new images size of generation is normalized to the input size of network;Particularly may be divided into three steps, the first step be determine with Machine cut size, second step determine random cropping center, and third step is normalized image size.It is assumed that input picture size is D × H ignores channel, random cropping size range are as follows:
εsize∈[εminmax] (1)
Then random cropping is sized to:
εsize=random (εminmax) (2)
Wherein random () indicates to take the random value in section.By random cropping size εsizeAnd input image size can Determine random cropping center εposAre as follows:
Wherein εpos-xpos-yRespectively indicate the transverse and longitudinal coordinate of random cropping center in the picture.It finally again will be after cutting Picture size is normalized to specified size.
Random brightness transformation is the linear transformation of the pixel value about original image, and effect intuitively is brightness change;If Determining luminance factor is εlight, original image pixel is vx,y, wherein subscript x, y indicate pixel coordinate, then the image slices after brightness changes Element is εlight·vx,y
Random Gaussian is superposition random Gaussian signal on the basis of original image, for imitating the generation of camera component Electrical noise;It is assumed that input image pixels are vx,y, wherein subscript x, y indicate pixel coordinate, Gaussian noise be f (mean, Sigma), the enhancing coefficient of noise is k, wherein mean, and sigma, k are the random value in given section, then after adding noise Image pixel be vx,y+k·f(mean,sigma)。
Then again with the probability output original image of k, image is enhanced with the probability output of 1-k, guarantees the instruction of form of ownership with this Neural metwork training can be passed through by practicing image;Model training, mould are carried out to convolutional neural networks with the training set data after the completion Type training is divided into stress model structure, optimization loss function and error back propagation;Specially lost using Adam algorithm optimization Function, and the backpropagation of error is carried out, update network parameter;
The loss function that neural network uses is Softmax cross entropy loss function, and Softmax function is by the defeated of network Numerical value conversion is probability out, it is easier to comparison and processing, formula are as follows:
Wherein yiIt is the corresponding Linear Score function of correct classification, SiIt is corresponding Softmax output of correctly classifying.And Softmax cross entropy loss function are as follows:
C=- ∑iy′ilogyi (5)
After the loss function of Current Situation of Neural Network is calculated, the parameter of neural network is carried out more by Back Propagation Algorithm Newly.
In neural network training process, learning rate lr and learning rate attenuation coefficient δ are individually stated in dynamic configuration file, And other hyper parameters are then stated directly in the form of variable in a program, and under automatic adjustment mode, learning rate lr and learning rate The relationship of attenuation coefficient δ are as follows:
Lr=lr* δepoch (6)
Wherein epoch is current iteration number, and under manual fine-tuning mode, program can access dynamic in each iterative process State configuration file, reads newest primary parameter, under this mode, the relationship of learning rate lr and learning rate attenuation coefficient δ are as follows:
Lr=lr* δ (7)
After model training of every completion, convolutional neural networks read training set data and carry out facial expression recognition test, It obtains the accuracy rate of facial expression recognition after the secondary model training and saves the secondary model file;Then proceed to model next time Training updates the preservation secondary model file after completing the secondary model training.
D, training/test accuracy rate data and training/test loss data are exported after each model training, and with curve Form display is updated in TensorBoard in real time, pass through observation training/test accuracy rate curve and training/test loss Curve, dynamic adjustment model parameter, and the model parameter that adjusts will be needed to be stored in individual configuration file, take automatically with The mode combined manually is adjusted;
E, when training/test accuracy rate no longer rises and stablizes, and training/test loss value no longer declines and stablizes, Terminate the model training of convolutional neural networks, and saves the model parameter of last time;
F, convolutional neural networks load last time model parameter, then read test collection data (i.e. FER2013'sPrivateTestData set), as shown in fig. 6, being pre-processed to test image, respectively on five positions (upper left, upper right, a left sides Under, bottom right and center) cut, and the picture size after cutting is normalized to network inputs size, a final test Image can be transformed to six images (image and original image after five cuttings) and be identified for neural network, finally lead to again Cross ballot decision-making mechanism obtain current test image expression classification and corresponding probability.
The Average Accuracy and F1-score of recognition result are calculated by confusion matrix, and confusion matrix is statistics Common evaluation means, also referred to as Error Matrix in habit and machine learning classification problem.Assuming that classification number is n, then square is obscured Battle array is exactly the square matrix of a n x n, and row indicates the actual distribution of sample, the classification results of column expression classifier, on each position Digital representation ratio, neural network prediction is such ratio in current category of test, then when the diagonal entry of matrix Closer to 1, illustrate that the performance of neural network is more excellent;The confusion matrix stage is being calculated, system is successively read every a kind of expression figure Picture counts recognition result, form is the vector of 7*7, as shown in fig. 7, on position (1,1) after neural network recognization 0.664 indicates that the expression picture all " Angry " classifications in training data passes through neural network recognization, wherein having 66.4% Picture network be identified as " Angry " classification, i.e. identification is correct, and the picture of residue 33.6% is then identified as other classifications.
As shown in table 2, the validity of method of the invention is illustrated, FER2013'sPrivateTestIt is obtained on data set Best recognition effect was obtained, the identification for greatly exceeding the mankind is horizontal.
Table 1: the neural network structure of building
Name Kernel size/stride Input size(Name)
Conv1 3*3*64/1 96*96*1(*)
Multi resolution feature extraction module M - 48*48*32(Conv1)
dConv2 3*3*64/1 48*48*64(M)
dConv3 3*3*128/2 48*48*64(dConv2)
dConv4 3*3*128/1 24*24*128(dConv3)
dConv5 3*3*256/2 24*24*128(dConv4)
dConv6 1*3*256/1 12*12*256(dConv5)
dConv7 3*1*256/1 12*12*256(dConv6)
dConv8 1*3*256/1 12*12*256(dConv7)
dConv9 3*1*256/1 12*12*256(dConv8)
dConv10 1*3*256/1 12*12*256(dConv9)
dConv11 3*1*256/1 12*12*256(dConv10)
dConv12 3*3*512/2 12*12*256(dConv11)
dConv13 3*3*512/1 6*6*512(dConv12)
Global pool layer GP1 6*6/1 6*6*512(dConv13)
Global pool layer GP2 12*12/1 12*12*256(dConv11)
Global pool layer GP3 12*12/1 12*12*256(dConv9)
Fused layer Concat - (GP1+GP2+GP3)
Conv14 1*1/1 1*1*1024(Concat)
Softmax classifier 1*1*7(Conv14)
DConv indicates that depth separates convolution
The experimental result that various network models and the mankind distinguish on the PrivateTest data set of table 2:FER2013
Note: existing method is document (lan J.Goodfellow, Yoshua Bengio, et al.Challenges in Representation Learning:Areport on three machine learning contests[J],Machine Learning, 2015,64:59-63.) in obtain.

Claims (5)

1. a kind of Multi resolution feature extraction and the facial expression recognizing method of global characteristics fusion, which is characterized in that specific steps Are as follows:
A, it selects a face expression data collection as initial data, initial data is divided into training set data and test set number According to training set data and test set data are converted into TFrecord format using TensorFlow artificial intelligence learning system File is saved;
B, the convolution mind of Multi resolution feature extraction and global characteristics fusion is constructed using TensorFlow artificial intelligence learning system Through network: Multi resolution feature extraction module is located at the inlet of convolutional neural networks, and convolutional neural networks are by multiple convolutional layers, more A global pool layer, fused layer and classification layer composition, multiple convolutional layers extract the characteristic pattern of different depth respectively, are completed Multiple global pool layers are passed to after extraction, multiple global pool layer combination fused layers are after global characteristics fusion pondization operation It is sent into classification layer, the expression classification in layer of classifying is divided into: it is angry, detest, fears, it is glad, it is sad, it is surprised, it is totally seven kinds tranquil;
C, convolutional neural networks read training set data, pre-process to training set data, pretreatment include random overturning, with Machine is cut, random brightness changes and random Gaussian, carries out model to convolutional neural networks with the training set data after the completion It trains, after model training of every completion, convolutional neural networks read training set data and carry out facial expression recognition test, obtain The accuracy rate of facial expression recognition, then proceedes to model training next time, during model training after the secondary model training Save intermediate output model file;
D, training/test accuracy rate data and training/test loss data are exported after each model training, and with the shape of curve Formula real-time update is shown, passes through observation training/test accuracy rate curve and training/test loss curve, dynamic adjustment model ginseng Number;
E, when training/test accuracy rate curve no longer rises and stablize, and training/test loss curve no longer decline and stabilization When, terminate the model training of convolutional neural networks, and save the model parameter of last time;
F, convolutional neural networks load the model parameter of last time, then read test collection data, in test set data Each facial expression image successively identifies affiliated expression classification, after the identification for completing institute's espressiove, calculates the average standard of institute's espressiove True rate and average F1-score index, are finally completed the process of facial expression recognition.
2. a kind of Multi resolution feature extraction according to claim 1 and the facial expression recognizing method of global characteristics fusion, It is characterized in that, the detailed process of step A are as follows: select FER2013 human face expression data set as initial data, image data text Part is .csv format, and by Usage attribute, data set is divided into Training, PrivateTest and PublicTest three parts, Wherein using Training data set as training set data, PrivateTest data set is as test set data;Using TensorFlow artificial intelligence learning system by training set data and test set data be converted into TFrecord formatted file into Row saves.
3. a kind of Multi resolution feature extraction according to claim 1 and the facial expression recognizing method of global characteristics fusion, It is characterized in that, in the step B pond layer by the characteristic pattern of different depth by global poolization operation after, be transformed to it is one-dimensional to Amount, then all vectors are spliced, finally it is sent into classification layer.
4. a kind of Multi resolution feature extraction according to claim 1 and the facial expression recognizing method of global characteristics fusion, It is characterized in that, the detailed process of the step C are as follows:
Convolutional neural networks read the training set data of TFrecord format using multithreading, carry out to training set data pre- Processing, pretreatment includes random overturning, random cropping, random brightness variation and random Gaussian, wherein random overturning includes Flip horizontal, flip vertical and horizontal vertical overturning;Random cropping be with the random center of selection for cut origin to original image into Row is cut, then the new images size of generation is normalized to the input size of network;Random brightness transformation is about original image The linear transformation of pixel value, effect intuitively are brightness change;Random Gaussian is superimposed at random on the basis of original image Gaussian signal, for imitating the electrical noise of camera component generation;After the completion with the training set data to convolutional neural networks Model training is carried out, model training is divided into stress model structure, optimization loss function and error back propagation;Loss function uses Softmax intersects entropy function, is optimized using Adam algorithm to loss function;After model training of every completion, convolution mind Training set data is read through network and carries out facial expression recognition test, obtains the standard of the facial expression recognition after the secondary model training True rate simultaneously saves the secondary model file;Model training next time is then proceeded to, updates this time of preservation after completing the secondary model training Model file.
5. a kind of Multi resolution feature extraction according to claim 1 and the facial expression recognizing method of global characteristics fusion, It is characterized in that, the step D detailed process are as follows: exported after each model training training/test accuracy rate data and training/ Loss data are tested, and real-time update is shown in graph form, passes through observation training/test accuracy rate curve and training/survey Loss curve, dynamic adjustment model parameter are tried, and the model parameter adjusted will be needed to be stored in individual configuration file, is taken Automatically it is adjusted with the mode combined manually.
CN201811167972.4A 2018-10-08 2018-10-08 A kind of Multi resolution feature extraction and the facial expression recognizing method of global characteristics fusion Pending CN109492529A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811167972.4A CN109492529A (en) 2018-10-08 2018-10-08 A kind of Multi resolution feature extraction and the facial expression recognizing method of global characteristics fusion

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811167972.4A CN109492529A (en) 2018-10-08 2018-10-08 A kind of Multi resolution feature extraction and the facial expression recognizing method of global characteristics fusion

Publications (1)

Publication Number Publication Date
CN109492529A true CN109492529A (en) 2019-03-19

Family

ID=65690023

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811167972.4A Pending CN109492529A (en) 2018-10-08 2018-10-08 A kind of Multi resolution feature extraction and the facial expression recognizing method of global characteristics fusion

Country Status (1)

Country Link
CN (1) CN109492529A (en)

Cited By (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110008876A (en) * 2019-03-26 2019-07-12 电子科技大学 A kind of face verification method based on data enhancing and Fusion Features
CN110110644A (en) * 2019-04-30 2019-08-09 重庆邮电大学 A kind of Multiscale Fusion facial feature extraction method and system
CN110163080A (en) * 2019-04-02 2019-08-23 腾讯科技(深圳)有限公司 Face critical point detection method and device, storage medium and electronic equipment
CN110188708A (en) * 2019-06-03 2019-08-30 西安工业大学 A kind of facial expression recognizing method based on convolutional neural networks
CN110321807A (en) * 2019-06-13 2019-10-11 南京行者易智能交通科技有限公司 A kind of convolutional neural networks based on multilayer feature fusion are yawned Activity recognition method and device
CN110348322A (en) * 2019-06-19 2019-10-18 西华师范大学 Human face in-vivo detection method and equipment based on multi-feature fusion
CN110472583A (en) * 2019-08-16 2019-11-19 广东工业大学 The micro- Expression Recognition system of face based on deep learning
CN110490134A (en) * 2019-08-20 2019-11-22 四川九洲电器集团有限责任公司 Signal recognition method, equipment and storage medium
CN110728179A (en) * 2019-09-04 2020-01-24 天津大学 Pig face identification method adopting multi-path convolutional neural network
CN110781760A (en) * 2019-05-24 2020-02-11 西安电子科技大学 Facial expression recognition method and device based on space attention
CN110991533A (en) * 2019-12-03 2020-04-10 Oppo广东移动通信有限公司 Image recognition method, recognition device, terminal device and readable storage medium
CN111079514A (en) * 2019-10-28 2020-04-28 湖北工业大学 Face recognition method based on CLBP and convolutional neural network
CN111444787A (en) * 2020-03-12 2020-07-24 江西赣鄱云新型智慧城市技术研究有限公司 Fully intelligent facial expression recognition method and system with gender constraint
CN111507381A (en) * 2020-03-31 2020-08-07 上海商汤智能科技有限公司 Image recognition method and related device and equipment
CN111539942A (en) * 2020-04-28 2020-08-14 中国科学院自动化研究所 Method for detecting face depth tampered image based on multi-scale depth feature fusion
CN111652171A (en) * 2020-06-09 2020-09-11 电子科技大学 Construction method of facial expression recognition model based on double branch network
CN111695407A (en) * 2020-04-23 2020-09-22 西安电子科技大学 Gender identification method, system, storage medium and terminal based on multispectral fusion
CN111709278A (en) * 2020-04-30 2020-09-25 北京航空航天大学 Method for identifying facial expressions of macaques
CN111967389A (en) * 2020-08-18 2020-11-20 厦门理工学院 Face attribute recognition method and system based on deep double-path learning network
CN112151071A (en) * 2020-09-23 2020-12-29 哈尔滨工程大学 Speech emotion recognition method based on mixed wavelet packet feature deep learning
CN112151040A (en) * 2020-09-27 2020-12-29 湖北工业大学 Robust speaker recognition method based on end-to-end joint optimization and decision
CN110321872B (en) * 2019-07-11 2021-03-16 京东方科技集团股份有限公司 Facial expression recognition method and device, computer equipment and readable storage medium
CN112507864A (en) * 2020-12-04 2021-03-16 河北地质大学 Credit archive identification method based on convolutional neural network
CN112784763A (en) * 2021-01-27 2021-05-11 南京邮电大学 Expression recognition method and system based on local and overall feature adaptive fusion
CN112986210A (en) * 2021-02-10 2021-06-18 四川大学 Scale-adaptive microbial Raman spectrum detection method and system
CN113011314A (en) * 2021-03-16 2021-06-22 华南理工大学 Facial expression recognition method based on frequency domain features and product neural network
CN113111940A (en) * 2021-04-13 2021-07-13 东南大学 Expression recognition method based on feature fusion
CN113158828A (en) * 2021-03-30 2021-07-23 华南理工大学 Facial emotion calibration method and system based on deep learning
CN113177558A (en) * 2021-04-13 2021-07-27 电子科技大学 Radiation source individual identification method based on feature fusion of small samples
CN113822229A (en) * 2021-10-28 2021-12-21 重庆科炬企业孵化器有限公司 Expression recognition-oriented user experience evaluation modeling method and device
CN114360009A (en) * 2021-12-23 2022-04-15 电子科技大学长三角研究院(湖州) Multi-scale characteristic face attribute recognition system and method under complex scene
CN114926877A (en) * 2022-05-10 2022-08-19 西北工业大学 Cross-domain facial expression recognition method based on contrast domain difference
WO2023009061A1 (en) * 2021-07-30 2023-02-02 脸萌有限公司 Method and device for evaluating effect of performing classification on fuzzy attribute

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106548160A (en) * 2016-11-09 2017-03-29 浙江博天科技有限公司 A kind of face smile detection method
CN107330420A (en) * 2017-07-14 2017-11-07 河北工业大学 The facial expression recognizing method of rotation information is carried based on deep learning
CN108197633A (en) * 2017-11-24 2018-06-22 百年金海科技有限公司 Deep learning image classification based on TensorFlow is with applying dispositions method

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106548160A (en) * 2016-11-09 2017-03-29 浙江博天科技有限公司 A kind of face smile detection method
CN107330420A (en) * 2017-07-14 2017-11-07 河北工业大学 The facial expression recognizing method of rotation information is carried based on deep learning
CN108197633A (en) * 2017-11-24 2018-06-22 百年金海科技有限公司 Deep learning image classification based on TensorFlow is with applying dispositions method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
余锐: ""基于深度学习的人脸表情特征分析"", 《现代计算机(专业版)》 *

Cited By (51)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110008876A (en) * 2019-03-26 2019-07-12 电子科技大学 A kind of face verification method based on data enhancing and Fusion Features
WO2020199931A1 (en) * 2019-04-02 2020-10-08 腾讯科技(深圳)有限公司 Face key point detection method and apparatus, and storage medium and electronic device
CN110163080A (en) * 2019-04-02 2019-08-23 腾讯科技(深圳)有限公司 Face critical point detection method and device, storage medium and electronic equipment
US11734851B2 (en) 2019-04-02 2023-08-22 Tencent Technology (Shenzhen) Company Limited Face key point detection method and apparatus, storage medium, and electronic device
CN110110644A (en) * 2019-04-30 2019-08-09 重庆邮电大学 A kind of Multiscale Fusion facial feature extraction method and system
CN110781760A (en) * 2019-05-24 2020-02-11 西安电子科技大学 Facial expression recognition method and device based on space attention
CN110188708A (en) * 2019-06-03 2019-08-30 西安工业大学 A kind of facial expression recognizing method based on convolutional neural networks
CN110321807A (en) * 2019-06-13 2019-10-11 南京行者易智能交通科技有限公司 A kind of convolutional neural networks based on multilayer feature fusion are yawned Activity recognition method and device
CN110348322A (en) * 2019-06-19 2019-10-18 西华师范大学 Human face in-vivo detection method and equipment based on multi-feature fusion
US11281895B2 (en) 2019-07-11 2022-03-22 Boe Technology Group Co., Ltd. Expression recognition method, computer device, and computer-readable storage medium
CN110321872B (en) * 2019-07-11 2021-03-16 京东方科技集团股份有限公司 Facial expression recognition method and device, computer equipment and readable storage medium
CN110472583A (en) * 2019-08-16 2019-11-19 广东工业大学 The micro- Expression Recognition system of face based on deep learning
CN110472583B (en) * 2019-08-16 2022-04-19 广东工业大学 Human face micro-expression recognition system based on deep learning
CN110490134A (en) * 2019-08-20 2019-11-22 四川九洲电器集团有限责任公司 Signal recognition method, equipment and storage medium
CN110728179A (en) * 2019-09-04 2020-01-24 天津大学 Pig face identification method adopting multi-path convolutional neural network
CN111079514A (en) * 2019-10-28 2020-04-28 湖北工业大学 Face recognition method based on CLBP and convolutional neural network
CN110991533B (en) * 2019-12-03 2023-08-04 Oppo广东移动通信有限公司 Image recognition method, recognition device, terminal device and readable storage medium
CN110991533A (en) * 2019-12-03 2020-04-10 Oppo广东移动通信有限公司 Image recognition method, recognition device, terminal device and readable storage medium
CN111444787B (en) * 2020-03-12 2023-04-07 江西赣鄱云新型智慧城市技术研究有限公司 Fully intelligent facial expression recognition method and system with gender constraint
CN111444787A (en) * 2020-03-12 2020-07-24 江西赣鄱云新型智慧城市技术研究有限公司 Fully intelligent facial expression recognition method and system with gender constraint
CN111507381B (en) * 2020-03-31 2024-04-02 上海商汤智能科技有限公司 Image recognition method, related device and equipment
CN111507381A (en) * 2020-03-31 2020-08-07 上海商汤智能科技有限公司 Image recognition method and related device and equipment
CN111695407A (en) * 2020-04-23 2020-09-22 西安电子科技大学 Gender identification method, system, storage medium and terminal based on multispectral fusion
CN111695407B (en) * 2020-04-23 2023-04-07 西安电子科技大学 Gender identification method, system, storage medium and terminal based on multispectral fusion
CN111539942A (en) * 2020-04-28 2020-08-14 中国科学院自动化研究所 Method for detecting face depth tampered image based on multi-scale depth feature fusion
CN111539942B (en) * 2020-04-28 2021-08-31 中国科学院自动化研究所 Method for detecting face depth tampered image based on multi-scale depth feature fusion
CN111709278A (en) * 2020-04-30 2020-09-25 北京航空航天大学 Method for identifying facial expressions of macaques
CN111652171B (en) * 2020-06-09 2022-08-05 电子科技大学 Construction method of facial expression recognition model based on double branch network
CN111652171A (en) * 2020-06-09 2020-09-11 电子科技大学 Construction method of facial expression recognition model based on double branch network
CN111967389A (en) * 2020-08-18 2020-11-20 厦门理工学院 Face attribute recognition method and system based on deep double-path learning network
CN111967389B (en) * 2020-08-18 2022-02-18 厦门理工学院 Face attribute recognition method and system based on deep double-path learning network
CN112151071B (en) * 2020-09-23 2022-10-28 哈尔滨工程大学 Speech emotion recognition method based on mixed wavelet packet feature deep learning
CN112151071A (en) * 2020-09-23 2020-12-29 哈尔滨工程大学 Speech emotion recognition method based on mixed wavelet packet feature deep learning
CN112151040B (en) * 2020-09-27 2023-04-28 湖北工业大学 Robust speaker recognition method based on end-to-end joint optimization and decision
CN112151040A (en) * 2020-09-27 2020-12-29 湖北工业大学 Robust speaker recognition method based on end-to-end joint optimization and decision
CN112507864A (en) * 2020-12-04 2021-03-16 河北地质大学 Credit archive identification method based on convolutional neural network
CN112784763B (en) * 2021-01-27 2022-07-29 南京邮电大学 Expression recognition method and system based on local and overall feature adaptive fusion
CN112784763A (en) * 2021-01-27 2021-05-11 南京邮电大学 Expression recognition method and system based on local and overall feature adaptive fusion
CN112986210A (en) * 2021-02-10 2021-06-18 四川大学 Scale-adaptive microbial Raman spectrum detection method and system
CN113011314A (en) * 2021-03-16 2021-06-22 华南理工大学 Facial expression recognition method based on frequency domain features and product neural network
CN113011314B (en) * 2021-03-16 2023-07-18 华南理工大学 Facial expression recognition method based on frequency domain characteristics and product neural network
CN113158828A (en) * 2021-03-30 2021-07-23 华南理工大学 Facial emotion calibration method and system based on deep learning
CN113158828B (en) * 2021-03-30 2024-04-09 华南理工大学 Facial emotion calibration method and system based on deep learning
CN113111940A (en) * 2021-04-13 2021-07-13 东南大学 Expression recognition method based on feature fusion
CN113177558A (en) * 2021-04-13 2021-07-27 电子科技大学 Radiation source individual identification method based on feature fusion of small samples
WO2023009061A1 (en) * 2021-07-30 2023-02-02 脸萌有限公司 Method and device for evaluating effect of performing classification on fuzzy attribute
US11978280B2 (en) 2021-07-30 2024-05-07 Lemon Inc. Method and device for evaluating effect of classifying fuzzy attribute
CN113822229A (en) * 2021-10-28 2021-12-21 重庆科炬企业孵化器有限公司 Expression recognition-oriented user experience evaluation modeling method and device
CN114360009A (en) * 2021-12-23 2022-04-15 电子科技大学长三角研究院(湖州) Multi-scale characteristic face attribute recognition system and method under complex scene
CN114926877A (en) * 2022-05-10 2022-08-19 西北工业大学 Cross-domain facial expression recognition method based on contrast domain difference
CN114926877B (en) * 2022-05-10 2024-02-20 西北工业大学 Cross-domain facial expression recognition method based on contrast domain difference

Similar Documents

Publication Publication Date Title
CN109492529A (en) A kind of Multi resolution feature extraction and the facial expression recognizing method of global characteristics fusion
CN110427867B (en) Facial expression recognition method and system based on residual attention mechanism
CN107273845B (en) Facial expression recognition method based on confidence region and multi-feature weighted fusion
CN106096538B (en) Face identification method and device based on sequencing neural network model
CN111401257B (en) Face recognition method based on cosine loss under non-constraint condition
CN103605972B (en) Non-restricted environment face verification method based on block depth neural network
CN107967456A (en) A kind of multiple neural network cascade identification face method based on face key point
CN108509839A (en) One kind being based on the efficient gestures detection recognition methods of region convolutional neural networks
CN106096535A (en) A kind of face verification method based on bilinearity associating CNN
CN112766158A (en) Multi-task cascading type face shielding expression recognition method
CN107871101A (en) A kind of method for detecting human face and device
CN106570521B (en) Multilingual scene character recognition method and recognition system
Ahranjany et al. A very high accuracy handwritten character recognition system for Farsi/Arabic digits using convolutional neural networks
CN106651915B (en) The method for tracking target of multi-scale expression based on convolutional neural networks
CN110378208B (en) Behavior identification method based on deep residual error network
CN108520213B (en) Face beauty prediction method based on multi-scale depth
CN113239784A (en) Pedestrian re-identification system and method based on space sequence feature learning
CN111652273B (en) Deep learning-based RGB-D image classification method
CN107704859A (en) A kind of character recognition method based on deep learning training framework
CN116386102A (en) Face emotion recognition method based on improved residual convolution network acceptance block structure
CN103942572A (en) Method and device for extracting facial expression features based on bidirectional compressed data space dimension reduction
CN111160327B (en) Expression recognition method based on lightweight convolutional neural network
CN108520539A (en) A kind of image object detection method based on sparse study variable model
CN113887509B (en) Rapid multi-modal video face recognition method based on image set
Chun-man et al. Face expression recognition based on improved MobileNeXt

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20190319