CN107423707A - A kind of face Emotion identification method based under complex environment - Google Patents

A kind of face Emotion identification method based under complex environment Download PDF

Info

Publication number
CN107423707A
CN107423707A CN201710612421.3A CN201710612421A CN107423707A CN 107423707 A CN107423707 A CN 107423707A CN 201710612421 A CN201710612421 A CN 201710612421A CN 107423707 A CN107423707 A CN 107423707A
Authority
CN
China
Prior art keywords
face
illumination
under
emotion identification
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710612421.3A
Other languages
Chinese (zh)
Inventor
陈援东
陈忠
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Paluo Artificial Intelligence Technology Co Ltd
Original Assignee
Shenzhen Paluo Artificial Intelligence Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Paluo Artificial Intelligence Technology Co Ltd filed Critical Shenzhen Paluo Artificial Intelligence Technology Co Ltd
Priority to CN201710612421.3A priority Critical patent/CN107423707A/en
Publication of CN107423707A publication Critical patent/CN107423707A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • G06V40/171Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/40Scaling the whole image or part thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/60Rotation of a whole image or part thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/60Analysis of geometric attributes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/174Facial expression recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning

Abstract

The invention discloses a kind of face Emotion identification method based under mobile embedded complex environment, face is divided into forehead, looks, cheek, nose, face, chin main region by this method, and it is further divided into 68 characteristic points, for features described above point, in order to realize the discrimination of face Emotion identification, the degree of accuracy and reliability under circumstances, the method of user's face portion and expressive features classification in the case of normality, uses methods of the Faster R CNN based on facial zone convolutional neural networks under illumination, reflection, shadowed condition;The method that Bayesian network, Markov Chain, variation reasoning in the case where moving, shake, rocking and moving complex situations are combined;The method being combined in the case where face shows incomplete, plurality of human faces environment, noisy background condition using depth convolutional neural networks and super resolution degree antagonism neutral net SRGANs, enhancing study, back-propagation algorithm, discard algorithm, it is effectively facilitated effect, the degree of accuracy and the reliability of expression recognition.

Description

A kind of face Emotion identification method based under complex environment
Technical field
The present invention relates to the technical field that the moving video image of artificial intelligence identifies, more particularly under a kind of complex environment Face identification method, this is an emerging field in current global range.
Background technology
Recognition of face based on video image is an emerging important research field in the world, is reached the U.S. is tall and handsome (NVIDIA) since releasing the deep learning graphics processor video card (GPU) based on server and host computer within 2006, the world Leading high-tech company such as Google, Apple Inc., Amazon Company, well-known scientific research institutions such as Massachusetts science and engineering, Berkeley And Stamford, one after another using tall and handsome application of the GPU video cards research and development based on image recognition reached.By contrast, the mobile insertion in the whole world The video image identification market of formula lag far behind the similar application based on server, until in mid-May, 2017 just have including It is tall and handsome reach including manufacturer start to provide the GPU modules of high-performance machine vision and deep learning, mobile embedded video figure When being in and save up strength to start out as identification application market.
Recognition of face mainly includes three facial zone division, feature point extraction and classification parts.Facial zone divides Face is divided into several interest regions (Region ofInterest);Feature point extraction is carried from each interest region The characteristic point of key is taken, then merges characteristic point, it is linear that feature extracting method generally comprises PCA principal component analysis, LDA The methods of analysis;Classification is the suitable discrimination standard of selection and categorised decision, designs the grader such as RVM for being adapted to face characteristic Method Using Relevance Vector Machine, Kernel nuclear technology, statistical nature or neutral net carry out Classification and Identification.In facial image identification, feature Point extraction and classification need design parameter, and determine the weights of parameter, are weighted processing, obtain image transform coefficients matrix, So as to realize the recognition of face with certain accuracy rate.In mobile embedded video human face identification, recognition of face needs gram Take the various information such as light and shade, operating state, face position, more people and the background noisy degree of light.Currently used for recognition of face Prior art has CNN convolutional neural networks, Bayesian network (Bayes Network), depth convolutional neural networks (Deep Convolutional Neural Network), antagonism neutral net GANs (Super-resolution Generative Adversarial Network), enhancing study (Reinforcement Learning), sparse features, random forest, action The methods of tracking, statistical model or SVMs.In field of face identification, current most companies are all to use above-mentioned What single method, single method can reach certain effect, but also have in the presence of very big black box, such as in camera The various complex environments such as illumination captured, fuzzy, shade, rock, shake, move, face show not entirely, plurality of human faces environment, noise Discrimination under miscellaneous background is not high, accuracy rate is not high, realizes cost prohibitive.In ARM central processing units and (SuSE) Linux OS structure Into mobile embedded calculating platform on, the method for in the market Emotion identification lacks the support of bottom core face recognition technology, Apply not or rarely the efficient quickly face Emotion identification case in mobile embedded type equipment video.Appointing more than therefore The face Emotion identification method of what monotechnics exploitation, all it is difficult to develop into skill on the basis of mobile embedded calculating platform Art scheme, therefore be difficult industrialization.
The content of the invention
For the weak point of generally existing in above-mentioned technology, primary and foremost purpose of the present invention is to provide one kind and is based on complex environment Under face Emotion identification method, this method solved in mobile embedded various complex environments such as illumination, fuzzy, shade, shifting Move, rock, shaking, face position is not complete, the face Emotion identification rate that is obtained in real time under plurality of human faces environment, noisy background, accurate Degree and reliability.
Another object of the present invention is to provide a kind of face Emotion identification method based under complex environment, and this method uses The image steganalysis (Image and Pattern Recognition) and deep learning (Deep of traditional artificial intelligence Learning the method) being combined, bases such as " happy, sad, angry, frightened, detest, surprised, contempt, neutrality " is effectively differentiated This mood and deeper into combination mood, promote it is man-machine between affective interaction, promote human-computer interaction amusement as in various types of games The four-dimensional dummy environment such as emotion stage property, augmented reality or virtual reality in virtual and simulation games, interpersonal long-range Emotive advisory, mood regulation and emotion communication in video exchange.
To achieve these goals, implementation method of the invention is described as follows.
A kind of face Emotion identification method based under complex environment, it is characterised in that face is divided into volume by this method Head, looks, cheek, nose, face, chin main region, and it is further divided into 68 characteristic points.For above-mentioned 68 features Point, in order to realize discrimination, the degree of accuracy and reliability in various complex environment human face Emotion identifications, in illumination, reflection, the moon The method being combined under the conditions of shadow using the quick regional area convolutional neural networks of Faster R-CNN and RVM Method Using Relevance Vector Machines; Bayesian network (Bayes Network) in the case where moving, shake, rocking and moving complex situations, Markov Chain, variation push away Manage the method being combined;Depth convolutional neural networks are used in the case where face shows incomplete, plurality of human faces environment, noisy background condition (Deep Convolutional Neural Network) and super resolution degree antagonism neutral net SRGANs (Super- Resolution Generative Adversarial Network), enhancing study (Reinforcement Learning), The method that back-propagation algorithm (Backpropagation), discard algorithm (Dropout) are combined, is effectively facilitated face table Effect, the degree of accuracy and the reliability of feelings identification.
Described characteristic point includes 68, forms the uniform characteristics distribution of face, and provides a set of testing index, the finger Mark includes traceable (Traceable), slight (Slight), significantly medium (Medium), (Marked), very big (Great), most (Maximum) totally 6 states by force.
Further, some combinational algorithms, including face characteristic and texture processing algorithm, light are developed in terms of application technology Processing Algorithm and plurality of human faces track algorithm etc. are imitated, and mobile embedded software is built based on C language and builds internet using soft Part.
Specifically, expression recognition needs to overcome the light and shade of light, operating state, face position, more people and background The various information such as noisy degree.For any of the above complexity face recognition process, feature and grader are subjected to combined optimization, hair Wave the function of feature representation and Classifier combination cooperation, and by (input layer, hidden layer and the output layer totally 3 of traditional artificial intelligence It is more than layer) non-linear the reflecting of Linear Mapping and deep-neural-network (including at least more than 20 layers of input layer, hidden layer, output layer) Penetrate and be combined, such as under conditions of illumination, fuzzy, shade, using traditional RWM Method Using Relevance Vector Machines feature point extraction and The tagsort of Faster R-CNN neutral nets is combined, and promotes the face Emotion identification degree of accuracy under these conditions, so It is afterwards to use the number with powerful learning ability and efficient feature representation ability (Representation) in deep learning Model is learned, changes picture to face from the initial data of Pixel-level, edge, texture, part, whole face picture, multiple faces Expression successively extracts information.
Further, such as continuous action, display do not use depth entirely, under the conditions of noisy environment under other complex situations Learning method, including Deep CNN depth convolutional neural networks, Bayes CNN Bayes convolutional neural networks, SRGANs supersolution Analysis degree confrontation generation network, RL enhancing learning method in 2 or 2 above in association with method, realize local feature region obtain, Global sign ability, the interference in recognition of face assorting process and decision method, so as to improve internet and environment of internet of things Lower movement regards the face Emotion identification degree of accuracy of embedded video captured in real-time.
Further, for face in identification process, interference can also be carried out and handled, i.e., the deep layer volume more than 20 layers In product neutral net, using superiority of the multilayer neural network in abstract, cluster, sign, mobile embedded type equipment is being ensured Video capture during face efficient process while, peeled off for various complicated non-face intrinsic features;I Combine super resolution degree confrontation network, it by the positive counter-example method of antagonistic training calculate positive counter-example facial image to damage-retardation Become estranged reconstruct high-definition picture and original high-resolution image loss weighted sum, obtain above complex situations under face table Feelings identify;Back-propagation algorithm is used in deep neural network learning process, checked the error of recognition of face, and is used Stochastic gradient descent (Stochastic Gradient Descent) method is to hidden layer-hidden layer, hidden layer-output layer Weights are updated, and are promoted the global optimization during neural network learning, are formed the video capture continuous picture based on prediction Expression recognition result method.
Specifically, fast area convolutional neural networks (Faster R-CNN) are used under illumination, reflection or shadow condition The method for being combined sum with traditional RVM graders.Pass through the Face datection fast positioning human face region from video frequency pick-up head rapidly And face features point information is obtained, specifically, by intensity of illumination and direction, reflection, the quantization of shade illumination, adopting With illumination mode is determined, illumination system is divided into light source, face and photoenvironment, determines the global context light in illumination model Intensity, point of observation position, the front and back of face and background calculates illumination, specular color calculates and its separation, to illumination, anti- Penetrate and carry out illumination elimination, reflection-absorption and illumination compensation with shade, eliminate bearing for the recognition of face under these illumination complex situations Face rings;Using the diffusing reflection model (Diffuse Reflectivity Model) of illumination model including ambient light, skin Specular reflectance model (Mirror Reflectivity Model) and lambert's illumination model (Lambert based on material and texture Illumination Model) in a kind of illumination to face handle, generate various training samples and the survey of light irradiation Sample sheet;The mode for taking size normalization, unitary of illumination to combine carries out abstract pretreatment to facial image.Size normalizes Algorithm is to be changed in given change to each position equiblibrium mass distribution of the feature with invariance such as face area, girth, face, is carried out Image rotation, zoom in or out, be met the face standard picture of needs;Illumination normalizing algorithm is to be imaged to manage in Retinex On the basis of, a kind of multiresolution fast algorithm based on Total Variation is established, for calculating the illumination component in image, And then calculate the reflectance factor image of reflection image texture, finally the reflectance factor of the illumination of low frequency and high frequency is separated, Obtain accurate face picture.Non-face intrinsic attribute is stripped out from face by more than;By the human face region after segmentation Image is sent in the fast area convolutional neural networks trained, the feature of each region portion of quick obtaining face;Series connection is all The feature of human face region, and using the grader of Method Using Relevance Vector Machine RVM (Relevance Vector Machine) supervised learning The classification of expression is carried out, forms face Emotion identification (the Facial Emotion in illumination, reflection or shade Recognition)。
Under movement, the complex situations rocked and obscured, according to the characteristic of division under static state, Bayes's convolution is used Neutral net (Bayes CNN) carries out unsupervised deep learning, the face picture continuously captured according to video, extraction face spy Sign, is classified, and solve data fitting problems automatically using Bayes classifier.Under general distribution and common dimension, adopt (Expectation Maximization) algorithm is maximized with EM desired values, sufficient desired value is calculated in statistics, Using the existing estimate to face parameter hidden variable, maximum likelihood estimator (Maximum Likelihood are calculated Estimation);People is calculated by MAP estimation (MaximumAPosteriori) on the basis of desired value is maximized The value of face parameter.By the method for data mining, to data being described property cluster analysis, according to similarity, space length or If density by data object automatically be respectively Ganlei, such as space length method in the space at each position of face have it is constant Shape.Descriptive cluster is mainly that the general property of face characteristic data is classified, it is intended to finds that facial recognition data collection owns Similitude, regularity and pattern in class between variable, for example judge the combination of the moving cell of face region of interest, form new spy Seek peace classification, and form trickleer mood.Using Naive Bayes Classifier (Bayes) method carry out above class and The reasoning of numeric data, eight kinds of moods such as pleasure, anger, sorrow, happiness of face are classified and segmented;In movement, rock and obscure Under conditions of, it is necessary to carry out video frequency object tracking, positioning and identification for dynamic human face object, this relates to high-dimensional condition The solution of distribution.Under these conditions, the picture convolution integral computing of posterior probability can not be directly calculated, therefore can using Ma Er Husband's chain Meng Teka multi-methods (MCMC).MCMC is one group and having for face and Emotional Picture sample is progressively produced from Posterior distrbutionp Efficacious prescriptions method, it obtains the inference on model parameter by the average value of the face sample under continuous acquisition above dynamic condition, Approximate Likelihood Function value is solved, the optimization problem for solving the face Emotion identification under conditions above.It is different in the parameter calculated in real time In the case of often huge, such as under the conditions of the dynamic human face crawl of 60 frame per second, every equipment picture meter of daily 24 hours Calculation amount is at 60*60*60*24=5184000 pictures (relative to iPhone average 3000 photos), and we are using becoming decibel Leaf this method processing video pictures data.Variational Bayesian method is a kind of method approximate solution Bayes by function approximation A kind of technology of the integration of difficult product in reasoning and machine learning, it is generally used for the complexity under the conditions of real-time video picture big data Statistical model, it can be considered as the extension of EM algorithms, and it introduces latent variable on the basis of observable variable and parameter value, Approached by optimal method, so as to solve the imponderable problem of more probability distribution under video dynamic frame environment, And solves the imponderable problem of big matrix inversion of picture convolution;With Backpropagation algorithms recall training with When upper video pictures data, in order to prevent the over-fitting problem of data, we are removed implicit at random using Dropout algorithms The neuron of layer and visible layer, Bayes's CNN networks are trained in multiple batch processing, and draw consensus forecast rate, so as to return Receive (or extensive) go out a reliable and stable Bayesian model;In this process, we are also using enhancing learning method (Reinforcement Learning), the intelligently mapping from environment to behavior under conditions of unsupervised so that reward letter Number functional value maximum, heighten the reliability of Bayesian network.Intensified learning under embedded environment mainly includes analysis policy The rewards and punishments assessment of the strategy process, value network of network, the algorithm of prototype network are recommended, and search for optimal decision behavior respectively, Analysis is moved, rocked, the video human face picture classification under fuzzy situation, for example judges to shake by the frame difference between analyzing per frame Dynamic program or the intelligent connecting for judging fuzzy state, carrying out Intelligent human-face picture by continuous time series, shadow The interference feature for ringing face and Expression Recognition is peeled off, and is realized movement, is rocked and the expression recognition under hazy condition Accuracy and reliability, environment is characterized using model, carried out in the movement of AMR mobile embedded type equipments crawl or continuous face Expression Recognition.
In the case where face shows incomplete, plurality of human faces environment, noisy background, in order to ensure the reliability of facial expression classification, Processing face show it is not full-time, employ Deep CNN depth convolutional neural networks, SRGANs super resolutions degree confrontation generate network, The method that RL enhancing study is combined, using the sparse feature of the moderate of the upper strata neuron response of deep neural network to face Attribute analyzed, in the case where detection recognizes face, the feature such as gray value or marginal value that obtain face are being protected Hold in the class for keeping minimum while change between class and change, there is very strong selectivity to face and the attribute of expression, to face In the case of partial occlusion, this method accurately can identify face by behavioral characteristics, have very strong robustness, and lead to Cross the face that the global information of face and the contextual information of successive video frames identify needs to judge, and call face system Shape of face in system is spliced to showing that infull face carries out completion, feature is described and expression is global, obtains the face for needing to identify Partial data;In the case of plurality of human faces, face snap lighting system is regarded simultaneously from the multiple faces of motion tracking by continuously capturing The residence time of frequency frame, judge to need the importance and precedence of multiple faces, and each face machine expression is carried out Classification is handled;Using scale invariant feature conversion SIFT (Scale Invariant Feature under noisy background Transformation) algorithm, detect and position the background object unrelated with face such as window, chair, landscape painting etc. simultaneously handle They are peeled off away from the face picture for needing to identify, only display needs the face and expression of Tracking Recognition.During this Using Backpropagation algorithms (algorithm introduction see before), in the case where output layer cannot get data, along boarding steps The weights of each neuron are changed in backtracking on the basis of degree declines, and obtain mean test value, optimize Deep CNN convolution models.We Network is resisted using SRGANs super resolutions degree, it calculates pair of positive counter-example facial image by the positive counter-example method of antagonistic training Damage-retardation become estranged reconstruct high-definition picture and original high-resolution image loss weighted sum, obtain above complex situations under people Face Expression Recognition.Antagonistic training mainly includes maker, arbiter and ultramagnifier, and wherein maker is responsible for exporting interference people The true and false that face picture, arbiter carry out face picture judges;It is reconstitution training be to original high definition and it is down-sampled after it is low The paired samples of resolution ratio is trained, by reconstructing Feature Mapping between high-definition picture and original high-resolution image Euclidean distance, it is fine definition picture low resolution picture reconstruction, and feeds back to Deep CNN and carry out complicated ring above Face and Emotion identification under border.
Present invention incorporates image and the methodology of pattern-recognition, neutral net and Bayes networking, has invented based on shifting Move Embedded complex environment human face Emotion identification method, with reference to the embedded software based on C, based on three-tier architecture because The various high-new skills such as working application software, figure and pattern-recognition, deep neural network, complex mathematical model, big data application Art, each fields such as the four-dimensional environment of robot, interaction entertainment, Internet of Things, internet, virtual reality and augmented reality have than Higher market value.
The present invention combines traditional images and simple direct and the efficient classification of neutral net and the automated reasoning of pattern-recognition Etc. advantage, and using super resolution degree antagonism neutral net, enhancing study and back-propagation algorithm, improve shallow-layer and depth Practise it is various under the conditions of mobile embedded environment in frame of video operation and the synchronism stability of expression recognition run, and can For moving, obscuring, rocking, shaking shade, illumination variation, face position show it is real under incomplete, plurality of human faces environment, noisy background Real-time human face expression processing, has reached high accuracy, high efficiency and reliability when real.
Brief description of the drawings
Fig. 1 is the expression recognition and classification process figure that the present invention is realized.
Fig. 2 is the flow chart that the present invention realizes quick local convolutional neural networks.
Fig. 3 is the flow chart that the present invention realizes Bayes's convolutional neural networks.
Fig. 4 is the flow chart that the present invention realizes Deep CNN depth convolutional neural networks.
Embodiment
In order to more clearly state the present invention, the present invention is further described below in conjunction with the accompanying drawings.
Shown in Fig. 1, for the overall flow figure based on the face Emotion identification method under complex environment realized of the present invention, The flow of expression recognition is:First video pictures are inputted, Face datection is then carried out, testing result is pre-processed, Interference (going interference to handle) is removed, then enters line edge and is mapped into pedestrian's face region recognition and human face region selection, is directed to simultaneously Data after mapping carry out carrying out 8 area-of-interests Gabor wavelet conversion, carry out uniform pattern to the image after conversion LBP feature extractions, dimension-reduction treatment is then carried out to the feature of extraction by PCA principal component analysis, carries out characteristic vector contrast, fortune Moving cell and neutral expression's contrast, finally draw 8 kinds of expressions.
Relative normality ARM screens shooting face under the conditions of, using image-recognizing method by Face datection rapidly from Fast positioning human face region and face features point information is obtained in video frequency pick-up head;According to face 68 characteristic points of face Position, 8 area-of-interests are carried out with Gabor wavelet conversion, the LBP features to the image progress uniform pattern after conversion carry Take, carry out dimension-reduction treatment to the feature of extraction by PCA principal component analysis, and by with the feature in training, test sample to Amount is contrasted, and forms face and the preliminary recognition result of expression;Basic exercise unit (the Basic showed according to human face region Action Unit) with the contrast of neutral expression (Neutral Facial Expression) moving cell, with reference to face and table The preliminary recognition result of feelings, human face expression is finely divided and finally draws 8 kinds of main facial expression classifications and segments classification, shape Into face Emotion identification.
Face is divided into forehead, looks, cheek, nose, face, chin main region, and is further divided into some Individual characteristic point, for features described above point, in order to realize the discrimination of face Emotion identification, the degree of accuracy and reliable under circumstances Property, the method for user's face portion and expressive features classification in the case of normality;Used under illumination, reflection, shadowed condition Methods of the Faster R-CNN based on facial zone convolutional neural networks;In the case where moving, shake, rocking and moving complex situations The method that Bayesian network (Bayes Network), Markov chain Monte-Carlo method, variation reasoning are combined;In face Display is not complete, uses depth convolutional neural networks (Deep Convolutional under plurality of human faces environment, noisy background condition Neural Network) and super resolution degree antagonism neutral net SRGANs (Super-resolution Generative Adversarial Network), enhancing study (Reinforcement Learning), back-propagation algorithm (Backpropagation), the method that discard algorithm (Dropout) is combined, be effectively facilitated expression recognition effect, The degree of accuracy and reliability.
Described characteristic point, it is 68 characteristic points, forms the characteristic point distribution of face, and a set of testing index is provided, institute Stating index includes traceable (Traceable), slight (Slight), significantly medium (Medium), (Marked), greatly (Great), most strong (Maximum) totally 6 states.
Expression recognition needs to overcome the light and shade of light, operating state, face position, more people and background noisy degree etc. each Kind information.For any of the above complexity face recognition process, feature and grader separate or combined optimization, performance are special Sign expression and Classifier combination cooperation function, and by traditional artificial intelligence (input layer, hidden layer and output layer totally 3 layers with On) Nonlinear Mapping of Linear Mapping and deep-neural-network (including at least more than 20 layers of input layer, hidden layer, output layer) enters Row combines
Under conditions of illumination, fuzzy, shade, using the feature point extraction and Faster R- of traditional RVM Method Using Relevance Vector Machines The tagsort of CNN neutral nets is combined, and promotes the face Emotion identification degree of accuracy under these conditions, then in depth Habit is to use the mathematical modeling with powerful learning ability and efficient feature representation ability (Representation), from picture The initial data of plain level, edge, texture, part, whole face successively extract information to human face expression.
Deep learning method, including Deep CNN depth are used under the conditions of continuous action, display not complete, noisy environment etc. Convolutional neural networks, Bayes CNN Bayes convolutional neural networks, SRGANs super resolutions degree confrontation generation network, RL enhancings are learned 2 in learning method or 2 above in association with method, realize local feature region acquisition, global sign ability, recognition of face classification During interference and decision method, regard embedded video captured in real-time so as to improve under internet and environment of internet of things movement The face Emotion identification degree of accuracy.
For face in identification process, interference can also be carried out and handled, i.e., the deep layer convolutional Neural more than 20 layers In network, using superiority of the multilayer neural network in abstract, cluster, sign, the video of mobile embedded type equipment is being ensured During candid photograph while face efficient process, peeled off for various complicated non-face intrinsic features;We combine Super resolution degree resists network, and it calculates the confrontation loss and again of positive counter-example facial image by the positive counter-example method of antagonistic training Structure high-definition picture and the weighted sum of original high-resolution image loss, the human face expression obtained under above complex situations are known Not;Back-propagation algorithm is used in deep neural network learning process, checked the error of recognition of face, and using random Gradient declines (Stochastic Gradient Descent) method to hidden layer-hidden layer, the weights of hidden layer-output layer It is updated, promotes the global optimization during neural network learning, forms the people of the video capture continuous picture based on prediction Face Expression Recognition result method.
As shown in Fig. 2 fast area convolutional neural networks (Faster R-CNN) are used under illumination, reflection or shadow condition The method for being combined sum with traditional RVM correlation classifiers.Pass through the Face datection fast positioning face from video frequency pick-up head rapidly Region simultaneously obtains face features point information, enters line light source and photoenvironment analysis, then carry out illumination elimination, absorption and benefit Repay, then establish photo-irradiation treatment model (including diffusing reflection, mirror-reflection and lambert's illumination model), according to model generation test Sample, then pre-processed, carry out size normalization and Retinex normalizeds, then peel off non-face feature, carried out Faster R-CNN facial images are trained, and the acquisition of face local feature and series connection are then carried out, finally by RVM correlation classifiers Carry out Expression Recognition.
Specifically, by intensity of illumination and direction, reflection, the quantization of shade illumination, using determining illumination mode, Illumination system is divided into light source, face and photoenvironment, determine global context light intensity in illumination model, point of observation position, Face and the front and back of background calculate illumination, specular color calculates and its separation, and illumination is carried out to illumination, reflection and shade Elimination, reflection-absorption and illumination compensation, eliminate the negative effect of the recognition of face under these illumination complex situations;Using illumination Model includes the diffusing reflection model (Diffuse Reflectivity Model) of ambient light, the specular reflectance model of skin (Mirror Reflectivity Model) and lambert's illumination model (Lambert based on material and texture Illumination Model) in a kind of illumination to face handle, generate various training samples and the survey of light irradiation Sample sheet;The mode for taking size normalization, unitary of illumination to combine carries out abstract pretreatment to facial image.Size normalizes Algorithm is to be changed in given change to each position equiblibrium mass distribution of the feature with invariance such as face area, girth, face, is carried out Image rotation, zoom in or out, be met the face standard picture of needs;Illumination normalizing algorithm is to be imaged to manage in Retinex On the basis of, a kind of multiresolution fast algorithm based on Total Variation is established, for calculating the illumination component in image, And then calculate the reflectance factor image of reflection image texture, finally the reflectance factor of the illumination of low frequency and high frequency is separated, Obtain accurate face picture.Non-face intrinsic attribute is stripped out from face by more than;By the human face region after segmentation Image is sent in the fast area convolutional neural networks trained, the feature of each region portion of quick obtaining face;Series connection is all The feature of human face region, and using the grader of Method Using Relevance Vector Machine RVM (Relevance Vector Machine) supervised learning The classification of expression is carried out, forms face Emotion identification (the Facial Emotion in illumination, reflection or shade Recognition)。
In the training of positive counter-example, the interference picture generation of antagonism neutral net and the advantage differentiated, pass through maker (Generator) face picture of input tape jamming pattern, maker face figure is verified by arbiter (Discriminator) The true and false of piece, genuine face picture will be determined as by ultramagnifier (Feedback Processor) and be transmitted to depth convolutional Neural net Network carries out the processing of human face expression;In reconstruct high-definition picture and original high-resolution image loss, mainly analysis is former The characteristic value of beginning resolution ratio and classification, the reconstruct of image is carried out in global and contextual information.Back-propagation algorithm is used In deep neural network learning process, the error of recognition of face is checked, and to hidden layer-hidden layer, hidden layer-output layer Weights be updated, promote neural network learning during global optimization, form the video capture sequential chart based on prediction The expression recognition result method of piece.
Under movement, the complex situations rocked and obscured, according to the characteristic of division under static state, as shown in figure 3, using The unsupervised deep learning of Bayes's convolutional neural networks (Bayes CNN) progress, the face picture continuously captured according to video, Face characteristic is extracted, is classified automatically using Bayes classifier, and solve data fitting problems.In general distribution and commonly Under dimension, (Expectation Maximization) algorithm is maximized using EM desired values, fully calculated in statistics Desired value, using the existing estimate to face parameter hidden variable, calculate maximum likelihood estimator (Maximum Likelihood Estimation);Pass through MAP estimation (Maximum A on the basis of desired value is maximized Posteriori the value of face parameter) is calculated.By the method for data mining, to data being described property cluster analysis, according to If data object is respectively in each portion of face in Ganlei, such as space length method automatically by similarity, space length or density Have in the space of position indeformable.Descriptive cluster is mainly that the general property of face characteristic data is classified, it is intended to is found Similitude, regularity and pattern in class between all variables of facial recognition data collection, for example judge the motion list of face region of interest The combination of member, new feature and classification are formed, and form trickleer mood.Using Naive Bayes Classifier ( Bayes) method carries out the reasoning of above class and numeric data, eight kinds of moods such as pleasure, anger, sorrow, happiness of face are carried out classification and Subdivision;In movement, rock and obscure under conditions of, it is necessary to carry out video frequency object tracking, positioning and knowledge for dynamic human face object Not, this solution for relating to high-dimensional condition distribution.Under these conditions, the picture convolution of posterior probability can not directly be calculated Integral operation, therefore using Markov chain Meng Teka multi-methods (MCMC).MCMC is one group of progressive real estate from Posterior distrbutionp The effective ways of stranger's face and Emotional Picture sample, it is averaged by the face sample under continuous acquisition above dynamic condition Value, obtains the inference on model parameter, solves Approximate Likelihood Function value, solves the excellent of face Emotion identification under conditions above Change problem.In the case where the abnormal parameters calculated in real time are huge, such as under the conditions of the dynamic human face crawl of 60 frame per second, often The platform equipment picture amount of calculation of daily 24 hours is in 60*60*60*24=5184000 pictures (relative to the flat of iPhone storages Equal 3000 photos), we are using variational Bayesian method processing video pictures data.Variational Bayesian method is that one kind passes through A kind of technology of difficult product integration in the method approximate solution Bayesian inference of function approximation and machine learning, it is generally used in real time Complex statistics model under the conditions of video pictures big data, it can be considered as the extension of EM algorithms, and it is in observable variable and ginseng Latent variable is introduced on the basis of numerical value, is approached by optimal method, so as to solve mobile video dynamic frame ring The imponderable problem of more probability distribution under border, and solve the imponderable problem of the big matrix inversion of picture convolution;With When Backpropagation algorithms backtracking training above video pictures data, in order to prevent the over-fitting problem of data, I Hidden layer and the neuron of visible layer are removed using Dropout algorithms at random, Bayes CNN is trained in multiple batch processing Network, and consensus forecast rate is drawn, go out a reliable and stable Bayesian model so as to conclude (or extensive);In this process In, we are also using enhancing learning method (Reinforcement Learning), intelligently from ring under conditions of unsupervised Mapping of the border to behavior so that prize signal functional value is maximum, heightens the reliability of Bayesian network.Under embedded environment Intensified learning mainly includes the rewards and punishments assessment of the strategy process, value network of analysis Policy Network, the algorithm of prototype network is recommended, Optimal decision behavior is searched for respectively, analysis is moved, rocked, the video human face picture classification under fuzzy situation, such as by dividing Frame difference between the every frame of analysis judges the program rocked or the state obscured by the judgement of continuous time series, carries out face The intelligent connecting of picture, the interference feature of influence face and Expression Recognition is peeled off, realize it is mobile, rock and hazy condition Under expression recognition accuracy and reliability, using model characterize environment, AMR mobile embedded type equipments crawl shifting Dynamic or continuous face carries out Expression Recognition.
In the case where face shows incomplete, plurality of human faces environment, noisy background, in order to ensure the reliability of facial expression classification, Processing face show it is not full-time, as shown in figure 4, employing Deep CNN depth convolutional neural networks, SRGANs super resolution degree pair The method that antibiosis is combined into network, RL enhancing study, the moderate using the upper strata neuron response of deep neural network are sparse Feature is analyzed the attribute of face, in the case where detection recognizes face, obtain face feature such as gray value or Marginal value keeps changing in minimum class while between keeping class changing, and has very strong selection to face and the attribute of expression Property, to face by the case of partial occlusion, this method accurately can identify face by behavioral characteristics, have very strong Shandong Rod, and the face identified by the global information of face and the contextual information of successive video frames to needs judges, and The shape of face in face system is called to splice to showing that infull face carries out completion, feature is described and expression is global, obtaining needs The partial data of the face of identification;In the case of plurality of human faces, face snap lighting system passes through simultaneously from the multiple faces of motion tracking The residence time of continuous crawl frame of video, judge to need the importance and precedence of multiple faces, and to each face machine Device expression carries out classification processing;Using scale invariant feature conversion SIFT (Scale Invariant under noisy background Feature Transformation) algorithm, detects and positions the background object unrelated with face such as window, chair, landscape Picture etc. is simultaneously peeled off away them from the face picture for needing to identify, only display needs the face and expression of Tracking Recognition. Using Backpropagation algorithms (algorithm introduction see before) during this, in the case where output layer cannot get data, The weights of each neuron are changed in backtracking on the basis of along stochastic gradient descent, obtain mean test value, optimize Deep CNN convolution Model., using SRGANs super resolutions degree confrontation network, it calculates positive counter-example people by the positive counter-example method of antagonistic training for we The confrontation of face image is lost and reconstruct high-definition picture and the weighted sum of original high-resolution image loss, more than acquisition complexity In the case of expression recognition.Antagonistic training mainly includes maker, arbiter and ultramagnifier, and wherein maker is responsible for defeated Go out interference face picture, the true and false of arbiter progress face picture judges;Reconstitution training is to original high definition and drop The paired samples of low resolution after sampling is trained, by reconstructing between high-definition picture and original high-resolution image Feature Mapping Euclidean distance, be fine definition picture low resolution picture reconstruction, and feed back to Deep CNN and enter Face and Emotion identification more than row under complicated ring.
Present invention incorporates image and pattern-recognition, the methodology of deep neural network and Bayesian network, base has been invented In the method for mobile embedded complex environment human face Emotion identification, with reference to the embedded software based on C, based on three-tier architecture Because working application software, figure and pattern-recognition, deep neural network, complex mathematical model, big data application etc. various height New technology, have in each fields such as the four-dimensional environment of robot, Internet of Things, internet, virtual reality and augmented reality higher Market value.
It is excellent that the present invention combines simple direct and neutral net the efficient classification and reasoning of traditional images and pattern-recognition etc. Gesture, and using super resolution degree antagonism neutral net, enhancing study and back-propagation algorithm, improve shallow-layer and deep learning is each The synchronism stability of frame of video operation and expression recognition in mobile embedded environment under the conditions of kind is run, and can be directed to Move, obscure, rocking, shaking shade, illumination variation, face position show realized under incomplete, plurality of human faces environment, noisy background it is real When the processing of real-time human face expression, reached high accuracy, high efficiency and reliability.
Disclosed above is only several specific embodiments of the present invention, but the present invention is not limited to this, any ability What the technical staff in domain can think change should all fall into protection scope of the present invention.

Claims (10)

  1. A kind of 1. face Emotion identification method based under complex environment, it is characterised in that this method by face be divided into forehead, Looks, cheek, nose, face, chin main region, and 68 characteristic points are further divided into, for features described above point, normal The method of user's face portion and expressive features classification in the case of state, uses Faster R- under illumination, reflection, shadowed condition Methods of the CNN based on facial zone convolutional neural networks;Bayesian network in the case where moving, shake, rocking and moving complex situations The method that network, Markov Chain, variation reasoning are combined;Make in the case where face shows incomplete, plurality of human faces environment, noisy background condition Calculated with depth convolutional neural networks and super resolution degree antagonism neutral net SRGANs, enhancing study, back-propagation algorithm, discarding The method that method is combined, it is effectively facilitated effect, the degree of accuracy and the reliability of expression recognition.
  2. 2. the face Emotion identification method according to claim 1 based under complex environment, it is characterised in that described spy Point is levied, is 68 characteristic points, the characteristic point for forming face is uniformly distributed, and provides a set of testing index, and the index includes Traceable (Traceable), slight (Slight), significantly medium (Medium), (Marked), greatly (Great), most strong (Maximum) totally 6 states.
  3. 3. the face Emotion identification method according to claim 2 based under complex environment, it is characterised in that more than being directed to Various complexity face recognition process, feature and grader are subjected to combined optimization, play feature representation and Classifier combination association Work energy, and the Nonlinear Mapping of the Linear Mapping of input layer, hidden layer and output layer and deep-neural-network is tied Close, under conditions of illumination, fuzzy, shade, using feature point extraction and Faster the R-CNN god of traditional RWM Method Using Relevance Vector Machines Tagsort through network is combined, and promotes the face Emotion identification degree of accuracy under these conditions.
  4. 4. the face Emotion identification method according to claim 3 based under complex environment, it is characterised in that for face In identification process, interference can also be carried out and handled, i.e., in the deep layer convolutional neural networks more than 20 layers, utilize multilayer god Superiority through network in abstract, cluster, sign, face is efficient during the video capture of mobile embedded type equipment is ensured While processing, peeled off for various complicated non-face intrinsic features;Network is resisted with reference to super resolution degree, by right The positive counter-example method of resistance training calculates confrontation loss and reconstruct high-definition picture and the original high score of positive counter-example facial image The weighted sum of resolution image impairment, obtain the expression recognition under above complex situations;Back-propagation algorithm is used in depth Spend during neural network learning, checked the error of recognition of face, and using stochastic gradient descent method to hidden layer-hide Layer, the weights of hidden layer-output layer are updated, and promote the global optimization during neural network learning.
  5. 5. the face Emotion identification method according to claim 4 based under complex environment, it is characterised in that illumination, reflection Or the method for using fast area convolutional neural networks and traditional RVM graders to be combined sum under shadow condition, examined by face Survey and fast positioning human face region and obtain face features point information from video frequency pick-up head rapidly, specifically, by right Intensity of illumination and direction, reflection, the quantization of shade illumination, using determine illumination mode, illumination system be divided into light source, face and Photoenvironment, determine the front and back meter of global context light intensity in illumination model, point of observation position, face and background Illumination, specular color calculating and its separation are calculated, illumination elimination, reflection-absorption and illumination compensation are carried out to illumination, reflection and shade, Eliminate the negative effect of the recognition of face under these illumination complex situations;Include the diffusing reflection mould of ambient light using illumination model At a kind of illumination to face in type, the specular reflectance model of skin and lambert's illumination model based on material and texture Reason, generate the various training samples and test sample of light irradiation;The mode that size normalization, unitary of illumination combine is taken to people Face image carries out abstract pretreatment;Non-face intrinsic attribute is stripped out from face by more than;By the face area after segmentation Area image is sent in the fast area convolutional neural networks trained, the feature of each region portion of quick obtaining face;Series connection institute There are the feature of human face region, and the classification of the grader progress expression using SVMs RVM supervised learnings, formed in light According to, reflection or shade face Emotion identification.
  6. 6. the face Emotion identification method according to claim 4 based under complex environment, it is characterised in that in movement, shake Under dynamic and fuzzy complex situations, according to the characteristic of division under static state, carried out using Bayes's convolutional neural networks without prison The deep learning superintended and directed, the face picture continuously captured according to video, face characteristic is extracted, carried out using Bayes classifier automatic Classification, and solve data fitting problems.
  7. 7. the face Emotion identification method according to claim 6 based under complex environment, it is characterised in that typically dividing Under cloth and common dimension, algorithm is maximized using EM desired values, desired value is calculated in statistics, using hidden to face parameter The existing estimate of variable is hidden, calculates maximum likelihood estimator;Pass through MAP estimation on the basis of desired value is maximized Calculate the value of face parameter;By the method for data mining, to data being described property cluster analysis, according to similarity, space If data object is respectively Ganlei automatically by distance or density, above class and numerical value are carried out using Naive Bayes Classifier method The reasoning of data, eight kinds of moods such as pleasure, anger, sorrow, happiness of face are classified and segmented.
  8. 8. the face Emotion identification method according to claim 7 based under complex environment, it is characterised in that in movement, shake , it is necessary to video frequency object tracking, positioning and identification be carried out for dynamic human face object, using markov under conditions of dynamic and fuzzy Chain Meng Teka multi-methods, by the average value of the face sample under continuous acquisition above dynamic condition, obtain on model parameter Inference, solve Approximate Likelihood Function value, solve conditions above under face Emotion identification optimization problem.
  9. 9. the face Emotion identification method according to claim 4 based under complex environment, it is characterised in that show in face Show under incomplete, plurality of human faces environment, noisy background, in order to ensure the reliability of facial expression classification, shown in processing face incomplete When, employing Deep CNN depth convolutional neural networks, SRGANs super resolutions degree confrontation generation network, RL strengthens study and is combined Method, the attribute of face is analyzed using the moderate of the upper strata neuron response of deep neural network sparse feature, In the case where detection recognizes face, the feature such as gray value or marginal value of face are obtained while changing between keeping class The minimum interior change of class is kept, and by the global information of face and the contextual information of successive video frames to needing to know others Face is judged, and calls the shape of face in face system to showing that infull face carries out completion, feature is described and the expression overall situation Splicing, obtain the partial data for the face for needing to identify.
  10. 10. the face Emotion identification method according to claim 9 based under complex environment, it is characterised in that in plurality of human faces In the case of, face snap lighting system judges simultaneously from the multiple faces of motion tracking, and by continuously capturing the residence time of frame of video The importance and precedence of multiple faces are needed, and classification processing is carried out to each face machine expression;In noisy background It is lower using scale invariant feature conversion SIFT algorithms, detect and position the background object unrelated with face and them from needing to know Peeled off away in other face picture, using Backpropagation algorithms, in the case where output layer cannot get data, The weights of each neuron are changed in backtracking on the basis of along stochastic gradient descent, obtain mean test value, optimize DeepCNN convolution moulds Type;Network is resisted using SRGANs super resolutions degree, positive counter-example facial image is calculated by the positive counter-example method of antagonistic training Confrontation loss and reconstruct high-definition picture and the weighted sum of original high-resolution image loss, are obtained under above complex situations Expression recognition;Antagonistic training mainly includes maker, arbiter and ultramagnifier, and wherein maker is responsible for exporting interference The true and false that face picture, arbiter carry out face picture judges;It is reconstitution training be to original high definition and it is down-sampled after The paired samples of low resolution is trained, and is reflected by the feature reconstructed between high-definition picture and original high-resolution image The Euclidean distance penetrated, it is fine definition picture low resolution picture reconstruction, and feeds back to Deep CNN and carry out above again Face and Emotion identification under heterocycle.
CN201710612421.3A 2017-07-25 2017-07-25 A kind of face Emotion identification method based under complex environment Pending CN107423707A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710612421.3A CN107423707A (en) 2017-07-25 2017-07-25 A kind of face Emotion identification method based under complex environment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710612421.3A CN107423707A (en) 2017-07-25 2017-07-25 A kind of face Emotion identification method based under complex environment

Publications (1)

Publication Number Publication Date
CN107423707A true CN107423707A (en) 2017-12-01

Family

ID=60430495

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710612421.3A Pending CN107423707A (en) 2017-07-25 2017-07-25 A kind of face Emotion identification method based under complex environment

Country Status (1)

Country Link
CN (1) CN107423707A (en)

Cited By (60)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107909061A (en) * 2017-12-07 2018-04-13 电子科技大学 A kind of head pose tracks of device and method based on incomplete feature
CN108177146A (en) * 2017-12-28 2018-06-19 北京奇虎科技有限公司 Control method, device and the computing device of robot head
CN108230239A (en) * 2017-12-25 2018-06-29 中国科学院自动化研究所 Facial expression synthesis device
CN108280413A (en) * 2018-01-17 2018-07-13 百度在线网络技术(北京)有限公司 Face identification method and device
CN108446609A (en) * 2018-03-02 2018-08-24 南京邮电大学 A kind of multi-angle human facial expression recognition method based on generation confrontation network
CN108447041A (en) * 2018-01-30 2018-08-24 中国航天电子技术研究院 A kind of multisource image anastomosing method based on enhancing study
CN108491816A (en) * 2018-03-30 2018-09-04 百度在线网络技术(北京)有限公司 The method and apparatus for carrying out target following in video
CN108509940A (en) * 2018-04-20 2018-09-07 北京达佳互联信息技术有限公司 Facial image tracking, device, computer equipment and storage medium
CN108765014A (en) * 2018-05-30 2018-11-06 中海云智慧(北京)物联网科技有限公司 A kind of intelligent advertisement put-on method based on access control system
CN108858245A (en) * 2018-08-20 2018-11-23 深圳威琳懋生物科技有限公司 A kind of shopping guide robot
CN108961272A (en) * 2018-07-02 2018-12-07 浙江工业大学 It is a kind of to fight the generation method for generating the skin disease image of network based on depth convolution
CN108985377A (en) * 2018-07-18 2018-12-11 太原理工大学 A kind of image high-level semantics recognition methods of the multiple features fusion based on deep layer network
CN109048934A (en) * 2018-08-20 2018-12-21 深圳威琳懋生物科技有限公司 A kind of intelligent shopping guide robot system
CN109086707A (en) * 2018-07-25 2018-12-25 电子科技大学 A kind of expression method for tracing based on DCNNs-LSTM model
CN109145992A (en) * 2018-08-27 2019-01-04 西安电子科技大学 Cooperation generates confrontation network and sky composes united hyperspectral image classification method
CN109299669A (en) * 2018-08-30 2019-02-01 清华大学 Video human face critical point detection method and device based on double intelligent bodies
CN109345770A (en) * 2018-11-14 2019-02-15 深圳市尼欧科技有限公司 A kind of child leaves in-vehicle alarm system and child leaves interior alarm method
CN109376625A (en) * 2018-10-10 2019-02-22 东北大学 A kind of human facial expression recognition method based on convolutional neural networks
CN109493297A (en) * 2018-11-01 2019-03-19 重庆中科云丛科技有限公司 Low quality facial image Enhancement Method, system, equipment and storage medium
CN109508783A (en) * 2018-12-28 2019-03-22 杭州翼兔网络科技有限公司 Mood incorporates roughly model construction and the automatic progress rough acquisition methods of mood into
CN109508689A (en) * 2018-11-28 2019-03-22 中山大学 A kind of expression recognition method that confrontation is strengthened
CN109598212A (en) * 2018-11-20 2019-04-09 北京知道创宇信息技术有限公司 A kind of method for detecting human face and device
CN109614611A (en) * 2018-11-28 2019-04-12 中山大学 A kind of fusion generates the sentiment analysis method of non-confrontation network and convolutional neural networks
CN109620269A (en) * 2019-01-28 2019-04-16 深圳市赛梅斯凯科技有限公司 Fatigue detection method, device, equipment and readable storage medium storing program for executing
CN109635727A (en) * 2018-12-11 2019-04-16 昆山优尼电能运动科技有限公司 A kind of facial expression recognizing method and device
CN109711392A (en) * 2019-01-24 2019-05-03 郑州市现代人才测评与考试研究院 A kind of talent's assessment method based on recognition of face
CN109829362A (en) * 2018-12-18 2019-05-31 深圳壹账通智能科技有限公司 Safety check aided analysis method, device, computer equipment and storage medium
CN109934150A (en) * 2019-03-07 2019-06-25 百度在线网络技术(北京)有限公司 A kind of meeting participation recognition methods, device, server and storage medium
CN109948441A (en) * 2019-02-14 2019-06-28 北京奇艺世纪科技有限公司 Model training, image processing method, device, electronic equipment and computer readable storage medium
CN110019732A (en) * 2017-12-27 2019-07-16 杭州华为数字技术有限公司 A kind of intelligent answer method and relevant apparatus
CN110059593A (en) * 2019-04-01 2019-07-26 华侨大学 A kind of human facial expression recognition method based on feedback convolutional neural networks
CN110084143A (en) * 2019-04-04 2019-08-02 广州大学 A kind of emotional information guard method and system for recognition of face
CN110147822A (en) * 2019-04-16 2019-08-20 北京师范大学 A kind of moos index calculation method based on the detection of human face action unit
CN110188604A (en) * 2019-04-18 2019-08-30 盎锐(上海)信息科技有限公司 Face identification method and device based on 2D and 3D image
CN110288677A (en) * 2019-05-21 2019-09-27 北京大学 It is a kind of based on can deformation structure pedestrian image generation method and device
CN110287895A (en) * 2019-04-17 2019-09-27 北京阳光易德科技股份有限公司 A method of emotional measurement is carried out based on convolutional neural networks
WO2019218111A1 (en) * 2018-05-14 2019-11-21 合刃科技(武汉)有限公司 Electronic device and photographing control method
CN110569741A (en) * 2019-08-19 2019-12-13 昆山琪奥智能科技有限公司 Expression recognition system based on artificial intelligence
CN110866962A (en) * 2019-11-20 2020-03-06 成都威爱新经济技术研究院有限公司 Virtual portrait and expression synchronization method based on convolutional neural network
WO2020052170A1 (en) * 2018-09-11 2020-03-19 深圳云天励飞技术有限公司 Target object identification method and device, and storage medium
CN111325243A (en) * 2020-02-03 2020-06-23 天津大学 Visual relation detection method based on regional attention learning mechanism
CN111368579A (en) * 2018-12-25 2020-07-03 中国电信股份有限公司 Image recognition method, apparatus and system, and computer-readable storage medium
CN111539476A (en) * 2020-04-24 2020-08-14 四川大学 Observation point deployment method for information source positioning based on naive Bayes
CN111753798A (en) * 2020-07-03 2020-10-09 重庆智者炎麒科技有限公司 Teaching auxiliary system based on image processing and method thereof
CN111967526A (en) * 2020-08-20 2020-11-20 东北大学秦皇岛分校 Remote sensing image change detection method and system based on edge mapping and deep learning
CN109829959B (en) * 2018-12-25 2021-01-08 中国科学院自动化研究所 Facial analysis-based expression editing method and device
TWI717008B (en) * 2019-09-09 2021-01-21 訊連科技股份有限公司 Method for establishing biometric database, method for facial recognition and system thereof
CN112487811A (en) * 2020-10-21 2021-03-12 上海旻浦科技有限公司 Cascading information extraction system and method based on reinforcement learning
CN112840359A (en) * 2018-10-12 2021-05-25 渊慧科技有限公司 Controlling agents on a long time scale by using time value delivery
CN113205504A (en) * 2021-05-12 2021-08-03 青岛大学附属医院 Artificial intelligence kidney tumor prediction system based on knowledge graph
CN113674373A (en) * 2021-07-02 2021-11-19 清华大学 Realistic face rendering method and device based on deep learning
CN115052193A (en) * 2022-05-25 2022-09-13 天翼爱音乐文化科技有限公司 Video recommendation method, system, device and storage medium
CN115641236A (en) * 2022-09-30 2023-01-24 广州宏途数字科技有限公司 Campus intelligent security management system based on big data
CN115690891A (en) * 2022-12-30 2023-02-03 苏州魔视智能科技有限公司 Face emotion recognition method, device, equipment and storage medium
US20230080164A1 (en) * 2017-12-20 2023-03-16 Alpvision S.A. Authentication machine learning from multiple digital presentations
CN116186594A (en) * 2023-04-26 2023-05-30 成都市环境应急指挥保障中心 Method for realizing intelligent detection of environment change trend based on decision network combined with big data
US11668684B2 (en) * 2019-07-23 2023-06-06 Landmark Graphics Corporation Stochastic realization of parameter inversion in physics-based empirical models
WO2023098912A1 (en) * 2021-12-02 2023-06-08 新东方教育科技集团有限公司 Image processing method and apparatus, storage medium, and electronic device
CN117315749A (en) * 2023-09-25 2023-12-29 惠州市沃生照明有限公司 Intelligent light regulation and control method and system for desk lamp
CN113674373B (en) * 2021-07-02 2024-04-26 清华大学 Real face rendering method based on deep learning

Cited By (79)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107909061B (en) * 2017-12-07 2021-03-30 电子科技大学 Head posture tracking device and method based on incomplete features
CN107909061A (en) * 2017-12-07 2018-04-13 电子科技大学 A kind of head pose tracks of device and method based on incomplete feature
US20230080164A1 (en) * 2017-12-20 2023-03-16 Alpvision S.A. Authentication machine learning from multiple digital presentations
CN108230239A (en) * 2017-12-25 2018-06-29 中国科学院自动化研究所 Facial expression synthesis device
CN110019732A (en) * 2017-12-27 2019-07-16 杭州华为数字技术有限公司 A kind of intelligent answer method and relevant apparatus
CN108177146A (en) * 2017-12-28 2018-06-19 北京奇虎科技有限公司 Control method, device and the computing device of robot head
CN108280413A (en) * 2018-01-17 2018-07-13 百度在线网络技术(北京)有限公司 Face identification method and device
CN108280413B (en) * 2018-01-17 2022-04-19 百度在线网络技术(北京)有限公司 Face recognition method and device
CN108447041A (en) * 2018-01-30 2018-08-24 中国航天电子技术研究院 A kind of multisource image anastomosing method based on enhancing study
CN108447041B (en) * 2018-01-30 2020-12-15 中国航天电子技术研究院 Multi-source image fusion method based on reinforcement learning
CN108446609A (en) * 2018-03-02 2018-08-24 南京邮电大学 A kind of multi-angle human facial expression recognition method based on generation confrontation network
CN108491816A (en) * 2018-03-30 2018-09-04 百度在线网络技术(北京)有限公司 The method and apparatus for carrying out target following in video
CN108509940A (en) * 2018-04-20 2018-09-07 北京达佳互联信息技术有限公司 Facial image tracking, device, computer equipment and storage medium
CN108509940B (en) * 2018-04-20 2019-11-05 北京达佳互联信息技术有限公司 Facial image tracking, device, computer equipment and storage medium
CN111279682A (en) * 2018-05-14 2020-06-12 合刃科技(武汉)有限公司 Electronic device and shooting control method
WO2019218111A1 (en) * 2018-05-14 2019-11-21 合刃科技(武汉)有限公司 Electronic device and photographing control method
CN108765014A (en) * 2018-05-30 2018-11-06 中海云智慧(北京)物联网科技有限公司 A kind of intelligent advertisement put-on method based on access control system
CN108961272A (en) * 2018-07-02 2018-12-07 浙江工业大学 It is a kind of to fight the generation method for generating the skin disease image of network based on depth convolution
CN108985377A (en) * 2018-07-18 2018-12-11 太原理工大学 A kind of image high-level semantics recognition methods of the multiple features fusion based on deep layer network
CN109086707A (en) * 2018-07-25 2018-12-25 电子科技大学 A kind of expression method for tracing based on DCNNs-LSTM model
CN109048934A (en) * 2018-08-20 2018-12-21 深圳威琳懋生物科技有限公司 A kind of intelligent shopping guide robot system
CN108858245A (en) * 2018-08-20 2018-11-23 深圳威琳懋生物科技有限公司 A kind of shopping guide robot
CN109145992B (en) * 2018-08-27 2021-07-20 西安电子科技大学 Hyperspectral image classification method for cooperatively generating countermeasure network and spatial spectrum combination
CN109145992A (en) * 2018-08-27 2019-01-04 西安电子科技大学 Cooperation generates confrontation network and sky composes united hyperspectral image classification method
CN109299669B (en) * 2018-08-30 2020-11-13 清华大学 Video face key point detection method and device based on double intelligent agents
CN109299669A (en) * 2018-08-30 2019-02-01 清华大学 Video human face critical point detection method and device based on double intelligent bodies
WO2020052170A1 (en) * 2018-09-11 2020-03-19 深圳云天励飞技术有限公司 Target object identification method and device, and storage medium
CN109376625A (en) * 2018-10-10 2019-02-22 东北大学 A kind of human facial expression recognition method based on convolutional neural networks
CN112840359A (en) * 2018-10-12 2021-05-25 渊慧科技有限公司 Controlling agents on a long time scale by using time value delivery
CN109493297A (en) * 2018-11-01 2019-03-19 重庆中科云丛科技有限公司 Low quality facial image Enhancement Method, system, equipment and storage medium
CN109345770A (en) * 2018-11-14 2019-02-15 深圳市尼欧科技有限公司 A kind of child leaves in-vehicle alarm system and child leaves interior alarm method
CN109598212A (en) * 2018-11-20 2019-04-09 北京知道创宇信息技术有限公司 A kind of method for detecting human face and device
CN109614611B (en) * 2018-11-28 2021-09-03 中山大学 Emotion analysis method for fusion generation of non-antagonistic network and convolutional neural network
CN109614611A (en) * 2018-11-28 2019-04-12 中山大学 A kind of fusion generates the sentiment analysis method of non-confrontation network and convolutional neural networks
CN109508689A (en) * 2018-11-28 2019-03-22 中山大学 A kind of expression recognition method that confrontation is strengthened
CN109508689B (en) * 2018-11-28 2023-01-03 中山大学 Face recognition method for strengthening confrontation
CN109635727A (en) * 2018-12-11 2019-04-16 昆山优尼电能运动科技有限公司 A kind of facial expression recognizing method and device
CN109829362A (en) * 2018-12-18 2019-05-31 深圳壹账通智能科技有限公司 Safety check aided analysis method, device, computer equipment and storage medium
CN111368579B (en) * 2018-12-25 2023-07-07 中国电信股份有限公司 Image recognition method, apparatus and system, and computer-readable medium
CN109829959B (en) * 2018-12-25 2021-01-08 中国科学院自动化研究所 Facial analysis-based expression editing method and device
CN111368579A (en) * 2018-12-25 2020-07-03 中国电信股份有限公司 Image recognition method, apparatus and system, and computer-readable storage medium
CN109508783B (en) * 2018-12-28 2021-07-20 沛誉(武汉)科技有限公司 Method for constructing rough emotion classifying model and automatically performing rough emotion acquisition
CN109508783A (en) * 2018-12-28 2019-03-22 杭州翼兔网络科技有限公司 Mood incorporates roughly model construction and the automatic progress rough acquisition methods of mood into
CN109711392A (en) * 2019-01-24 2019-05-03 郑州市现代人才测评与考试研究院 A kind of talent's assessment method based on recognition of face
CN109620269A (en) * 2019-01-28 2019-04-16 深圳市赛梅斯凯科技有限公司 Fatigue detection method, device, equipment and readable storage medium storing program for executing
CN109620269B (en) * 2019-01-28 2021-10-22 锦图计算技术(深圳)有限公司 Fatigue detection method, device, equipment and readable storage medium
CN109948441A (en) * 2019-02-14 2019-06-28 北京奇艺世纪科技有限公司 Model training, image processing method, device, electronic equipment and computer readable storage medium
CN109934150A (en) * 2019-03-07 2019-06-25 百度在线网络技术(北京)有限公司 A kind of meeting participation recognition methods, device, server and storage medium
CN110059593B (en) * 2019-04-01 2022-09-30 华侨大学 Facial expression recognition method based on feedback convolutional neural network
CN110059593A (en) * 2019-04-01 2019-07-26 华侨大学 A kind of human facial expression recognition method based on feedback convolutional neural networks
CN110084143B (en) * 2019-04-04 2021-09-28 广州大学 Emotion information protection method and system for face recognition
CN110084143A (en) * 2019-04-04 2019-08-02 广州大学 A kind of emotional information guard method and system for recognition of face
CN110147822A (en) * 2019-04-16 2019-08-20 北京师范大学 A kind of moos index calculation method based on the detection of human face action unit
CN110287895B (en) * 2019-04-17 2021-08-06 北京阳光易德科技股份有限公司 Method for measuring emotion based on convolutional neural network
CN110287895A (en) * 2019-04-17 2019-09-27 北京阳光易德科技股份有限公司 A method of emotional measurement is carried out based on convolutional neural networks
CN110188604A (en) * 2019-04-18 2019-08-30 盎锐(上海)信息科技有限公司 Face identification method and device based on 2D and 3D image
CN110288677A (en) * 2019-05-21 2019-09-27 北京大学 It is a kind of based on can deformation structure pedestrian image generation method and device
US11668684B2 (en) * 2019-07-23 2023-06-06 Landmark Graphics Corporation Stochastic realization of parameter inversion in physics-based empirical models
CN110569741A (en) * 2019-08-19 2019-12-13 昆山琪奥智能科技有限公司 Expression recognition system based on artificial intelligence
TWI717008B (en) * 2019-09-09 2021-01-21 訊連科技股份有限公司 Method for establishing biometric database, method for facial recognition and system thereof
CN110866962B (en) * 2019-11-20 2023-06-16 成都威爱新经济技术研究院有限公司 Virtual portrait and expression synchronization method based on convolutional neural network
CN110866962A (en) * 2019-11-20 2020-03-06 成都威爱新经济技术研究院有限公司 Virtual portrait and expression synchronization method based on convolutional neural network
CN111325243B (en) * 2020-02-03 2023-06-16 天津大学 Visual relationship detection method based on regional attention learning mechanism
CN111325243A (en) * 2020-02-03 2020-06-23 天津大学 Visual relation detection method based on regional attention learning mechanism
CN111539476A (en) * 2020-04-24 2020-08-14 四川大学 Observation point deployment method for information source positioning based on naive Bayes
CN111753798A (en) * 2020-07-03 2020-10-09 重庆智者炎麒科技有限公司 Teaching auxiliary system based on image processing and method thereof
CN111967526B (en) * 2020-08-20 2023-09-22 东北大学秦皇岛分校 Remote sensing image change detection method and system based on edge mapping and deep learning
CN111967526A (en) * 2020-08-20 2020-11-20 东北大学秦皇岛分校 Remote sensing image change detection method and system based on edge mapping and deep learning
CN112487811A (en) * 2020-10-21 2021-03-12 上海旻浦科技有限公司 Cascading information extraction system and method based on reinforcement learning
CN113205504A (en) * 2021-05-12 2021-08-03 青岛大学附属医院 Artificial intelligence kidney tumor prediction system based on knowledge graph
CN113674373B (en) * 2021-07-02 2024-04-26 清华大学 Real face rendering method based on deep learning
CN113674373A (en) * 2021-07-02 2021-11-19 清华大学 Realistic face rendering method and device based on deep learning
WO2023098912A1 (en) * 2021-12-02 2023-06-08 新东方教育科技集团有限公司 Image processing method and apparatus, storage medium, and electronic device
CN115052193A (en) * 2022-05-25 2022-09-13 天翼爱音乐文化科技有限公司 Video recommendation method, system, device and storage medium
CN115641236A (en) * 2022-09-30 2023-01-24 广州宏途数字科技有限公司 Campus intelligent security management system based on big data
CN115690891B (en) * 2022-12-30 2023-04-07 苏州魔视智能科技有限公司 Face emotion recognition method, device, equipment and storage medium
CN115690891A (en) * 2022-12-30 2023-02-03 苏州魔视智能科技有限公司 Face emotion recognition method, device, equipment and storage medium
CN116186594A (en) * 2023-04-26 2023-05-30 成都市环境应急指挥保障中心 Method for realizing intelligent detection of environment change trend based on decision network combined with big data
CN117315749A (en) * 2023-09-25 2023-12-29 惠州市沃生照明有限公司 Intelligent light regulation and control method and system for desk lamp

Similar Documents

Publication Publication Date Title
CN107423707A (en) A kind of face Emotion identification method based under complex environment
Zahara et al. The facial emotion recognition (FER-2013) dataset for prediction system of micro-expressions face using the convolutional neural network (CNN) algorithm based Raspberry Pi
Li et al. Deep reinforcement learning for robust emotional classification in facial expression recognition
Zhou et al. A lightweight convolutional neural network for real-time facial expression detection
Suzuki et al. Recognition and mapping of facial expressions to avatar by embedded photo reflective sensors in head mounted display
Peng et al. Towards facial expression recognition in the wild: A new database and deep recognition system
Tian et al. Ear recognition based on deep convolutional network
Pathar et al. Human emotion recognition using convolutional neural network in real time
Singh et al. Deep learning and machine learning based facial emotion detection using CNN
Ali et al. Facial emotion detection using neural network
CN109815920A (en) Gesture identification method based on convolutional neural networks and confrontation convolutional neural networks
Zhang et al. A survey on face anti-spoofing algorithms
Rusia et al. An efficient CNN approach for facial expression recognition with some measures of overfitting
Patil et al. Performance analysis of static hand gesture recognition approaches using artificial neural network, support vector machine and two stream based transfer learning approach
Shanthi et al. Algorithms for face recognition drones
Guo et al. Dynamic facial expression recognition based on ResNet and LSTM
Tunc et al. Age group and gender classification using convolutional neural networks with a fuzzy logic-based filter method for noise reduction
Moran Classifying emotion using convolutional neural networks
Hua et al. Collaborative Generative Adversarial Network with Visual perception and memory reasoning
ALISAWI et al. Real-Time Emotion Recognition Using Deep Learning Methods: Systematic Review
Kousalya et al. Prediction of Best Optimizer for Facial Expression Detection using Convolutional Neural Network
Sultana et al. A Deep CNN based Kaggle Contest Winning Model to Recognize Real-Time Facial Expression
Tiwari et al. Personality prediction from Five-Factor Facial Traits using Deep learning
Wang Facial affect detection using convolutional neural networks
Andrian et al. Comparative analysis of deep convolutional neural networks architecture in facial expression recognition: A survey

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20171201

WD01 Invention patent application deemed withdrawn after publication