CN107194347A - A kind of method that micro- expression detection is carried out based on Facial Action Coding System - Google Patents
A kind of method that micro- expression detection is carried out based on Facial Action Coding System Download PDFInfo
- Publication number
- CN107194347A CN107194347A CN201710356981.7A CN201710356981A CN107194347A CN 107194347 A CN107194347 A CN 107194347A CN 201710356981 A CN201710356981 A CN 201710356981A CN 107194347 A CN107194347 A CN 107194347A
- Authority
- CN
- China
- Prior art keywords
- layer
- expression
- micro
- training
- network
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/174—Facial expression recognition
- G06V40/176—Dynamic expression
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- General Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Image Analysis (AREA)
Abstract
A kind of method that micro- expression detection is carried out based on Facial Action Coding System proposed in the present invention, its main contents are included:Visualize CNN wave filters, the network architecture and training, transfer learning, the detection of micro- expression, its process is, initially set up sound mood taxonomy model, analysis proposes the model of e-learning, the wave filter of the network training proposed is visualized in different emotional semantic classification tasks, model is applied into micro- expression detects.The present invention improves discrimination of the existing method in micro- expression detection, illustrate the feature that is produced by unsupervised learning process and for the strong correlation between motor unit in facial expression analysis method, provide high-precision fraction across data and demonstrate in terms of task the generalization ability of the function based on FACS, and improve the discrimination of micro- expression detection, more accurately recognize facial expression and infer emotional state, its validity and accuracy rate applied in every field is improved, the development of artificial intelligence is promoted.
Description
Technical field
The present invention relates to Expression Recognition field, more particularly, to a kind of based on the micro- expression of Facial Action Coding System progress
The method of detection.
Background technology
Expression Recognition is usually used in man-machine interaction, social gaming, psychological research, aids in the fields such as driving, automatic identification face
Express one's feelings and infer emotional state.Specifically, automatically snapped as detecting person taken picture smiling face starts, the automatic expression of game player is more
Change, the senior application such as user's viewing effect analysis of multimedia advertising detects painful and unfortunate, the inspection that driver is drowsiness of patient
Survey.Facial expression plays a significant role in terms of human communication and behavior, although existing method is in object of observation feature and analysis
Certain accuracy has been met, but method is most at present only considers local message, and ignore Space Consistency, so as to guide
Evaluated error, causes that the partial target in special scenes accurately can not be recognized and detect.
The present invention proposes a kind of method that micro- expression detection is carried out based on Facial Action Coding System, visual using CNN
Change the characteristic pattern of emotion detection.Sound mood taxonomy model is initially set up, analysis proposes the model of e-learning, will carried
The wave filter of the network training gone out is visualized in different emotional semantic classification tasks, then, in the high-precision fraction of offer across number
According to the generalization ability that the function based on Facial Action Coding System (FACS) is demonstrated with across task aspect, model is applied to micro-
Expression detection.The present invention improves discrimination of the existing method in micro- expression detection, illustrates and is produced by unsupervised learning process
Raw feature and for the strong correlation between motor unit in facial expression analysis method, provide high-precision fraction across number
According to demonstrating the generalization ability of the function based on FACS with across task aspect, and the discrimination of micro- expression detection is improved, more
Facial expression is recognized exactly and emotional state is inferred, is improved its validity and accuracy rate applied in every field, is promoted people
The development of work intelligence.
The content of the invention
The problem of for existing method discrimination deficiency, the present invention improves identification of the existing method in micro- expression detection
Rate, illustrate the feature that is produced by unsupervised learning process with for the strong phase between motor unit in facial expression analysis method
Guan Xing, provide high-precision fraction across data and demonstrate in terms of task the generalization ability of the function based on FACS, and
The discrimination of micro- expression detection is improved, facial expression is more accurately recognized and infers emotional state, improve it in every field
The validity and accuracy rate of application, promote the development of artificial intelligence.
To solve the above problems, the present invention provides a kind of side that micro- expression detection is carried out based on Facial Action Coding System
Method, its main contents include:
(1) CNN wave filters are visualized;
(2) network architecture and training;
(3) transfer learning;
(4) micro- expression detection.
Wherein, described visualization CNN wave filters, set up after sound mood taxonomy model, and analysis proposes network science
The model of habit, the wave filter of the network training proposed is visualized in different emotional semantic classification tasks, and lower floor provides rudimentary
Other class Gabor filter, and provide high level human body close to the intermediate layer of output and higher level and feature can be read, by making
In aforementioned manners, it can be seen that the feature of institute's training network, wave filter needed for feature visualization shows maximization by input
With the activation for the pixel for being responsible for the response, from trained model is analyzed as can be seen that characteristic pattern and the specific face of network
There is very big similitude between portion region and motion, and these regions and motion are with defining Facial Action Coding System (FACS)
There is significant correlation the part of moving cell.
Further, described FACS, be Facial Action Coding System, it is first determined 7 main universal moods, full
Foot constant characteristic of expressed meaning under different cultural environments, them are marked with corresponding affective state, i.e., happy, sad
Wound, it is pleasantly surprised, it is frightened, detest, indignation and despise, be widely used in cognitive calculating, and FACS is a kind of to be based on anatomical system
System, all observable face actions for describing every kind of mood, using FACS as methodology measuring system, can be retouched
State any expression of motor unit (AU) activation and its enliven intensity, each motor unit describes one group of facial muscles, together altogether
With one specific motion of composition.
Further, described CNN wave filters, are represented the doubtful AU of wave filter and actual data using following methods
AU labels are concentrated to match:
(1) convolutional layer l and wave filter j is given, activation output is marked as FL, j;
(2) maximum N number of input picture i=arg is extractedimax FL, j(i);
(3) for each input i, the AU labels of manual annotations areIf motor unit u exists in i, AI, u
For 1;
(4) wave filter j and the correlation of motor unit u presence are PJ, uAnd byDefinition;
A large amount of top neurons are found to be itself and do not produce effective output, last convolution for any input
Layer in enliven neuron quantity be about characteristic pattern size 30% (having 60 in 256), the quantity of formal neuron and
The vocabulary size of FACS motor unit is approximate, can identify corresponding facial expression.
Wherein, described Gabor filter, it is characterised in that Gabor filter, which is one, is used for the linear of rim detection
Wave filter, the frequency of Gabor filter and direction represent the expression for frequency and direction close to human visual system, and it
It is standing be used for texture representation and description, in spatial domain, the Gabor filter of one 2 dimension is a sinusoidal plane wave and Gaussian kernel
The product of function, with the characteristic for obtaining optimal partial simultaneously in spatial domain and frequency domain, with human biological's visual characteristic very
It is similar, therefore, it is possible to describe the partial structurtes letter corresponding to spatial frequency (yardstick), locus and set direction well
Breath, Gabor filter is self similarity, that is to say, that all Gabor filters can from morther wavelet by expansion and
Rotation is produced, in practical application, Gabor filter can frequency domain different scale, extract correlated characteristic on different directions.
Wherein, the described network architecture and training, realize a simple classical feedforward convolutional neural networks, each net
The structure of network is as follows:Input layer, receives gray-scale map or RGB image, and input includes wave filter by 3 convolutional layer blocks, each block
Layer, non-linear (or activation) and maximum pond layer composition, wherein 3 convolution blocks, each block has amendment linear unit (ReLU)
The pond layer of activation primitive and 2x2, convolutional layer has filter graph, and wave filter (neuron) number is more, and layer is deeper, respectively obtains
64,128 and 256 filter graph sizes, each filter supports 5x5 pixels, and convolution block is one hidden with 512 afterwards
That hides neuron is fully connected layer, and the output of hidden layer is transferred to output layer, and Output Size size is influenceed by task, and 8
Individual to be used for emotional semantic classification, up to 50 are used for AU labels, and output layer can change in activation, in order to reduce over-fitting, use
Layer is abandoned, using layer is abandoned between last convolutional layer and the layer being fully connected, its probability is respectively 0.25 and 0.5,
It is p to abandon layer probability, it is meant that the output of each neuron has Probability p to be arranged to 0.
Further, described network training, using ADAM optimizer training networks, learning rate is 10-3, attenuation rate is
10-5, in order to make model generalization to greatest extent, using random upset and the combination of affine transformation, for example, rotate, change, contracting
Put, carry out data extending, generated data is generated on image and amplifies training set.
Wherein, described transfer learning, transfer learning is intended to use and trained in advance in different pieces of information for new task
Model, neural network model usually requires larger training set, however, in some cases, the size of training set is not enough to
Reach correct training, transfer learning allows the feature extractor for using convolutional layer as pre-training, only output layer is according to working as
Preceding task is altered or modified, i.e. first layer is considered as predefined feature, and defines the final layer of task by based on can
It is adjusted with the study of training set.
Wherein, described micro- expression detection, micro- expression is a kind of more spontaneous and delicate facial movement, by identical face
Motion composition, these motions define FACS motor units and intensity is different, and micro- expression often only continues 0.5 second, institute
Think and detect implication therein, be 3 steps by each micro- expression decomposition:Starting, summit and skew, respectively description are moved
Start, the end spied on and acted, FACS category features extractor is applied to the task of the micro- expression of automatic detection, therefore, using
Data set includes the 256 spontaneous micro- expressions shot with 200fps, and all videos are collectively labeled as starting, summit and skew, and
The expression passed on, is summit frame addition AU codings, expression is captured by showing the theme video-frequency band of the required response of triggering.
Further, described micro- expression detection network, network instruction is carried out first from training data sequence to selected frame
Practice, for each video, starting, summit and first and last frame of skew frame, and sequence are only taken, to explain neutral appearance
Gesture, trains CNN to detect mood first, then, future self-training network convolutional layer and shot and long term memory network (LSTM) group
Close, it, which is inputted, is connected to first of feature extractor CNN and is fully connected layer, used LSTM only comprising one LSTM layer with
One output layer, layer is abandoned after LSTM layers using circulation.
Brief description of the drawings
Fig. 1 is a kind of system flow chart for the method that micro- expression detection is carried out based on Facial Action Coding System of the present invention.
Fig. 2 is a kind of wave filter visualization for the method that micro- expression detection is carried out based on Facial Action Coding System of the present invention
Process.
Fig. 3 is a kind of main expression for the method that micro- expression detection is carried out based on Facial Action Coding System of the present invention.
Fig. 4 is that a kind of motor unit for the method that micro- expression detection is carried out based on Facial Action Coding System of the present invention is compiled
Code.
Fig. 5 is a kind of data set legend for the method that micro- expression detection is carried out based on Facial Action Coding System of the present invention.
Embodiment
It should be noted that in the case where not conflicting, the feature in embodiment and embodiment in the application can phase
Mutually combine, the present invention is described in further detail with specific embodiment below in conjunction with the accompanying drawings.
Fig. 1 is a kind of system flow chart for the method that micro- expression detection is carried out based on Facial Action Coding System of the present invention.
It is main to include visualization CNN wave filters, the network architecture and training, transfer learning, the detection of micro- expression.
Wherein, described visualization CNN wave filters, set up after sound mood taxonomy model, and analysis proposes network science
The model of habit, the wave filter of the network training proposed is visualized in different emotional semantic classification tasks, and lower floor provides rudimentary
Other class Gabor filter, and provide high level human body close to the intermediate layer of output and higher level and feature can be read, by making
In aforementioned manners, it can be seen that the feature of institute's training network, wave filter needed for feature visualization shows maximization by input
With the activation for the pixel for being responsible for the response, from trained model is analyzed as can be seen that characteristic pattern and the specific face of network
There is very big similitude between portion region and motion, and these regions and motion are with defining Facial Action Coding System (FACS)
There is significant correlation the part of moving cell.
Further, described FACS, be Facial Action Coding System, it is first determined 7 main universal moods, full
Foot constant characteristic of expressed meaning under different cultural environments, them are marked with corresponding affective state, i.e., happy, sad
Wound, it is pleasantly surprised, it is frightened, detest, indignation and despise, be widely used in cognitive calculating, and FACS is a kind of to be based on anatomical system
System, all observable face actions for describing every kind of mood, using FACS as methodology measuring system, can be retouched
State any expression of motor unit (AU) activation and its enliven intensity, each motor unit describes one group of facial muscles, together altogether
With one specific motion of composition.
Further, described CNN wave filters, are represented the doubtful AU of wave filter and actual data using following methods
AU labels are concentrated to match:
(1) convolutional layer l and wave filter j is given, activation output is marked as FL, j;
(2) maximum N number of input picture i=arg is extractedimax FL, j(i);
(3) for each input i, the AU labels of manual annotationsIf motor unit u exists in i, AI, uFor
1;
(4) wave filter j and the correlation of motor unit u presence are PJ, uAnd byDefinition;
A large amount of top neurons are found to be itself and do not produce effective output, last convolution for any input
Layer in enliven neuron quantity be about characteristic pattern size 30% (having 60 in 256), the quantity of formal neuron and
The vocabulary size of FACS motor unit is approximate, can identify corresponding facial expression.
Wherein, described Gabor filter, it is characterised in that Gabor filter, which is one, is used for the linear of rim detection
Wave filter, the frequency of Gabor filter and direction represent the expression for frequency and direction close to human visual system, and it
It is standing be used for texture representation and description, in spatial domain, the Gabor filter of one 2 dimension is a sinusoidal plane wave and Gaussian kernel
The product of function, with the characteristic for obtaining optimal partial simultaneously in spatial domain and frequency domain, with human biological's visual characteristic very
It is similar, therefore, it is possible to describe the partial structurtes letter corresponding to spatial frequency (yardstick), locus and set direction well
Breath, Gabor filter is self similarity, that is to say, that all Gabor filters can from morther wavelet by expansion and
Rotation is produced, in practical application, Gabor filter can frequency domain different scale, extract correlated characteristic on different directions.
Wherein, the described network architecture and training, realize a simple classical feedforward convolutional neural networks, each net
The structure of network is as follows:Input layer, receives gray-scale map or RGB image, and input includes wave filter by 3 convolutional layer blocks, each block
Layer, non-linear (or activation) and maximum pond layer composition, wherein 3 convolution blocks, each block has amendment linear unit (ReLU)
The pond layer of activation primitive and 2x2, convolutional layer has filter graph, and wave filter (neuron) number is more, and layer is deeper, respectively obtains
64,128 and 256 filter graph sizes, each filter supports 5x5 pixels, and convolution block is one hidden with 512 afterwards
That hides neuron is fully connected layer, and the output of hidden layer is transferred to output layer, and Output Size size is influenceed by task, and 8
Individual to be used for emotional semantic classification, up to 50 are used for AU labels, and output layer can change in activation, in order to reduce over-fitting, use
Layer is abandoned, using layer is abandoned between last convolutional layer and the layer being fully connected, its probability is respectively 0.25 and 0.5,
It is p to abandon layer probability, it is meant that the output of each neuron has Probability p to be arranged to 0.
Further, described network training, using ADAM optimizer training networks, learning rate is 10-3, attenuation rate is
10-5, in order to make model generalization to greatest extent, using random upset and the combination of affine transformation, for example, rotate, change, contracting
Put, carry out data extending, generated data is generated on image and amplifies training set.
Wherein, described transfer learning, transfer learning is intended to use and trained in advance in different pieces of information for new task
Model, neural network model usually requires larger training set, however, in some cases, the size of training set is not enough to
Reach correct training, transfer learning allows the feature extractor for using convolutional layer as pre-training, only output layer is according to working as
Preceding task is altered or modified, i.e. first layer is considered as predefined feature, and defines the final layer of task by based on can
It is adjusted with the study of training set.
Wherein, described micro- expression detection, micro- expression is a kind of more spontaneous and delicate facial movement, by identical face
Motion composition, these motions define FACS motor units and intensity is different, and micro- expression often only continues 0.5 second, institute
Think and detect implication therein, be 3 steps by each micro- expression decomposition:Starting, summit and skew, respectively description are moved
Start, the end spied on and acted, FACS category features extractor is applied to the task of the micro- expression of automatic detection, therefore, using
Data set includes the 256 spontaneous micro- expressions shot with 200fps, and all videos are collectively labeled as starting, summit and skew, and
The expression passed on, is summit frame addition AU codings, expression is captured by showing the theme video-frequency band of the required response of triggering.
Further, described micro- expression detection network, network instruction is carried out first from training data sequence to selected frame
Practice, for each video, starting, summit and first and last frame of skew frame, and sequence are only taken, to explain neutral appearance
Gesture, trains CNN to detect mood first, then, future self-training network convolutional layer and shot and long term memory network (LSTM) group
Close, it, which is inputted, is connected to first of feature extractor CNN and is fully connected layer, used LSTM only comprising one LSTM layer with
One output layer, layer is abandoned after LSTM layers using circulation.
Fig. 2 is a kind of wave filter visualization for the method that micro- expression detection is carried out based on Facial Action Coding System of the present invention
Process.Set up after sound mood taxonomy model, analysis proposes the model of e-learning, by the filtering of proposed network training
Device is visualized in different emotional semantic classification tasks.Lower floor provides the Gabor-like wave filters of low level, and close to defeated
The intermediate layer gone out and higher level provide high level human body and feature can be read.It is described along being responsible for by inputting in feature visualization
The activation of wave filter needed for the pixel of response is maximized.
Fig. 3 is a kind of main expression for the method that micro- expression detection is carried out based on Facial Action Coding System of the present invention.From
Left-to-right is respectively to detest, frightened, joyful, surprised, and sad and indignation, is the main expression on facial expression, its generality
The implication of expression will not be changed because of Different Culture, simplicity and the requirement to generality is met.
Fig. 4 is that a kind of motor unit for the method that micro- expression detection is carried out based on Facial Action Coding System of the present invention is compiled
Code.Facial Action Coding System (FACS) is a kind of based on anatomical system, all observables for describing every kind of mood
The face action arrived.Using FACS as methodology measuring system, any expression and its work of motor unit activation can be described
Jump intensity.Each motor unit describes one group of facial muscles, cooperatively constitutes a specific motion.It is dynamic including 44 faces
Make unit, description such as " is dehisced ", the action such as " narrowing eye " now also added 20 other motor units, count head and eye in
The motion of eyeball.
Fig. 5 is a kind of data set legend for the method that micro- expression detection is carried out based on Facial Action Coding System of the present invention.
A common model structure is obtained on various data sets using the method based on CNN, and studies these models and FACS
Relation.In order to check the generalization ability of learning model, these models are understood using transfer learning method how in other data
Performed on collection.In order to understand the predicable based on the state-of-the-art models of CNN in FER, these methods are applied to numerous numbers
According to concentration, figure is the part legend of selection.
For those skilled in the art, the present invention is not restricted to the details of above-described embodiment, in the essence without departing substantially from the present invention
In the case of refreshing and scope, the present invention can be realized with other concrete forms.In addition, those skilled in the art can be to this hair
Bright to carry out various changes and modification without departing from the spirit and scope of the present invention, these improvement and modification also should be regarded as the present invention's
Protection domain.Therefore, appended claims are intended to be construed to include preferred embodiment and fall into all changes of the scope of the invention
More and modification.
Claims (10)
1. a kind of method that micro- expression detection is carried out based on Facial Action Coding System, it is characterised in that main to include visualization
CNN wave filters (one);The network architecture and training (two);Transfer learning (three);Micro- expression detection (four).
2. based on the visualization CNN wave filters (one) described in claims 1, it is characterised in that set up sound mood classification
After framework, analysis proposes the model of e-learning, and the wave filter of the network training proposed is appointed in different emotional semantic classifications
Visualized in business, lower floor provides the class Gabor filter of low level, and provides high-level close to the intermediate layer of output and higher level
Human body feature can be read, by using the above method, it can be seen that the feature of institute's training network, feature visualization pass through input
The activation of pixel of the wave filter needed for maximizing with being responsible for the response is shown, be can be seen that from the model for analyzing trained
There are very big similitude, and these regions and motion and definition between the characteristic pattern of network and specific facial zone and motion
There is significant correlation the part of Facial Action Coding System (FACS) moving cell.
3. based on the FACS described in claims 2, it is characterised in that Facial Action Coding System, it is first determined 7 main
Universal mood, meet the constant characteristic of expressed meaning under different cultural environments, him marked with corresponding affective state
, i.e., it is happy, it is sad, it is pleasantly surprised, it is frightened, detest, indignation and despise, be widely used in cognitive calculating, and FACS is a kind of base
In anatomical system, all observable face actions for describing every kind of mood are surveyed using FACS as methodology
Amount system, can describe any expression of motor unit (AU) activation and its enliven intensity, each motor unit describes one group of face
Portion's muscle, cooperatively constitutes a specific motion.
4. based on the CNN wave filters described in claims 2, it is characterised in that use following methods by the doubtful AU of wave filter
Represent to match with AU labels in actual data set:
(1) convolutional layer l and wave filter j is given, activation output is marked as FL, j;
(2) maximum N number of input picture i=arg is extractedimax FL, j(i);
(3) for each input i, the AU labels of manual annotations areIf motor unit u exists in i, AI, uFor 1;
(4) wave filter j and the correlation of motor unit u presence are PJ, uAnd byDefinition;
A large amount of top neurons are found to be itself and not produced for any input in effective output, last convolutional layer
The quantity for enlivening neuron is about 30% (having 60 in 256) of characteristic pattern size, the quantity of formal neuron and FACS's
The vocabulary size of motor unit is approximate, can identify corresponding facial expression.
5. based on the Gabor filter described in claims 2, it is characterised in that Gabor filter, which is one, is used for edge inspection
The linear filter of survey, the frequency of Gabor filter and direction are represented close to human visual system for frequency and the table in direction
Show, and they are standing for texture representation and description, and in spatial domain, the Gabor filter of one 2 dimension is a sinusoidal plane wave
With the product of gaussian kernel function, with the characteristic for obtaining optimal partial simultaneously in spatial domain and frequency domain, regarded with human biological
Feel that characteristic is much like, therefore, it is possible to describe the office corresponding to spatial frequency (yardstick), locus and set direction well
Portion's structural information, Gabor filter is self similarity, that is to say, that all Gabor filters can be from a morther wavelet warp
Cross expansion and rotation produced, in practical application, Gabor filter can frequency domain different scale, extract phase on different directions
Close feature.
6. based on the network architecture described in claims 1 and training (two), it is characterised in that realize a simple classics
Feedover convolutional neural networks, and the structure of each network is as follows:Input layer, receives gray-scale map or RGB image, and input passes through 3 volumes
Lamination block, each block includes filter layer, non-linear (or activation) and maximum pond layer composition, wherein 3 convolution blocks, each block
Pond layer with amendment linear unit (ReLU) activation primitive and 2x2, convolutional layer has filter graph, wave filter (neuron)
Number is more, and layer is deeper, respectively obtains 64,128 and 256 filter graph sizes, and each filter supports 5x5 pixels, convolution
Be after block one there is 512 hidden neurons be fully connected layer, the output of hidden layer is transferred to output layer, exports chi
Very little size is influenceed by task, and 8 are used for emotional semantic classification, and up to 50 are used for AU labels, and output layer can become in activation
Change, in order to reduce over-fitting, using layer is abandoned, using discarding layer between last convolutional layer and the layer being fully connected,
Its probability is respectively 0.25 and 0.5, and it is p to abandon layer probability, it is meant that the output of each neuron has Probability p to be arranged to
0。
7. based on the network training described in claims 6, it is characterised in that utilize ADAM optimizer training networks, learning rate
For 10-3, attenuation rate is 10-5, in order to make model generalization to greatest extent, use random upset and the combination of affine transformation, example
As rotated, change, scaling carries out data extending, generated data is generated on image and amplifies training set.
8. based on the transfer learning (three) described in claims 1, it is characterised in that transfer learning, which is intended to use, is directed to new task
The model trained in advance in different pieces of information, neural network model usually requires larger training set, however, in some situations
Under, the size of training set is not enough to reach correct training, and transfer learning allows the feature using convolutional layer as pre-training to carry
Device is taken, only output layer is altered or modified according to current task, i.e., first layer is considered as predefined feature, and defines and appoint
The final layer of business is adjusted by the study based on available training set.
9. (four) are detected based on micro- expression described in claims 1, it is characterised in that micro- expression is a kind of more spontaneous and delicate
Facial movement, be made up of identical facial movement, these motions define FACS motor units and intensity is different, micro-
Expression often only continues 0.5 second, so being 3 steps by each micro- expression decomposition to detect implication therein:Starting, top
Point and skew, describe the beginning of motion respectively, and FACS category features extractor is applied to automatic detection by the end spied on and acted
The task of micro- expression, therefore, 256 spontaneous micro- expressions for including shooting with 200fps using data set, all videos are all marked
For starting, summit and skew, and the expression passed on, it is summit frame addition AU codings, response needed for by showing triggering
Theme video-frequency band captures expression.
10. network is detected based on micro- expression described in claims 9, it is characterised in that right from training data sequence first
Selected frame carries out network training, for each video, only takes starting, summit and skew frame, and sequence first and finally
One frame, to explain neutral position, trains CNN to detect mood first, then, future self-training network convolutional layer and shot and long term
Memory network (LSTM) is combined, and it, which is inputted, is connected to first of feature extractor CNN and is fully connected layer, and used LSTM is only
Comprising one LSTM layers and an output layer, layer is abandoned using circulation after LSTM layers.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710356981.7A CN107194347A (en) | 2017-05-19 | 2017-05-19 | A kind of method that micro- expression detection is carried out based on Facial Action Coding System |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710356981.7A CN107194347A (en) | 2017-05-19 | 2017-05-19 | A kind of method that micro- expression detection is carried out based on Facial Action Coding System |
Publications (1)
Publication Number | Publication Date |
---|---|
CN107194347A true CN107194347A (en) | 2017-09-22 |
Family
ID=59874841
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710356981.7A Withdrawn CN107194347A (en) | 2017-05-19 | 2017-05-19 | A kind of method that micro- expression detection is carried out based on Facial Action Coding System |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107194347A (en) |
Cited By (29)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107679526A (en) * | 2017-11-14 | 2018-02-09 | 北京科技大学 | A kind of micro- expression recognition method of face |
CN107862598A (en) * | 2017-09-30 | 2018-03-30 | 平安普惠企业管理有限公司 | Long-range the interview measures and procedures for the examination and approval, server and readable storage medium storing program for executing |
CN107862292A (en) * | 2017-11-15 | 2018-03-30 | 平安科技(深圳)有限公司 | Personage's mood analysis method, device and storage medium |
CN108052982A (en) * | 2017-12-22 | 2018-05-18 | 北京联合网视文化传播有限公司 | A kind of emotion detection method and system based on textures expression |
CN108416235A (en) * | 2018-03-30 | 2018-08-17 | 百度在线网络技术(北京)有限公司 | The anti-peeping method, apparatus of display interface, storage medium and terminal device |
CN108876812A (en) * | 2017-11-01 | 2018-11-23 | 北京旷视科技有限公司 | Image processing method, device and equipment for object detection in video |
CN109165608A (en) * | 2018-08-30 | 2019-01-08 | 深圳壹账通智能科技有限公司 | The micro- expression recognition method of multi-angle of view, device, storage medium and computer equipment |
CN109583431A (en) * | 2019-01-02 | 2019-04-05 | 上海极链网络科技有限公司 | A kind of face Emotion identification model, method and its electronic device |
CN109711310A (en) * | 2018-12-20 | 2019-05-03 | 北京大学 | A kind of infant's attachment type automatic Prediction system and its prediction technique |
CN109766461A (en) * | 2018-12-15 | 2019-05-17 | 深圳壹账通智能科技有限公司 | Photo management method, device, computer equipment and medium based on micro- expression |
CN109800771A (en) * | 2019-01-30 | 2019-05-24 | 杭州电子科技大学 | Mix spontaneous micro- expression localization method of space-time plane local binary patterns |
CN109840513A (en) * | 2019-02-28 | 2019-06-04 | 北京科技大学 | A kind of micro- expression recognition method of face and identification device |
CN110175565A (en) * | 2019-05-27 | 2019-08-27 | 北京字节跳动网络技术有限公司 | The method and apparatus of personage's emotion for identification |
CN110232102A (en) * | 2019-06-13 | 2019-09-13 | 哈尔滨工程大学 | A kind of personnel's relational model modeling method based on transfer learning |
WO2019183758A1 (en) * | 2018-03-26 | 2019-10-03 | Intel Corporation | Methods and apparatus for multi-task recognition using neural networks |
CN110472564A (en) * | 2019-08-14 | 2019-11-19 | 成都中科云集信息技术有限公司 | A kind of micro- Expression Recognition depression method of two-way LSTM based on feature pyramid network |
CN110830849A (en) * | 2018-08-10 | 2020-02-21 | 三星电子株式会社 | Electronic device, method of controlling electronic device, and method for controlling server |
CN111439267A (en) * | 2020-03-30 | 2020-07-24 | 上海商汤临港智能科技有限公司 | Method and device for adjusting cabin environment |
CN111666911A (en) * | 2020-06-13 | 2020-09-15 | 天津大学 | Micro-expression data expansion method and device |
CN111783543A (en) * | 2020-06-02 | 2020-10-16 | 北京科技大学 | Face activity unit detection method based on multitask learning |
CN111950373A (en) * | 2020-07-13 | 2020-11-17 | 南京航空航天大学 | Method for recognizing micro-expressions through transfer learning based on optical flow input |
CN112017986A (en) * | 2020-10-21 | 2020-12-01 | 季华实验室 | Semiconductor product defect detection method and device, electronic equipment and storage medium |
CN112115779A (en) * | 2020-08-11 | 2020-12-22 | 浙江师范大学 | Interpretable classroom student emotion analysis method, system, device and medium |
CN112147573A (en) * | 2020-09-14 | 2020-12-29 | 山东科技大学 | Passive positioning method based on amplitude and phase information of CSI (channel State information) |
CN112380924A (en) * | 2020-10-26 | 2021-02-19 | 华南理工大学 | Depression tendency detection method based on facial micro-expression dynamic recognition |
CN113515702A (en) * | 2021-07-07 | 2021-10-19 | 北京百度网讯科技有限公司 | Content recommendation method, model training method, device, equipment and storage medium |
CN113935377A (en) * | 2021-10-13 | 2022-01-14 | 燕山大学 | Pipeline leakage aperture identification method combining feature migration with time-frequency diagram |
CN114565964A (en) * | 2022-03-03 | 2022-05-31 | 网易(杭州)网络有限公司 | Emotion recognition model generation method, recognition method, device, medium and equipment |
WO2023098912A1 (en) * | 2021-12-02 | 2023-06-08 | 新东方教育科技集团有限公司 | Image processing method and apparatus, storage medium, and electronic device |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105031938A (en) * | 2015-07-07 | 2015-11-11 | 安徽瑞宏信息科技有限公司 | Intelligent toy based on target expression detection and recognition method |
-
2017
- 2017-05-19 CN CN201710356981.7A patent/CN107194347A/en not_active Withdrawn
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105031938A (en) * | 2015-07-07 | 2015-11-11 | 安徽瑞宏信息科技有限公司 | Intelligent toy based on target expression detection and recognition method |
Non-Patent Citations (1)
Title |
---|
RAN BREUER等: ""A Deep Learning Perspective on the Origin of Facial Expressions"", 《网页在线公开:HTTPS://ARXIV.ORG/ABS/1705.01842》 * |
Cited By (42)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107862598A (en) * | 2017-09-30 | 2018-03-30 | 平安普惠企业管理有限公司 | Long-range the interview measures and procedures for the examination and approval, server and readable storage medium storing program for executing |
CN108876812A (en) * | 2017-11-01 | 2018-11-23 | 北京旷视科技有限公司 | Image processing method, device and equipment for object detection in video |
CN107679526A (en) * | 2017-11-14 | 2018-02-09 | 北京科技大学 | A kind of micro- expression recognition method of face |
CN107862292B (en) * | 2017-11-15 | 2019-04-12 | 平安科技(深圳)有限公司 | Personage's mood analysis method, device and storage medium |
CN107862292A (en) * | 2017-11-15 | 2018-03-30 | 平安科技(深圳)有限公司 | Personage's mood analysis method, device and storage medium |
WO2019095571A1 (en) * | 2017-11-15 | 2019-05-23 | 平安科技(深圳)有限公司 | Human-figure emotion analysis method, apparatus, and storage medium |
CN108052982A (en) * | 2017-12-22 | 2018-05-18 | 北京联合网视文化传播有限公司 | A kind of emotion detection method and system based on textures expression |
CN108052982B (en) * | 2017-12-22 | 2021-09-03 | 深圳市云网拜特科技有限公司 | Emotion detection method and system based on chartlet expression |
US11106896B2 (en) | 2018-03-26 | 2021-08-31 | Intel Corporation | Methods and apparatus for multi-task recognition using neural networks |
WO2019183758A1 (en) * | 2018-03-26 | 2019-10-03 | Intel Corporation | Methods and apparatus for multi-task recognition using neural networks |
CN108416235A (en) * | 2018-03-30 | 2018-08-17 | 百度在线网络技术(北京)有限公司 | The anti-peeping method, apparatus of display interface, storage medium and terminal device |
US11388465B2 (en) | 2018-08-10 | 2022-07-12 | Samsung Electronics Co., Ltd. | Electronic apparatus and method for upscaling a down-scaled image by selecting an improved filter set for an artificial intelligence model |
US11825033B2 (en) | 2018-08-10 | 2023-11-21 | Samsung Electronics Co., Ltd. | Apparatus and method with artificial intelligence for scaling image data |
CN110830849A (en) * | 2018-08-10 | 2020-02-21 | 三星电子株式会社 | Electronic device, method of controlling electronic device, and method for controlling server |
CN109165608A (en) * | 2018-08-30 | 2019-01-08 | 深圳壹账通智能科技有限公司 | The micro- expression recognition method of multi-angle of view, device, storage medium and computer equipment |
CN109766461A (en) * | 2018-12-15 | 2019-05-17 | 深圳壹账通智能科技有限公司 | Photo management method, device, computer equipment and medium based on micro- expression |
CN109711310A (en) * | 2018-12-20 | 2019-05-03 | 北京大学 | A kind of infant's attachment type automatic Prediction system and its prediction technique |
CN109583431A (en) * | 2019-01-02 | 2019-04-05 | 上海极链网络科技有限公司 | A kind of face Emotion identification model, method and its electronic device |
CN109800771A (en) * | 2019-01-30 | 2019-05-24 | 杭州电子科技大学 | Mix spontaneous micro- expression localization method of space-time plane local binary patterns |
CN109840513A (en) * | 2019-02-28 | 2019-06-04 | 北京科技大学 | A kind of micro- expression recognition method of face and identification device |
CN110175565A (en) * | 2019-05-27 | 2019-08-27 | 北京字节跳动网络技术有限公司 | The method and apparatus of personage's emotion for identification |
CN110232102A (en) * | 2019-06-13 | 2019-09-13 | 哈尔滨工程大学 | A kind of personnel's relational model modeling method based on transfer learning |
CN110232102B (en) * | 2019-06-13 | 2020-12-04 | 哈尔滨工程大学 | Personnel relation model modeling method based on transfer learning |
CN110472564A (en) * | 2019-08-14 | 2019-11-19 | 成都中科云集信息技术有限公司 | A kind of micro- Expression Recognition depression method of two-way LSTM based on feature pyramid network |
WO2021196721A1 (en) * | 2020-03-30 | 2021-10-07 | 上海商汤临港智能科技有限公司 | Cabin interior environment adjustment method and apparatus |
CN111439267A (en) * | 2020-03-30 | 2020-07-24 | 上海商汤临港智能科技有限公司 | Method and device for adjusting cabin environment |
CN111439267B (en) * | 2020-03-30 | 2021-12-07 | 上海商汤临港智能科技有限公司 | Method and device for adjusting cabin environment |
CN111783543A (en) * | 2020-06-02 | 2020-10-16 | 北京科技大学 | Face activity unit detection method based on multitask learning |
CN111783543B (en) * | 2020-06-02 | 2023-10-27 | 北京科技大学 | Facial activity unit detection method based on multitask learning |
CN111666911A (en) * | 2020-06-13 | 2020-09-15 | 天津大学 | Micro-expression data expansion method and device |
CN111950373B (en) * | 2020-07-13 | 2024-04-16 | 南京航空航天大学 | Method for micro expression recognition based on transfer learning of optical flow input |
CN111950373A (en) * | 2020-07-13 | 2020-11-17 | 南京航空航天大学 | Method for recognizing micro-expressions through transfer learning based on optical flow input |
CN112115779A (en) * | 2020-08-11 | 2020-12-22 | 浙江师范大学 | Interpretable classroom student emotion analysis method, system, device and medium |
CN112147573A (en) * | 2020-09-14 | 2020-12-29 | 山东科技大学 | Passive positioning method based on amplitude and phase information of CSI (channel State information) |
CN112017986A (en) * | 2020-10-21 | 2020-12-01 | 季华实验室 | Semiconductor product defect detection method and device, electronic equipment and storage medium |
CN112380924B (en) * | 2020-10-26 | 2023-09-15 | 华南理工大学 | Depression tendency detection method based on facial micro expression dynamic recognition |
CN112380924A (en) * | 2020-10-26 | 2021-02-19 | 华南理工大学 | Depression tendency detection method based on facial micro-expression dynamic recognition |
CN113515702A (en) * | 2021-07-07 | 2021-10-19 | 北京百度网讯科技有限公司 | Content recommendation method, model training method, device, equipment and storage medium |
CN113935377A (en) * | 2021-10-13 | 2022-01-14 | 燕山大学 | Pipeline leakage aperture identification method combining feature migration with time-frequency diagram |
CN113935377B (en) * | 2021-10-13 | 2024-05-07 | 燕山大学 | Pipeline leakage aperture identification method combining characteristic migration with time-frequency diagram |
WO2023098912A1 (en) * | 2021-12-02 | 2023-06-08 | 新东方教育科技集团有限公司 | Image processing method and apparatus, storage medium, and electronic device |
CN114565964A (en) * | 2022-03-03 | 2022-05-31 | 网易(杭州)网络有限公司 | Emotion recognition model generation method, recognition method, device, medium and equipment |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107194347A (en) | A kind of method that micro- expression detection is carried out based on Facial Action Coding System | |
Breuer et al. | A deep learning perspective on the origin of facial expressions | |
Linsley et al. | Learning what and where to attend | |
Elgammal et al. | Can: Creative adversarial networks, generating" art" by learning about styles and deviating from style norms | |
Ambadar et al. | Deciphering the enigmatic face: The importance of facial dynamics in interpreting subtle facial expressions | |
El Hammoumi et al. | Emotion recognition in e-learning systems | |
Dewan et al. | A deep learning approach to detecting engagement of online learners | |
Pathar et al. | Human emotion recognition using convolutional neural network in real time | |
Lake | Towards more human-like concept learning in machines: Compositionality, causality, and learning-to-learn | |
Carvalhais et al. | Recognition and use of emotions in games | |
Tavares et al. | Crowdsourcing facial expressions for affective-interaction | |
Glowinski et al. | Body, space, and emotion: A perceptual study | |
ALISAWI et al. | Real-Time Emotion Recognition Using Deep Learning Methods: Systematic Review | |
Chandrasekharan et al. | Ideomotor Design: using common coding theory to derive novel video game interactions | |
Xu et al. | Spontaneous visual database for detecting learning-centered emotions during online learning | |
Madhu et al. | Convolutional Siamese networks for one-shot malaria parasite recognition in microscopic images | |
Kousalya et al. | Prediction of Best Optimizer for Facial Expression Detection using Convolutional Neural Network | |
Liliana et al. | The Fuzzy Emotion Recognition Framework Using Semantic-Linguistic Facial Features | |
Sadek et al. | Intelligent real-time facial expression recognition from video sequences based on hybrid feature tracking algorithms | |
Sun | Neural Networks for Emotion Classification | |
Melgare et al. | Investigating Emotion Style in Human Faces Using Clustering Methods | |
Veesam et al. | Deep neural networks for automatic facial expression recognition | |
Portaz et al. | Towards personalised learning of psychomotor skills with data mining | |
Sharmila | Depression level calculation for predicting child psychometric retardation using DepressNet approach through GPU accelerated Google cloud platform | |
Nawaf et al. | Human Emotion Identification based on Deep Learning |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WW01 | Invention patent application withdrawn after publication |
Application publication date: 20170922 |
|
WW01 | Invention patent application withdrawn after publication |