CN113792572A - Facial expression recognition method based on local representation - Google Patents
Facial expression recognition method based on local representation Download PDFInfo
- Publication number
- CN113792572A CN113792572A CN202110670264.8A CN202110670264A CN113792572A CN 113792572 A CN113792572 A CN 113792572A CN 202110670264 A CN202110670264 A CN 202110670264A CN 113792572 A CN113792572 A CN 113792572A
- Authority
- CN
- China
- Prior art keywords
- facial
- local
- expressions
- regions
- face
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 30
- 230000008921 facial expression Effects 0.000 title claims abstract description 29
- 230000014509 gene expression Effects 0.000 claims abstract description 48
- 230000001815 facial effect Effects 0.000 claims abstract description 20
- 238000013527 convolutional neural network Methods 0.000 claims abstract description 16
- 238000000605 extraction Methods 0.000 claims abstract description 13
- 210000003205 muscle Anatomy 0.000 claims abstract description 10
- 210000000056 organ Anatomy 0.000 claims abstract description 7
- 210000001097 facial muscle Anatomy 0.000 claims description 23
- 230000006870 function Effects 0.000 claims description 18
- 210000004709 eyebrow Anatomy 0.000 claims description 12
- 210000001508 eye Anatomy 0.000 claims description 8
- 230000000694 effects Effects 0.000 claims description 7
- 238000005192 partition Methods 0.000 claims description 6
- 210000000214 mouth Anatomy 0.000 claims description 5
- 210000001061 forehead Anatomy 0.000 claims description 4
- 238000013459 approach Methods 0.000 claims description 2
- 238000012512 characterization method Methods 0.000 claims description 2
- 230000003247 decreasing effect Effects 0.000 claims description 2
- 210000001331 nose Anatomy 0.000 claims description 2
- 238000011524 similarity measure Methods 0.000 claims description 2
- 238000012549 training Methods 0.000 claims description 2
- 230000008451 emotion Effects 0.000 abstract description 2
- 238000012545 processing Methods 0.000 abstract description 2
- 210000000744 eyelid Anatomy 0.000 description 8
- 238000011176 pooling Methods 0.000 description 8
- 230000003993 interaction Effects 0.000 description 3
- 238000011840 criminal investigation Methods 0.000 description 2
- 239000000284 extract Substances 0.000 description 2
- 230000007935 neutral effect Effects 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 230000037303 wrinkles Effects 0.000 description 2
- 230000007547 defect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 239000003814 drug Substances 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 210000000887 face Anatomy 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 238000000638 solvent extraction Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/047—Probabilistic or stochastic networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Molecular Biology (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Probability & Statistics with Applications (AREA)
- Image Analysis (AREA)
Abstract
The invention relates to a facial expression recognition method based on local representation, which belongs to the field of face recognition, wherein expressions are important embodiment of human internal emotion change, and the current expression recognition method usually relies on global facial features for processing and ignores local feature extraction. In order to motivate psychologists to point out different facial expressions to different local muscle motion regions, the paper proposes an expression recognition algorithm based on local tokens, called expression motion unit convolutional neural network. To extract the local features of the face, the text first divides the whole face image into 43 sub-regions according to the 68 feature points of the acquired human face, and then selects the muscle movement region and 8 local candidate regions covered by the facial salient organs as the input of the convolutional neural network. To equalize the features of the local candidate regions, 8 parallel feature extraction branches are employed, each branch dominating a different dimensional fully connected layer. The outputs of the branches are adaptively connected according to attention to highlight the importance degree of different local candidate areas, and finally expressions are classified into seven categories of neutrality, anger, disgust, surprise, happiness, sadness and fear through a Softmax function.
Description
Technical Field
The invention belongs to the field of face recognition, and particularly relates to a facial expression recognition method based on local representation.
Background
Facial expression recognition may recognize human facial expressions, such as: surprisingly, sadness, happiness, anger and the like, and facial expression recognition has wide potential application, and can be used for man-machine interaction, shopping recommendation, criminal investigation, medical assistance and the like. For example: recommending commodities and inferring the preference according to the expression change of the commodities browsed by the user during man-machine interaction; in criminal investigation, facial expressions of criminal suspects are used for inferring psychological changes of the criminal suspects; and the facial expression of the patient is observed during medical aid to adjust the dosage of the medicine and the like. Mehrabian proposes that 55% of human emotion information is transmitted through human faces, and if the face information can be read by using a computer, human-computer interaction has better experience. Based on this, american psychologist Ekman et al proposed a Facial motion Coding System (FACS) that elaborated and specified common Facial muscle motion units (AU) into codes in 1987, while they classified Facial expressions into 6 basic expressions: anger, disgust, surprise, happiness, sadness, fear. Most methods generally divide facial expression recognition into two steps: and extracting and classifying the features, extracting and analyzing the face image and acquiring potential features, and performing optimal classification by the classifier according to the acquired potential features. Feature extraction is a key step of expression recognition, directly influences the effect of expression recognition, and is divided into two types for the existing research of feature extraction: geometric features and textural features. With the advent of Convolutional Neural Networks (CNN), some scholars propose to use CNN features for expression recognition, extract face global feature iterative regression through labeling in advance, and obtain good results, but only face global features ignore local detail features, thereby causing difficulty in distinguishing similar expressions, for example: surprise and fear. Routing adopts face key point positioning and extracts LBP characteristics of a proper amount of face Salient blocks (SFP) around the key points; kaleekal et al extracted Hahn features of 8 SFPs around 68 feature points on the face and cascaded using SVM classification with the highest average accuracy of 91.33% and 93.16% in the CK + and JAFFE data sets. The SFP method highlights the expression local features, however, the existing research method has no basis for SFP selection, and meanwhile, the different proportions of the different expression local features are ignored. According to the study of Li and the like, 6 types of facial basic expressions all have corresponding facial muscle movement units AU in FACS coding, the 6 types of facial expression AU are classified finely, and the expressions are classified through a subsequent Bayesian algorithm by identifying and counting AU, however, the method needs to identify a large number of AUs and has low accuracy in identifying a single AU.
Disclosure of Invention
The invention relates to a facial expression recognition method based on local representation, which realizes facial expression recognition of neutrality, anger, disgust, surprise, happiness, sadness and fear, and the specific technical scheme comprises the following 4 parts.
(1) Face partitioning: the 68 characteristic points of the human face divide the whole facial image into 43 sub-regions, and the 6 types of basic expressions of the face have corresponding single or combined facial muscle movement units.
(2) A convolutional neural network: the convolutional neural network features are used for expression recognition, and face global feature iterative regression is extracted through calibrating labels in advance. The 8 local candidate regions covered by the muscle motor region and the facial salient organs are selected as the input of the convolutional neural network.
(3) Feature extraction: to equalize the features of the local candidate regions, 8 parallel feature extraction branches are employed, each branch dominating a different dimensional fully connected layer. The outputs of the branches are adaptively connected according to attention to highlight the importance of different local candidate regions.
(4) Classification by Softmax function: expressions are classified into seven categories of neutral, angry, disgust, surprise, happy, sad and fear by the Softmax function.
Compared with other face recognition methods, the invention has the following advantages: 1. based on the facial muscle movement unit, the local features are extracted in parallel and evenly, and are adaptively connected according to the attention, so that the accuracy rate of facial expression recognition can be greatly improved. 2. The method solves the disadvantages that facial features are not prominent and facial muscle movements are ignored by the SFP method. 3. And the accuracy of face recognition can be effectively improved by adopting AU partitions.
Drawings
FIG. 1 is a general flowchart of a facial expression recognition method based on local characterization according to the present invention
FIG. 2 is a facial region partition
FIG. 3 is a schematic view of a minimum rectangular area
FIG. 4 CK + data set EAU-CNN algorithm effect graph
FIG. 5 JAFFE data set EAU-CNN algorithm effect graph
Detailed Description
The present invention is directed to a method for recognizing facial expressions based on local representations, and in order to make the technical solutions and effects of the present invention clearer and clearer, specific embodiments of the present invention are described in detail below with reference to the accompanying drawings.
As shown in FIG. 1, the present invention relates to a flow chart of a facial expression recognition method based on local representation. To extract the local features of the face, the text first divides the whole face image into 43 sub-regions according to the 68 feature points of the acquired human face, and then selects the muscle movement region and 8 local candidate regions covered by the facial salient organs as the input of the convolutional neural network. To equalize the features of the local candidate regions, 8 parallel feature extraction branches are employed, each branch dominating a different dimensional fully connected layer. The outputs of the branches are adaptively connected according to attention to highlight the importance degree of different local candidate areas, and finally expressions are classified into seven categories of neutrality, anger, disgust, surprise, happiness, sadness and fear through a Softmax function.
1. Face partition
Facial muscle movements form Facial expressions, which are classified into 45 types in the prior art, and a Facial motion Coding System (FACS) can constitute various expressions. Facial 6-type basic expressions each have a corresponding single or combined facial muscle movement single FACS descriptor. When angry, eyebrows will wrinkle together, vertical wrinkles appear between eyebrows, lower eyelid is tense and is lifted or not lifted, and the like, and the face isMuscle motor units are represented by one or more of AU4, AU5, AU7, AU23, AU24 respectively. When the eyebrow is pressed down, the upper lip is lifted, and the lower part of the lower eyelid appears with striation, the facial muscle movement unit is expressed as AU9 and AU 17. When the eyebrows are wrinkled and raised during fear, the upper eyelid is raised, the lower eyelid is tensed, the lips are slightly tense, and the like, facial muscle movement units are expressed as AU4, AU1+ AU5, AU5+ AU7 respectively. In the elevation, eyebrows may be bent downwards, lower edges of lower eyelids may be bulged or wrinkled, corners of the mouth may be pulled backwards and raised, teeth may be exposed, and the like, and the facial muscle movement units are expressed as AU6, AU12 and AU25, respectively. When sadness exists, the internal angles of eyebrows are wrinkled together and raised, the corners of mouths are pulled down, the eyelids on the internal angles of eyes are raised, and the like, and the facial muscle movement units are respectively expressed as AU1, AU4, AU15 and AU 17. Surprisingly, when the eyebrows are raised and bent, the skin under the eyebrows is stretched, the eyes are opened wide, the eyelids fall when the upper eyelids are raised, the mouth is opened, the lips and the teeth are separated, and the like, facial muscle movement units are respectively represented as AU5, AU26, AU27 and AU1+ AU 2. According to the FACS definition, regional statistics are generated on facial muscle movement units with 6 basic facial expressions, and it is found that facial muscle movement generation regions are concentrated on eyebrows, partial forehead, eyes and lower half of the face, and for other parts of the face, such as forehead, temples, partial side faces and the like, facial muscle movement units do not appear in the regions, but the characteristics of the regions are too similar when different expressions are used, and the overall expression recognition accuracy is affected. Aiming at the problem, the existing method constructs a certain amount of SFP extraction features for expression recognition, however, FACS is not considered in SFP selection, and the extracted local facial features are not outstanding enough. For this purpose, the face is divided into 43 feature regions according to the generation regions of the 6-type basic expression muscle movement units, and 8 local candidate regions AUg are constructed according to the regions of the eyebrows, eyes, nose and mouth of the facial organs as shown in fig. 2iAnd i represents 1 to 8. Each candidate region contains a certain feature region and is responsible for extracting only the muscle motor unit features belonging to the group.
2. Face recognition
Based on 68 characteristic points of the human face and the partition, 8 local candidate regions are formed, and the image is intercepted by adopting the thought of a minimum rectangular frameEach AUg, as shown in FIG. 3iThe images are normalized to different fixed sizes, and after the characteristics are extracted through each parallel CNN network, the images are spliced into a 4096-size full-connection layer. The spliced full-connection layer is multiplied by different expression weight values to highlight different expression local features, and expressions are divided into 7 types through subsequent feature extraction and a Softmax function: neutral, angry, disgust, surprise, happy, sad, fear.
AUg1The method comprises splicing two regions of a left eye and a right eye into 45 x 90 pixels, extracting features through 3 convolution layers and pooling layers to generate a 64-channel feature map with a size of 3 x 9, and performing AUg-step extractioniDominates the output 129-dimensional fully connected layer. AUg2The image is normalized to 45 × 220 pixels, and a 64-channel 3 × 25 feature map is generated by similarly passing through 3 convolutional layers and pooling layers, and 290-dimensional fully-connected layers are output according to the area proportion of the local candidate region. AUg3The image is normalized to 110 × 220 pixels, a 64-channel 5 × 11-size feature map is generated through 5 convolutional layers and 4 pooling layers, and 716-dimensional fully-connected layers are output according to the area proportion. AUg4The image is normalized to 140 × 170 pixels, passes through 5 convolutional layers and 4 pooling layers to generate a 64-channel 6 × 8 size feature map, and outputs 704-dimensional fully-connected layers according to area ratio. AUg5The image is normalized to 120 × 220 pixels, a 64-channel 5 × 11 feature map is generated through 5 convolutional layers and 4 pooling layers, and 782-dimensional fully-connected layers are output according to the area proportion. AUg6The image is normalized to 110 x 220 pixels, and is subjected to 5 convolution layers and 4 pooling layers to generate a 64-channel 5 x 11 feature map, and a 717-dimensional full-connected layer is output according to an area ratio. AUg7The image is normalized to 80 × 190 pixels, a 64-channel 3 × 10 feature map is generated through 5 convolutional layers and 4 pooling layers, and 451-dimensional fully-connected layers are output according to the area ratio. AUg8The image was normalized to 80 × 130 pixels, passed through 3 convolutional and pooling layers to generate a 64-channel 8 × 14 feature map, dominated by area scale 758 dimensional fully connected layers. And outputting and splicing all the candidate areas to obtain a 4096-dimensional full-connected layer.
3. Loss function design
Is a protrusionLocal features of different expressions are obtained, and the spliced 4096-dimensional full-connected layer is multiplied by different expression weight values WfjAnd j represents 1 to 7 different expressions. Based on the characteristic value XiSum weight value WfjNormalizing to obtain cos thetajFor each class as WfjXi. ComputingAnd obtaining a characteristic value XiAnd true weight WfjThe angle therebetween. In fact, WfjA hub is provided for each class. Then, at the target (ground truth) anglePlus an angular margin penalty m. Calculate cos (θ)yi+ m) and all logits are multiplied by the characteristic scale s and then logarithmically passed the softmax function and resulted in cross-entropy loss. Applying margin to the angle, the loss function is shown in equation 1:
whether a cosine margin should be used depends on the similarity measure (or distance) that the final loss function is optimizing. Obviously, the modified softmax loss function is to optimize cosine similarity, not angle. This may not be a problem if we use the traditional softmax penalty, since the decision boundaries of both forms are the same (cos θ @)1=cosθ2→θ1=θ2). However, when we try to push the boundary, we will face the problem that these two similarities (distances) have different densities. When the angle is close to 0 or n, the cosine values are denser. If we want to optimize the angle, we get the inner product WTThe value of f may then need to be inverted cosine. It may be more computationally expensive. In general, the angle margin is conceptually better than the cosine margin, but the cosine margin is more attractive in view of the computational cost because it can achieve the same goal with less effort. Intra-class distance and classMathematical description of the inter-distance, as shown in equations 2 and 3:
w is the parameter to be learned, feature XiIs also obtained through weight learning of the front layer, and X is obtained in the training processiAnd W will both change, both being driven by the gradient in the direction of decreasing Loss. The introduction of margin intentionally lowers the value of the component corresponding to the class label, the potential of 'squeezing' the model is reduced as much as possible, the position which can be originally converged in softmax needs to be continuously lowered, and the lowering can be realized by improving the value of the component corresponding to the class label or reducing the values of other components. XiIn the direction ofAt the same time as the approach is made,or may be far away from XiMay be moved in the direction of (c), the effect ultimately achieved may be XiAs close as possible toAnd Wj,j≠yiFar away from
To verify AUgiAnd if the selection is reasonable, selecting another 2 groups of local candidate regions for comparison and verification, wherein the selected regions are shown in the following table 1. To test the accuracy of the method, CK + and JAFFE data sets were used, the effect graphs of which are shown in fig. 4 and 5. The average accuracy of the CK + data set reaches 99.85 percent, and is 6.83 percent higher and 9.03 percent higher than that of the selected characteristic region 1 and the characteristic region 2Precision; the average accuracy of the JAFFE data set reaches 96.61%, and compared with the selected characteristic region 1 and the characteristic region 2, the accuracy is 11.87% and 18.65%.
Experiments show that the method has higher accuracy for expression recognition, overcomes the defects that the overall facial features are not prominent and facial muscle movement is neglected by an SFP (small form-factor pluggable) method, adopts AU (AU) partition, enables the recognition accuracy to be higher, has higher processing speed and can better meet the real-time monitoring requirement.
Claims (6)
1. A facial expression recognition method based on local characterization, comprising:
step 1: the face is divided, 68 characteristic points of the human face divide the whole face image into 43 sub-regions, and 6 basic expressions of the face have corresponding single or combined facial muscle movement units.
Step 2: the 8 local candidate regions covered by the muscle motor region and the facial salient organs are selected as the input of the convolutional neural network.
And step 3: and (3) feature extraction, wherein 8 parallel feature extraction branches are adopted, and each branch governs a different-dimension full-connection layer. The outputs of the branches are adaptively connected according to attention to highlight the importance of different local candidate regions.
And 4, step 4: and (4) feature classification, namely classifying expressions into seven classes of neutrality, anger, disgust, surprise, happiness, sadness and fear through a Softmax function.
And 5: a loss function for highlighting local features of different expressions, and multiplying the spliced 4096-dimensional full-connected layer by different expression weight values Wfj。
2. The method for recognizing Facial expressions based on local characteristics according to claim 1, wherein Facial muscle movements form Facial expressions, which are divided into 45 types by the existing practice, and a Facial motion Coding System (FACS) can constitute a plurality of expressions. Facial 6-type basic expressions each have a corresponding single or combined facial muscle movement single FACS descriptor. The regional statistics are generated on facial muscle movement units with 6 types of basic facial expressions, the facial muscle movement generation regions are found to be concentrated on eyebrows, partial forehead, eyes and lower half parts of the face, for other parts of the face, such as the forehead, the temples, partial side faces and the like, the facial muscle movement units do not appear in the regions, however, the features of the regions are too similar when the facial muscle movement units are in different expressions, and the overall expression recognition accuracy is affected. Aiming at the problem, the existing method constructs a certain amount of SFP extraction features for expression recognition, however, FACS is not considered in SFP selection, and the extracted local facial features are not outstanding enough. For this purpose, the face is divided into 43 feature regions according to the 6-type basic expression muscle movement unit generation regions, and 8 local candidate regions are constructed according to the regions where the eyebrows, eyes, nose and mouth of facial organs are located at the same time, wherein i represents 1 to 8. Each candidate region contains a certain feature region and is responsible for extracting only the muscle motor unit features belonging to the group.
3. The local token-based facial expression recognition method of claim 2, wherein: the 8 local candidate regions covered by the muscle motor region and the facial salient organs are selected and input into the convolutional neural network. Convolutional layers are the core of CNN and are used to convolve the input layers to extract higher level features.
4. The local token-based facial expression recognition method of claim 3, wherein: based on 68 characteristic points of the human face and the above partitions, 8 local candidate regions are formed, and the image is intercepted by adopting the idea of minimum rectangular box, such as AUg eachiThe images are normalized to different fixed sizes, and after the characteristics are extracted through each parallel CNN network, the images are spliced into a 4096-size full-connection layer. And multiplying the spliced full connecting layer by different expression weight values to highlight different expression local features.
5. The local token-based facial expression recognition method of claim 4, wherein the feature classification comprises: the expressions are divided into seven categories of neutrality, anger, disgust, surprise, happiness, sadness and fear by using a Softmax function, the prediction result of the model is converted into an exponential function in the first step by using Softmax, the nonnegativity of the probability is ensured, and the converted result is normalized.
6. The local token-based facial expression recognition method of claim 5, wherein: the loss function comprises the following steps:
in order to highlight local features of different expressions, the spliced 4096-dimensional full-connected layer is multiplied by different expression weight values WfjAnd j represents 1 to 7 different expressions. Based on the characteristic value XiSum weight value WfjNormalizing to obtain cos thetajFor each class as WfjXi. ComputingAnd obtaining a characteristic value XiAnd true weight WfjThe angle therebetween. In fact, WfjA hub is provided for each class. Then, at the target (ground truth) anglePlus an angular margin penalty m. Calculate cos (θ)yi+ m) and all logits are multiplied by the characteristic scale s and then logarithmically passed the softmax function and resulted in cross-entropy loss. Applying margin to the angle, the loss function is shown in equation 1:
whether a cosine margin should be used depends on the similarity measure (or distance) that the final loss function is optimizing. Obviously, the modified softmax loss function is to optimize cosine similarity, not angle. This may not be a problem if we use the traditional softmax penalty, since the decision boundaries of both forms are the same (cos θ @)1=cosθ2→θ1=θ2). However, when we try to push the boundary, we will face the problem that these two similarities (distances) have different densities. When the angle approaches0 or n, the cosine values are more dense. If we want to optimize the angle, we get the inner product WTThe value of f may then need to be inverted cosine. It may be more computationally expensive. In general, the angle margin is conceptually better than the cosine margin, but the cosine margin is more attractive in view of the computational cost because it can achieve the same goal with less effort. The mathematical description of the intra-class distance and the inter-class distance is shown in equations 2 and 3:
w is the parameter to be learned, feature XiIs also obtained through weight learning of the front layer, and X is obtained in the training processiAnd W will both change, both being driven by the gradient in the direction of decreasing Loss. The introduction of margin intentionally lowers the value of the component corresponding to the class label, the potential of 'squeezing' the model is reduced as much as possible, the position which can be originally converged in softmax needs to be continuously lowered, and the lowering can be realized by improving the value of the component corresponding to the class label or reducing the values of other components. XiIn the direction ofAt the same time as the approach is made,or may be far away from XiMay be moved in the direction of (c), the effect ultimately achieved may be XiAs close as possible toAnd Wj,j≠yiFar away from
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110670264.8A CN113792572B (en) | 2021-06-17 | 2021-06-17 | Facial expression recognition method based on local characterization |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110670264.8A CN113792572B (en) | 2021-06-17 | 2021-06-17 | Facial expression recognition method based on local characterization |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113792572A true CN113792572A (en) | 2021-12-14 |
CN113792572B CN113792572B (en) | 2024-08-02 |
Family
ID=78876952
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110670264.8A Active CN113792572B (en) | 2021-06-17 | 2021-06-17 | Facial expression recognition method based on local characterization |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113792572B (en) |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2009251634A (en) * | 2008-04-01 | 2009-10-29 | Seiko Epson Corp | Image processor, image processing method, and program |
CN107491726A (en) * | 2017-07-04 | 2017-12-19 | 重庆邮电大学 | A kind of real-time expression recognition method based on multi-channel parallel convolutional neural networks |
CN109271930A (en) * | 2018-09-14 | 2019-01-25 | 广州杰赛科技股份有限公司 | Micro- expression recognition method, device and storage medium |
CN110188615A (en) * | 2019-04-30 | 2019-08-30 | 中国科学院计算技术研究所 | A kind of facial expression recognizing method, device, medium and system |
CN111626113A (en) * | 2020-04-20 | 2020-09-04 | 北京市西城区培智中心学校 | Facial expression recognition method and device based on facial action unit |
WO2020252903A1 (en) * | 2019-06-18 | 2020-12-24 | 平安科技(深圳)有限公司 | Au detection method and apparatus, electronic device, and storage medium |
CN112668378A (en) * | 2019-10-16 | 2021-04-16 | 哈尔滨理工大学 | Facial expression recognition method based on combination of image fusion and convolutional neural network |
-
2021
- 2021-06-17 CN CN202110670264.8A patent/CN113792572B/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2009251634A (en) * | 2008-04-01 | 2009-10-29 | Seiko Epson Corp | Image processor, image processing method, and program |
CN107491726A (en) * | 2017-07-04 | 2017-12-19 | 重庆邮电大学 | A kind of real-time expression recognition method based on multi-channel parallel convolutional neural networks |
CN109271930A (en) * | 2018-09-14 | 2019-01-25 | 广州杰赛科技股份有限公司 | Micro- expression recognition method, device and storage medium |
CN110188615A (en) * | 2019-04-30 | 2019-08-30 | 中国科学院计算技术研究所 | A kind of facial expression recognizing method, device, medium and system |
WO2020252903A1 (en) * | 2019-06-18 | 2020-12-24 | 平安科技(深圳)有限公司 | Au detection method and apparatus, electronic device, and storage medium |
CN112668378A (en) * | 2019-10-16 | 2021-04-16 | 哈尔滨理工大学 | Facial expression recognition method based on combination of image fusion and convolutional neural network |
CN111626113A (en) * | 2020-04-20 | 2020-09-04 | 北京市西城区培智中心学校 | Facial expression recognition method and device based on facial action unit |
Non-Patent Citations (6)
Title |
---|
JASON C. DESKA等: "The Face of Fear and Anger: Facial Width-to-Height Ratio Biases Recognition of Angry and Fearful Expressions", 《JOURNAL INFORMATION JOURNAL TOC》, vol. 18, no. 4, 1 April 2018 (2018-04-01), pages 453 - 464 * |
YONGQIANG LI等: "Simultaneous Facial Feature Tracking and Facial Expression Recognition", IEEE TRANSACTIONS ON IMAGE PROCESSING, vol. 22, no. 7, pages 2559 - 2573, XP011510290, DOI: 10.1109/TIP.2013.2253477 * |
日拱一卒: "人脸识别损失函数疏理与分析", 《HTTPS://WWW.CNBLOGS.COM/SHINE-LEE/P/13435621.HTML#ARCFACE-LOSS---CVPR2019》, 4 August 2020 (2020-08-04), pages 8 - 10 * |
李帅: "基于卷积神经网络的人脸表情识别研究", 《万方数据》, 4 February 2021 (2021-02-04), pages 3 - 5 * |
李辉等: "基于卷积神经网络的人脸识别算法", 《软件导刊》, vol. 16, no. 3, 15 March 2017 (2017-03-15), pages 26 - 29 * |
苏志明等: "基于角度距离损失与小尺度核网络的表情识别", 《电讯技术》, vol. 64, no. 4, pages 1 * |
Also Published As
Publication number | Publication date |
---|---|
CN113792572B (en) | 2024-08-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Sun et al. | A visual attention based ROI detection method for facial expression recognition | |
Wen et al. | Ensemble of deep neural networks with probability-based fusion for facial expression recognition | |
Zhang et al. | Adaptive facial point detection and emotion recognition for a humanoid robot | |
CN112800903B (en) | Dynamic expression recognition method and system based on space-time diagram convolutional neural network | |
Ng et al. | A review of facial gender recognition | |
Shan | Learning local features for age estimation on real-life faces | |
CN111523462A (en) | Video sequence list situation recognition system and method based on self-attention enhanced CNN | |
CN110188708A (en) | A kind of facial expression recognizing method based on convolutional neural networks | |
Liu et al. | A 3 GAN: an attribute-aware attentive generative adversarial network for face aging | |
Huang et al. | Multimodal framework for analyzing the affect of a group of people | |
Ali et al. | Facial emotion detection using neural network | |
CN108830237A (en) | A kind of recognition methods of human face expression | |
Wang | Effect of subject's age and gender on face recognition results | |
Fan et al. | Hierarchical scale convolutional neural network for facial expression recognition | |
Prabhu et al. | Facial Expression Recognition Using Enhanced Convolution Neural Network with Attention Mechanism. | |
Ullah et al. | Improved deep CNN-based two stream super resolution and hybrid deep model-based facial emotion recognition | |
Upadhyay et al. | A review on different facial feature extraction methods for face emotions recognition system | |
CN113743389B (en) | Facial expression recognition method and device and electronic equipment | |
Tautkutė et al. | Classifying and visualizing emotions with emotional DAN | |
Ullah et al. | Emotion recognition from occluded facial images using deep ensemble model | |
Albrici et al. | G2-VER: Geometry guided model ensemble for video-based facial expression recognition | |
Alrabiah et al. | Computer-based approach to detect wrinkles and suggest facial fillers | |
Guo et al. | Facial expression recognition: a review | |
Dimlo et al. | Innovative method for face emotion recognition using hybrid deep neural networks | |
Aslam et al. | Gender classification based on isolated facial features and foggy faces using jointly trained deep convolutional neural network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |