CN115393944A - Micro-expression identification method based on multi-dimensional feature fusion - Google Patents

Micro-expression identification method based on multi-dimensional feature fusion Download PDF

Info

Publication number
CN115393944A
CN115393944A CN202211123830.4A CN202211123830A CN115393944A CN 115393944 A CN115393944 A CN 115393944A CN 202211123830 A CN202211123830 A CN 202211123830A CN 115393944 A CN115393944 A CN 115393944A
Authority
CN
China
Prior art keywords
micro
layer
feature fusion
features
expression
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202211123830.4A
Other languages
Chinese (zh)
Inventor
张家波
徐光辉
甘海洋
黄钟玉
高洁
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing University of Post and Telecommunications
Original Assignee
Chongqing University of Post and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing University of Post and Telecommunications filed Critical Chongqing University of Post and Telecommunications
Priority to CN202211123830.4A priority Critical patent/CN115393944A/en
Publication of CN115393944A publication Critical patent/CN115393944A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/174Facial expression recognition
    • G06V40/175Static expression
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/80Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
    • G06V10/806Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • Databases & Information Systems (AREA)
  • Software Systems (AREA)
  • Medical Informatics (AREA)
  • Artificial Intelligence (AREA)
  • Computing Systems (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a micro-expression recognition method based on multi-dimensional feature fusion, which comprises four steps of image preprocessing, extraction of optical flow feature components, construction of a multi-dimensional feature fusion network, and classification of micro-expressions by using a model obtained through training. In the invention, in order to compensate the facial detail information lost by the model, a feature fusion module is constructed, and the shallow feature extracted by the feature fusion module is fused with the abstract feature extracted by the double-current convolution network and is commonly used for model classification. For the extracted high-dimensional fusion features, different weights are given to the channels through the channel attention module, so that the model focuses more on the channels with high contribution degree, and the precision of micro-expression recognition is further improved.

Description

Micro-expression identification method based on multi-dimensional feature fusion
Technical Field
The invention relates to the technical field of expression recognition, in particular to a micro expression recognition method based on multi-dimensional feature fusion.
Background
Micro-expressions are special expressions of very short duration and very low expression intensity, and usually occur involuntarily when people attempt to mask their own true mind. Since the time that a micro-expression stays on the face is typically only 1/25 second to 1/3 second, it is difficult for most people to recognize its appearance. This particular expression is thought to be related to the human self-defense mechanism, reflecting the real idea of the mind of the human. Accurate identification of micro-expressions helps people make appropriate judgments and decisions, and thus micro-expression identification is extremely important. The micro-expression recognition is essentially an image classification problem, and in recent years, with the development of deep learning, people begin to perform micro-expression analysis recognition by means of computers, and the micro-expression recognition method has the advantages that as long as an accurate and stable model is trained, a large-scale micro-expression recognition task can be automatically and efficiently processed at low cost, and manpower and material resources are saved.
Document 1, gan y.s., liong s.t., yau w.c., et al. Off-extranet on micro-expression recognition system [ J ]. Signal Processing: image Communication,2019,74,. On the basis, the Genetic Algorithm (Genetic Algorithm, GA) is introduced into the model learning characteristics of the [ Genetic Algorithm, GA ] 2[ jin Qiushi, xu Huanggchao, liu Kunhong, et al, ga-apexnet: genetic Algorithm in apex frame network for micro-expression recognition system [ C ]// Proceedings of the Journal of Physics: reference series.Suzhou: IOP Press,2020,1544 (1): 012149.], the Genetic Algorithm, GA) reduces the dimensions of the model learning characteristics, only retains the characteristics beneficial to classification, and finally, the precision of the tested model is further improved. Document 3[ Nie xuan, takalkar M.A., duan Mengyang, et al.GEME Dual-stream multi-task generator-based micro-expression recognition [ J ]. Neuro expression, 2021,427 ] ], document 4[ Zhou Ling, mao Qirong, xue Luoyang. Dual-input network for cross-database micro-expression recognition [ C ]// Proceedings of the 2019 11 IEEE International Conference Automatic facility and Gestregistration. Lille. In contrast, document [3] introduces a multitask learning method, which assists in model classification by detecting gender through a network leg. While the overall idea of the document [4] is the same as that of the document [1], optical flow characteristic components of a start frame and a vertex frame are extracted as input, but the document [4] uses an initiation block module when constructing a model, namely, convolution kernels of various sizes are used in parallel on the same layer.
The existing models are not sufficient in attention to face detail information, and meanwhile, the micro expression recognition effect needs to be improved.
Disclosure of Invention
The invention discloses a micro-expression recognition method based on multi-dimensional feature fusion, and aims to solve the problem that an existing model is not enough in attention to face detail information. In order to compensate face detail information lost by the model, a feature fusion module is constructed, and the shallow features extracted by the feature fusion module are fused with the abstract features extracted by the double-current convolution network and are used for model classification. For the extracted high-dimensional fusion features, different weights are given to the channels through the channel attention module, so that the model focuses more on the channels with high contribution degree, and the precision of micro-expression recognition is further improved.
In view of this, the technical scheme adopted by the invention is as follows: a micro expression identification method based on multi-dimensional feature fusion is characterized by comprising the following steps:
the method comprises the steps of firstly, image preprocessing, including image graying, face key point detection, face alignment cutting and scale normalization.
And step two, extracting an optical flow horizontal component u and an optical flow vertical component v between the initial frame and the vertex frame by using an iterative Lucas-Kanade algorithm.
Step three, constructing a multi-dimensional feature fusion network, which specifically comprises the following steps: taking an optical flow horizontal component u and an optical flow vertical component v as the input of a double-current convolutional neural network, wherein the backbone network adopts a symmetrical structure, convolution layers are used in a first layer, a third layer and a fourth layer and are used for extracting texture information and edge features of optical flow features, and maximum pooling layers are used in a second layer and a fifth layer and are used for down-sampling input feature dimensions; the output of the first layer convolution layer in the backbone network is subjected to feature fusion through a feature fusion module, and then is fused with the maximum pooling layer output feature of the fifth layer of the two branches again to obtain the output containing the facial detail information and the abstract feature; after the multi-dimensional features are fused, a channel attention module is introduced to endow different channels with different importance, and useful features for model classification and discrimination are highlighted; two fully-connected layers are then introduced, and features output from the fully-connected layers are passed to the output layer, sorted by softmax function.
And step four, classifying the micro-expressions by using the model obtained by training.
The invention has the following beneficial technical effects:
according to the invention, the horizontal component and the vertical component of the optical flow are used as the input of the model, the multi-dimensional feature fusion module is constructed, the extracted shallow feature of the facial details is fused with the abstract feature extracted by the double-flow convolution network, the channel attention module is introduced, and the recognition effect is good. Specifically, the method comprises the following steps:
(1) And (3) obtaining optical flow characteristic components by using a fast and robust iterative Lucas-Kanade (iLK) algorithm, wherein the optical flow characteristics can better highlight the slight movement of the face.
(2) According to the model of the double-current convolutional neural network, the horizontal component of the optical flow and the vertical component of the optical flow are learned and fused respectively, and the built network model is high in identification precision.
(3) The shallow layer features extracted by the feature fusion module are fused with the abstract features extracted by the double-current convolution network and are jointly used for model classification, so that the features are richer.
(4) And a channel attention mechanism is introduced at a proper position, different channels are endowed with different importance, and the characteristics useful for model classification and discrimination are highlighted.
Drawings
FIG. 1 is a multi-dimensional feature fusion model;
FIG. 2 is a feature fusion Module FFM;
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, embodiments of the present invention will be described in further detail below.
A micro-expression recognition method based on multi-dimensional feature fusion comprises four steps of image preprocessing, optical flow feature component extraction, multi-dimensional feature fusion network construction and micro-expression classification by using a model obtained through training.
Step one, image preprocessing
Compared with common expressions, micro-expressions have three remarkable characteristics, namely short duration, low intensity and local motion. Aiming at the characteristics of the micro expression, the image preprocessing is particularly important for micro expression recognition, and specifically comprises image graying, face key point detection, face alignment cutting and scale normalization.
Because the number of micro expression samples is small, if a color image is used in an experiment, a series of problems such as difficulty in training and the like can be caused, and therefore the gray processing is carried out on the original color image by using the formula (1).
Gray(x,y)=0.299×R(x,y)+0.587×G(x,y)+0.114×B(x,y) (1)
Wherein, gray is the pixel Gray value, x and y represent the spatial position index of the target pixel point, and R, G and B are the values of three channels of the RGB image.
The original image in the micro-expression dataset contains a lot of background image noise, including laboratory background walls and the participants 'headphones, and only the subject's facial region helps the model to recognize the micro-expression, so the invention aligns and cuts the facial region.
Before the facial region is cut, the invention adopts a method based on cascade shape regression to detect the key points of the face in the micro expression data set. The method based on the cascade shape regression can conveniently perform data enhancement by adjusting the initialized shape, and has the advantages of high calculation efficiency, strong universality and the like.
According to the key points of the face extracted by positioning, the distance d between two eyes and the included angle theta between the central coordinate connecting line of the left eye and the right eye and the horizontal direction are calculated by using a formula (2) and a formula (3).
Figure BDA0003847520010000031
Figure BDA0003847520010000032
In the formula, x 1 ,y 1 Is the inner canthus coordinate of the left eye, x 2 ,y 2 Is the coordinate of the eye corner in the right eye, d is the distance between the two eyes, and theta is the included angle between the connecting line of the left eye and the right eye and the horizontal direction.
Then calculating the midpoint (x) of the distance between the two eyes by the coordinates of the two eyes in the formula (4) 0 ,y 0 ),
Figure BDA0003847520010000033
And taking the center of the eye distance as a rotation center, and calculating the positions of the key points of the human face after rotation according to a formula (5):
Figure BDA0003847520010000034
wherein, x and y are indexes of spatial positions of the target pixels before adjustment, and x 'and y' are indexes of spatial positions of the target pixels after adjustment. After face alignment, a face cropping operation is performed, typically cropping the face area of the participant into a rectangle.
And finally, carrying out scale normalization. In the invention, all the micro expression images are uniformly adjusted to be 112 multiplied by 112 in the image preprocessing stage through the formula (6).
Figure BDA0003847520010000041
Wherein g ' is the pixel value after the image adjustment, g is the pixel value before the image adjustment, w, h is the width and height before the image adjustment, and w ', h ' is the width and height after the image adjustment.
Step two, extracting characteristic components of the optical flow
The optical flow can extract representative motion characteristics, and after image preprocessing is completed in the first step, the invention extracts an optical flow horizontal component u and an optical flow vertical component v between a start frame and a vertex frame by using an iterative Lucas-Kanade (iLK) algorithm. Wherein, the initial frame is the moment of micro expression, the vertex frame is the moment of maximum micro expression, and is a frame with most abundant face information.
The main idea of the optical flow method is to calculate the change of the same pixel point in time through two adjacent frames of images, so as to obtain the motion change of the same pixel point between the two adjacent frames of images, therefore, the optical flow method is often used for describing a moving object in image processing. An iterative Lucas-Kanade (iLK) solver is applied to each level of the image pyramid. iLK is a fast and robust alternative to the TVL1 algorithm.
Step three, constructing a multi-dimensional feature fusion model
As shown in FIG. 1, the multi-dimensional feature fusion model takes an optical flow horizontal component u and an optical flow vertical component v as inputs of a double-current convolution neural network, a main part adopts a symmetrical structure, convolution layers are used in a first layer, a third layer and a fourth layer and used for extracting texture information and edge features of optical flow features, all convolution kernels in the model are set to be 5 x 5, and the step length is set to be 1. The number of kernels of the convolutional layer is 64, 128 and 128, respectively, and the Relu function is adopted as the activation function. The second and fifth layers of the backbone network use the largest pooling layer, and downsample the input feature dimensions to reduce the parameters by half. The window size of the pooling layer is 2 x 2 with the step size set to 1.
The output of the first layer convolutional layer in the backbone network is subjected to Feature Fusion through a Feature Fusion Module (FFM), and then is fused with the maximum pooling layer output features of the 5 th layer of the two branches again to obtain the output containing the facial detail information and the abstract features, so that the output is used for the final classification and judgment of the model.
After the multi-dimensional feature fusion, the feature dimension of the output is relatively high, wherein the feature graph which does not contribute much to the classification interferes with the judgment of the model and influences the final classification precision. The invention introduces a channel attention module (SENET) to solve the problem, endows different channels with different importance and highlights the characteristics useful for model classification and discrimination. Two fully connected layers were then introduced to reduce the amount of parameters in the model to prevent overfitting. The number of nodes per layer is 256. Features output from the fully connected layer are passed to the output layer, sorted by the softmax function. The detailed parameters of the model are shown in Table 1.
Specifically, the probability of being classified as the i-th expression is:
Figure BDA0003847520010000042
in the formula, p, x, L, L and y are respectively a model prediction label, an input image, a predicted emotion label, an emotion category total number and a real label. x is the number of j ,x n Respectively representing the output values of j and n nodes.
The invention adopts an Adam optimizer to update and optimize model parameters, and the Loss function adopts a classical cross entropy Loss function (Cross entropy Loss, CELoss), and the calculation formula is as follows:
Figure BDA0003847520010000051
wherein y is i Is a label and y i ∈{0,1},p i Is a predicted value of the model output.
TABLE 1 model parameter Table
Figure BDA0003847520010000052
The FFM module is described in detail below.
The detailed structure of the FFM designed by the invention is shown in figure 2.
In fig. 2, conv denotes a convolutional layer, and M denotes a max pooling layer. Suppose the input is characterized by x m I.e. of the network m The layer output, including the characteristics of c channels, then the input characteristics can be expressed as
Figure BDA0003847520010000056
Before fusion, one-dimensional convolution operation is carried out on two features from a shallow layer to make the dimensions of the two features uniform, and the transformed features can be expressed as follows:
Figure BDA0003847520010000053
in the formula, x mm′ The input feature and the output feature are used as input features, c is a channel of the input feature and the output feature, w is convolution kernel weight, and b is a bias term.
Figure BDA0003847520010000054
Representing the input characteristics of a certain channel.
After one-dimensional convolution operation, the two feature maps are superimposed on the basis of the channel, i.e.
Figure BDA0003847520010000055
χ n′ Showing the output of the FFM from the lower arm through the convolutional layer (see figure 2 for details).
And then activating the fused high-dimensional features by using a Relu function and inputting the features into a maximum pooling layer, reducing feature dimensions and extracting more representative features.
Step four, using the trained model to classify the micro-expression
A Leave-One-Subject-Out (LOSO) method is adopted on a CASME II data set, namely samples of K participants exist in the data set, image data of One person is selected to serve as a test set, data of the remaining K-1 persons serve as a training set, and samples of each person are taken as a primary test set.
The deep learning frameworks used were TensorFlow 2.4.0 and Keras 2.4.3, a learning rate of 0.0001, a batch size of 32, and a training round of 100.
Unweighted F1-score (Unweighted F1-score, UF 1) values and Unweighted Average Recall (UAR) values of test results were calculated by formulas (11) - (14):
Figure BDA0003847520010000061
wherein:
Figure BDA0003847520010000062
Figure BDA0003847520010000063
wherein:
Figure BDA0003847520010000064
wherein C is the total number of emotion categories, TP i 、TN i 、FP i 、FN i The numbers of true positive, true negative, false positive and false negative of the i-th category are indicated respectively.
Compared with the current method, the leave-one-out cross validation method (LOSO) is adopted on the CASME II micro-expression data set, the identification effect is good, and UF1 and UAR are compared as shown in the following table.
Method UF1 UAR
LBP-top[ 1 ] 0.7026 0.7429
Bi-WOOF[ 2 ] 0.7805 0.8026
OFF-Apex [3] 0.8764 0.8681
Ga-apexnet [4] 0.8634 0.9232
MERFGR[ 5 ] 0.871 0.8798
MJFN[ 6 ] 0.9151 0.8871
MERASTC[ 7 ] 0.854 0.862
MERSiamC3D[ 8 ] 0.8818 0.8763
This example 0.9406 0.9502
The references referred to in the table above are as follows:
[1]Zhao Guoying,Pietikainen M.Dynamic texture recognition using local binary patterns with an application to facial expressions[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2007,29(6):915-928.
[2]Liong S.T.,SeeJ.,Wong K.S.,et al.Less is more:Micro-expression recognition from video using apex frame[J].Signal Processing:Image Communication,2018,62:82-92.
[3]Gan Y.S.,Liong S.T.,Yau W.C.,et al.Off-apexnet on micro-expression recognition system[J]. Signal Processing:Image Communication,2019,74:129-139.
[4]Jin Qiushi,Xu Huangchao,Liu Kunhong,et al.Ga-apexnet:Genetic algorithm in apex frame network for micro-expression recognition system[C]//Proceedings of the Journal of Physics: Conference Series.Suzhou:IOP Press,2020,1544(1):012149.
[5]Lei Ling,Chen Tong,Li Shigang,et al.Micro-expression recognition based on facial graph representation learning and facial action unit fusion[OL]URL:[2021-6-25]. https://openaccess.thecvf.com/content/CVPR2021W/AUVi/papers/Lei_Micro-Expression_Rec ognition_Based_on_Facial_Graph_Representation_Learning_and_Facial_CVPRW_2021_paper .pdf.
[6]Li Xinyu,Wei Guangshun,Wang Jie,et al.Multi-scale joint feature network for micro-expression recognition[J].Computational Visual Media,2021,7(3):407-417.
[7]Gupta P.MERASTC:Micro-expression recognition using effective feature encodings and 2D convolutional neural network[J].IEEE Transactions on Affective Computing,doi: 10.1109/TAFFC.2021.3061967,2021.
[8]Zhao Sirui,Tao Hanqing,Zhang Yangsong,et al.A two-stage 3D CNN based learning method for spontaneous micro-expression recognition[J].Neurocomputing,2021,448:276-289。

Claims (7)

1. a micro expression identification method based on multi-dimensional feature fusion is characterized by comprising the following steps:
the method comprises the steps of firstly, image preprocessing, including image graying, face key point detection, face alignment cutting and scale normalization;
secondly, extracting an optical flow horizontal component u and an optical flow vertical component v between the initial frame and the vertex frame by using an iterative Lucas-Kanade algorithm;
step three, constructing a multi-dimensional feature fusion network;
and step four, carrying out micro-expression classification by using the model obtained by training.
2. The micro-expression recognition method based on multi-dimensional feature fusion as claimed in claim 1, wherein: firstly, the facial key point detection adopts a cascade shape regression-based method to detect the facial key points in the micro expression data set.
3. The micro-expression recognition method based on multi-dimensional feature fusion as claimed in claim 2, wherein: step one the face alignment cropping includes:
calculating the distance d between the two eyes and the included angle theta between the central coordinate connecting line of the left eye and the right eye and the horizontal direction by using a formula (1) and a formula (2);
Figure FDA0003847520000000011
Figure FDA0003847520000000012
in the formula, x 1 ,y 1 Is the inner canthus coordinate of the left eye, x 2 ,y 2 Is the coordinate of the eye corner in the right eye, d is the distance between the two eyes, and theta is the included angle between the connecting line of the left eye and the right eye and the horizontal direction;
then calculating the midpoint (x) of the distance between the two eyes by the coordinates of the two eyes in the formula (3) 0 ,y 0 ),
Figure FDA0003847520000000013
And taking the center of the eye distance as a rotation center, and calculating the positions of the key points of the human face after rotation according to a formula (4):
Figure FDA0003847520000000014
wherein, x and y are indexes of spatial positions of the target pixels before adjustment, and x 'and y' are indexes of spatial positions of the target pixels after adjustment; and after the face is aligned, performing face cropping operation.
4. The micro-expression recognition method based on multi-dimensional feature fusion according to claim 1, characterized in that: and step two, the initial frame is the moment when the micro expression appears, and the vertex frame is the moment when the micro expression amplitude is maximum.
5. The micro-expression recognition method based on multi-dimensional feature fusion according to claim 1, characterized in that: step three, the multidimensional feature fusion network specifically comprises:
taking an optical flow horizontal component u and an optical flow vertical component v as the input of a double-current convolutional neural network, wherein the backbone network adopts a symmetrical structure, convolution layers are used in a first layer, a third layer and a fourth layer and are used for extracting texture information and edge features of optical flow features, and maximum pooling layers are used in a second layer and a fifth layer and are used for down-sampling input feature dimensions; the output of the first layer convolution layer in the backbone network is subjected to feature fusion through a feature fusion module, and then is fused with the maximum pooling layer output feature of the fifth layer of the two branches again to obtain the output containing the facial detail information and the abstract feature; after the multi-dimensional features are fused, a channel attention module is introduced to endow different channels with different importance, and useful features for classifying and distinguishing the models are highlighted; two fully-connected layers are then introduced, and features output from the fully-connected layers are passed to the output layer, sorted by the softmax function.
6. The micro-expression recognition method based on multi-dimensional feature fusion as claimed in claim 5, wherein: the feature fusion module specifically comprises:
let the characteristic of the input be x m I.e. the m-th layer output of the network, contains the characteristics of c channels, the input characteristics can be expressed as
Figure FDA0003847520000000021
Before fusion, one-dimensional convolution operation is carried out on two characteristics from a shallow layer, so that the dimensions of the two characteristics are unified, and the converted characteristics are represented as follows:
Figure FDA0003847520000000022
in the formula, chi mm′ For the input features and the output features, c is the channel of the input features and the output features, w is the convolution kernel weight, b is the bias term,
Figure FDA0003847520000000023
representing the input characteristics of a certain channel,
after one-dimensional convolution operation, the two feature maps are superimposed on the basis of the channel, i.e.
Figure FDA0003847520000000024
χ n′ The output of the lower branch of the FFM module passing through the convolution layer is shown;
the fused high-dimensional features are then activated using the Relu function and input to the max pooling layer.
7. The micro-expression recognition method based on multi-dimensional feature fusion according to any one of claims 1-6, characterized in that: the micro expressions are classified, and the probability of the micro expressions classified into the ith expression is as follows:
Figure FDA0003847520000000025
in the formula, p, x, L, L and y are respectively a model prediction label, an input image, a predicted emotion label, an emotion category total number and a real label, and x j ,x n Respectively representing the output values of j and n nodes.
CN202211123830.4A 2022-09-15 2022-09-15 Micro-expression identification method based on multi-dimensional feature fusion Pending CN115393944A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211123830.4A CN115393944A (en) 2022-09-15 2022-09-15 Micro-expression identification method based on multi-dimensional feature fusion

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211123830.4A CN115393944A (en) 2022-09-15 2022-09-15 Micro-expression identification method based on multi-dimensional feature fusion

Publications (1)

Publication Number Publication Date
CN115393944A true CN115393944A (en) 2022-11-25

Family

ID=84126963

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211123830.4A Pending CN115393944A (en) 2022-09-15 2022-09-15 Micro-expression identification method based on multi-dimensional feature fusion

Country Status (1)

Country Link
CN (1) CN115393944A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116189311A (en) * 2023-04-27 2023-05-30 成都愚创科技有限公司 Protective clothing wears standardized flow monitoring system
CN116824280A (en) * 2023-08-30 2023-09-29 安徽爱学堂教育科技有限公司 Psychological early warning method based on micro-expression change

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116189311A (en) * 2023-04-27 2023-05-30 成都愚创科技有限公司 Protective clothing wears standardized flow monitoring system
CN116824280A (en) * 2023-08-30 2023-09-29 安徽爱学堂教育科技有限公司 Psychological early warning method based on micro-expression change
CN116824280B (en) * 2023-08-30 2023-11-24 安徽爱学堂教育科技有限公司 Psychological early warning method based on micro-expression change

Similar Documents

Publication Publication Date Title
Yuan et al. Fingerprint liveness detection using an improved CNN with image scale equalization
CN115393944A (en) Micro-expression identification method based on multi-dimensional feature fusion
Kumar et al. Automatic live facial expression detection using genetic algorithm with haar wavelet features and SVM
CN112766158A (en) Multi-task cascading type face shielding expression recognition method
Setiowati et al. A review of optimization method in face recognition: Comparison deep learning and non-deep learning methods
CN107918772B (en) Target tracking method based on compressed sensing theory and gcForest
CN111126307B (en) Small sample face recognition method combining sparse representation neural network
CN112434732A (en) Deep learning classification method based on feature screening
Liu et al. Micro-expression recognition using advanced genetic algorithm
CN111476178A (en) Micro-expression recognition method based on 2D-3D CNN
CN112613479B (en) Expression recognition method based on light-weight streaming network and attention mechanism
CN114360071A (en) Method for realizing off-line handwritten signature verification based on artificial intelligence
Fueten et al. An artificial neural net assisted approach to editing edges in petrographic images collected with the rotating polarizer stage
Shanthi et al. Algorithms for face recognition drones
Lee et al. Face and facial expressions recognition system for blind people using ResNet50 architecture and CNN
Sakthimohan et al. Detection and Recognition of Face Using Deep Learning
Ma et al. Noise-against skeleton extraction framework and application on hand gesture recognition
CN111160327B (en) Expression recognition method based on lightweight convolutional neural network
Tunc et al. Age group and gender classification using convolutional neural networks with a fuzzy logic-based filter method for noise reduction
Karilingappa et al. Human emotion detection and classification using modified Viola-Jones and convolution neural network
Yu et al. Research on face recognition method based on deep learning
Abayomi-Alli et al. Facial image quality assessment using an ensemble of pre-trained deep learning models (EFQnet)
CN115797827A (en) ViT human body behavior identification method based on double-current network architecture
CN114049675A (en) Facial expression recognition method based on light-weight two-channel neural network
CN113887509A (en) Rapid multi-modal video face recognition method based on image set

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination