CN116861211A - Electroencephalogram emotion recognition method and system integrating space-time interaction neural network - Google Patents

Electroencephalogram emotion recognition method and system integrating space-time interaction neural network Download PDF

Info

Publication number
CN116861211A
CN116861211A CN202310674320.4A CN202310674320A CN116861211A CN 116861211 A CN116861211 A CN 116861211A CN 202310674320 A CN202310674320 A CN 202310674320A CN 116861211 A CN116861211 A CN 116861211A
Authority
CN
China
Prior art keywords
time
emotion
space
electroencephalogram
emotion recognition
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202310674320.4A
Other languages
Chinese (zh)
Other versions
CN116861211B (en
Inventor
郭浩
陆轲熠
顾志浩
齐菲
孙承峰
孙立宁
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Suzhou University
Original Assignee
Suzhou University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Suzhou University filed Critical Suzhou University
Priority to CN202310674320.4A priority Critical patent/CN116861211B/en
Publication of CN116861211A publication Critical patent/CN116861211A/en
Application granted granted Critical
Publication of CN116861211B publication Critical patent/CN116861211B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/16Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
    • A61B5/165Evaluating the state of mind, e.g. depression, anxiety
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/24Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
    • A61B5/316Modalities, i.e. specific diagnostic methods
    • A61B5/369Electroencephalography [EEG]
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/24Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
    • A61B5/316Modalities, i.e. specific diagnostic methods
    • A61B5/369Electroencephalography [EEG]
    • A61B5/372Analysis of electroencephalograms
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Psychiatry (AREA)
  • Animal Behavior & Ethology (AREA)
  • Evolutionary Biology (AREA)
  • Medical Informatics (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Public Health (AREA)
  • Veterinary Medicine (AREA)
  • Pathology (AREA)
  • Surgery (AREA)
  • Psychology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computational Linguistics (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Social Psychology (AREA)
  • Hospice & Palliative Care (AREA)
  • Educational Technology (AREA)
  • Developmental Disabilities (AREA)
  • Child & Adolescent Psychology (AREA)
  • Measurement And Recording Of Electrical Phenomena And Electrical Characteristics Of The Living Body (AREA)

Abstract

The application relates to the technical field of emotion recognition, in particular to an electroencephalogram emotion recognition method integrating space-time interaction neural networks, which comprises the following steps: inputting electroencephalogram data into an electroencephalogram emotion recognition model, and constructing a 2D matrix representation of the electroencephalogram according to the position of an electroencephalogram acquisition electrode; capturing common characteristics of a plurality of continuous time steps from a low-dimensional layer and a high-dimensional layer through time step sharing convolution, and enhancing the information expression of the original input; by downsampling convolution, the time dimension is reduced and time-space features highly correlated to emotion categories are extracted; information interaction among a plurality of subspace regions is realized through a space interaction multi-layer perceptron, and the space dependence of the features is enhanced; and constructing emotion classification tasks, obtaining the probability of emotion categories in each task, and completing emotion recognition. The application is based on fewer electroencephalogram acquisition electrodes, enhances the time expression and potential space dependence of electroencephalogram signals, and further improves the classification precision and generalization capability of electroencephalogram emotion recognition.

Description

Electroencephalogram emotion recognition method and system integrating space-time interaction neural network
Technical Field
The application relates to the technical field of emotion recognition, in particular to an electroencephalogram emotion recognition method and system integrating space-time interaction neural networks.
Background
Emotion is one of the highest levels of cognitive activity in humans, directly affecting aspects of public daily life, including learning, decision making, and social interactions. The physiological signals truly and objectively reflect the emotional state of a person, wherein brain waves (EEG) directly record the signals of brain cortex-related mood-related neural activity. In the emotion recognition task, the brain electrical signal can accurately reflect the emotional state of the human. With the rapid development of brain-computer interface technology, electroencephalogram signals are widely applied to emotion recognition at present.
The electroencephalogram signal is a time sequence of multiple channels and multiple electrodes, so that the electroencephalogram signal has the characteristics of time, space, frequency spectrum and the like. In the prior art, methods such as convolutional neural networks, recurrent neural networks, and Transformer architectures, which are popular in deep learning, have been successfully applied to electroencephalogram emotion recognition. However, convolutional neural networks can effectively capture local information, but ignore global information; the recurrent neural network and transducer architecture can effectively mine the time information of the EEG, but ignores the spatial information.
Therefore, the existing deep learning-based electroencephalogram emotion recognition method has a defect of utilizing space information and time information of electroencephalogram signals, so that emotion recognition accuracy is limited, complexity of a model is increased, and arrangement and application in a real scene are not utilized. Thus, in the electroencephalogram emotion recognition method, how to adequately capture the temporal and spatial features of the EEG remains a challenge to be solved.
Disclosure of Invention
Therefore, the technical problem to be solved by the application is to overcome the defects of the prior art that the deep learning-based electroencephalogram emotion recognition method has the defects of space information and time information of the electroencephalogram, so that the emotion recognition precision is limited, the complexity of a model is increased, and the problems of arrangement and application in a real scene are not utilized.
In order to solve the technical problems, the application provides an electroencephalogram emotion recognition method integrating space-time interaction neural networks, which comprises the following specific steps:
s1, inputting electroencephalogram data into an electroencephalogram emotion recognition model, and according to the positions of electroencephalogram acquisition electrodes, putting the electroencephalogram acquired by each electroencephalogram electrode into a 2D matrix of h multiplied by w to construct a 2D matrix representation of the electroencephalogram; dividing the electroencephalogram signal of each channel into samples with the time length T according to the time window length T to obtain the size of each sample as T multiplied by h multiplied by w, and carrying out 0-1 normalization on the electroencephalogram data of each channel;
s2, dividing each sample into M continuous time steps according to the time dimension, and marking the continuous time steps as P t ,t∈[1,M]The method comprises the steps of carrying out a first treatment on the surface of the Will P t Respectively sequentially inputting to a time step sharing convolution, capturing common features of a plurality of continuous time steps from a low-dimensional layer and a high-dimensional layer to obtain common features F of continuous M time steps in the sample t The method comprises the steps of carrying out a first treatment on the surface of the Splicing the common features F according to the time dimension t Taking the sample as a residual error, and adding the residual error and the spliced characteristic to obtain a characteristic MF;
s3, constructing a downsampling convolution, inputting the feature MF into the downsampling convolution, reducing the time dimension of data, and extracting the time-space feature highly related to emotion typesObtaining a time-space characteristic diagram C epsilon R (T/8)×h×w
S4, operating the space dimension of the time-space characteristic diagram C to divide the time-space characteristic diagram C into N sub-areas with the size of p multiplied by p, wherein the sub-areas are expressed asWhere i=1, 2,..n, j=1, 2,., T/8; according to the time dimension of the time-space feature map C, (T/8) feature maps are spliced together and are expressed as C i ∈R 1×[(T/8)×p×p] Therefore, N spliced feature maps C i Final constituent feature map C 0 ∈R N×[(T/8 ) ×p×p]
Map the characteristic diagram C 0 Transposing, and performing multi-layer perceptron operation on each transposed sub-region to enable spatial information interaction between each sub-region, further enhancing spatial expression of features, and obtaining a feature map C 1 The method comprises the steps of carrying out a first treatment on the surface of the Again for the characteristic diagram C 1 Transposed to obtain a feature map C 2 The characteristic diagram C is processed 0 As residual and feature map C 2 Adding and carrying out layer normalization to obtain a characteristic diagram C A
S5, constructing three emotion classification tasks according to the valence, awakening and dominant three dimensions of emotion, wherein each emotion classification task comprises two categories; using full connection layer to the feature map C A And respectively carrying out two classification in three emotion classification tasks to obtain the probability of emotion classification in each task, and completing emotion recognition.
In one embodiment of the present application, in S1, the electroencephalogram data selects 14 channels to be input into an electroencephalogram emotion recognition model, including channels AF3, AF4, F7, F3, F4, F8, FC5, FC6, T7, T8, P7, P8, O1, O2.
In one embodiment of the present application, in S2, the time-step shared convolution includes two 2D convolution layers, a first 2D convolution reducing the time dimension and a second 2D convolution enlarging the time dimension to an initial size.
In one embodiment of the application, in S3, the downsampling convolution comprises 3 2D convolution layers, each 2D convolution layer reduces the temporal dimension of the feature MF to 1/2 of the initial state, leaves the spatial dimension unchanged, and simultaneously extracts the temporal-spatial feature highly correlated to the emotion classification.
In one embodiment of the application, each of the 2D convolution layers is provided with an activation function, a batch normalization layer, and a Dropout layer.
In one embodiment of the application, the multi-layered perceptron includes two fully connected layers and a GELU activation function.
In one embodiment of the application, in S5, the feature map C is used with a full connection layer A When two classification is respectively carried out in three classification tasks, the emotion category with the highest probability is calculated by using a Softmax function, and the formula is as follows:
wherein x is i Representing the output value of the ith neuron, x c The output value of the C-th neuron is represented, C is the number of neurons,representing the output values of all neurons.
In one embodiment of the present application, the step of training the electroencephalogram emotion recognition model includes:
collecting brain electrical information of a subject, constructing an brain electrical emotion data set, enabling the subject to evaluate emotion of the subject, and scoring from three dimensions of valence, awakening and domination of the emotion, so as to determine a real label of emotion type; dividing the electroencephalogram emotion data set into a training set and a testing set according to a preset proportion;
inputting electroencephalogram signal data of the training set into an electroencephalogram emotion recognition model to obtain probability of emotion categories in each task; during training, the cross entropy loss function is used for minimizing the difference between the classification result and the real label, and the Adam optimizer is used for optimizing the loss function and updating the learning rate;
and performing performance evaluation on the electroencephalogram emotion recognition model by using classification accuracy, standard deviation, training and test time on a test set.
In one embodiment of the present application, the cross entropy loss function L is formulated as:
where N represents the number of samples and M represents the number of categories; if the classification category is the same as the true category, y ic 1, otherwise 0; p is p ic Representing the probability that sample i belongs to category c.
The application also provides an electroencephalogram emotion recognition system fusing the space-time interaction neural network, which comprises the following steps:
a memory for storing a computer program;
and the processor is used for realizing the step of the electroencephalogram emotion recognition method fusing the space-time interaction neural network when executing the computer program.
Compared with the prior art, the technical scheme of the application has the following advantages:
according to the electroencephalogram emotion recognition method integrating the space-time interaction neural network, the time step sharing convolution is used, the time dimension is divided into a plurality of continuous time steps, common features of the continuous time steps are captured from the low-dimension layer and the high-dimension layer and spliced, and the time expression of electroencephalogram signals is enhanced. And the space interaction multi-layer perceptron is used for dividing the feature map from the space dimension and splicing the feature map from the time dimension so as to perform information interaction among a plurality of subspace regions, further strengthen the space expression of the features and strengthen the space dependence of the electroencephalogram signals.
Therefore, the method is based on a time-step shared convolution and space interaction multi-layer perceptron, so that the time and space characteristics of the electroencephalogram emotion characteristics are further mined, the problem that the utilization of the electroencephalogram space information and the time information is insufficient in the existing electroencephalogram emotion recognition method is solved to a certain extent, and the method has the advantages of being high in calculation efficiency, high in recognition precision and strong in generalization capability.
In addition, fewer electroencephalogram acquisition electrodes are used, so that the training and testing time of the model is shortened while the recognition accuracy is ensured, and the realizability of the model in a practical application scene is improved.
Drawings
In order that the application may be more readily understood, a further description of the application will be rendered by reference to specific embodiments thereof that are illustrated in the appended drawings, in which
FIG. 1 is a flow chart of an electroencephalogram emotion recognition method fused with a space-time interaction neural network;
FIG. 2 is an overall frame diagram of an electroencephalogram emotion recognition method fused with a space-time interaction neural network;
FIG. 3 is a schematic diagram of a specific method for constructing a 2D matrix representation of an electroencephalogram signal in the present application;
FIG. 4 is a schematic illustration of the spatial zone division of the present application;
FIG. 5 is a block diagram of a spatially interactive multilayer perceptron of the present application;
fig. 6 is a comparative graph of the results of an ablation experiment in an embodiment of the present application.
Detailed Description
The present application will be further described with reference to the accompanying drawings and specific examples, which are not intended to be limiting, so that those skilled in the art will better understand the application and practice it.
Referring to fig. 1 and 2, the application discloses an electroencephalogram emotion recognition method fusing a space-time interaction neural network, which comprises the following specific steps:
s1, inputting electroencephalogram data into an electroencephalogram emotion recognition model, and referring to FIG. 3, 14 channels which are highly related to emotion are selected, wherein the 14 channels comprise channels AF3, AF4, F7, F3, F4, F8, FC5, FC6, T7, T8, P7, P8, O1 and O2, and the aim is to reduce the number of the channels so as to improve the realizability in an actual application scene. And carrying out 0.1-45hz band-pass filtering on the electroencephalogram signals, and then placing the electroencephalogram signals acquired by each electroencephalogram electrode into a 2D matrix of h multiplied by w according to the positions of the electroencephalogram acquisition electrodes to construct a 2D matrix representation of the electroencephalogram signals.
In this embodiment, the time window T takes 1s. Dividing the electroencephalogram signals of each channel into samples with the time length T according to the time window length T, namely, arranging the electroencephalogram signals of all time nodes into T h×w 2D matrixes according to the time sequence, obtaining the size of each sample as T×h×w, and carrying out 0-1 normalization on the electroencephalogram data of each channel.
S2, extracting common features of the continuous time steps of the sample through time step sharing convolution, capturing the common features of a plurality of continuous time steps from the low-dimensional and high-dimensional layers, and enhancing the information expression of the original input.
Dividing each sample into M continuous time steps according to the time dimension, and marking the time steps as P t Wherein t is [1, M ]]. Will P t Respectively sequentially inputting to a time step sharing convolution, capturing common features of a plurality of continuous time steps from a low-dimensional layer and a high-dimensional layer to obtain common features F of continuous M time steps in the sample t . Splicing the common features F according to the time dimension t And taking the sample as a residual error, adding the residual error with the spliced characteristic, and enhancing the information expression of the original input to obtain the characteristic MF.
The time-step shared convolution includes two 2D convolution layers, a first 2D convolution reducing the time dimension and a second 2D convolution enlarging the time dimension to an initial size. Each 2D convolution layer is provided with an activation function, a batch normalization layer and a Dropout layer.
S3, constructing downsampling convolution, reducing the time dimension, and extracting time-space characteristics highly relevant to emotion categories.
Inputting the feature MF into downsampling convolution, reducing the time dimension of data, and extracting time-space features highly related to emotion categories to obtain a time-space feature map C epsilon R (T/8)×h×w
The downsampling convolution comprises 3 2D convolution layers, each 2D convolution layer reduces the time dimension of the feature MF to 1/2 of the initial state, keeps the space dimension unchanged, and simultaneously extracts time-space features highly correlated to emotion categories. Each 2D convolution layer is provided with an activation function, a batch normalization layer and a Dropout layer.
S4, space region division is carried out on the time-space feature map C, a space interaction multi-layer perceptron is constructed, information interaction in a plurality of subspace regions is achieved, and the space dependence of the features is enhanced.
And carrying out space region division on the time-space characteristic diagram C. Referring to FIG. 4, the spatial dimension of the time-space feature map C is operated on to divide it into N sub-regions of p size, denoted asWhere i=1, 2,..n, j=1, 2,..t/8. According to the time dimension of the time-space feature map C, (T/8) feature maps are spliced together and are expressed as C i ∈R 1×[(T/8)×p×p] . Thus, N spliced feature maps C i Final constituent feature map C 0 ∈R N ×[(T/8)×p×p]
And constructing a space interaction multi-layer perceptron. Referring to FIG. 5, the characteristic diagram C 0 Transposing, and performing multi-layer perceptron operation on each transposed sub-region, namely performing spatial information interaction between each sub-region, further enhancing spatial expression of the features, and obtaining a feature map C 1 . The multi-layer perceptron comprises two full connection layers and a GELU activation function. Again for the characteristic diagram C 1 Transposed to obtain a feature map C 2 The characteristic diagram C is processed 0 As residual and feature map C 2 Adding and carrying out layer normalization to obtain a characteristic diagram C with enhanced spatial dependence A
S5, constructing emotion classification tasks, obtaining the probability of emotion categories in each task, and completing emotion recognition.
According to the valence, awakening and dominant three dimensions of emotion, three emotion classification tasks are constructed, and each emotion classification task comprises two categories; using full connection layer to the feature map C A And respectively carrying out two classification in three emotion classification tasks to obtain the probability of emotion classification in each task, and completing emotion recognition.
Using full connection layer to the feature map C A When two classification is respectively carried out in three classification tasks, the emotion category with the highest probability is calculated by using a Softmax function, and the formula is as follows:
wherein x is i Representing the output value of the ith neuron, x c The output value of the C-th neuron is represented, C is the number of neurons,representing the output values of all neurons.
In order to solve the problems that in the existing electroencephalogram emotion recognition method, the time features and the space features of electroencephalogram data are not captured sufficiently, so that the recognition accuracy is limited, the complexity of a model is increased, and arrangement and application in a real scene are not utilized, the method further excavates the time and the space features of the emotion features in the electroencephalogram data, the time expression of the electroencephalogram is increased by capturing common features of continuous time steps, the space dependence of the electroencephalogram is enhanced by local space region interaction of the electroencephalogram emotion features, the classification performance of the electroencephalogram emotion recognition model is improved, and the accurate recognition of the electroencephalogram emotion is realized. In addition, the application is based on fewer electroencephalogram acquisition electrodes, and the realizability of the application in the actual application scene is improved.
The application also provides another specific embodiment for clarifying the process of training the electroencephalogram emotion recognition model, and comparing the training result with the current advanced deep learning algorithms MLF-CapsNet, NAS and ECNN-C to further verify the beneficial effects of the application.
The electroencephalogram data for training the electroencephalogram emotion recognition model can be an existing electroencephalogram data set based on emotion recognition recorded in literature data, or can be electroencephalogram information of a tested person collected from a site.
The embodiment uses the DEAP brain electrical emotion data set to train the brain electrical emotion recognition model. The dataset was collected from 32 subjects, with a 50% male and female ratio each, aged 19-37 years. The dataset recorded brain electrical signals, which were recorded primarily by 32 electrode channels, of 40 60s of music video viewed by each subject, along with peripheral physiological signals, with a baseline signal of 3s before viewing the video. After viewing the video, the subject scored 1 to 9 from three dimensions of mood, arousal and dominance to evaluate his mood and determine the true signature of the mood category.
The original brain electrical signal recorded mainly 32 electrode channels, sampled at 512Hz, and passed through noise of muscle movements, 4-45Hz band pass filter and 128Hz downsampling.
Dividing the electroencephalogram emotion data set into a training set and a testing set according to a preset proportion.
And inputting the electroencephalogram data of the training set into the electroencephalogram emotion recognition model to obtain the probability of emotion categories in each task.
The cross entropy loss function is used for minimizing the difference between the classification result and the corresponding real label during training, and the Adam optimizer is used for optimizing the loss function and updating the learning rate.
Using the cross entropy as a loss function, calculating a loss index in the training process, wherein the loss function L is defined as:
where N represents the number of samples and M represents the number of categories; if the classification category is the same as the true category, y ic 1, otherwise 0; p is p ic Representing the probability that sample i belongs to category c.
And finally, performing performance evaluation on the electroencephalogram emotion recognition model by using classification accuracy, standard deviation, training and test time as indexes on a test set.
To further illustrate the benefits of the present application, in this embodiment, the method of the present application is compared with the current advanced deep learning algorithms MLF-CapsNet, NAS and ECNN-C.
Parameters set in the experiment include: adam optimizer is used to minimize class cross-over entropy loss function with a learning rate of 5e-4. The number of heads of the multi-head self-attention mechanism was set to 4, and a total of 80 epochs were trained. When training is completed for 5 epochs and loss is not reduced, the learning rate is reduced by 80%, the value of Dropout takes 0.3, and an early-stop strategy is used in the training process.
The classification result of the method of the application is compared with an advanced deep learning algorithm based on DEAP emotion recognition brain electricity data set, and the specific result is shown in table 1.
TABLE 1 classification results of DEAP-based different methods for emotion recognition of brain electrical data sets
Table 1 shows the average classification accuracy and standard deviation over 32 subjects for the current advanced deep learning model and DEAP-based emotion recognition electroencephalogram dataset of the method of the present application. The application can be found that the highest average classification precision of 98.56%,98.70% and 98.63% are obtained in all three classification tasks, and meanwhile, the lower standard deviation is also obtained. Compared with the current advanced deep learning model, the method provided by the application improves the classification performance of the electroencephalogram emotion recognition model and realizes more accurate recognition of the electroencephalogram emotion.
Meanwhile, in the embodiment, the effectiveness analysis is carried out on the time-step sharing convolution and space interaction multi-layer perceptron by using an ablation experiment method.
The model containing only downsampled convolutions was designated as baseline and was named Base-CNN. The results of the ablation experiments are shown in FIG. 6, which shows the average classification accuracy and standard deviation of Base-CNN and the application in three classification tasks. Compared with a baseline model, the model provided by the application has higher classification precision and lower standard deviation, and proves that the time-step sharing convolution and space interaction multi-layer perceptron can more comprehensively capture the time characteristics and the space characteristics of the electroencephalogram signals, and further improve the classification precision.
Experimental results show that the electroencephalogram emotion recognition method based on the time-step sharing convolution and space interaction multi-layer perceptron has good classification performance and strong generalization capability.
In addition, we also compare the training and testing time of the method of the present application with the current advanced deep learning model, and the specific results are shown in table 2.
TABLE 2 training and test time for different methods based on DEAP dataset
Table 2 shows training and testing times of the present advanced deep learning model and the method of the present application on the DEAP dataset. It can be found that the method of the application has shorter training and testing time, which makes it possible to apply the method to actual scenes.
The experiment shows that the application realizes accurate identification of the brain electricity emotion, improves the classification performance and the realizability in the practical application scene, and can provide reference for the development of the brain electricity emotion identification technical field.
The specific embodiment of the application also provides an electroencephalogram emotion recognition system fusing the space-time interaction neural network, which comprises the following steps:
a memory for storing a computer program;
and the processor is used for realizing the step of the electroencephalogram emotion recognition method fusing the space-time interaction neural network when executing the computer program.
It will be appreciated by those skilled in the art that embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flowchart illustrations and/or block diagrams, and combinations of flows and/or blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
It is apparent that the above examples are given by way of illustration only and are not limiting of the embodiments. Other variations and modifications of the present application will be apparent to those of ordinary skill in the art in light of the foregoing description. It is not necessary here nor is it exhaustive of all embodiments. And obvious variations or modifications thereof are contemplated as falling within the scope of the present application.

Claims (10)

1. The brain electricity emotion recognition method integrating the space-time interaction neural network is characterized by comprising the following specific steps of:
s1, inputting electroencephalogram data into an electroencephalogram emotion recognition model, and according to the positions of electroencephalogram acquisition electrodes, putting the electroencephalogram acquired by each electroencephalogram electrode into a 2D matrix of h multiplied by w to construct a 2D matrix representation of the electroencephalogram; dividing the electroencephalogram signal of each channel into samples with the time length T according to the time window length T to obtain the size of each sample as T multiplied by h multiplied by w, and carrying out 0-1 normalization on the electroencephalogram data of each channel;
s2, dividing each sample into M continuous time steps according to the time dimension, and marking the continuous time steps as P t ,t∈[1,M]The method comprises the steps of carrying out a first treatment on the surface of the Will P t Respectively sequentially inputting to a time step sharing convolution, capturing common features of a plurality of continuous time steps from a low-dimensional layer and a high-dimensional layer to obtain common features F of continuous M time steps in the sample t The method comprises the steps of carrying out a first treatment on the surface of the Splicing the common features F according to the time dimension t Taking the sample as a residual error, and adding the residual error and the spliced characteristic to obtain a characteristic MF;
s3, constructing a downsampling convolution, inputting the feature MF into the downsampling convolution, reducing the time dimension of data, extracting time-space features highly related to emotion categories, and obtaining a time-space feature map C epsilon R (T/8)×h×w
S4, operating the space dimension of the time-space characteristic diagram C to divide the time-space characteristic diagram C into N sub-areas with the size of p multiplied by p, wherein the sub-areas are expressed asWhere i=1, 2,..n, j=1, 2,., T/8; according to the time dimension of the time-space feature map C, (T/8) feature maps are spliced together and are expressed as C i ∈R 1×[(T/8)×p×p] Therefore, N spliced feature maps C i Final constituent feature map C 0 ∈R N×[T/8)×p×p]
Map the characteristic diagram C 0 Transpose and pair transposeThen each sub-region is operated by a multi-layer perceptron to enable the space information interaction between each sub-region to further strengthen the space expression of the features and obtain a feature map C 1 The method comprises the steps of carrying out a first treatment on the surface of the Again for the characteristic diagram C 1 Transposed to obtain a feature map C 2 The characteristic diagram C is processed 0 As residual and feature map C 2 Adding and carrying out layer normalization to obtain a characteristic diagram C A
S5, constructing three emotion classification tasks according to the valence, awakening and dominant three dimensions of emotion, wherein each emotion classification task comprises two categories; using full connection layer to the feature map C A And respectively carrying out two classification in three emotion classification tasks to obtain the probability of emotion classification in each task, and completing emotion recognition.
2. The brain-electrical emotion recognition method based on the fusion space-time interaction neural network according to claim 1, wherein in S1, 14 channels are selected for inputting brain-electrical signal data to a brain-electrical emotion recognition model, and the brain-electrical emotion recognition model comprises channels AF3, AF4, F7, F3, F4, F8, FC5, FC6, T7, T8, P7, P8, O1 and O2.
3. The method for recognizing brain waves fused with space-time interactive neural network according to claim 1, wherein in S2, the time-step sharing convolution comprises two 2D convolution layers, the first 2D convolution reduces the time dimension, and the second 2D convolution enlarges the time dimension to the initial size.
4. The method for recognizing brain waves fused with space-time interactive neural network according to claim 1, wherein in S3, the downsampling convolution comprises 3 2D convolution layers, each 2D convolution layer reduces the time dimension of the feature MF to 1/2 of the initial state, keeps the space dimension unchanged, and simultaneously extracts the time-space feature highly related to the emotion classification.
5. The method for recognizing brain electric emotion integrating space-time interaction neural network according to any one of claims 3 and 4, wherein each 2D convolution layer is provided with an activation function, a batch normalization layer and a Dropout layer.
6. The method for recognizing brain electricity emotion fusing space-time interaction neural network as claimed in claim 1, wherein in S4, said multi-layer perceptron comprises two full connection layers and a gel activating function.
7. The brain-electrical emotion recognition method based on the fusion of space-time interaction neural network according to claim 1, wherein in S5, the feature map C is obtained by using a full connection layer A When two classification is respectively carried out in three classification tasks, the emotion category with the highest probability is calculated by using a Softmax function, and the formula is as follows:
wherein x is i Representing the output value of the u-th neuron, x c The output value of the C-th neuron is represented, C is the number of neurons,representing the output values of all neurons.
8. The method for recognizing brain waves by fusing a space-time interaction neural network according to claim 1, wherein the step of training the brain wave recognition model comprises the steps of:
collecting brain electrical information of a subject, constructing an brain electrical emotion data set, enabling the subject to evaluate emotion of the subject, and scoring from three dimensions of valence, awakening and domination of the emotion, so as to determine a real label of emotion type; dividing the electroencephalogram emotion data set into a training set and a testing set according to a preset proportion;
inputting electroencephalogram signal data of the training set into an electroencephalogram emotion recognition model to obtain probability of emotion categories in each task; during training, the cross entropy loss function is used for minimizing the difference between the classification result and the real label, and the Adam optimizer is used for optimizing the loss function and updating the learning rate;
and performing performance evaluation on the electroencephalogram emotion recognition model by using classification accuracy, standard deviation, training and test time on a test set.
9. The brain electrical emotion recognition method based on the fusion space-time interaction neural network of claim 8, wherein the cross entropy loss function L formula is:
where N represents the number of samples and M represents the number of categories; if the classification category is the same as the true category, y ic 1, otherwise 0; p is p ic Representing the probability that sample i belongs to category c.
10. An electroencephalogram emotion recognition system integrating space-time interaction neural networks, which is characterized by comprising:
a memory for storing a computer program;
a processor for implementing the steps of a method for identifying brain waves fusing space-time interaction neural networks according to any one of claims 1 to 9 when executing the computer program.
CN202310674320.4A 2023-06-08 2023-06-08 Electroencephalogram emotion recognition method and system integrating space-time interaction neural network Active CN116861211B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310674320.4A CN116861211B (en) 2023-06-08 2023-06-08 Electroencephalogram emotion recognition method and system integrating space-time interaction neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310674320.4A CN116861211B (en) 2023-06-08 2023-06-08 Electroencephalogram emotion recognition method and system integrating space-time interaction neural network

Publications (2)

Publication Number Publication Date
CN116861211A true CN116861211A (en) 2023-10-10
CN116861211B CN116861211B (en) 2024-09-06

Family

ID=88218070

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310674320.4A Active CN116861211B (en) 2023-06-08 2023-06-08 Electroencephalogram emotion recognition method and system integrating space-time interaction neural network

Country Status (1)

Country Link
CN (1) CN116861211B (en)

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111297380A (en) * 2020-02-12 2020-06-19 电子科技大学 Emotion recognition method based on space-time convolution core block
EP3716152A1 (en) * 2019-03-26 2020-09-30 Robert Bosch GmbH Reusing of neural network training for higher dimensional data
CN113128353A (en) * 2021-03-26 2021-07-16 安徽大学 Emotion sensing method and system for natural human-computer interaction
CN113723557A (en) * 2021-09-08 2021-11-30 山东大学 Depression electroencephalogram classification system based on multiband time-space convolution network
CN113749657A (en) * 2021-09-10 2021-12-07 合肥工业大学 Brain wave emotion recognition method based on multitask capsules
CN114209342A (en) * 2022-01-28 2022-03-22 南京邮电大学 Electroencephalogram signal motor imagery classification method based on space-time characteristics
CN114224342A (en) * 2021-12-06 2022-03-25 南京航空航天大学 Multi-channel electroencephalogram emotion recognition method based on space-time fusion feature network
CN115474899A (en) * 2022-08-17 2022-12-16 浙江大学 Basic taste perception identification method based on multi-scale convolution neural network
WO2023060721A1 (en) * 2021-10-11 2023-04-20 北京工业大学 Emotional state displaying method, apparatus and system based on resting-state cerebral functional image

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3716152A1 (en) * 2019-03-26 2020-09-30 Robert Bosch GmbH Reusing of neural network training for higher dimensional data
CN111297380A (en) * 2020-02-12 2020-06-19 电子科技大学 Emotion recognition method based on space-time convolution core block
CN113128353A (en) * 2021-03-26 2021-07-16 安徽大学 Emotion sensing method and system for natural human-computer interaction
CN113723557A (en) * 2021-09-08 2021-11-30 山东大学 Depression electroencephalogram classification system based on multiband time-space convolution network
CN113749657A (en) * 2021-09-10 2021-12-07 合肥工业大学 Brain wave emotion recognition method based on multitask capsules
WO2023060721A1 (en) * 2021-10-11 2023-04-20 北京工业大学 Emotional state displaying method, apparatus and system based on resting-state cerebral functional image
CN114224342A (en) * 2021-12-06 2022-03-25 南京航空航天大学 Multi-channel electroencephalogram emotion recognition method based on space-time fusion feature network
CN114209342A (en) * 2022-01-28 2022-03-22 南京邮电大学 Electroencephalogram signal motor imagery classification method based on space-time characteristics
CN115474899A (en) * 2022-08-17 2022-12-16 浙江大学 Basic taste perception identification method based on multi-scale convolution neural network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
PHAM TD ET AL.,: "Enhancing Performance of EEG-based Emotion Recognition Systems Using Feature Smoothing", 《SPRINGER-VERLAG BERLIN》, vol. 9492, 31 December 2015 (2015-12-31), pages 95 - 102 *
张慧港,: "基于脉冲神经网络和脑功能分析的情绪识别研究", 《中国优秀硕士学位论文全文数据库 (基础科学辑)》, vol. 2023, no. 2, 15 February 2023 (2023-02-15), pages 006 - 1377 *

Also Published As

Publication number Publication date
CN116861211B (en) 2024-09-06

Similar Documents

Publication Publication Date Title
CN112932502B (en) Electroencephalogram emotion recognition method combining mutual information channel selection and hybrid neural network
CN113158964B (en) Sleep stage method based on residual error learning and multi-granularity feature fusion
Kan et al. Self-supervised group meiosis contrastive learning for eeg-based emotion recognition
CN117195099A (en) Electroencephalogram signal emotion recognition algorithm integrating multi-scale features
CN115659207A (en) Electroencephalogram emotion recognition method and system
Bai et al. Sect: A method of shifted eeg channel transformer for emotion recognition
Xu et al. AMDET: Attention based multiple dimensions EEG transformer for emotion recognition
Li et al. GCNs–FSMI: EEG recognition of mental illness based on fine-grained signal features and graph mutual information maximization
Li et al. Dynamical graph neural network with attention mechanism for epilepsy detection using single channel EEG
CN116861211B (en) Electroencephalogram emotion recognition method and system integrating space-time interaction neural network
CN117407748A (en) Electroencephalogram emotion recognition method based on graph convolution and attention fusion
CN110507288A (en) Vision based on one-dimensional convolutional neural networks induces motion sickness detection method
CN116115240A (en) Electroencephalogram emotion recognition method based on multi-branch chart convolution network
CN115316955A (en) Light-weight and quick decoding method for motor imagery electroencephalogram signals
CN115859185A (en) Electroencephalogram emotion recognition method based on pulse convolution neural network
CN114638253A (en) Identity recognition system and method based on emotion electroencephalogram feature fusion optimization mechanism
Wu et al. A novel emotion recognition method based on 1D-DenseNet
Gao et al. Multi-Head Self-Attention Enhanced Convolutional Neural Network for Driver Fatigue Detection using EEG Signals
Zhou et al. Dual-Branch Convolution Network with Efficient Channel Attention for EEG-Based Motor Imagery Classification
CN118467934A (en) Brain region characteristic-based electroencephalogram signal representation learning method and system
Neverlien et al. Decoding Emotions From EEG Responses Elicited by Videos Using Machine Learning Techniques on Two Datasets
Wang et al. EEG Emotion Classification Using 2D-3DCNN
CN117158912B (en) Sleep stage detection system based on graph attention mechanism and space-time graph convolution
CN117332317B (en) EEG emotion recognition method combining attention residual error network with LSTM
CN118490232B (en) Brain depression diagnosis method based on multi-frequency domain decomposition

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant