CN111832431A - Emotional electroencephalogram classification method based on CNN - Google Patents
Emotional electroencephalogram classification method based on CNN Download PDFInfo
- Publication number
- CN111832431A CN111832431A CN202010582404.1A CN202010582404A CN111832431A CN 111832431 A CN111832431 A CN 111832431A CN 202010582404 A CN202010582404 A CN 202010582404A CN 111832431 A CN111832431 A CN 111832431A
- Authority
- CN
- China
- Prior art keywords
- emotion
- data
- layer
- samples
- convolutional neural
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 21
- 230000002996 emotional effect Effects 0.000 title claims abstract description 8
- 230000008451 emotion Effects 0.000 claims abstract description 49
- 238000013527 convolutional neural network Methods 0.000 claims abstract description 38
- 238000002474 experimental method Methods 0.000 claims abstract description 26
- 238000012549 training Methods 0.000 claims abstract description 14
- 238000012216 screening Methods 0.000 claims abstract description 5
- 230000037007 arousal Effects 0.000 claims description 16
- 238000012360 testing method Methods 0.000 claims description 12
- 238000011156 evaluation Methods 0.000 claims description 9
- 238000011176 pooling Methods 0.000 claims description 6
- 230000036651 mood Effects 0.000 claims description 5
- 238000012935 Averaging Methods 0.000 claims description 4
- 230000008909 emotion recognition Effects 0.000 claims description 4
- 238000002372 labelling Methods 0.000 claims description 4
- 238000010606 normalization Methods 0.000 claims description 4
- 210000002569 neuron Anatomy 0.000 claims description 3
- 206010062519 Poor quality sleep Diseases 0.000 claims 2
- PIZHFBODNLEQBL-UHFFFAOYSA-N 2,2-diethoxy-1-phenylethanone Chemical compound CCOC(OCC)C(=O)C1=CC=CC=C1 PIZHFBODNLEQBL-UHFFFAOYSA-N 0.000 abstract description 7
- 239000013589 supplement Substances 0.000 abstract description 2
- 238000013528 artificial neural network Methods 0.000 description 4
- 210000004556 brain Anatomy 0.000 description 3
- 238000007781 pre-processing Methods 0.000 description 3
- 238000012545 processing Methods 0.000 description 3
- 238000010008 shearing Methods 0.000 description 3
- 238000013145 classification model Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000004880 explosion Methods 0.000 description 2
- 230000008447 perception Effects 0.000 description 2
- 230000001502 supplementing effect Effects 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000007635 classification algorithm Methods 0.000 description 1
- 230000008034 disappearance Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2218/00—Aspects of pattern recognition specially adapted for signal processing
- G06F2218/08—Feature extraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2415—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/047—Probabilistic or stochastic networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2218/00—Aspects of pattern recognition specially adapted for signal processing
- G06F2218/12—Classification; Matching
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/15—Biometric patterns based on physiological signals, e.g. heartbeat, blood flow
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computing Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Molecular Biology (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Probability & Statistics with Applications (AREA)
- Signal Processing (AREA)
- Measurement And Recording Of Electrical Phenomena And Electrical Characteristics Of The Living Body (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
Abstract
The invention discloses a CNN-based emotion electroencephalogram classification method. The method comprises the steps of taking a Russell emotion dimension model in a continuous emotion dimension model as a reference, taking a DEAP data set as a sample, removing a base line of emotion electroencephalograms in the emotion electroencephalograms, normalizing data, extracting Pearson coefficients of three frequency bands of the electroencephalograms, converting the Pearson coefficients into a 2D picture format, screening experiments valuable for emotion electroencephalograms classification by taking the experiments as a unit through SBS, inputting screened experimental data into CWAGAN-GP to perform data enhancement so as to supplement a training set, and inputting the screened experimental data into an integrated convolutional neural network in a frame mode. The method can effectively classify the emotional electroencephalogram signals, provides considerable classification precision, and can effectively prevent overfitting of the convolutional neural network by inputting the emotional electroencephalogram signals into the integrated convolutional neural network in a frame form.
Description
Technical Field
The invention relates to an emotion electroencephalogram classification method, in particular to a method for classifying emotion electroencephalograms by using CNN (CNN) after preprocessing and optimization.
Background
A Convolutional Neural Network (CNN) is a feedforward type neural network that has excellent performance in large-scale image processing, and has been widely used in the fields of image classification, localization, and the like. Compared with other neural network structures, the convolutional neural network requires relatively few parameters, so that the convolutional neural network can be widely applied.
The generative countermeasure network (GAN) passes through two modules in the framework: the mutual game learning of the generative model and the discriminant model generates quite good output, and new data which are false and spurious can be generated according to the original data set. The WGAN-GP is proposed aiming at the problems of the WGAN, the WGAN still has the problems of difficult training and slow convergence rate in the real experimental process, and compared with the traditional GAN, the WGAN-GP is not obviously improved experimentally. The problem with WGAN is to use the clipping weight directly in dealing with the continuity constraint, i.e. check the absolute values of all the parameters of the arbiter for exceeding a threshold, e.g. 0.01, each time the parameters of the arbiter have been updated, and return these parameters to the range of-0.01, 0.01. In this case, the optimal strategy is to make all parameters as extreme as possible, either take the maximum value (e.g. 0.01) or take the minimum value (e.g. -0.01), the fitting capability of the deep neural network cannot be fully exerted for the deep neural network, and it is also found that forcing the shearing weight easily causes gradient disappearance or gradient explosion, both of which are caused by selection of the shearing range, and if the shearing range is selected to be too small, the gradient becomes larger slightly after passing through one layer of network, and then gradient explosion occurs after multiple layers. To solve this problem and find a suitable way to satisfy the continuity condition, the WGAN-GP uses a gradient penalty way to satisfy this continuity condition. CWGAN-GP is a condition constraint, namely a label, added on the basis of WGAN-GP, so that the CWGAN-GP can generate data of the specified label.
And (4) Selecting Backward (SBS) sequences, starting from the feature complete set A, and evaluating a function value to be optimal after one feature is removed from the feature set A each time.
Convolutional neural networks are a very efficient model of the image classification task. For such models, the present study attempts to convert the DEAP data into a 2D image format so that the CNN model can learn how to classify them efficiently.
Disclosure of Invention
The method adopts a Russell emotion dimension model in a continuous emotion dimension model as a reference, uses a DEAP data set as a sample, performs preprocessing data normalization, extracts Pearson coefficients of three frequency bands most relevant to emotion recognition in an electroencephalogram channel, converts the Pearson coefficients into a 2D picture format, screens valuable experiments for emotion electroencephalogram classification through SBS by taking the experiments as a unit, inputs the screened experimental data into CWAGAN-GP for data enhancement so as to supplement a training set, and finally inputs convolutional neural network classification.
The invention classifies the preprocessed emotion electroencephalogram by using CNN, and specifically comprises the following steps:
And 2, extracting the baseline data of 3 seconds before each experiment, segmenting the baseline data by using a selected time window, adding the segmented baseline data, averaging the segmented baseline data, segmenting the rest electroencephalogram data by using the selected time window, subtracting the processed baseline data to obtain the difference value of the electroencephalogram data, and performing data normalization. 3 frequency bands most relevant to emotion recognition are extracted from the difference values of the electroencephalogram data, and then Pearson coefficients of the 3 frequency bands are calculated and converted into a picture format.
And 3, for each subject, taking an experiment as a unit, converting the data of each subject into the experiment times, the segment number and the segment number, the residual number after screening is set, sending the data into SBS to screen out the experiment combination with the highest accuracy rate, sending the data into CWGAN-GP to carry out data enhancement through a labeled generating type confrontation network, and supplementing the data to balance the emotion data of each category because the emotion of some categories is reduced after screening.
And 4, processing the processed data and labels according to a training sample and a test sample 8: 2, and inputting the frame format into the integrated convolutional neural network to perform classification test on the processed samples.
In the step 1, when a Russell emotion dimension model is taken as a reference, each emotion state is placed on a two-dimensional plane, the arousal degree and the pleasure degree form two dimensions of a coordinate axis, the valence (value) represents the subjective evaluation of a main body (person) on emotion, and the negative emotion is changed into the positive emotion on the coordinate axis; arousal (arousal) represents the degree of emotion perception by a subject (human), varying on the numerical axis from calm to excited. The values of both dimensions vary from 1-9, with the magnitude of the values representing the degree of emotion. When selecting emotion dichotomy, the rating values on valence and arousal are divided with a median of 5 as a threshold, greater than 5 being labeled as 1, representing high valence or high arousal (HV, HA); less than or equal to 5 is labeled 0, representing low titer or low arousal (LV, LA); when selecting mood quartiles, HV, HA, LV, LA may be combined to generate four classes of tags for HVHA, HVLA, LVHA, LVLA.
In the step 2, the electroencephalogram signals are converted into a 2D picture form, and the method specifically comprises the following steps:
2-2. for 8-12Hz (alpha band), 12-30Hz (beta band), 30-60Hz (gamma band), the signal is filtered using a fourth order butterworth filter.
And 2-3, calculating 32 electroencephalogram channels for each frequency band, calculating a Pearson coefficient between every two channels, converting the Pearson coefficient into a format of 3X 1, and combining the results of 3 frequency bands to obtain a 2D picture format of 32X 3.
In step 4, the processed data and labels are processed according to the training sample and the test sample 8: 2, and inputting the frame format into the integrated convolutional neural network to perform classification test on the processed samples. The method specifically comprises the following steps:
the time window size is set to 1s, 40 × 60 × 32 ═ 76800 samples are segmented in total, for 60 segments of each experiment, every 3 segments, namely every 3 seconds of samples are combined into one frame, no repeated samples exist between frames, one experiment can generate 20 frames, the total number of samples becomes 40 × 20 × 32 ═ 25600, the sample format is 25600 × 3 × 32 × 3, 3 pictures are input to 3 convolutional neural networks at the same time, and the average of 3 output results is the final result of the input. All samples were run as 8: 2 training samples and testing samples.
The integrated convolutional neural network is composed of three convolutional neural networks with completely same structures, and the structure of each convolutional neural network is as follows:
the C1 layer is a convolutional layer with RELU, with 32 filters, convolution kernel set to 3 x 3, step size 1 x 1.
The S2 level is an average pooling level with a size of 2 x 2 and steps of 2 x 2.
The B3 layer is a BN layer.
The C4 layer is also a convolutional layer with RELU, with 64 filters, with the convolutional kernel set to 3 x 3, with steps of 1 x 1.
The S5 level is an average pooling level with a size of 2 x 2 and steps of 2 x 2.
The B6 layer is a BN layer.
The C7 layer is a convolutional layer with RELU, with 128 filters, convolution kernel set to 3 x 3, step size 1 x 1.
The D8 layer is a fully connected layer with an output dimension of 128.
The DO9 layer is a Dropout layer that randomly turns off some of the neurons with a probability of 0.5 during the training process.
Finally, the output dimension of the finally concatenated Softmax layer is 2.
The invention has the following beneficial effects:
the method adopts a Russell emotion dimension model in a continuous emotion dimension model as a reference, uses a DEAP data set as a sample, removes a base line of emotion electroencephalogram in the sample, performs data normalization to extract Pearson coefficients of three frequency bands of the electroencephalogram, converts the Pearson coefficients into a 2D picture format, screens through SBS data, enhances the data, and inputs the data into an integrated convolutional neural network in a frame form, so that the emotion classification accuracy can be effectively improved, and overfitting of the convolutional neural network is prevented.
Drawings
FIG. 1 is a Russell emotional dimension model;
FIG. 2 is a data type of a DEAP data set;
FIG. 3 is a schematic diagram of brain electrical data preprocessing;
fig. 4 is a basic configuration diagram of a CNN network;
FIG. 5 is a flow chart of emotion electroencephalogram classification;
Detailed Description
The present invention is further illustrated by the following specific examples. The following description is exemplary and explanatory only and is not restrictive of the invention in any way.
As shown in fig. 5, a CNN-based emotion electroencephalogram classification method specifically includes the following steps:
The Russell emotion dimension model is an emotion space continuous two-dimensional emotion space with arousal degree and valence as coordinate axes. Valence (value) represents the subjective evaluation of the mood by the subject (person), from negative to positive mood on the numerical axis; arousal (arousal) represents the degree of emotion perception by a subject (human), varying on the numerical axis from calm to excited. The values of both dimensions vary from 1-9, with the magnitude of the values representing the degree of emotion. When selecting emotion dichotomy, the rating values on valence and arousal are divided with a median of 5 as a threshold, greater than 5 being labeled as 1, representing high valence or high arousal (HV, HA); less than or equal to 5 is labeled 0, representing low titer or low arousal (LV, LA); when selecting mood quartiles, HV, HA, LV, LA may be combined to generate four classes of tags for HVHA, HVLA, LVHA, LVLA.
where X, Y represent channels selected from 32 brain electrical channels.
The pearson coefficients are calculated to convert the image into a 32 x 1 format, and the results of 3 bands are combined to obtain a 32 x 3 2D image format.
And 3, for each subject, taking an experiment as a unit, converting the data of each subject into 40 × 60 × 32 × 3 (experiment times × segment number × picture), setting the number of the rest experiments to be 30, sending the data into SBS to screen out the experiment combination with the highest accuracy, sending the screened experiment data into CWGAN-GP to pass through a labeled generative countermeasure network to perform data enhancement to generate data of a specified label, and supplementing the data to balance the emotion data of each category.
And 4, combining samples of every 3 segments, namely every 3 seconds, into one frame for 60 segments of each experiment, wherein the samples are not repeated from frame to frame, one experiment can generate 20 frames, the total number of the samples becomes 40 × 20 × 32 × 25600, the format of the samples is 25600 × 3 × 32 × 3, and all the samples are divided into 8: 2 into training samples and test samples. And simultaneously inputting 3 pictures to 3 convolutional neural networks each time, wherein the convolutional neural networks are the same in configuration, and averaging the 3 output results to obtain the final input result. All samples were run as 8: 2 training samples and testing samples.
As shown in fig. 4;
layer 4-1.C1 is a convolutional layer with RELU, with 32 filters, convolutional kernel set to 3 x 3, step size 1 x 1.
The S2 level is an average pooling level of 2 x 2 in size with steps of 2 x 2.
The layer 4-3.B3 is a BN layer.
Layer C4 is also a convolutional layer with RELU, with 64 filters, convolution kernel set to 3 x 3, step size 1 x 1.
The S5 level is an average pooling level of 2 x 2 in size with steps of 2 x 2.
Layer 4-6.B6 is a BN layer.
Layer C7 is a convolutional layer with RELU, with 128 filters, convolutional kernel set to 3 x 3, step size 1 x 1.
The layer 4-8.D8 is a fully connected layer with an output dimension of 128.
The DO9 layer is a Dropout layer that randomly turns off some of the neurons with a probability of 0.5 during the training process.
4-10. finally, the output dimension of the finally concatenated Softmax layer is 2.
Table 1 is a comparison of the two classification effects of a traditional classification model based on DEAP data set and Russell emotion dimension model:
TABLE 1 comparison of classification accuracy for each classification model
The samples extracted in proportion are used as training samples, the rest samples are used as test samples, the accuracy in the process of classifying the CNN model respectively according to two standards of the pleasure degree and the arousal degree is input into the CNN model in a frame mode, considerable results are obtained, the average accuracy reaches 83.3% and 84.2%, and the method is obviously improved compared with other classification algorithms.
Claims (2)
1. A CNN-based emotion electroencephalogram classification method is characterized by comprising the following steps:
step 1, establishing an emotion dimension model represented by Russell, and labeling evaluation indexes labels in a data set according to required categories;
step 2, extracting the baseline data 3 seconds before each experiment, segmenting the baseline data by a selected time window, adding the segmented baseline data, averaging the segmented baseline data, segmenting the rest electroencephalogram data by the selected time window, subtracting the processed baseline data to obtain the difference value of the electroencephalogram data, and performing data normalization; extracting 3 frequency bands most relevant to emotion recognition from the difference values of the electroencephalogram data, calculating Pearson coefficients of the 3 frequency bands, and converting the Pearson coefficients into a 32 × 3 picture format;
step 3, for each subject, taking an experiment as a unit, converting the data of each subject into the experiment times, the segment number and the segment number 3, setting the residual number after screening, sending the residual number into a sequence, and then selecting the experiment combination with the highest screening accuracy rate;
step 4, distributing the processed data and the labels according to the proportion of the set training sample and the set test sample, and inputting the data and the labels into the integrated convolutional neural network in a frame form to perform classification test on the processed samples;
inputting the data obtained in the step 3 into a convolutional neural network in a frame form, specifically: for 60 segments of each experiment, combine every 3 segments, i.e., every 3 seconds of samples into one frame, with no repeated samples from frame to frame, one experiment generates 20 frames, and the total number of samples becomes: number of experiments 20 number of subjects, sample format: the total number of samples is 3 x 32 x 3, and all the samples are in proportion to the set training samples and the set testing samples; simultaneously inputting 3 pictures to 3 convolutional neural networks each time, wherein the convolutional neural networks are the same in configuration, and averaging the 3 output results to obtain the final input result;
the integrated convolutional neural network is composed of three convolutional neural networks with completely same structures, and the structure of each convolutional neural network is as follows:
the C1 layer is a convolutional layer with RELU, with 32 filters, convolutional kernel set to 3 x 3, step size 1 x 1;
the S2 level is an average pooling level with a size of 2 x 2 and a step size of 2 x 2;
the B3 layer is a BN layer;
c4 layer is also a convolutional layer with RELU, with 64 filters, convolutional kernel set to 3 x 3, step size 1 x 1;
the S5 level is an average pooling level with a size of 2 x 2 and a step size of 2 x 2;
the B6 layer is a BN layer;
the C7 layer is a convolutional layer with RELU, with 128 filters, convolutional kernel set to 3 x 3, step size 1 x 1;
the D8 layer is a fully connected layer with an output dimension of 128;
the DO9 layer is a Dropout layer and randomly turns off part of neurons with a probability of 0.5 in the training process;
the output dimension of the finally concatenated Softmax layer is 2.
2. The CNN-based emotional feature classification method according to claim 1, wherein: in the first step, labeling the evaluation indexes labels in the data set according to the required categories, specifically: when the Russell emotion dimension model is used as a reference, each emotion state is placed on a two-dimensional plane, the arousal degree and the pleasure degree form two dimensions of a coordinate axis, the valence represents the subjective evaluation of a main body on emotion, and the passive emotion is changed into the active emotion on the coordinate axis; the wakefulness degree represents the degree of feeling of the main body on the emotion and changes from calmness to excitement on a numerical axis; the numerical values of the two dimensions are changed from 1 to 9, and the degree of emotion is represented by the size of the numerical value; when selecting emotion classification, dividing evaluation values on valence and arousal degree by taking a median 5 as a threshold, marking the evaluation values larger than 5 as 1, and representing high valence HV or high arousal HA; less than or equal to 5 is labeled 0, representing low titer LV or low wakefulness LA; when selecting mood quartiles, HV, HA, LV, LA were combined to generate four classes of tags for HVHA, HVLA, LVHA, LVLA.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010582404.1A CN111832431A (en) | 2020-06-23 | 2020-06-23 | Emotional electroencephalogram classification method based on CNN |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010582404.1A CN111832431A (en) | 2020-06-23 | 2020-06-23 | Emotional electroencephalogram classification method based on CNN |
Publications (1)
Publication Number | Publication Date |
---|---|
CN111832431A true CN111832431A (en) | 2020-10-27 |
Family
ID=72898051
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010582404.1A Pending CN111832431A (en) | 2020-06-23 | 2020-06-23 | Emotional electroencephalogram classification method based on CNN |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111832431A (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112633365A (en) * | 2020-12-21 | 2021-04-09 | 西安理工大学 | Mirror convolution neural network model and motor imagery electroencephalogram recognition algorithm |
CN112966566A (en) * | 2021-02-05 | 2021-06-15 | 武汉中旗生物医疗电子有限公司 | Electrocardiosignal baseline filtering method and device |
CN112990342A (en) * | 2021-04-08 | 2021-06-18 | 重庆大学 | Semi-supervised SAR target recognition method |
CN113191429A (en) * | 2021-04-29 | 2021-07-30 | 国网河北省电力有限公司电力科学研究院 | Power transformer bushing fault diagnosis method and device |
CN114403877A (en) * | 2022-01-21 | 2022-04-29 | 中山大学 | Multi-physiological-signal emotion quantitative evaluation method based on two-dimensional continuous model |
CN114745675A (en) * | 2022-04-28 | 2022-07-12 | 重庆邮电大学 | Wi-Fi indoor positioning method based on improved GAN combined hypothesis test |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20190035368A (en) * | 2017-09-26 | 2019-04-03 | 연세대학교 산학협력단 | Method and Apparatus for Recognizing Emotion Based on Image Converted from Brain Signal |
CN110110584A (en) * | 2019-03-14 | 2019-08-09 | 杭州电子科技大学 | A kind of emotional characteristics classification method based on CNN |
CN110353702A (en) * | 2019-07-02 | 2019-10-22 | 华南理工大学 | A kind of emotion identification method and system based on shallow-layer convolutional neural networks |
-
2020
- 2020-06-23 CN CN202010582404.1A patent/CN111832431A/en active Pending
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20190035368A (en) * | 2017-09-26 | 2019-04-03 | 연세대학교 산학협력단 | Method and Apparatus for Recognizing Emotion Based on Image Converted from Brain Signal |
CN110110584A (en) * | 2019-03-14 | 2019-08-09 | 杭州电子科技大学 | A kind of emotional characteristics classification method based on CNN |
CN110353702A (en) * | 2019-07-02 | 2019-10-22 | 华南理工大学 | A kind of emotion identification method and system based on shallow-layer convolutional neural networks |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112633365A (en) * | 2020-12-21 | 2021-04-09 | 西安理工大学 | Mirror convolution neural network model and motor imagery electroencephalogram recognition algorithm |
CN112633365B (en) * | 2020-12-21 | 2024-03-19 | 西安理工大学 | Mirror convolution neural network model and motor imagery electroencephalogram recognition algorithm |
CN112966566A (en) * | 2021-02-05 | 2021-06-15 | 武汉中旗生物医疗电子有限公司 | Electrocardiosignal baseline filtering method and device |
CN112966566B (en) * | 2021-02-05 | 2023-07-07 | 武汉中旗生物医疗电子有限公司 | Electrocardiosignal baseline filtering method and device |
CN112990342A (en) * | 2021-04-08 | 2021-06-18 | 重庆大学 | Semi-supervised SAR target recognition method |
CN112990342B (en) * | 2021-04-08 | 2023-09-19 | 重庆大学 | Semi-supervised SAR target recognition method |
CN113191429A (en) * | 2021-04-29 | 2021-07-30 | 国网河北省电力有限公司电力科学研究院 | Power transformer bushing fault diagnosis method and device |
CN114403877A (en) * | 2022-01-21 | 2022-04-29 | 中山大学 | Multi-physiological-signal emotion quantitative evaluation method based on two-dimensional continuous model |
CN114745675A (en) * | 2022-04-28 | 2022-07-12 | 重庆邮电大学 | Wi-Fi indoor positioning method based on improved GAN combined hypothesis test |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111832431A (en) | Emotional electroencephalogram classification method based on CNN | |
CN110069958B (en) | Electroencephalogram signal rapid identification method of dense deep convolutional neural network | |
CN111329474B (en) | Electroencephalogram identity recognition method and system based on deep learning and information updating method | |
CN114052735B (en) | Deep field self-adaption-based electroencephalogram emotion recognition method and system | |
CN110353702A (en) | A kind of emotion identification method and system based on shallow-layer convolutional neural networks | |
Alqudah et al. | Segmented and non-segmented skin lesions classification using transfer learning and adaptive moment learning rate technique using pretrained convolutional neural network | |
CN111709267B (en) | Electroencephalogram signal emotion recognition method of deep convolutional neural network | |
CN111832416A (en) | Motor imagery electroencephalogram signal identification method based on enhanced convolutional neural network | |
Kaziha et al. | A convolutional neural network for seizure detection | |
CN110135244B (en) | Expression recognition method based on brain-computer collaborative intelligence | |
CN113128552A (en) | Electroencephalogram emotion recognition method based on depth separable causal graph convolution network | |
CN112932501B (en) | Method for automatically identifying insomnia based on one-dimensional convolutional neural network | |
CN110717423A (en) | Training method and device for emotion recognition model of facial expression of old people | |
CN115221969A (en) | Motor imagery electroencephalogram signal identification method based on EMD data enhancement and parallel SCN | |
CN111920420A (en) | Patient behavior multi-modal analysis and prediction system based on statistical learning | |
CN113180659A (en) | Electroencephalogram emotion recognition system based on three-dimensional features and cavity full convolution network | |
Thomas et al. | Artificial neural network for diagnosing autism spectrum disorder | |
Micheli-Tzanakou et al. | Neural networks and blood cell identification | |
CN113128353A (en) | Emotion sensing method and system for natural human-computer interaction | |
Malviya | Tumor detection in MRI images using modified multi-level Otsu Thresholding (MLOT) and cross-correlation of principle components | |
CN116421200A (en) | Brain electricity emotion analysis method of multi-task mixed model based on parallel training | |
CN112438741A (en) | Driving state detection method and system based on electroencephalogram feature transfer learning | |
Immanuel et al. | Analysis of different emotions with bio-signals (EEG) using deep CNN | |
CN115527275A (en) | Behavior identification method based on P2CS _3DNet | |
Suttapakti et al. | Multi-directional texture feature extraction for glaucoma classification from color retinal images |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |