CN114089834A - Electroencephalogram identification method based on time-channel cascade Transformer network - Google Patents

Electroencephalogram identification method based on time-channel cascade Transformer network Download PDF

Info

Publication number
CN114089834A
CN114089834A CN202111614470.3A CN202111614470A CN114089834A CN 114089834 A CN114089834 A CN 114089834A CN 202111614470 A CN202111614470 A CN 202111614470A CN 114089834 A CN114089834 A CN 114089834A
Authority
CN
China
Prior art keywords
time
module
layer
output
channel
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202111614470.3A
Other languages
Chinese (zh)
Other versions
CN114089834B (en
Inventor
周文晖
王宇涵
莫良言
孔万增
戴国骏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Dianzi University
Original Assignee
Hangzhou Dianzi University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hangzhou Dianzi University filed Critical Hangzhou Dianzi University
Priority to CN202111614470.3A priority Critical patent/CN114089834B/en
Publication of CN114089834A publication Critical patent/CN114089834A/en
Application granted granted Critical
Publication of CN114089834B publication Critical patent/CN114089834B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/015Input arrangements based on nervous system activity detection, e.g. brain waves [EEG] detection, electromyograms [EMG] detection, electrodermal response detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/047Probabilistic or stochastic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/08Feature extraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/12Classification; Matching

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Biophysics (AREA)
  • Computing Systems (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Probability & Statistics with Applications (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Dermatology (AREA)
  • Neurology (AREA)
  • Neurosurgery (AREA)
  • Human Computer Interaction (AREA)
  • Measurement And Recording Of Electrical Phenomena And Electrical Characteristics Of The Living Body (AREA)

Abstract

The invention discloses an electroencephalogram identification method based on a time-channel cascade transducer network. The invention comprises the following steps: 1: acquiring ideogram English character electroencephalogram data, and constructing a preprocessing module; 2: constructing a time module of a time-channel cascade transducer network, wherein input data of the time module is preprocessed electroencephalogram data, and output is extracted time characteristics; 3: constructing a brain electrical channel module of a time-channel cascade transducer network, wherein input data of the brain electrical channel module is time characteristics, and output is extracted time-space fusion characteristics; 4: and constructing a classification module of the time-channel cascade Transformer network, wherein the input of the classification module is a space-time fusion characteristic, and the output of the classification module is a classification result. The invention can effectively improve the identification accuracy of the character imagination electroencephalogram signal.

Description

Electroencephalogram identification method based on time-channel cascade Transformer network
Technical Field
The invention belongs to the field of brain-computer interfaces and deep learning, and particularly relates to an electroencephalogram identification method based on a time-channel cascade transducer network.
Background
Brain Computer Interface (BCI) is a cross technology involving multiple disciplines such as neuroscience, signal processing, pattern recognition, and the like. The human intention is recognized by detecting the brain nerve activity, the brain nerve activity is converted into a command for driving external equipment, the human body is replaced by limbs or language organs of a human body to realize the communication between the human body and the outside and the control over the external environment, and the human-computer interaction mode is novel. With the rapid development of the related fields, the brain-computer interface technology and the theoretical research have made remarkable progress, have attracted wide attention internationally, and become one of the research hotspots in the fields of biomedical engineering, computer technology, communication and the like.
In order to realize high-resolution activity mapping of the neuron network, brain neuron potential activity at corresponding positions of cerebral cortex can be recorded for a long time on a sub-millisecond time scale by implanting a group of closely-arranged Microelectrode Arrays (MEA) in the brain. A key advantage of MEA is the ability to simultaneously record and stimulate neurons at multiple sites. The electrode arrangement with high sensitivity and high stability is used for recording the signal of the target neuron cluster, the signal-to-noise ratio is higher, the time and the spatial resolution are good, the precise issuing time and the waveform of the neuron action potential can be simultaneously ensured to be recorded in a large range and at high precision, and a solid foundation is laid for fully extracting neural information and reading the activity of a cranial nerve network.
The nerve of a patient with nervous system diseases cannot be effectively communicated with the outside due to nerve damage of a certain part of a body, and the existing research shows that the mode of implanting a microelectrode array into a damaged brain area can be used for recovering lost functions through electrical stimulation, such as mechanical arm control, ideographic typing, ideographic speech control, touch arousing and the like, so that electroencephalogram signal recognition based on ideology is an important development direction of BCI. Through the analysis of the mind imagination electroencephalogram signals, the neural activity of the human brain in the imagination process can be identified, so that the thought and the intention of a patient with mobility disability can be transmitted to the outside, and the neural decoding is further realized. Therefore, the research on the ideogram electroencephalogram signal identification can promote the exploration on the cerebral nerve cognition and the cerebral disease rehabilitation, and has important research value and practical significance in the novel human-computer interaction field. The invention mainly aims at a character imagination electroencephalogram signal identification task in the field of idea imagination.
In recent years, with the wide application of deep learning in the research fields of computer vision, natural language processing and the like, the neural network has strong capability of processing nonlinear and high-dimensional data, and therefore, the neural network is also applied to data analysis of brain-computer interfaces. The invasive BCI has higher signal-to-noise ratio and good time and space resolution, and records signals of target neuron clusters by using an electrode arrangement with high sensitivity and high stability. The conventional algorithm only focuses on electroencephalogram characteristics under a time sequence when identifying electroencephalogram signals, and rarely focuses on importance of different characteristic channels, a used method such as a Convolutional Neural Network (CNN) has the problem of depending on selection of a convolutional kernel, and a Recurrent Neural Network (RNN) has the problem of being incapable of processing sequence parallelization and only focuses on previous records and current states. Therefore, the invention provides an electroencephalogram identification method based on a time-channel cascade transducer network, which utilizes a self-attention mechanism to extract characteristics of time dimension and electroencephalogram channel dimension information, and further extracts electroencephalogram channel characteristic information by fusing in a residual error cascade mode. The invention can effectively improve the recognition performance of character imagination electroencephalogram signals.
Disclosure of Invention
The invention provides an electroencephalogram identification method based on a time-channel cascade transducer network, which can effectively improve the identification accuracy of character imagination electroencephalogram signals.
The technical scheme adopted by the invention for solving the technical problem comprises the following steps as shown in figure 1:
step S1: acquiring ideogram English character electroencephalogram data, constructing a preprocessing module, and performing data preprocessing operation by using a time alignment technology, wherein the input of the preprocessing module is the ideogram English character electroencephalogram data, and the output of the preprocessing module is the preprocessed electroencephalogram data;
step S2: constructing a time module of the time-channel cascade Transformer network, wherein the input data of the time module is the preprocessed electroencephalogram data output in the step S1, and the output of the time module is the extracted time characteristics;
step S3: constructing a brain electrical channel module of the time-channel cascade transducer network, wherein the input data of the brain electrical channel module is the time characteristic output in the step S2, and the output of the brain electrical channel module is the extracted space-time fusion characteristic;
step S4: and constructing a classification module of the time-channel cascade Transformer network, wherein the input of the classification module is the space-time fusion characteristics output by the step S3, and the output of the classification module is the classification result.
The invention has the following beneficial effects:
the invention provides a electroencephalogram identification method based on a time-channel cascade transducer network, and a network block diagram of the method is shown in figure 2. The time-channel cascade transducer network comprises a time module, an electroencephalogram channel module and a classification module which are respectively used for acquiring time characteristics, space-time fusion characteristics and classification results. Meanwhile, the time-channel cascade transducer network adopts random position coding and increases classification identification bits so as to realize high-precision character imagination electroencephalogram classification performance.
Drawings
FIG. 1 is a schematic flow chart of the main steps of the present invention
FIG. 2 is a block diagram of a time-channel cascaded Transformer network
FIG. 3 is an internal structure diagram of a time-channel cascaded Transformer network module
Detailed Description
The invention is further illustrated by the following figures and examples.
As shown in fig. 1-3, a method for electroencephalogram identification based on a time-channel cascade transducer network includes the following steps:
step S1: acquiring ideogram English character electroencephalogram data, constructing a preprocessing module, and performing data preprocessing operation by using a time alignment technology, wherein the input of the preprocessing module is the ideogram English character electroencephalogram data, and the output of the preprocessing module is the preprocessed electroencephalogram data;
step S2: constructing a time module of the time-channel cascade Transformer network, wherein the input data of the time module is the preprocessed electroencephalogram data output in the step S1, and the output of the time module is the extracted time characteristics;
step S3: constructing a brain electrical channel module of the time-channel cascade transducer network, wherein the input data of the brain electrical channel module is the time characteristic output in the step S2, and the output of the brain electrical channel module is the extracted space-time fusion characteristic;
step S4: and constructing a classification module of the time-channel cascade Transformer network, wherein the input of the classification module is the space-time fusion characteristics output by the step S3, and the output of the classification module is the classification result.
The step S1 includes:
the method for acquiring the brain electrical data of the English characters by the aid of the ideological imagery is a mature technology, and the size of the data is 201 × 192 generally.
The preprocessing module adopts a time alignment technology (F.R. Willett, D.T. Avansino, L.R. Hochberg, et al.high-performance bridge-to-text communication video writing) as a mature technology, and is used for eliminating the problem of inconsistency of the writing speed of ideographic imagination characters, and the size of output data is consistent with that of input data.
The step S2 includes:
the time module sequentially comprises the following structures: the preprocessed data output in step S1 is used as input feature- > position coding layer- > multi-head self-attention mechanism module- > residual connecting layer- > LN regularization layer- > feed-forward network- > residual connecting layer- > LN regularization layer- > output time feature i. The time module executes 2 times in a co-loop manner, and the time characteristic I- > multi-head self-attention mechanism module- > residual error connecting layer- > LN regularization layer- > feedforward network- > residual error connecting layer- > LN regularization layer- > output time characteristic II for the first time.
Referring to fig. 3, the position coding layer adopts a random position coding method, and is used for learning the position relationship of the electroencephalogram data in the time dimension in the network training. The position coding layer outputs a random number matrix with the same format as the input data of the position coding layer, and the random number matrix is added with the input data of the position coding layer to be used as the input data of the multi-head self-attention mechanism module.
The multi-head self-attention mechanism module maps the input features to different subspaces, and then performs point multiplication operation on all the subspaces to calculate the attention vector. And finally, splicing and mapping the attention vectors calculated in all the subspaces to an original input space to obtain a final attention vector as an output. This may allow the model to learn the relevant information in different representation subspaces. The expression of the multi-head self-attention mechanism module is as follows (1):
Figure BDA0003436052220000041
wherein
Figure BDA0003436052220000042
i represents different subspaces, query vector Q, key vector K and value vector V as input of multi-headed self-attention module, Wi QMapping matrices for Q in different subspaces, Wi KMapping matrix for K in different subspaces, Wi VMapping matrix for V in different subspaces, WOFrom W in all subspacesi VAnd (4) splicing to obtain the finished product. The calculation mode of the attention vector on the independent subspace is as follows in sequence: firstly, the query vector Q and the key vector K are subjected to dot product operation, and then the dot product operation is divided by the dimensional square root of the key vector K
Figure BDA0003436052220000051
Obtaining a fractional matrix of the query vector Q, wherein the Softmax function has good perception capability, utilizing the Softmax function to normalize to obtain a weight matrix,and multiplying the value vector V to obtain the attention vector of a subspace, wherein the expression is as the following formula (2):
Figure BDA0003436052220000052
for the data set adopted by the invention, the dimension d of the parameter matrix of Q, K, Vq,dkAnd d andvare all 128, the number of heads is 16, dmodelIs 256. We transform the query vector Q from d by a linear transformationmodelDimension mapping as dqHead, from the key vector KmodelDimension mapping as dkHead, vector of values V from dmodelDimension mapping as dvHead. By implicitly increasing the number of attention heads without reducing the hidden dimension assigned to each attention head, global features can be efficiently extracted and classification accuracy can be improved. The residual connecting layer is mainly used for residual connection, and the method effectively solves the problems of gradient disappearance and gradient explosion; the LN regularization layer is mainly used for normalizing input data; the Feed-Forward Network module (FFN) is composed of two layers of Feed-Forward neural networks, wherein the first layer of Feed-Forward neural Network is used for inputting features from dmodelThe dimensionality is mapped into 512 dimensionality, the activation function is a GELU function, and the second-layer feedforward neural network is mapped back to d from the 512 dimensionalitymodelDimension, no activation function is used. The expression of the feedforward network is as follows (3):
FFN(x)=max(0,xW1+b1)W2+b2 (3)
wherein W1And W2Is a randomly initialized weight vector, b1And b2Is a randomly initialized bias.
The interior of a time module based on a time-channel cascade transducer is completely dependent on an attention machine mechanism for modeling, a multi-head attention machine mechanism is utilized to solve the long-distance dependence problem existing in a Recurrent Neural Network (RNN) and a variant thereof, the importance of each time step of a brain electrical signal in a global sequence is concerned, global context information is effectively captured, and the method has better feature extraction capability. The structure of the model avoids the mode that CNN stacks convolution layers to obtain global information, and the model can have good performance.
The step S3 includes:
and constructing a time-channel cascade Transformer electroencephalogram channel module, wherein input data of the electroencephalogram channel module are time characteristics, the size of the time-channel cascade Transformer electroencephalogram channel module is 201 x 192, and output of the electroencephalogram channel module is extracted space-time fusion characteristics, and the size of the time-channel cascade Transformer electroencephalogram channel module is 196 x 256.
Fusing the time characteristics output by the step S2 and the preprocessed electroencephalogram data output by the step S1 in a residual cascade mode, wherein the size of the fused characteristics is 201 x 192,
the electroencephalogram channel module structure based on time-channel cascade Transformer sequentially comprises the following steps: the time characteristic output in the step S2 is used as an input characteristic- > characteristic cascade layer- > real layer- > Linear layer- > classification identification bit layer (performing splicing class token operation) - > multi-head self-attention mechanism module- > residual error connecting layer- > LN regularization layer- > feedforward network- > residual error connecting layer- > LN regularization layer- > output space-time fusion characteristic i, and the module locally circulates for 2 times; and during the second local circulation, the space-time fusion characteristic I- > a multi-head self-attention mechanism module- > a residual error connecting layer- > an LN regularization layer- > a feedforward network- > a residual error connecting layer- > an LN regularization layer- > outputs a space-time fusion characteristic II.
The feature cascade layer is formed by connecting the time features output in the step S2 with the preprocessed electroencephalogram data output in the step S1 through residual operation, and the size of the obtained fusion features is 201 x 192; the Rearrange layer is dimensional transformation and converts the fused features from 201 by 192 to 192 by 201; the Linear layer performs feature mapping for Linear operation, and maps the fusion features from 201 dimension to 256 dimension; the classification identification bit layer adopts a method for classifying identification bits in a paper (Dosovitskiy A, Beyer L, Kolesnikov A, et al. an image is 16x16 words: transformations for image recognition at scale), which is a mature technology. The method has the function that the information of the whole channel sequence can be acquired through an attention mechanism in the training process, so that the influence of the original sequence node can be effectively avoided. The four bits of the flag are concatenated in front of the dimension of the fused feature 192, at which time the fused feature is transformed into 196 x 256. The subsequent structures are kept consistent with the time module in step S2. And further extracting the spatial characteristics of the electroencephalogram data through a channel module to finally obtain the space-time fusion characteristics.
The step S4 includes:
and constructing a time-channel cascade Transformer classification module, wherein the input of the classification module is the space-time fusion characteristics output by the step S3, the size of the classification module is 196 × 256, and the output of the classification module is a classification result.
The structure of a time-channel cascade transform-based classification module sequentially comprises the following steps: the space-time fusion feature output in step S3 is used as an input feature- > capture classification flag- > Reduce layer- > LN regularization layer- > Linear layer (3 layers) - > output classification result.
Intercepting four classification identification bits with the size of 4 x 256 from the space-time fusion feature obtained in the step S3; reduce layer is reduced to 1 × 256 on 4 × 256; the Linear layer maps input features from 256 to 256 x 2 dimensions, the second layer maps the input features from 256 x 2 dimensions back to 256 dimensions, and the third layer maps the input features from 256 dimensions to 26 dimensions for classification. The activation function used between the two layers is a GELU function. The corresponding classification label is output through the classification module, and the loss function is calculated through comparison with the real label, the cross entropy loss function is adopted in the invention, and the specific formula is as follows:
Figure BDA0003436052220000071
where M is the number of trials, N is the number of categories,
Figure BDA0003436052220000072
the true label of the m-th trial,
Figure BDA0003436052220000073
the predicted probability of class n trial m is shown. When the method is used specifically, Adam with high convergence rate is used as an optimizer, and the initial learning rate is setSet to 7e-5 and batch size 8.
The comparative results are shown in table 1;
table 1: accuracy enhancement of form parameters compared to existing methods
Figure BDA0003436052220000074

Claims (10)

1. A electroencephalogram identification method based on a time-channel cascade transducer network is characterized by comprising the following steps:
step S1: acquiring ideogram English character electroencephalogram data, constructing a preprocessing module, and performing data preprocessing operation by using a time alignment technology, wherein the input of the preprocessing module is the ideogram English character electroencephalogram data, and the output of the preprocessing module is the preprocessed electroencephalogram data;
step S2: constructing a time module of the time-channel cascade Transformer network, wherein the input data of the time module is the preprocessed electroencephalogram data output in the step S1, and the output of the time module is the extracted time characteristics;
step S3: constructing a brain electrical channel module of the time-channel cascade transducer network, wherein the input data of the brain electrical channel module is the time characteristic output in the step S2, and the output of the brain electrical channel module is the extracted space-time fusion characteristic;
step S4: and constructing a classification module of the time-channel cascade Transformer network, wherein the input of the classification module is the space-time fusion characteristics output by the step S3, and the output of the classification module is the classification result.
2. The electroencephalogram identification method based on the time-channel cascade transducer network, according to claim 1, characterized in that the preprocessing module adopts a time alignment technique to make the size of the output data consistent with the size of the input data.
3. The electroencephalogram identification method based on the time-channel cascade transducer network according to claim 1, characterized in that the time modules sequentially have the following structures: the preprocessed data output by the step S1 is used as input features- > position coding layer- > multi-head self-attention mechanism module- > residual connecting layer- > LN regularization layer- > feed-forward network- > residual connecting layer- > LN regularization layer- > output time features i; the time module is executed for 2 times in a circulating mode, and the time characteristic I- > multi-head self-attention mechanism module- > residual error connecting layer- > LN regularization layer- > feedforward network- > residual error connecting layer- > LN regularization layer- > output time characteristic II for the first time.
4. The electroencephalogram identification method based on the time-channel cascade transducer network, according to claim 1, characterized in that a multi-head self-attention mechanism module maps input features to different subspaces, and then performs point multiplication operation on all the subspaces to calculate attention vectors; finally, splicing and mapping the attention vectors calculated in all the subspaces to an original input space to obtain a final attention vector as output; the expression of the multi-head self-attention mechanism module is as follows (1):
Figure FDA0003436052210000021
wherein,
Figure FDA0003436052210000022
representing different subspaces, the query vector Q, the key vector K, and the value vector V as inputs to a multi-headed self-attention module,
Figure FDA0003436052210000023
for the mapping matrix of Q in different subspaces,
Figure FDA0003436052210000024
for the mapping matrix of K in the different subspaces,
Figure FDA0003436052210000025
mapping matrix for V in different subspaces, WOFrom all subspaces
Figure FDA0003436052210000026
Splicing to obtain the finished product; the calculation mode of the attention vector on the independent subspace is as follows in sequence: firstly, the query vector Q and the key vector K are subjected to dot product operation, and then the dot product operation is divided by the dimensional square root of the key vector K
Figure FDA0003436052210000027
Obtaining a fractional matrix of the query vector Q, wherein the Softmax function has good perception capability, normalizing by using the fractional matrix to obtain a weight matrix, and multiplying by a value vector V to obtain an attention vector of a subspace, wherein the expression is as the following formula (2):
Figure FDA0003436052210000028
wherein Q, K, V parameter matrix dimension dq、dkAnd dvAre all 128, the number of heads is 16, dmodelIs 256; by linear transformation, the query vector Q is converted from dmodelDimension mapping as dqHead, from the key vector KmodelDimension mapping as dkHead, vector of values V from dmodelDimension mapping as dv*head。
5. The EEG identification method based on time-channel cascade Transformer network as claimed in claim 3 or 4, wherein the input data of EEG channel module is time characteristic with size of 201 × 192, and the output of EEG channel module is extracted time-space fusion characteristic with size of 196 × 256.
6. The electroencephalogram identification method based on the time-channel cascade Transformer network, according to claim 5, characterized in that the time features output in the step S2 and the preprocessed electroencephalogram data output in the step S1 are fused in a residual cascade mode, and the size of the fused features is 201 x 192.
7. The electroencephalogram identification method based on the time-channel cascade transducer network, according to claim 5, is characterized in that the electroencephalogram channel module structure sequentially comprises: the time characteristic output by the step S2 is used as an input characteristic- > characteristic cascade layer- > real layer- > Linear layer- > classification identification bit layer- > multi-head self-attention machine system module- > residual error connecting layer- > LN regularization layer- > feed-forward network- > residual error connecting layer- > LN regularization layer- > output space-time fusion characteristic i, and the module locally circulates for 2 times; and during the second local circulation, the space-time fusion characteristic I- > a multi-head self-attention mechanism module- > a residual error connecting layer- > an LN regularization layer- > a feedforward network- > a residual error connecting layer- > an LN regularization layer- > outputs a space-time fusion characteristic II.
8. The electroencephalogram identification method based on the time-channel cascade transducer network, according to claim 5, characterized in that the feature cascade layer connects the time features output in the step S2 with the preprocessed electroencephalogram data output in the step S1 through residual error operation, and the size of the obtained fusion features is 201 x 192; the Rearrange layer is dimensional transformation and converts the fused features from 201 by 192 to 192 by 201; the Linear layer performs feature mapping for Linear operation, and maps the fusion features from 201 dimension to 256 dimension; the classification identification bit layer can acquire the information of the whole channel sequence; splicing four identification bits in front of the dimension of the fused feature 192, wherein the fused feature is transformed into 196 × 256; the subsequent structures are all consistent with the time module in step S2; and further extracting the spatial characteristics of the electroencephalogram data through a channel module to finally obtain the space-time fusion characteristics.
9. The EEG identification method based on time-channel cascade Transformer network as claimed in claim 7 or 8, wherein the input of the classification module is the space-time fusion feature output in step S3, the size is 196 × 256, and the output of the classification module is the classification result;
the structure of a time-channel cascade transform-based classification module sequentially comprises the following steps: the space-time fusion feature output in step S3 is used as an input feature- > capture classification flag- > Reduce layer- > LN regularization layer- > Linear layer (3 layers) - > output classification result.
10. The electroencephalogram recognition method based on the time-channel cascade transducer network, according to claim 9, characterized in that four classification identification bits with a size of 4 × 256 are intercepted from the space-time fusion features obtained in step S3; reduce layer is reduced to 1 × 256 on 4 × 256; linear is used for Linear operation to carry out feature mapping, a first layer of Linear layer maps input features from 256 to 256 x 2 dimensions, a second layer of Linear layer maps the 256 x 2 dimensions back to the 256 dimensions, and a third layer of Linear layer maps the 256 dimensions to 26 dimensions for classification; an activation function is used between the two layers as a GELU function; outputting a corresponding classification label through a classification module, calculating a loss function by comparing the classification label with a real label, and adopting a cross entropy loss function, wherein a specific formula is as follows:
Figure FDA0003436052210000041
where M is the number of trials, N is the number of categories,
Figure FDA0003436052210000042
the true label of the m-th trial,
Figure FDA0003436052210000043
the predicted probability of class n trial m is shown.
CN202111614470.3A 2021-12-27 2021-12-27 Electroencephalogram identification method based on time-channel cascade converter network Active CN114089834B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111614470.3A CN114089834B (en) 2021-12-27 2021-12-27 Electroencephalogram identification method based on time-channel cascade converter network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111614470.3A CN114089834B (en) 2021-12-27 2021-12-27 Electroencephalogram identification method based on time-channel cascade converter network

Publications (2)

Publication Number Publication Date
CN114089834A true CN114089834A (en) 2022-02-25
CN114089834B CN114089834B (en) 2024-07-12

Family

ID=80308061

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111614470.3A Active CN114089834B (en) 2021-12-27 2021-12-27 Electroencephalogram identification method based on time-channel cascade converter network

Country Status (1)

Country Link
CN (1) CN114089834B (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114767120A (en) * 2022-04-25 2022-07-22 上海韶脑传感技术有限公司 Depth learning-based selection method for motor imagery electroencephalogram channels of unilateral limb patients
CN115169227A (en) * 2022-07-04 2022-10-11 四川大学 Design concept generation network construction method and concept scheme automatic generation method
CN115192045A (en) * 2022-09-16 2022-10-18 季华实验室 Destination identification/wheelchair control method, device, electronic device and storage medium
CN116127364A (en) * 2023-04-12 2023-05-16 上海术理智能科技有限公司 Integrated transducer-based motor imagery decoding method and system
CN116502069A (en) * 2023-06-25 2023-07-28 四川大学 Haptic time sequence signal identification method based on deep learning
CN117272002A (en) * 2023-11-23 2023-12-22 中国电建集团西北勘测设计研究院有限公司 Solar radiation amount estimation method and device, electronic equipment and storage medium
CN118490232A (en) * 2024-07-17 2024-08-16 东北电力大学 Brain depression diagnosis method based on multi-frequency domain decomposition

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110309797A (en) * 2019-07-05 2019-10-08 齐鲁工业大学 Merge the Mental imagery recognition methods and system of CNN-BiLSTM model and probability cooperation
CN111681636A (en) * 2020-06-16 2020-09-18 深圳市华创技术有限公司 Technical term sound generation method based on brain-computer interface, medical system and terminal

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110309797A (en) * 2019-07-05 2019-10-08 齐鲁工业大学 Merge the Mental imagery recognition methods and system of CNN-BiLSTM model and probability cooperation
CN111681636A (en) * 2020-06-16 2020-09-18 深圳市华创技术有限公司 Technical term sound generation method based on brain-computer interface, medical system and terminal

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
CHAO LI: ""HIERARCHICAL ATTENTION-BASED TEMPORAL CONVOLUTIONAL NETWORKS FOR EEG-BASED EMOTION RECOGNITION"", 《ICASSP 2021》, 13 May 2021 (2021-05-13), pages 1 - 5 *
YI AN: ""Leveraging spatial-temporal convolutional features for EEG-based emotion recognition"", 《ELSEVIER》, 29 June 2021 (2021-06-29), pages 1 - 8 *

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114767120A (en) * 2022-04-25 2022-07-22 上海韶脑传感技术有限公司 Depth learning-based selection method for motor imagery electroencephalogram channels of unilateral limb patients
CN114767120B (en) * 2022-04-25 2024-05-10 上海韶脑传感技术有限公司 Single-side limb patient motor imagery electroencephalogram channel selection method based on deep learning
CN115169227A (en) * 2022-07-04 2022-10-11 四川大学 Design concept generation network construction method and concept scheme automatic generation method
CN115192045A (en) * 2022-09-16 2022-10-18 季华实验室 Destination identification/wheelchair control method, device, electronic device and storage medium
CN116127364A (en) * 2023-04-12 2023-05-16 上海术理智能科技有限公司 Integrated transducer-based motor imagery decoding method and system
CN116502069A (en) * 2023-06-25 2023-07-28 四川大学 Haptic time sequence signal identification method based on deep learning
CN116502069B (en) * 2023-06-25 2023-09-12 四川大学 Haptic time sequence signal identification method based on deep learning
CN117272002A (en) * 2023-11-23 2023-12-22 中国电建集团西北勘测设计研究院有限公司 Solar radiation amount estimation method and device, electronic equipment and storage medium
CN117272002B (en) * 2023-11-23 2024-02-20 中国电建集团西北勘测设计研究院有限公司 Solar radiation amount estimation method and device, electronic equipment and storage medium
CN118490232A (en) * 2024-07-17 2024-08-16 东北电力大学 Brain depression diagnosis method based on multi-frequency domain decomposition
CN118490232B (en) * 2024-07-17 2024-09-10 东北电力大学 Brain depression diagnosis method based on multi-frequency domain decomposition

Also Published As

Publication number Publication date
CN114089834B (en) 2024-07-12

Similar Documents

Publication Publication Date Title
CN114089834B (en) Electroencephalogram identification method based on time-channel cascade converter network
Aneja et al. Transfer learning using CNN for handwritten devanagari character recognition
CN108491077B (en) Surface electromyographic signal gesture recognition method based on multi-stream divide-and-conquer convolutional neural network
Xiao et al. An efficient temporal network with dual self-distillation for electroencephalography signal classification
CN112861604B (en) Myoelectric action recognition and control method irrelevant to user
CN114176607B (en) Electroencephalogram signal classification method based on vision transducer
CN112488205A (en) Neural network image classification and identification method based on optimized KPCA algorithm
CN114298216A (en) Electroencephalogram vision classification method based on time-frequency domain fusion Transformer
CN115381466A (en) Motor imagery electroencephalogram signal classification method based on AE and Transformer
Ma et al. A novel hybrid CNN-transformer model for EEG motor imagery classification
US20230101539A1 (en) Physiological electric signal classification processing method and apparatus, computer device and storage medium
CN115349860A (en) Multi-modal emotion recognition method, system, device and medium
CN114209342A (en) Electroencephalogram signal motor imagery classification method based on space-time characteristics
CN112932504B (en) Dipole imaging and identifying method
Li et al. A novel motor imagery EEG recognition method based on deep learning
CN111967326B (en) Gait recognition method based on lightweight multi-scale feature extraction
CN113180695A (en) Brain-computer interface signal classification method, system, device and storage medium
CN110432899B (en) Electroencephalogram signal identification method based on depth stacking support matrix machine
Gao et al. Chinese fingerspelling sign language recognition using a nine-layer convolutional neural network
Sun et al. Training-free deep generative networks for compressed sensing of neural action potentials
Wang et al. Calibration-Free Transfer Learning for EEG-Based Cross-Subject Motor Imagery Classification
Liu et al. Spatial-temporal convolutional attention for mapping functional brain networks
CN116458896A (en) Electrocardiogram classification method and device based on time sequence feature diagram and attention mechanism
CN114936583A (en) Teacher-student model-based two-step field self-adaptive cross-user electromyogram pattern recognition method
CN114944002B (en) Text description-assisted gesture-aware facial expression recognition method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant