CN111539331A - Visual image reconstruction system based on brain-computer interface - Google Patents

Visual image reconstruction system based on brain-computer interface Download PDF

Info

Publication number
CN111539331A
CN111539331A CN202010330551.XA CN202010330551A CN111539331A CN 111539331 A CN111539331 A CN 111539331A CN 202010330551 A CN202010330551 A CN 202010330551A CN 111539331 A CN111539331 A CN 111539331A
Authority
CN
China
Prior art keywords
electroencephalogram
image
feature
picture
vector
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010330551.XA
Other languages
Chinese (zh)
Other versions
CN111539331B (en
Inventor
潘红光
董娜
温帆
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xian University of Science and Technology
Original Assignee
Xian University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xian University of Science and Technology filed Critical Xian University of Science and Technology
Priority to CN202010330551.XA priority Critical patent/CN111539331B/en
Publication of CN111539331A publication Critical patent/CN111539331A/en
Application granted granted Critical
Publication of CN111539331B publication Critical patent/CN111539331B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/02Preprocessing
    • G06F2218/04Denoising
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/08Feature extraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/12Classification; Matching

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Signal Processing (AREA)
  • Measurement And Recording Of Electrical Phenomena And Electrical Characteristics Of The Living Body (AREA)

Abstract

The invention discloses a visual image reconstruction system based on a brain-computer interface, which can realize real-time image reconstruction based on an electroencephalogram signal and comprises the following steps: an electroencephalogram signal acquisition subsystem; an electroencephalogram signal feature extraction and classification subsystem; an image encoder; an image decoder; a discriminator; an electroencephalogram feature mapper; an image quality evaluation subsystem. The electroencephalogram signal feature extraction and classification subsystem extracts and classifies electroencephalogram signal features; the image encoder encodes the graphics. The image decoder performs decoding of the feature vector. The discriminator discriminates the decoded picture by convolution. The electroencephalogram feature mapper converts the electroencephalogram signal feature vector into a feature vector of a potential space in an image decoder through an electroencephalogram feature mapping network. The image quality evaluation is to perform non-reference image quality evaluation on the reconstructed picture of the electroencephalogram signal. The invention can restore and reconstruct the image information seen by vision and understand the electroencephalogram signals induced by specific stimulation.

Description

Visual image reconstruction system based on brain-computer interface
Technical Field
The invention relates to the technical field of brain-computer interface visual image reconstruction, in particular to a visual image reconstruction system based on a brain-computer interface.
Background
Generally, vision is one of basic functions of human interaction with the environment, and 80% of information externally received by human is visual information, so that the visual information occupies an important position. In recent decades, with the continuous development of cognitive neuroscience, people can observe brain activation modes under different cognitive tasks through a neuroimaging technology, so that a neurocoding mechanism of a cerebral cortex for processing visual information and decoding corresponding cognitive states of the cerebral cortex are researched. And the machine learning provides powerful theoretical and technical support for the encoding and decoding of images. At present, visual image recognition and reconstruction in the field are suitable for decoding based on image similarity, but accurate information about objects seen or imagined by people is rarely provided, and the research of image reconstruction based on electroencephalogram signals is still in the primary stage and does not form a systematic result.
Disclosure of Invention
In order to solve the problems in the background art, the invention provides a visual image reconstruction system based on a brain-computer interface.
The invention has the beneficial effects that: the image information seen by the vision is restored and reconstructed, and the electroencephalogram signals induced by specific stimulation are understood.
In order to achieve the purpose, the invention adopts the technical scheme that:
a brain-computer interface based visual image reconstruction system, comprising:
the electroencephalogram signal acquisition subsystem mainly comprises an electroencephalogram cap, an electroencephalogram signal amplifier, a wavelet converter, a band-pass filter and electroencephalogram signal recording software, wherein the electroencephalogram cap is used for acquiring electroencephalogram signals generated by a brain, the electroencephalogram signal amplifier is used for amplifying the electroencephalogram signals, the wavelet converter is used for denoising the electroencephalogram signals, the band-pass filter is used for filtering the electroencephalogram signals, and the electroencephalogram signal recording software is used for recording the electroencephalogram signals;
the electroencephalogram signal feature extraction and classification subsystem mainly comprises an electroencephalogram signal feature extraction module and an electroencephalogram signal feature classification module, and is used for extracting and classifying the characteristics of the electroencephalogram signals output by the electroencephalogram signal acquisition subsystem;
the image encoder is used for clipping the video causing the visual stimulation and converting the video into an input picture, and then encoding the input picture to obtain an image feature vector;
the electroencephalogram feature mapper converts the electroencephalogram feature vectors output by the electroencephalogram feature extraction and classification subsystem into feature vectors of a potential space in an image decoder, namely electroencephalogram feature mapping vectors, through an electroencephalogram feature mapping network;
the first image decoder decodes the image feature vector output by the image encoder to obtain a decoded picture, and the second image decoder decodes the electroencephalogram feature mapping vector to obtain a reconstructed picture of the electroencephalogram signal;
the discriminator judges the decoded picture through convolution and judges whether the decoded picture meets the picture output condition or not;
and the image quality evaluation subsystem is used for performing non-reference image quality evaluation on the reconstructed picture of the electroencephalogram signal according to the decoded picture meeting the picture output condition.
The visual reconstruction of the invention comprises the following steps:
acquiring electroencephalogram signals of a video watched by a subject, and performing feature extraction and classification on the electroencephalogram signals to obtain electroencephalogram feature vectors;
inputting the electroencephalogram feature vector into a feature mapper of an LSTM neural network, so that the electroencephalogram feature vector subjected to electroencephalogram feature mapping is equal to a potential space vector as much as possible, and an electroencephalogram feature mapping vector, namely an image feature vector is obtained;
firstly, editing a video watched by a subject to obtain an input picture of an encoder, and encoding the input picture through a VGG-11 neural network to obtain a feature vector of a potential space;
decoding the feature vector by deconvolution;
judging the decoded picture to see whether the decoded picture meets the output condition or not;
and reconstructing a video picture seen by a person, and performing non-parameter quality evaluation on the reconstructed picture.
Furthermore, the acquiring of the electroencephalogram signals, namely acquiring of the electroencephalogram signals, comprises acquiring electric waves by using an electroencephalogram cap with 256 channels, amplifying the electroencephalogram waves by using an electroencephalogram signal amplifier, denoising the electroencephalogram waves by using a wavelet transform method, filtering the electroencephalogram waves by using a band-pass filter, and recording the electroencephalogram signals by using electroencephalogram signal recording software.
Furthermore, the electroencephalogram feature extraction is a method combining principal component analysis and linear discrimination, the acquired electroencephalogram signals are firstly transformed into another group of electroencephalogram feature variables which are independent from each other by a projection transformation method through an orthogonal principle from originally related electroencephalogram signal variables, the group of electroencephalogram signals which are independent from each other are subjected to minimum intra-class spacing and maximum inter-class spacing calculation to obtain projection space vectors, and then the final electroencephalogram feature vectors are obtained. And carrying out electroencephalogram signal feature classification on the electroencephalogram feature vector by a naive Bayes method. Meanwhile, the video signal and the electroencephalogram signal are synchronized through the photoelectric sensor, so that the electroencephalogram signal generated when a subject watches the video can be recorded on a computer in real time;
further, inputting the electroencephalogram feature mapping vector into an image decoder to obtain a reconstructed picture;
further, the picture feature vector is input into a decoder, the picture processed by the image decoder is input into a discriminator, the input picture processed by the video is also input into the discriminator at the moment, and the two pictures are processed by the discriminator to obtain a decoded picture;
further, the discriminator is a neural network formed by combining 5 layers of convolution, 4 Relu activation functions, 3 BN and 1 full connection layer.
Furthermore, the picture seen by people is reconstructed in real time, and the reconstructed picture is displayed on a computer.
That is, the process of acquiring the reconstructed picture according to the present invention can be summarized as follows:
the testee watches the video, the brain generates an electroencephalogram signal, and the electroencephalogram signal is collected by an electroencephalogram collection system and subjected to feature extraction and classification. And simultaneously, video processing is carried out on the video watched by the subject, the video is converted into an input picture, and the input picture is decoded into an image feature vector through an image decoder. The electroencephalogram feature mapper can find out the mapping relation between the vector of electroencephalogram feature extraction and classification and the image feature vector through training, and the electroencephalogram feature mapping vector is obtained in a one-to-one correspondence mode. And when the output of the decoder meets the output standard, the decoded picture is output. And obtaining a reconstructed picture by the electroencephalogram feature vector through a decoder. The quality of the reconstructed picture is improved by calculating the mean square loss of the image feature vector and the electroencephalogram feature mapping vector, calculating the pixel loss between the decoded picture and the reconstructed picture, and calculating the peak signal-to-noise ratio loss and the structural similarity loss of the reconstructed picture and the input picture. And finally, carrying out quality evaluation on the reconstructed picture.
Compared with the prior art, the system can visualize the brain activity of a human, can simulate the image observed by the human in real time, can effectively restore and reconstruct the image information seen by the vision, and can understand the electroencephalogram signals induced by specific stimulation. The invention can reconstruct the image observed by the brain on the screen in real time by using the clear image, and provides a brand-new solution for the information of the brain activity record containing the visual object type. Compared with other traditional electroencephalogram signal image reconstruction methods, the method has the advantages that the quality of the reconstructed image is greatly improved no matter in the aspects of an image encoder, an image decoder or an electroencephalogram mapper, and a loss function is newly designed. Provides a new solution for the reconstruction of natural images based on electroencephalogram signals in the future.
Drawings
FIG. 1 is a schematic diagram of the system of the present invention.
Fig. 2 is a schematic diagram of a model structure of an image encoder according to the present invention.
FIG. 3 is a schematic diagram of the structure of the discriminator according to the invention.
Detailed Description
The embodiments of the present invention will be described in detail below with reference to the drawings and examples.
The invention relates to a visual image reconstruction system based on a brain-computer interface, which can realize a real-time image reconstruction function based on an electroencephalogram signal and mainly comprises: (1) an electroencephalogram signal acquisition subsystem; (2) an electroencephalogram signal feature extraction and classification subsystem; (3) an image encoder; (4) an image decoder; (5) an electroencephalogram feature mapper; (6) a discriminator; (7) an image quality evaluation subsystem.
The electroencephalogram signal acquisition subsystem mainly comprises an electroencephalogram cap (256 channels) for acquiring electroencephalogram signals, an electroencephalogram signal amplifier for amplifying the electroencephalogram signals, a wavelet transformer for denoising the electroencephalogram signals, a band-pass filter for filtering the electroencephalogram signals and electroencephalogram signal recording software for recording the electroencephalogram signals when videos stimulate subjects visually.
The electroencephalogram signal feature extraction and classification subsystem mainly comprises an electroencephalogram signal feature extraction module and an electroencephalogram signal feature classification module, and carries out feature extraction and classification on the electroencephalogram signals output by the electroencephalogram signal acquisition subsystem. The electroencephalogram feature extraction adopts a method of combining principal component analysis and linear discrimination, and the electroencephalogram feature classification adopts a naive Bayes classification method.
The image encoder is used for clipping and converting a video causing the visual stimulation into an input picture, and as shown in fig. 2, the input picture is encoded by adopting a VGG-11 neural network to obtain an image feature vector;
the electroencephalogram feature mapper converts the electroencephalogram feature vectors output by the electroencephalogram feature extraction and classification subsystem into feature vectors of a potential space in an image decoder, namely electroencephalogram feature mapping vectors, through an electroencephalogram feature mapping network;
the first image decoder decodes the image feature vector output by the image encoder to obtain a decoded picture, and the second image decoder decodes the electroencephalogram feature mapping vector to obtain a reconstructed picture of the electroencephalogram signal;
the discriminator judges the decoded picture through convolution and judges whether the decoded picture meets the picture output condition or not;
and the image quality evaluation subsystem is used for performing non-reference image quality evaluation on the reconstructed picture of the electroencephalogram signal according to the decoded picture meeting the picture output condition.
Referring to fig. 1, a method for performing visual reconstruction using the visual reconstruction system of the present invention comprises the steps of:
step S1: collecting electroencephalogram signals;
inviting a certain number of testees to wear 256-channel electroencephalogram caps, enabling the testees to watch videos, utilizing a photoelectric sensor to accurately synchronize video segments and electroencephalogram signals, utilizing an electroencephalogram signal amplifier to amplify the electroencephalogram signals, utilizing wavelet transformer independent component analysis to denoise the electroencephalogram signals, utilizing a band-pass filter to filter the electroencephalogram signals, transmitting the processed electroencephalogram signals to a computer, and recording the electroencephalogram signals through electroencephalogram signal recording software.
Step S2: and extracting and classifying the characteristics of the electroencephalogram signals.
Step S21: and extracting the characteristics of the electroencephalogram signals by adopting a method combining principal component analysis and linear discrimination.
The principal component analysis is a method mainly used in multivariate statistical analysis of analysis data, and describes a sample by using a small number of features so as to achieve the purposes of feature extraction and dimension reduction. The basic idea is as follows: the original related independent variables are converted into another group of independent variables through projection by utilizing the orthogonal principle, namely, the so-called principal component, the important components in the principal component are reserved, the other part of unimportant components are discarded, and finally, the model parameters of the selected principal component are evaluated by utilizing the least square method. In this way, both the data set information can be retained virtually without losses and the dimensionality can be reduced.
Given electroencephalogram input signal S, Sij∈S,(1≤i≤Nj,1≤j≤K),NjRepresents the total number of all samples of the jth class, K represents the total number of classes,Sijrepresenting the ith sample in the jth class in S. The principle of principal component analysis is as follows:
1) calculating a sample mean value u of the electroencephalogram input signal S;
Figure BDA0002463076170000061
n denotes the total number of samples in all classes, N ═ N1+N2+…+Nj+…+NK
2) Calculating a covariance matrix C;
Figure BDA0002463076170000062
3) the eigenvector phi can be obtained by the covariance matrix CzAnd the eigenvalue lambda corresponding to the eigenvectorz
C×Φz=λZ×Φz
Eigenvalues λ of the covariance matrix CZThe solution is as follows:
ZE-C|=0
substituting the obtained characteristic value into the following formula to obtain characteristic vector phiz
ZE-C)Φz=0
Where E is the unit vector and z is the number of eigenvalues.
4) The characteristic values are obtained through the formula and are sequenced, the data are specifically reduced to several dimensions, the data are mainly determined by the finally selected principal component, and the selection of the principal component is mainly realized through accumulating the contribution rate. In general, the cumulative contribution rate can be calculated by the following formula:
Figure BDA0002463076170000071
wherein r represents the cumulative contribution rate, and the contribution rate of the ith principal component is represented by piTo indicate, d is the sample initial dimension and h is the sample dimension to be retained. According to the cumulative contribution rate r, the sum ofThe effective m eigenvalues are eigenvectors in one-to-one correspondence, and the most effective m eigenvalues are represented as:
λ1≥λ2≥…λm≥…λZ
5) selecting a projection space M by accumulating a contribution ratio rPCA
MPCA=[φ1,φ2,…,φm]
φmRepresents the mth eigenvalue lambdamThe corresponding feature vector;
6) determining a projection space MPCAThen, the brain electricity is input into the sample SijProjection to MPCAObtaining the EEG signal S after PCA conversionPCANamely, the electroencephalogram signal training sample with the preliminary dimension reduction;
SPCA=SijMPCA
obtaining S through principal component analysisPCAThen, the brain electrical signal S with the preliminary dimension reduction is carried outPCAAnd performing linear discrimination processing.
The specific steps of the linear discrimination method are as follows:
1) if S isPCAThere are b EEG signal samples, each of v1,v2,v3,…,vi,…,vb,viRepresents the ith brain electrical signal training sample, and each brain electrical signal training sample is an n-dimensional vector. If the number of the j-th electroencephalogram signal samples is njWhen c represents the total number of categories, n1+n2+…nj+…ncB, then the sample mean of class j is uj
Figure BDA0002463076170000081
2)SPCAThe mean value of the total electroencephalogram signal training sample is uL
Figure BDA0002463076170000082
In the formula: v. ofiRepresents the ith sample;
3) respectively calculating the dispersion matrixes between classes and in classes of the overall electroencephalogram signal training samples as follows:
Figure BDA0002463076170000083
Figure BDA0002463076170000084
in the formula: sbRepresenting the matrix of dispersion between classes, SwRepresenting a matrix of dispersion, v, within a classkRepresenting the kth class in the total categories of the electroencephalogram signals;
4) projection space vector:
Figure BDA0002463076170000085
wherein mi1, 2 … p, where p is equal to or less than K, that is, the characteristic dimension of the data after the principal component analysis dimensionality reduction is less than the category number K; w is the optimal mapping vector of the projection space, p is the dimension of the projection space vector, miIs the ith vector in the projection space vector;
5) will SPCALinear discrimination processing is carried out to obtain the following electroencephalogram signal feature vector X:
X=SPCA*Wopt
the principal component analysis is an unsupervised algorithm, only the effective direction of the data is searched, the linear discrimination is a supervised algorithm, the effective classification direction of the data can be searched, the data characteristics can be more effectively discriminated by combining the two algorithms, and the next electroencephalogram classification is more benefited.
Step S22: and (4) carrying out feature classification on the electroencephalogram signals.
The invention adopts a naive Bayes classification method, the core idea of the invention is to respectively calculate the probability of each class of the electroencephalogram signal feature vector to be classified, the classification with the maximum probability is selected as the classification of the feature, and the specific calculation process is as follows:
1) electroencephalogram signal feature vector X ═ { α ═1,a2,…,ai…,amThe classification items are X, which is any one of the EEG signal feature vectors, wherein aiThe ith characteristic attribute of the X electroencephalogram signal characteristic vector is obtained;
2) if the category set of all the electroencephalogram feature vectors is U ═ y1,y2,…,yi…,yn},yiRepresents the ith category;
3) calculating the probability that X is the electroencephalogram category:
P(y1|X),P(y2|X),…,P(yi|X)…,P(yn| X), wherein P (y)i| X) is X belongs to yiThe probability of (d);
4) if P (y)k|X)=max{P(y1|X),P(y2|X)…P(yn| X) }, the EEG type of X is yk
Step S3: the image coding comprises the following specific processes:
the image coding training model adopts a VGG-11 neural network, and the specific coding process is as follows:
step S31: a WXH 3 picture is input, and after one convolution of 64 convolution kernels, max-posing pooling is adopted once, wherein W is the width of the image, and H is the height of the image.
Step S32: after one convolution of 128 convolution kernels, adopting one max-pooling;
step S33: after the convolution of 256 convolution kernels for two times, adopting max-posing pooling for one time;
step S34: repeating the convolution kernel convolution twice for 512 times, and then performing max-pooling once;
after the input picture is convolved with the convolution kernel, the number of channels is increased, and the number of the channels is equal to the number of the convolution kernels. After the characteristic diagram is subjected to maximum pooling, the length and the width are changed into 1/2. The model structure of the image encoder of the present invention is shown in fig. 2, wherein the calculation process of convolution is as follows:
Ij,k,j∈[0,x),k∈[0,x)
Wq,m∈l∈[0,y),k∈[0,y)
Figure BDA0002463076170000101
wherein Ij,kRepresenting the input image, j, k each representing the position coordinates of a pixel point on the image, Wq,mRepresenting the weight corresponding to VGG-11 neural network convolution, q and m are the positions of convolution kernels corresponding to the weight, sigma is a ReLU activation function, phi is an output value after one-time convolution calculation, bias is bias, the size of a convolution kernel is 3 × 3, the max-posing pooling size is 2 × 2, stride step size is 2, and the length and width of the picture after the maximum pooling are respectively reduced by half.
Step S4: the image decoding comprises the following specific processes:
1) the image decoder and the image encoder have the same weight, the first image decoder is connected with the output end of the image decoder, the second image decoder is connected with the output end of the electroencephalogram feature mapper, the input of the first decoder is an image feature vector which is finally output by the image encoder, namely a feature map, the deconvolution layer is adopted to double the size of the feature map, and meanwhile, the number of channels is reduced by half; connecting the deconvoluted output characteristic diagram with the corresponding output characteristic diagram in the image encoder, performing convolution operation on the obtained characteristic diagram to keep the symmetry of the channel number and the image encoder, repeating the up-sampling for 5 times, and performing 5 times of maximum pooling in the corresponding image encoder; the second image decoder decodes the electroencephalogram feature vector to obtain a reconstructed picture, and trains an electroencephalogram feature mapper by calculating pixel loss between the decoded picture and the reconstructed picture, and peak signal-to-noise ratio loss and structural similarity loss between the reconstructed picture and an input picture.
2) Inputting an image I to an encoding-decoding model1,I2Converting the similarity between a pair of images into a corresponding vector Z in n-dimensional potential space1,Z2The mutual distance between them. A pair of images I1,I2As input samples, and according to oneContrast loss function dcUpdating the weight of VGG-11 neural network to make the model learn to judge whether the inputs are of the same class, dcThe calculation formula of (a) is as follows:
Figure BDA0002463076170000111
n is the dimension of the potential space vector.
The vectors of each type of image should be uniformly distributed in the potential space, and the invention provides a loss function L which is distance loss LdAngle loss LaContent loss LcWeighted sum of three components. Distance loss weight W is obtained by training to minimize loss function LaAngular loss weight WaContent loss weight WcThese three parameters are tuned to be optimal.
L=WdLd+WaLa+WcLc
The specific calculation formulas of these three parameters are as follows:
wherein the distance loss function LdFor controlling the mutual distance between the feature vectors in the potential spatial representation, the calculation formula is as follows:
Ld=tdc+(1-t)S2(m-dc)
Figure BDA0002463076170000112
Figure BDA0002463076170000113
in the formula: t is a target coefficient, S is a sigmood function, η is a distance weighting parameter, m is an edge for distinguishing a cluster in a potential space, and l is used for judging whether the input is of the same class, if the images belong to the same class, l is 1, otherwise l is 0; if the images are similar images in the same category, the target coefficient t is close to zero;
angle loss function LaFor maintaining the position of the whole potential space feature vector clusterThe angle loss function is such that the cluster-like is not formed into a linear distribution in the latent space, but a kind of polygon. In such a distribution, the mutual distances of the cluster centers are similar, with no a priori preference for any class. The calculation formula is as follows:
Figure BDA0002463076170000121
z1.z2are respectively an input image I1,I2A vector of the corresponding potential space;
content loss function LcControlling the quality of the decoded pictures, for the input image IiOutput picture I of an image decoder in a coding-decoding model0,LcBeing the original error per pixel between the input picture and the output picture, another useful property of content loss is that it causes similar images from the same video clip to have similar positions in potential space. The calculation formula is as follows:
Figure BDA0002463076170000122
in the formula: and N is the number of pixels of the picture.
The combination of the image encoder and the first image decoder is equivalent to a generator, the discriminator can be connected behind the first image decoder, the input picture enters the discriminator, the discriminator takes the input picture as a standard and is used for judging whether the decoded picture meets the output standard, if so, the decoded picture is output, and if not, the generator continues to be trained until the output decoded picture meets the output requirement of the discriminator.
The specific process is as follows: the image encoder plus (the first) image decoder plus the discriminator form a countermeasure network, and the training process is divided into three steps:
firstly, fixing generator parameters and training discriminator parameters to enable the discriminator to better distinguish data sources.
In a second step, the discriminators are then fixed and the generator is trained to generate samples that pass through the discriminators.
And thirdly, repeating the process until the network can reach the balance of the schnabel.
The discriminator is formed by combining 5 layers of convolution, 4 Relu activation functions, 3 BN and 1 full connecting layer. The block diagram is shown in fig. 3.
The loss function of the generator of the present invention is composed of two parts: respectively mean square loss LMSEAnd to combat the loss LagThe mean square loss is the sum of squares of pixel differences between the generated image and the input image of the generator, and is calculated as follows:
Figure BDA0002463076170000131
Iifor inputting the image, G (I)i) To decode the output image, i.e. the generator output image, x, y are the rows and columns, respectively, of image pixels, and G is the generator.
The countermeasure loss is a loss function of the generator G, which is calculated as follows:
Figure BDA0002463076170000132
v represents the output value of the loss function, pz(z) represents the data distribution generated by the generator, E is the mean value found, D represents the discriminator, G represents the generator, the discriminator loss function LGComprises the following steps:
LG=LMSE+Lag
loss function L of discriminatorDComprises the following steps:
Figure BDA0002463076170000133
pdata(x) Representing the true data distribution;
Lagand LDThe solving process of (2) is specifically as follows:
for the discriminator, it is desirable to distinguish the data sources well, judge the real data as 1, and judge the generator data as 0. The loss function that directs discriminant training is:
Figure BDA0002463076170000134
where V represents the output value of the loss function, D represents the network discriminator, G is the network generator, pdata(x) Representing the true data distribution, pz(z) represents the data distribution generated by the generator, and E is the average value obtained.
For the generated data obtained by the generator, the generator wants to be recognized by the discriminator as 1, so that it can be obtained that the corresponding loss function of the generator has the following form:
Figure BDA0002463076170000135
in summary, the loss functions of the guidance generator and the arbiter can obtain the following form of the loss function for generating the countermeasure network:
Figure BDA0002463076170000141
the above formula is varied as follows:
Figure BDA0002463076170000142
the best effect of the network arbiter is to distinguish the true source of the data, and the above formula should be maximized, that is:
Figure BDA0002463076170000143
pg(x) The representation generates a distribution of the data,
Figure BDA0002463076170000144
is a discrimination curve of the discriminator, and represents real data and generated data when the value of the curve is equal to 1/2The distribution of (A) is consistent.
It can be seen that the network achieves the best training results when the generator's generated data distribution and the real data distribution coincide. Substituting (2) into (1) can obtain:
Figure BDA0002463076170000145
wherein KL represents the Kullback-Leible divergence, in the form:
Figure BDA0002463076170000146
wherein P is1And P2Two probability distributions;
JSD is Jensen-Shamon divergence expressed as:
Figure BDA0002463076170000151
if and only if P1=P2When is, i.e. pdata=pgWhen the formula (3) is equal to zero, the minimum value of V (D, G) is obtained, and the generator effect is best. It can be seen that the network achieves the best training effect when the generated data distribution of the generator is consistent with the real data distribution.
Step S5: the electroencephalogram feature mapping comprises the following specific processes:
the goal of the electroencephalogram Feature Mapper (FM) is to convert the electroencephalogram feature vector f into an image feature vector f' through an LSTM neural network. Ideally, the image observed by the subject and the feature vector of the electroencephalogram signal recorded during the observation will eventually be converted into the same potential feature vector of the spatial image, so that the decoder can generate proper visualization effect for what the subject just sees or imagines. The calculation formula of the electroencephalogram feature mapper is as follows:
FM(f)=f′≈z
and (3) sending the electroencephalogram signal feature vector f into an LSTM neural network to obtain an electroencephalogram feature mapping vector f'. By calculation ofAnd training the electroencephalogram feature mapper by the mean square error of the electroencephalogram feature mapping vector and the image feature vector. Changing the EEG signal feature vector f into xtExpressed, the specific function of the LSTM neural network is as follows:
ft=σ(Wf·[ht-1,xt]+bf)
it=σ(Wi·[ht-1,xt]+bi)
Figure BDA0002463076170000152
Figure BDA0002463076170000153
ot=σ(Wo·[ht-1,xt]+bo)
ht=ot*tanh(ct)
wherein x istFor the input of electroencephalogram signals, ht-1Is the output of the last neuron electroencephalogram signal ftTo forget the door, control the degree of forgetting the previous cell, itIn order to input the information into the gate,
Figure BDA0002463076170000154
a new candidate vector generated for tanh, and itTogether with controlling how much new information is added. c. CtIs the new state of the memory cell. otIs an output gate that controls how much of the current cell state is filtered. h istIs the output of the unit. Wi,Wf,Wc,WoRespectively the weights of the input gate, the forgetting gate, the memory gate and the output gate, bi,bf,bc,boRespectively the offset of the input gate, the forgetting gate, the memory gate and the output gate, and the final htThe signal output is the electroencephalogram feature mapping vector f'.
For the feature mapper, the invention proposes a loss function LfmIt is a feature vectorLoss LtReconstructed picture pixel loss LrpReconstructed picture to noise ratio loss LPSNRAnd loss of structural similarity L of reconstructed picturesSSIMA weighted sum of these four loss functions.
Lfm=WtLt+WrLrp+WPSLPSNR+WSSLSSIM
WtIs a eigenvector loss parameter, WrIs a reconstructed picture pixel loss parameter, WPSIs the reconstructed picture to noise ratio loss parameter, WSSIs the reconstructed picture structure similarity loss parameter.
LtThe mean square error of the latent space vector Z and the electroencephalogram feature mapping vector f'.
Figure BDA0002463076170000161
In the formula: n is the dimension of the potential spatial feature vector.
LrpFor decoding picture IoReconstructed picture I of electroencephalogram feature mapping vector frMean square error of'.
Figure BDA0002463076170000162
In the formula: and N is the pixel number of the picture.
LPSNRFor the peak signal-to-noise ratio, the larger the PSNR value is, the smaller the distortion of the reconstructed picture is, and the calculation formula is as follows:
Figure BDA0002463076170000163
in the formula: n' is the number of bits to reconstruct a picture pixel.
LSSIMFor structural similarity, it calculates the loss of picture from three aspects of brightness, contrast and structure. The larger the value is, the smaller the distortion of the image is, and the calculation formula is as follows:
Figure BDA0002463076170000171
Figure BDA0002463076170000172
Figure BDA0002463076170000173
in the formula: c1,C2,C3Is set to be constant in order to avoid the case where the denominator is 0, typically C1=(K1×L)2,C2=(K2×L)2,C3=C2/2, setting constant K in general1=0.01,K2=0.03,L=225。μIAnd
Figure BDA0002463076170000174
respectively representing images I and Ir' the calculation formula is as follows:
Figure BDA0002463076170000175
Figure BDA0002463076170000176
in the formula: i denotes an input picture, Ir' denotes a reconstructed picture, where j, k each denote the position coordinates of a pixel point on the picture.
σI
Figure BDA0002463076170000177
Respectively representing images I, Ir' the variance is calculated as follows:
Figure BDA0002463076170000178
Figure BDA0002463076170000179
Figure BDA00024630761700001710
representing image I and image IrThe covariance of' is calculated as follows:
Figure BDA00024630761700001711
LSSIM=l(I,Ir′)×c(I,Ir′)×s(I,Ir′)
Lfmthe loss function of the feature mapper is used, the loss function of the feature mapper is minimized through training of a training set, and the LSTM parameters are adjusted to be optimal.
Step S6: the image quality evaluation comprises the following specific processes:
the quality of the reconstructed image is evaluated, and the purpose is to evaluate the quality of the reconstruction of the electroencephalogram signal image. The invention utilizes the average gradient
Figure BDA0002463076170000181
The objective evaluation index is used for evaluating the quality of the reconstructed image. Mean gradient
Figure BDA0002463076170000182
The image sharpness is measured and reflects the characteristics of micro detail contrast and texture transformation in the image, and generally, the higher the value is, the more the image layers are, the clearer the image is. The calculation formula is as follows:
Figure BDA0002463076170000183
where m is the width of the picture, n is the height of the picture, fi,jIs the gray value, x, of the picture pixel (i, j)iIs the pixel gray value in the horizontal direction, xjIs the pixel gray value in the vertical direction.
The previous description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present invention. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the invention. Thus, the present invention is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.

Claims (10)

1. A brain-computer interface based visual image reconstruction system, comprising:
the electroencephalogram signal acquisition subsystem mainly comprises an electroencephalogram cap, an electroencephalogram signal amplifier, a wavelet converter, a band-pass filter and electroencephalogram signal recording software, wherein the electroencephalogram cap is used for acquiring electroencephalogram signals generated by a brain, the electroencephalogram signal amplifier is used for amplifying the electroencephalogram signals, the wavelet converter is used for denoising the electroencephalogram signals, the band-pass filter is used for filtering the electroencephalogram signals, and the electroencephalogram signal recording software is used for recording the electroencephalogram signals;
the electroencephalogram signal feature extraction and classification subsystem mainly comprises an electroencephalogram signal feature extraction module and an electroencephalogram signal feature classification module, and is used for extracting and classifying the characteristics of electroencephalogram signals output by the electroencephalogram signal acquisition subsystem;
the image encoder clips and converts a video causing visual stimulation into an input picture, and then encodes the input picture to obtain an image feature vector;
the electroencephalogram feature mapper converts electroencephalogram feature vectors output by the electroencephalogram feature extraction and classification subsystem into feature vectors of a potential space in an image decoder, namely electroencephalogram feature mapping vectors, through an electroencephalogram feature mapping network;
the first image decoder decodes the image feature vector output by the image encoder to obtain a decoded picture, and the second image decoder decodes the electroencephalogram feature mapping vector to obtain a reconstructed picture of the electroencephalogram signal;
the discriminator judges the decoded picture through convolution and judges whether the decoded picture meets the picture output condition or not;
and the image quality evaluation subsystem carries out non-reference image quality evaluation on the reconstructed picture of the electroencephalogram signal according to the decoded picture meeting the picture output condition.
2. The brain-computer interface-based visual image reconstruction system according to claim 1, wherein the electroencephalogram signal feature extraction adopts a method combining principal component analysis and linear discrimination to extract the electroencephalogram signal feature, and the method comprises the following steps:
1) calculating a sample mean value u of the electroencephalogram input signal S;
Figure FDA0002463076160000021
wherein S isijRepresents the ith sample in the jth class of S, Sij∈S,,1≤f≤Nj,1≤j≤K,NjRepresents the total number of all samples in class j, K represents the total number of classes, N represents the total number of samples in all classes, N ═ N1+N2+…+Nj+…+NK
2) Calculating a covariance matrix C;
Figure FDA0002463076160000022
3) finding a feature vector phizAnd its corresponding eigenvalue lambdaZ
zE-C|=0
zE-C)Φz=0
E is a unit vector, and z is the number of eigenvalues;
4) calculating the cumulative contribution rate r, and calculating the obtained characteristic value lambdazSorting and reducing the dimension of the data;
Figure FDA0002463076160000023
where ρ isiRepresenting the contribution rate of the ith principal component, d being the initial dimension of the sample, h being the dimension of the sample to be preserved, preserving the eigenvectors corresponding to the most effective m eigenvalues one by one according to the cumulative contribution rate r, wherein the most effective m eigenvalues are represented as:
λ1≥λ2≥…λm≥…λz
5) selecting a projection space M by accumulating a contribution ratio rPCA
MPCA=[φ1,φ2,…,φm]
φmRepresents the mth eigenvalue lambdamThe corresponding feature vector;
6) inputting the brain electricity into a sample SijProjection to MPCAObtaining the EEG signal S after PCA conversionPCANamely, the electroencephalogram signal training sample with the preliminary dimension reduction;
SPCA=SijMPCA
7)SPCAthere are b training samples of EEG signal, v1,v2,v3,…,vi,…,vb,viRepresenting the f-th electroencephalogram signal training sample, wherein each electroencephalogram signal training sample is an n-dimensional vector, and the number of the j-th electroencephalogram signal training samples is njWhen c represents the total number of categories, n1+n2+…nj+…ncClass j has a sample mean value of uj
Figure FDA0002463076160000031
8)SPCAThe mean value of the total electroencephalogram signal training sample is uL
Figure FDA0002463076160000032
9) Respectively calculating the dispersion matrixes between classes and in classes of the overall electroencephalogram signal training samples as follows:
Figure FDA0002463076160000033
Figure FDA0002463076160000034
Sbrepresenting the matrix of dispersion between classes, SwRepresenting a matrix of dispersion, v, within a classkRepresenting the kth class in the total categories of the electroencephalogram signals;
10) projection space vector:
Figure FDA0002463076160000035
wherein mi1, 2.. p }, wherein p is less than or equal to K, namely the characteristic dimensionality of the data subjected to principal component analysis dimensionality reduction is less than the category number K; w is the optimal mapping vector of the projection space, p is the dimension of the projection space vector, miIs the ith vector in the projection space vector;
11) will SPCALinear discrimination processing is carried out to obtain the following electroencephalogram signal feature vector X:
X=SPCA*Wopt
the electroencephalogram signal feature classification module carries out feature classification on electroencephalogram signals by adopting a naive Bayes classification method, and the method comprises the following steps:
1) electroencephalogram signal feature vector X ═ { a ═ a1,a2,…,ai…,amThe classification items are X, which is any one of the EEG signal feature vectors, wherein aiThe ith characteristic attribute of the X electroencephalogram signal characteristic vector is obtained;
2) if the category set of all the electroencephalogram feature vectors is U ═ y1,y2,…,yi…,yn},yiRepresents the ith category;
3) calculating the probability that X is the electroencephalogram category: p (y)1|X),P(y2|X),…,P(yi|X)…,P(yn| X), wherein P (y)i| X) is X belongs to yiThe probability of (d);
4) if P (y)k|X)=max{P(y1|X),P(y2|X)…P(yn| X) }, the EEG type of X is yk
3. The brain-computer interface-based visual image reconstruction system of claim 1, wherein the image encoder encodes the input picture using a VGG-11 neural network by:
1) a WXH 3 picture is input, and after one convolution of 64 convolution kernels, max-posing pooling is adopted once, wherein W is the width of the image, and H is the height of the image.
2) After one convolution of 128 convolution kernels, one max-pooling is adopted;
3) after two convolutions with 256 convolution kernels, once max-posing pooling is adopted;
4) after two convolutions with 512 convolution kernels, one max-pooling is adopted;
the convolution calculation process is as follows:
Ij,k,j∈[0,x),k∈[0,x)
Wq,m∈l∈[0,y),k∈[0,y)
Figure FDA0002463076160000041
wherein Ij,kRepresenting the input image, j, k each representing the position coordinates of a pixel point on the image, Wq,mRepresenting the weight corresponding to VGG-11 neural network convolution, q and m are positions of convolution cores corresponding to the weight, sigma is a ReLU activation function, phi is an output value after one-time convolution calculation, and bias is bias.
4. The brain-computer interface based visual image reconstruction system according to claim 3, wherein the size of the convolution kernel is 3 x 3, the max-posing pooling size is 2 x 2, stride step size is 2, and the length and width of the picture after the maximal pooling is respectively reduced by half.
5. The brain-computer interface-based visual image reconstruction system according to claim 3 or 4, wherein the image decoder decodes the image feature vector or the electroencephalogram feature mapping vector by deconvolution, which comprises:
1) the image decoder and the image encoder have the same weight, the first image decoder is connected with the output end of the image decoder, the second image decoder is connected with the output end of the electroencephalogram feature mapper, the input of the first decoder is an image feature vector which is finally output by the image encoder, namely a feature map, the deconvolution layer is adopted to double the size of the feature map, and meanwhile, the number of channels is reduced by half; connecting the deconvoluted output characteristic diagram with the corresponding output characteristic diagram in the image encoder, performing convolution operation on the obtained characteristic diagram to keep the symmetry of the channel number and the image encoder, repeating the up-sampling for 5 times, and performing 5 times of maximum pooling in the corresponding image encoder; the second image decoder decodes the electroencephalogram feature vector to obtain a reconstructed picture, and trains an electroencephalogram feature mapper by calculating pixel loss between the decoded picture and the reconstructed picture, and peak signal-to-noise ratio loss and structural similarity loss of the reconstructed picture and an input picture;
2) inputting an image I to an encoding-decoding model1,I2Converting the similarity between a pair of images into a corresponding vector Z in n-dimensional potential space1,Z2A pair of images I representing the mutual distance between1,I2As input samples and according to a contrast loss function dcUpdating the weight of VGG-11 neural network to make the model learn to judge whether the inputs are of the same class, dcThe calculation formula of (a) is as follows:
Figure FDA0002463076160000051
the vectors of each type of picture should be uniformly distributed in the potential space, and the loss function L is distance lossLoss function LdAngle loss function LaContent loss function LcWeighted sum of three components:
L=WdLd+WaLa+WcLc
distance loss weight W is obtained by training to minimize loss function LdAngular loss weight WaContent loss weight WcThe three parameters are adjusted to be optimal;
wherein the distance loss function LdFor controlling the mutual distance between the feature vectors in the potential spatial representation, the calculation formula is as follows:
Ld=tdc+(1-t)S2(m-dc)
Figure FDA0002463076160000061
Figure FDA0002463076160000062
in the formula: t is a target coefficient, S is a sigmood function, η is a distance weighting parameter, m is an edge for distinguishing a cluster in a potential space, and l is used for judging whether input is of the same class, if pictures belong to the same class, l is 1, otherwise l is 0; if the pictures are similar pictures in the same category, the target coefficient t is close to zero;
angle loss function LaFor maintaining uniformity of the positions of the whole potential space feature vector cluster, the calculation formula is as follows:
Figure FDA0002463076160000063
z1.z2are respectively an input image I1,I2A vector of the corresponding potential space;
content loss function LcFor controlling the quality of the decoded image, the calculation formula is as follows:
Figure FDA0002463076160000064
in the formula: i isiFor inputting pictures, IoThe N is the output picture of an image decoder in the coding-decoding model, and is the pixel number of the picture.
6. The brain-computer interface-based visual image reconstruction system according to claim 5, wherein the combination of the image encoder and the first image decoder is equivalent to a generator, the output of the first image decoder is connected to a discriminator, the input picture enters the discriminator, the discriminator uses the input picture as a standard to determine whether the decoded picture meets the output standard, if yes, the decoded picture is output, and if not, the generator continues to be trained until the output decoded picture meets the output requirement of the discriminator.
7. The brain-computer interface based visual image reconstruction system according to claim 6, wherein the image encoder, the image decoder and the discriminator form a confrontation network, and the training process is divided into three steps:
1) firstly, fixing generator parameters and training discriminator parameters to enable the discriminator to better distinguish data sources;
2) then fixing the discriminator, training the generator to generate a sample passing through the discriminator;
3) finally, the above process is repeated until the network can reach the schner balance.
8. The brain-computer interface based visual image reconstruction system according to claim 6, wherein the discriminator is formed by combining 5 layers of convolution, 4 Relu activation functions, 3 BN, and 1 full connection layer;
the loss function of the generator consists of two parts: respectively mean square loss LMSEAnd to combat the loss LaqMean square loss LMSEIs the sum of squares of pixel differences of the generated image of the generator corresponding to the input image, calculated asThe following:
Figure FDA0002463076160000071
G(Ii) To generate the output image of the generator, x, y are the rows and columns of image pixels, respectively;
against loss LagI.e. the loss function of the generator, is calculated as follows:
Figure FDA0002463076160000072
v represents the output value of the loss function, pz(z) represents the data distribution generated by the generator, E is the mean value obtained, G represents the generator, D represents the discriminator, the loss function L thereofGComprises the following steps:
LG=LMSE+Lag
d represents the discriminator, its loss function LDComprises the following steps:
Figure FDA0002463076160000073
pdata(x) Representing the true data distribution;
the network achieves the best training effect when the generated data distribution of the generator is consistent with the real data distribution.
9. The brain-computer interface-based visual image reconstruction system according to claim 1, wherein the electroencephalogram feature mapper converts the electroencephalogram feature vector f into an image feature vector f' through an LSTM neural network, and the calculation formula is as follows:
FM(f)=f′≈z
sending the electroencephalogram signal feature vector f into an LSTM neural network to obtain an electroencephalogram feature mapping vector f';
the electroencephalogram feature mapper is trained by calculating the mean square error of the electroencephalogram feature mapping vector and the image feature vector.
10. The brain-computer interface based visual image reconstruction system of claim 1, wherein said image quality assessment subsystem utilizes mean gradient
Figure FDA0002463076160000081
The objective evaluation index is used for evaluating the quality of the reconstructed picture and averaging the gradient
Figure FDA0002463076160000082
The method is used for measuring the definition of the picture, the definition reflects the characteristics of micro detail contrast and texture transformation in the picture, the larger the value is, the more picture layers are, the clearer the picture is, and the calculation formula is as follows:
Figure FDA0002463076160000083
where m is the width of the picture, n is the height of the picture, fi,jIs the gray value, x, of the picture pixel (i, j)iIs the pixel gray value in the horizontal direction, xjIs the pixel gray value in the vertical direction.
CN202010330551.XA 2020-04-23 2020-04-23 Visual image reconstruction system based on brain-computer interface Active CN111539331B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010330551.XA CN111539331B (en) 2020-04-23 2020-04-23 Visual image reconstruction system based on brain-computer interface

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010330551.XA CN111539331B (en) 2020-04-23 2020-04-23 Visual image reconstruction system based on brain-computer interface

Publications (2)

Publication Number Publication Date
CN111539331A true CN111539331A (en) 2020-08-14
CN111539331B CN111539331B (en) 2023-03-28

Family

ID=71967541

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010330551.XA Active CN111539331B (en) 2020-04-23 2020-04-23 Visual image reconstruction system based on brain-computer interface

Country Status (1)

Country Link
CN (1) CN111539331B (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112001306A (en) * 2020-08-21 2020-11-27 西安交通大学 Electroencephalogram signal decoding method for generating neural network based on deep convolution countermeasure
CN112133311A (en) * 2020-09-18 2020-12-25 科大讯飞股份有限公司 Speaker recognition method, related device and readable storage medium
CN112180310A (en) * 2020-08-20 2021-01-05 山东省医学影像学研究所 Magnetic resonance imaging method combining parallel imaging and principal component analysis dynamic denoising
CN112381180A (en) * 2020-12-09 2021-02-19 杭州拓深科技有限公司 Power equipment fault monitoring method based on mutual reconstruction single-class self-encoder
CN113112017A (en) * 2021-04-16 2021-07-13 唐山市工人医院 Electroencephalogram grading and prognosis FPGA decoding system based on neural manifold
CN113408697A (en) * 2021-05-31 2021-09-17 杭州电子科技大学 Electroencephalogram data reconstruction method based on reinforcement learning self-encoder
CN115708687A (en) * 2022-11-21 2023-02-24 重庆邮电大学 Stimulated image reconstruction method and system based on artificial neural network

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130184558A1 (en) * 2009-03-04 2013-07-18 The Regents Of The University Of California Apparatus and method for decoding sensory and cognitive information from brain activity
CN108573512A (en) * 2018-03-21 2018-09-25 电子科技大学 A kind of complicated visual pattern reconstructing method based on depth encoding and decoding veneziano model
CN109271964A (en) * 2018-10-11 2019-01-25 刘仕琪 Emotion identification method and system based on deep learning model Yu length memory network
US20190046068A1 (en) * 2017-08-10 2019-02-14 Siemens Healthcare Gmbh Protocol independent image processing with adversarial networks

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130184558A1 (en) * 2009-03-04 2013-07-18 The Regents Of The University Of California Apparatus and method for decoding sensory and cognitive information from brain activity
US20190046068A1 (en) * 2017-08-10 2019-02-14 Siemens Healthcare Gmbh Protocol independent image processing with adversarial networks
CN108573512A (en) * 2018-03-21 2018-09-25 电子科技大学 A kind of complicated visual pattern reconstructing method based on depth encoding and decoding veneziano model
CN109271964A (en) * 2018-10-11 2019-01-25 刘仕琪 Emotion identification method and system based on deep learning model Yu length memory network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
官金安等: "基于3D卷积神经网络的IR-BCI脑电视频解码研究", 《中南民族大学学报(自然科学版)》 *
雷煜等: "基于脑功能磁共振影像的视觉信息解读技术", 《计算机应用研究》 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112180310A (en) * 2020-08-20 2021-01-05 山东省医学影像学研究所 Magnetic resonance imaging method combining parallel imaging and principal component analysis dynamic denoising
CN112001306A (en) * 2020-08-21 2020-11-27 西安交通大学 Electroencephalogram signal decoding method for generating neural network based on deep convolution countermeasure
CN112133311A (en) * 2020-09-18 2020-12-25 科大讯飞股份有限公司 Speaker recognition method, related device and readable storage medium
CN112133311B (en) * 2020-09-18 2023-01-17 科大讯飞股份有限公司 Speaker recognition method, related device and readable storage medium
CN112381180A (en) * 2020-12-09 2021-02-19 杭州拓深科技有限公司 Power equipment fault monitoring method based on mutual reconstruction single-class self-encoder
CN113112017A (en) * 2021-04-16 2021-07-13 唐山市工人医院 Electroencephalogram grading and prognosis FPGA decoding system based on neural manifold
CN113408697A (en) * 2021-05-31 2021-09-17 杭州电子科技大学 Electroencephalogram data reconstruction method based on reinforcement learning self-encoder
CN115708687A (en) * 2022-11-21 2023-02-24 重庆邮电大学 Stimulated image reconstruction method and system based on artificial neural network

Also Published As

Publication number Publication date
CN111539331B (en) 2023-03-28

Similar Documents

Publication Publication Date Title
CN111539331B (en) Visual image reconstruction system based on brain-computer interface
CN107085716B (en) Cross-view gait recognition method based on multi-task generation countermeasure network
US8599209B2 (en) Image processing apparatus, image processing method, and computer readable medium
US8538139B2 (en) Image processing apparatus and method, data processing apparatus and method, and program and recording medium
US8306316B2 (en) Image processing apparatus and method, data processing apparatus and method, and program and recording medium
US8421885B2 (en) Image processing system, image processing method, and computer readable medium
JP5506274B2 (en) Image processing apparatus and method, data processing apparatus and method, and program
CN109410177B (en) Image quality analysis method and system for super-resolution image
JP7357010B2 (en) Facial image recognition using pseudo images
WO2011013606A1 (en) Image processing device and method, data processing device and method, program, and recording medium
WO2011013608A1 (en) Image processing device and method, data processing device and method, program, and recording medium
CN109993068B (en) Non-contact human emotion recognition method based on heart rate and facial features
Kantarcı et al. Thermal to visible face recognition using deep autoencoders
CN111210415B (en) Method for detecting facial expression hypo of Parkinson patient
CN111967354B (en) Depression tendency identification method based on multi-mode characteristics of limbs and micro-expressions
CN109146785A (en) A kind of image super-resolution method based on the sparse autocoder of improvement
Du et al. Weakly supervised rppg estimation for respiratory rate estimation
CN113592769B (en) Abnormal image detection and model training method, device, equipment and medium
Colores-Vargas et al. Video images fusion to improve iris recognition accuracy in unconstrained environments
CN116421200A (en) Brain electricity emotion analysis method of multi-task mixed model based on parallel training
CN105930878B (en) Micro-expression recognition method based on differential slice energy diagram and sparse coding
Emara et al. A Hybrid Compressive Sensing and Classification Approach for Dynamic Storage Management of Vital Biomedical Signals
Yang et al. Model-agnostic Method: Exposing Deepfake using Pixel-wise Spatial and Temporal Fingerprints
CN107491739B (en) Robust face recognition method combining multivariate ellipse distribution of smooth matrix
Park et al. GMDM: A generalized multi-dimensional distribution overlap metric for data and model quality evaluation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant