US12106204B2 - Adaptive brain-computer interface decoding method based on multi-model dynamic integration - Google Patents

Adaptive brain-computer interface decoding method based on multi-model dynamic integration Download PDF

Info

Publication number
US12106204B2
US12106204B2 US17/777,956 US202117777956A US12106204B2 US 12106204 B2 US12106204 B2 US 12106204B2 US 202117777956 A US202117777956 A US 202117777956A US 12106204 B2 US12106204 B2 US 12106204B2
Authority
US
United States
Prior art keywords
model
state
selecting
neural
function
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US17/777,956
Other versions
US20230244909A1 (en
Inventor
Yu Qi
Yueming Wang
Xinyun Zhu
Kedi Xu
Gang Pan
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Assigned to ZHEJIANG UNIVERSITY reassignment ZHEJIANG UNIVERSITY ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: PAN, GANG, QI, YU, WANG, YUEMING, XU, Kedi, ZHU, Xinyun
Publication of US20230244909A1 publication Critical patent/US20230244909A1/en
Application granted granted Critical
Publication of US12106204B2 publication Critical patent/US12106204B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/015Input arrangements based on nervous system activity detection, e.g. brain waves [EEG] detection, electromyograms [EMG] detection, electrodermal response detection
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/047Probabilistic or stochastic networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N7/00Computing arrangements based on specific mathematical models
    • G06N7/01Probabilistic graphical models, e.g. probabilistic networks

Definitions

  • the present invention relates to the field of motor neural signal decoding, in particular to an adaptive brain-computer interface decoding method based on multi-model dynamic ensemble.
  • Brain-computer interfaces directly decode motion intentions from neural signals for external device control.
  • Intracortical brain-computer interfaces use neural signals recorded by implanted electrode arrays to extract motion-related information.
  • Advances in the intracortical brain-computer interfaces have led to the significant development of exoskeleton devices and computer cursor control.
  • neural decoding algorithms play an important role.
  • Scholars have proposed many algorithms to decode motion information from the neural signals, including population vector methods, optimal linear estimation methods, deep neural networks, and recursive Bayesian decoders.
  • the Kalman filter combines the evolution process of a trajectory as prior knowledge to make prediction more accurate, so it is widely used in online cursor control and exoskeleton control, achieving a best performance so far.
  • Chinese patent document with publication number CN107669416A discloses a wheelchair system and control method based on continuous-brisk motor imagery neural decoding, comprising an electroencephalogram signal acquisition system, a decoding device and an electric wheelchair connected in sequence; the decoding device comprises a continuous motor imagery electroencephalogram signal processing module and a brisk motor imagery electroencephalogram signal processing module that are arranged in series, where the two modules are respectively used to decode the continuous motor imagery electroencephalogram signal and the brisk motor imagery electroencephalogram signal transmitted by the electroencephalogram signal acquisition system.
  • the unstability of the neural signals is a great challenge in the neural decoding process.
  • Most of the neural decoders used in current intracortical brain-computer interfaces assume a stable functional relationship between the neural signals and kinematics, and therefore use a fixed decoding model.
  • signals generated by neurons occasionally introduce noise or disappear altogether; at the same time, because of the plasticity of the neurons, brain activity may also change over time.
  • a neural signal-to-kinematic mapping function may be unstable and change continuously over time.
  • a fixed decoding function can lead to an unstable and inaccurate decoding result, so it needs to be retrained every time period to maintain a certain performance.
  • Decoders proposed for coping with neural variabilities can be classified into two categories.
  • the first category of the decoders which are the most common decoders today, still uses fixed models, relying on periodic retraining or incrementally updating model parameters online to maintain performance.
  • the second category of the decoders uses dynamic models to track the changes in the neural signals, which can avoid the cost of retraining and are more suitable for long-term decoding tasks.
  • few studies in such methods dare to directly model unstable neural signals.
  • the present invention provides an adaptive brain-computer interface decoding method based on multi-model dynamic ensemble, which can greatly reduce the influence of unstability of the neural signals on decoding performance and improve the robustness of a decoder.
  • An adaptive brain-computer interface decoding method based on multi-model dynamic ensemble where the method comprises the following steps:
  • the observation function in a Kalman filter is improved and replaced with a pool of models that can be dynamically assembled online, so as to better adapt to the changes of the neural signals.
  • these models are automatically selected and combined according to the Bayesian update mechanism, thereby greatly reducing the influence of the unstability of the neural signals on the decoding performance and improving the robustness of the decoder.
  • step (1) the specific process of the pre-processing is as follows:
  • the data in paradigm preparation and return phases can be removed, and an actual operation phase can be selected for analysis.
  • a z-score function and a filter function can be used to normalize and smooth each dimension of the motion signal, and smooth the firing rate of each neuron of the neural signal.
  • the specific smoothing coefficients can be selected according to actual needs.
  • the Bayesian update mechanism is as follows:
  • y 0:k ) is the posterior probability of selecting the m-th model; M represents the number of the candidate models.
  • the Bayesian formula can be used to calculate the posterior probability of selecting the m-th model at the moment k.
  • the formula thereof is as follows:
  • p( M
  • y 0:k-1 ) is a priorprior probability of selecting the m-th model
  • y 0:k-1 ) is a marginal likelihood of selecting the m-th model
  • a forgetting factor ⁇ can be determined to retain part of the historical information, that is, the prior probability of selecting the m-th model at the current moment is influenced by the posterior probability of selecting the m-th model at the previous moment.
  • the priorprior probability of selecting the m-th model is calculated as follows:
  • p( m
  • y 0:k-1 ) is the probability of selecting the m-th model at the k-1-th moment; a (0 ⁇ 1) is a forgetting factor.
  • the calculation method of the marginal likelihood of selecting the m-th model is as follows: p m ( y k
  • y 0:k-1 ) ⁇ p m ( y k
  • x k ) is a likelihood function for the m-th model.
  • step (4) adopting a particle filter algorithm for the state estimation, and calculating p(x k
  • the present invention proposes a multi-model dynamic ensemble method applied to the neural signal decoding, which reduces the influence caused by the unstability of the neural signal to a certain extent. It outperforms the well-known Kalman filter algorithm in the test, proving the effectiveness of the method.
  • FIG. 1 is a schematic figure of a multi-model dynamic ensemble process in the method of the present invention
  • FIG. 2 is a comparison chart of the success rate of a Kalman filter and multi-model dynamic ensemble for five consecutive days in an online cursor control experiment in an embodiment of the present invention
  • FIG. 3 is a trajectory comparison figure of a Kalman filter and multi-model dynamic ensemble in an online cursor control experiment on Dec. 14, 2020, according to an embodiment of the present invention.
  • Two 96-channel Utah intracortical microelectrode arrays (4 mm ⁇ 4 mm, 1.4 mm long Utah arrays, Blackrock Microsystems, Salt Lake City, Utah, USA) were implanted in the left primary motor cortex of the volunteer, and the other was located in the central hand knob area 2 mm away. Electron Computed Tomography and functional MRI were used to guide the implantation. During the implantation, the participant was asked to perform flexion and extension motions of the elbow, and the scanning of the functional MRI was used to confirm an activation area in the motor cortex. The operation used a robotic arm to assist in implanting electrodes during the surgery. The participant had a week of rehabilitation before starting the neural signal recording and BCI training tasks.
  • the participant performed the BCI training tasks each weekday and rested on weekends.
  • the training time was approximately 3 hours per day and comprised of preparation for signal recording, impedance testing, spike classification and paradigm tasks.
  • the experiment was stopped once the participant felt tired or developed physical abnormalities, such as a fever or a urinary tract infection. The entire experiment lasted half a year, recording the neural signals for 12 weeks.
  • a 2D cursor control task was performed on a computer monitor 1.5 m in front of the volunteer.
  • the task contained eight targets at the top, bottom, left, right and four corners of the screen.
  • the targets can be displayed sequentially in a clockwise or pseudo-random fashion.
  • the relevant parameters of the task can be configured in task settings, comprising a distance from the target to the center of the screen, the diameter of a target ball, the maximum trial time, an arrival threshold, the minimum dwell time and the number of repetitions after failure.
  • the volunteer was asked to control a blue cursor ball from the center of the screen to a red target ball.
  • the distance between the centers of the blue and red balls should be less than a preset arrival threshold, and dwell time should not be shorter than a preset dwell time.
  • the maximum duration of each trial is set from 2 to 5 seconds. An audible cue is given at the end of the trial to indicate its success. A maximum of 4 attempts were made after each trial failed.
  • the default setting was to display the target balls at 15 cm from the center of the screen, with a diameter of 5 cm.
  • Step 1 motor neural signal pre-processing: obtaining an original motor neural signal recorded from the electrodes, selecting an appropriate window size to calculate the firing rate of the neural signal, intercepting a valid data segment according to a state label, normalizing and smoothing the data, and obtaining the pre-processed motor neural signal; dividing the data into a training set, a validation set, and a test set according to a proper proportion;
  • the experiment recorded the neural signals by using a Neuroport system (NSP, Blackrock Microsystems).
  • the neural activity is amplified, digitized, and recorded at 30 KHz.
  • Thresholds for neural action potential detection were set from ⁇ 6.5 RMS to ⁇ 5.5 RMS for each high-pass filtered (250 Hz cutoff frequency) electrode by using a Central software package (Blackrock Microsystem).
  • the researchers manually sort the spikes, which takes about 25 to 35 minutes.
  • a peak activity was converted to firing rate in 20 ms and low-pass filtered by using an exponential smoothing function with a 450 ms window.
  • Step 2 a state-space model based on the dynamic ensemble:
  • An improved observation model varying a traditional state-space model, using a set of functions (i.e., candidate models) instead of a fixed function to dynamically characterize the relationship between observation variables and state variables;
  • k represents a discrete time step
  • x k ⁇ R d x represents the state that is concerned about
  • y k ⁇ R d y represents the observation variable
  • v k , n k represents a state transition noise and an observation noise which are independent and identically distributed.
  • states and observations represent motion trajectories and neural signals, respectively.
  • the task of the neural decoding is to iteratively estimate the probability density of x k .
  • the Kalman filter can provide an analytically optimal solution when both the neural signal and the motion signal conform to a linear Gaussian assumption.
  • the observation function h( ⁇ ) of the traditional state-space model is determined in advance and cannot adapt to changing neural signals.
  • the observation model in step 2 improves upon it by allowing the observation function to be adjusted online.
  • k represents a discrete time step
  • x k ⁇ R d x represents a state of interest
  • y k ⁇ R d y represents the observation variable
  • n k represents an independent and identically distributed observation noise
  • ⁇ 1, 2, . . . , M ⁇ represents a model index in the observation equation.
  • M ⁇ represents a model index in the observation equation.
  • m represents the model in action at the moment k is h m .
  • the models in a model set can convert input kinematic parameters x t into the neural signals y t .
  • the multi-model dynamic ensemble method we learn a pool of forms of encoding models, such as linear functions, polynomial functions, and neural networks, to enhance the ensemble effectiveness of online models.
  • Past research has demonstrated an excellent decoding ability of linear models, while non-linear models have strong robustness in the presence of noise.
  • the multi-model dynamic ensemble method demonstrates the advantages of different models to deal with unstable neural signals.
  • the model pool of this example contains four models.
  • the third model H nn1 ( ⁇ ) and the fourth model H nn2 ( ⁇ ) are two neural networks with different parameter sizes. Both the neural networks contain one hidden layer.
  • the number of neurons in the input layer is equal to the dimensions of the motion parameters x t
  • the number of neurons in the output layer is equal to the dimensions of the neural signal y t .
  • the number of neurons in the hidden layers are 30 and 50, respectively.
  • H l ( ⁇ ) and H p ( ⁇ ) the parameters are estimated by using the least square method.
  • H nn1 ( ⁇ ) and H nn2 ( ⁇ ) the parameters are optimized by using the Adam algorithm, a learning rate is set to 0.01, and a weight decay is set to 1e-4.
  • an early-stopping method is adopted to alleviate the overfitting problem.
  • step 2 a set of candidate models are used to characterize the functional relationship between the observation signal and the state to be predicted.
  • the Bayesian update mechanism dynamically selects dominant models among these models in a data-driven manner. Given an observation sequence y 0:k , the posterior probability of the state at the moment k is: p ( x k
  • the Bayesian formula can be used to calculate the posterior probability of selecting the m-th model at the moment k.
  • p( m
  • y 0:k-1 ) is a priorprior probability of selecting the m-th model
  • y 0:k-1 ) is a marginal likelihood of selecting the m-th model
  • a forgetting factor ⁇ can be determined to retain part of the historical information, that is, the prior probability of selecting the m-th model at the current moment is influenced by the posterior probability of selecting the m-th model at the previous moment:
  • p( m
  • y 0:k-1 ) is the probability of selecting the m-th model at the k-1-th moment; ⁇ (0 ⁇ 1) is a forgetting factor.
  • the forgetting factor is set from 0.1 to 0.9.
  • the influence of the previous kinematic state is considered to a great extent, so the decoded kinematic state is smoother.
  • the calculation method of the marginal likelihood of selecting the m-th model is as follows: p m ( y k
  • y 0:k-1 ) ⁇ p m ( y k
  • x k ) is a likelihood function for the m-th model.
  • Step 3 candidate model learning and state estimation: using the training set and the validation set to obtain a set of different candidate models;
  • estimating the kinematic state for each candidate model by using a weighted particle set at each time step, and calculating p(x k
  • m, y 0:k ) based on the particle filtering.
  • x k i is obtained from the state transition prior p(x k
  • x k-1 i ),i 1, . . . , N S , and then according to the principle of importance sampling, we can get: p ( x k
  • ⁇ m,k i ⁇ k-1 i p m (y k
  • ⁇ m,k i represents the normalized importance weight of the i-th particle when the m-th model is selected.
  • the training set is used as the particle set to avoid the unstability caused by randomly scattering the particles.
  • p( m
  • y 0:k ) based on the particle filtering.
  • the posterior probability p m (y k m
  • y 0:k-1 ), m 1, . . . , M is known.
  • the state transition prior is used as the importance function, i.e., q(x k
  • x k-1 , y 0:k ) p(x k
  • the distribution of x k can be approximated by the particles, i.e, p(x k
  • y 0:k-1 ) ⁇ i 1 i ⁇ x k i . Further, we get: p m ( y k
  • y 0:k-1 ) ⁇ i 1 N s ⁇ k-1 i p m ( y k
  • the particle filtering usually suffers from particle degradation, and after a few iterations, only a small number of particles have high weights. Therefore, we adopt a resampling method to remove particles with too small or too large weights to alleviate the particle degradation problem.
  • Step 4 performance evaluation of the multi-model dynamic ensemble method: using this method to compare with other methods in the test set data to evaluate the performance and robustness of this method.
  • This example was an online experiment with closed-loop calibration. Two phases of calibration and testing were contained in each experiment. Each session contains several blocks, and each block consists of 16 trials, where 2 trials for each target.
  • the neural decoder was trained during the calibration phase. There were two observation blocks in the calibration phase, and the computer automatically completed the cursor control task. The neural signals of the volunteer during the observation phase were used to initialize the decoder. Subsequently, the decoder was further adjusted by using computer-assistant and ortho-impedance assistance, and the assistant ratio decreased with the increase of the number of blocks. The calibration process lasted about 10 minutes, and then the neural signals coming in the test phase were decoded by using the trained decoder.
  • FIG. 2 is the trajectory diagram of the ball test phase on Dec. 14, 2020. It can be seen that the trajectory solved by the Kalman filter has a strong bias and cannot reach each target point on time.
  • the trajectory distribution solved by the multi-model dynamic ensemble method is relatively uniform and successfully reaches each target point. It can be seen from the above results that the performance of the multi-model dynamic ensemble method is more stable.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Software Systems (AREA)
  • Evolutionary Computation (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • Computing Systems (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Neurology (AREA)
  • Neurosurgery (AREA)
  • Dermatology (AREA)
  • Probability & Statistics with Applications (AREA)
  • Algebra (AREA)
  • Computational Mathematics (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Pure & Applied Mathematics (AREA)
  • Measurement And Recording Of Electrical Phenomena And Electrical Characteristics Of The Living Body (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The present invention discloses an adaptive brain-computer interface decoding method based on multi-model dynamic ensemble, where a traditional state-space model is improved, and a set of measurement functions instead of one fixed measurement function are used to dynamically characterize a relationship between observation variables and state variables; and, by using a pool of linear and nonlinear decoders, and in a decoding process of a brain-computer interface system, decoders are automatically switched according to the data, so as to realize adaptive brain signal decoding. Through the above multi-model ensemble strategy, linear and nonlinear decoder capabilities can be integrated, the accuracy and stability of the brain-computer interface system can be improved, and decoding unstability caused by the non-stationary neural signal of the brain-computer interface system can be solved to a certain extent.

Description

This is a U.S. national stage application of PCT Application No. PCT/CN2021/126580 under 35 U.S.C. 371, filed Oct. 27, 2021 in Chinese, claiming priority to Chinese Patent Applications No. 202011593629.3, filed Dec. 29, 2020, all of which are hereby incorporated by reference.
FIELD OF TECHNOLOGY
The present invention relates to the field of motor neural signal decoding, in particular to an adaptive brain-computer interface decoding method based on multi-model dynamic ensemble.
BACKGROUND TECHNOLOGY
Brain-computer interfaces directly decode motion intentions from neural signals for external device control. Intracortical brain-computer interfaces use neural signals recorded by implanted electrode arrays to extract motion-related information. Advances in the intracortical brain-computer interfaces have led to the significant development of exoskeleton devices and computer cursor control.
In the intracortical brain-computer interface systems, neural decoding algorithms play an important role. Scholars have proposed many algorithms to decode motion information from the neural signals, including population vector methods, optimal linear estimation methods, deep neural networks, and recursive Bayesian decoders. Among these methods, the Kalman filter combines the evolution process of a trajectory as prior knowledge to make prediction more accurate, so it is widely used in online cursor control and exoskeleton control, achieving a best performance so far.
For example, Chinese patent document with publication number CN107669416A discloses a wheelchair system and control method based on continuous-brisk motor imagery neural decoding, comprising an electroencephalogram signal acquisition system, a decoding device and an electric wheelchair connected in sequence; the decoding device comprises a continuous motor imagery electroencephalogram signal processing module and a brisk motor imagery electroencephalogram signal processing module that are arranged in series, where the two modules are respectively used to decode the continuous motor imagery electroencephalogram signal and the brisk motor imagery electroencephalogram signal transmitted by the electroencephalogram signal acquisition system.
However, the unstability of the neural signals is a great challenge in the neural decoding process. Most of the neural decoders used in current intracortical brain-computer interfaces assume a stable functional relationship between the neural signals and kinematics, and therefore use a fixed decoding model. However, during online decoding, signals generated by neurons occasionally introduce noise or disappear altogether; at the same time, because of the plasticity of the neurons, brain activity may also change over time. Because of the noise and change described above, a neural signal-to-kinematic mapping function may be unstable and change continuously over time. A fixed decoding function can lead to an unstable and inaccurate decoding result, so it needs to be retrained every time period to maintain a certain performance.
Decoders proposed for coping with neural variabilities can be classified into two categories. The first category of the decoders, which are the most common decoders today, still uses fixed models, relying on periodic retraining or incrementally updating model parameters online to maintain performance. The second category of the decoders uses dynamic models to track the changes in the neural signals, which can avoid the cost of retraining and are more suitable for long-term decoding tasks. However, few studies in such methods dare to directly model unstable neural signals.
Therefore, in motor neural decoding, how to use dynamic models to model the unstability of the neural signals to achieve a stable decoding performance is an important problem to be solved in the current motor neural decoding field.
SUMMARY OF THE INVENTION
The present invention provides an adaptive brain-computer interface decoding method based on multi-model dynamic ensemble, which can greatly reduce the influence of unstability of the neural signals on decoding performance and improve the robustness of a decoder.
An adaptive brain-computer interface decoding method based on multi-model dynamic ensemble, where the method comprises the following steps:
(1) obtaining the original motor neural signal from electrodes, and after pre-processing the original motor neural signal, dividing the signal into a training set, a validation set and a test set;
(2) building a state-space decoder based on the multi-model dynamic ensemble, which is specified as follows:
    • (2-1) with an improved state-space model, using a pool of candidate models to dynamically characterize a relationship between observation variables and state variables, where the pool of candidate models comprises a linear function, a second-order polynomial function, and two neural networks. Here, the observation variables are neural signals, and the state variables are motion signals;
    • (2-2) dynamically combining the candidate models according to a Bayesian update mechanism as an observation function of the state-space model;
    • (3) using the training set and the validation set to train and evaluate the improved state-space model, and obtaining the parameters of different candidate models by training;
    • (4) using the test set to test the performance and robustness of the model.
In the method of the present invention, the observation function in a Kalman filter is improved and replaced with a pool of models that can be dynamically assembled online, so as to better adapt to the changes of the neural signals. During the decoding process, these models are automatically selected and combined according to the Bayesian update mechanism, thereby greatly reducing the influence of the unstability of the neural signals on the decoding performance and improving the robustness of the decoder.
In the step (1), the specific process of the pre-processing is as follows:
    • selecting an appropriate window size to calculate the firing rate of the neural signal, intercepting a valid data segment according to a state label, normalizing and smoothing the data respectively, and obtaining the pre-processed motor neural signal.
Preferably, the data in paradigm preparation and return phases can be removed, and an actual operation phase can be selected for analysis.
Preferably, when performing the normalizing and smoothing operations, a z-score function and a filter function can be used to normalize and smooth each dimension of the motion signal, and smooth the firing rate of each neuron of the neural signal. The specific smoothing coefficients can be selected according to actual needs.
In the step (2), the expression of the improved state-space model is as follows:
x k=ƒ(x k-1)+v k-1
y k =h H k (x k)+n k
where, k represents a discrete time step; xk∈Rd x represents a state of interest; ƒ(·) represents a state transition function; yk ∈Rd y represents an observation or measurement variable; nk represents an independent and identically distributed observation noise;
Figure US12106204-20241001-P00001
∈{1, 2, . . . , M} represents a model index in the observation function h(·),
Figure US12106204-20241001-P00002
=m represents that the model in action at the moment k is hm.
In the step (2), the Bayesian update mechanism is as follows:
dynamically selecting and switching among a pool of candidate models in a data-driven manner, where given an observation sequence y0:k, the posterior probability of the state at the moment k is:
p(x k |y 0:k)=Σm=1 M p(x k |
Figure US12106204-20241001-P00002
=m,y 0:k)p(
Figure US12106204-20241001-P00002
=m|y 0:k)
Where, p(xk|
Figure US12106204-20241001-P00002
=m, y0:k) is the posterior of the state when
Figure US12106204-20241001-P00002
=m; p(
Figure US12106204-20241001-P00002
=m|y0:k) is the posterior probability of selecting the m-th model; M represents the number of the candidate models.
Considering that p(
Figure US12106204-20241001-P00002
=m|y0:k) is obtained by iteration of p(
Figure US12106204-20241001-P00003
=m|y0:k-1), preferably, the Bayesian formula can be used to calculate the posterior probability of selecting the m-th model at the moment k. The formula thereof is as follows:
p ( = m | y 0 : k ) = p ( = m | y 0 : k - 1 ) p m ( y k | y 0 : k - 1 ) Σ j = 1 M p ( = j | y 0 : k - 1 ) p j ( y k | y 0 : k - 1 )
Where, p(
Figure US12106204-20241001-P00002
=M|y0:k-1) is a priorprior probability of selecting the m-th model; pm(yk|y0:k-1) is a marginal likelihood of selecting the m-th model.
Preferably, a forgetting factor α can be determined to retain part of the historical information, that is, the prior probability of selecting the m-th model at the current moment is influenced by the posterior probability of selecting the m-th model at the previous moment. The priorprior probability of selecting the m-th model is calculated as follows:
p ( = m | y 0 : k - 1 ) = p ( = m | y 0 : k - 1 ) α Σ j = 1 M p ( = j | y 0 : k - 1 ) α
where, p(
Figure US12106204-20241001-P00004
=m|y0:k-1) is the probability of selecting the m-th model at the k-1-th moment; a (0<α<1) is a forgetting factor.
Preferably, at the moment k, the calculation method of the marginal likelihood of selecting the m-th model is as follows:
p m(y k |y 0:k-1)=∫p m(y k |x k)p(x k |y 0:k-1)dx k
where pm(yk|xk) is a likelihood function for the m-th model.
In step (4), adopting a particle filter algorithm for the state estimation, and calculating p(xk|
Figure US12106204-20241001-P00005
=m, y0:k) and p(
Figure US12106204-20241001-P00006
=m|y0:k) based on particles.
Based on the traditional state-space model and particle filtering, the present invention proposes a multi-model dynamic ensemble method applied to the neural signal decoding, which reduces the influence caused by the unstability of the neural signal to a certain extent. It outperforms the well-known Kalman filter algorithm in the test, proving the effectiveness of the method.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a schematic figure of a multi-model dynamic ensemble process in the method of the present invention;
FIG. 2 is a comparison chart of the success rate of a Kalman filter and multi-model dynamic ensemble for five consecutive days in an online cursor control experiment in an embodiment of the present invention;
FIG. 3 is a trajectory comparison figure of a Kalman filter and multi-model dynamic ensemble in an online cursor control experiment on Dec. 14, 2020, according to an embodiment of the present invention.
DETAILED DESCRIPTION OF THE EMBODIMENTS
The present invention will be further described in detail below with reference to the accompanying drawings and embodiments. It should be pointed out that the following embodiments are intended to facilitate the understanding of the present invention, but do not have any limiting effect on it.
All clinical and experimental procedures in this embodiment were approved by the Medical Ethics Committee of the Second Affiliated Hospital of Zhejiang University (ethical review number 2019-158). Informed consent was obtained orally by the participants and their immediate family members and signed by their legal representatives. The volunteer was a 74-year-old male who suffered a car accident three years ago and was paralyzed after a C4 cervical spine trauma. The volunteer is only able to move the parts above the neck and has normal language skills and comprehension ability of all tasks. For a limb motion behavior, the patient has a skeletal muscle strength score of 0/5 and a complete loss of limb motion control.
Two 96-channel Utah intracortical microelectrode arrays (4 mm×4 mm, 1.4 mm long Utah arrays, Blackrock Microsystems, Salt Lake City, Utah, USA) were implanted in the left primary motor cortex of the volunteer, and the other was located in the central hand knob area 2 mm away. Electron Computed Tomography and functional MRI were used to guide the implantation. During the implantation, the participant was asked to perform flexion and extension motions of the elbow, and the scanning of the functional MRI was used to confirm an activation area in the motor cortex. The operation used a robotic arm to assist in implanting electrodes during the surgery. The participant had a week of rehabilitation before starting the neural signal recording and BCI training tasks.
The participant performed the BCI training tasks each weekday and rested on weekends. The training time was approximately 3 hours per day and comprised of preparation for signal recording, impedance testing, spike classification and paradigm tasks. The experiment was stopped once the participant felt tired or developed physical abnormalities, such as a fever or a urinary tract infection. The entire experiment lasted half a year, recording the neural signals for 12 weeks.
A 2D cursor control task was performed on a computer monitor 1.5 m in front of the volunteer. The task contained eight targets at the top, bottom, left, right and four corners of the screen. The targets can be displayed sequentially in a clockwise or pseudo-random fashion. The relevant parameters of the task can be configured in task settings, comprising a distance from the target to the center of the screen, the diameter of a target ball, the maximum trial time, an arrival threshold, the minimum dwell time and the number of repetitions after failure. For each experimental trial, the volunteer was asked to control a blue cursor ball from the center of the screen to a red target ball. In a successful trial, the distance between the centers of the blue and red balls should be less than a preset arrival threshold, and dwell time should not be shorter than a preset dwell time. The maximum duration of each trial is set from 2 to 5 seconds. An audible cue is given at the end of the trial to indicate its success. A maximum of 4 attempts were made after each trial failed. The default setting was to display the target balls at 15 cm from the center of the screen, with a diameter of 5 cm.
For the adaptive brain-computer interface decoding method based on multi-model dynamic ensemble proposed by the present invention, the specific implementation was as follows:
Step 1, motor neural signal pre-processing: obtaining an original motor neural signal recorded from the electrodes, selecting an appropriate window size to calculate the firing rate of the neural signal, intercepting a valid data segment according to a state label, normalizing and smoothing the data, and obtaining the pre-processed motor neural signal; dividing the data into a training set, a validation set, and a test set according to a proper proportion;
Specifically, the experiment recorded the neural signals by using a Neuroport system (NSP, Blackrock Microsystems). The neural activity is amplified, digitized, and recorded at 30 KHz. Thresholds for neural action potential detection were set from −6.5 RMS to −5.5 RMS for each high-pass filtered (250 Hz cutoff frequency) electrode by using a Central software package (Blackrock Microsystem). At the beginning of each routine task, the researchers manually sort the spikes, which takes about 25 to 35 minutes. A peak activity was converted to firing rate in 20 ms and low-pass filtered by using an exponential smoothing function with a 450 ms window.
Step 2, a state-space model based on the dynamic ensemble:
2-1. An improved observation model: varying a traditional state-space model, using a set of functions (i.e., candidate models) instead of a fixed function to dynamically characterize the relationship between observation variables and state variables;
The traditional state-space model consists of a state transition function ƒ(·) and an observation function h(·):
x k=ƒ(x k-1)+v k-1
y k =h(x k)+n k
Where, k represents a discrete time step; xk∈Rd x represents the state that is concerned about; yk∈Rd y represents the observation variable; vk, nk represents a state transition noise and an observation noise which are independent and identically distributed.
In the field of the neural decoding, states and observations represent motion trajectories and neural signals, respectively. Given a neural signal sequence y0:k, the task of the neural decoding is to iteratively estimate the probability density of xk. The Kalman filter can provide an analytically optimal solution when both the neural signal and the motion signal conform to a linear Gaussian assumption.
The observation function h(·) of the traditional state-space model is determined in advance and cannot adapt to changing neural signals. The observation model in step 2 improves upon it by allowing the observation function to be adjusted online. The improved observation function is expressed as follows:
y k=
Figure US12106204-20241001-P00007
(x k)+n k
Where, k represents a discrete time step; xk∈ Rd x represents a state of interest; yk∈Rd y represents the observation variable; nk represents an independent and identically distributed observation noise;
Figure US12106204-20241001-P00008
∈{1, 2, . . . , M} represents a model index in the observation equation. In particular,
Figure US12106204-20241001-P00009
=m represents the model in action at the moment k is hm.
The models in a model set
Figure US12106204-20241001-P00010
, similar to neural encoders, can convert input kinematic parameters xt into the neural signals yt. In the multi-model dynamic ensemble method, we learn a pool of forms of encoding models, such as linear functions, polynomial functions, and neural networks, to enhance the ensemble effectiveness of online models. Past research has demonstrated an excellent decoding ability of linear models, while non-linear models have strong robustness in the presence of noise. As shown in FIG. 1 , the multi-model dynamic ensemble method demonstrates the advantages of different models to deal with unstable neural signals.
Specifically, the model pool of this example contains four models. The first model is a linear function Hl(x)=Wx+b, which is similar to the observation model in the Kalman filter. The second model is a second-order polynomial function Hp (x)=a0+a1x+a2x2. The third model Hnn1(·) and the fourth model Hnn2 (·) are two neural networks with different parameter sizes. Both the neural networks contain one hidden layer. The number of neurons in the input layer is equal to the dimensions of the motion parameters xt, and the number of neurons in the output layer is equal to the dimensions of the neural signal yt. The number of neurons in the hidden layers are 30 and 50, respectively.
For Hl(·) and Hp (·), the parameters are estimated by using the least square method. For Hnn1(·) and Hnn2 (·), the parameters are optimized by using the Adam algorithm, a learning rate is set to 0.01, and a weight decay is set to 1e-4. At the same time, an early-stopping method is adopted to alleviate the overfitting problem.
2-2 Dynamically combining the candidate models according to a Bayesian update mechanism as an observation function of the state-space model;
In step 2, a set of candidate models are used to characterize the functional relationship between the observation signal and the state to be predicted. The Bayesian update mechanism dynamically selects dominant models among these models in a data-driven manner. Given an observation sequence y0:k, the posterior probability of the state at the moment k is:
p(x k |y 0:k)=Σm=1 M p(x k |
Figure US12106204-20241001-P00011
=m,y 0:k)p(
Figure US12106204-20241001-P00011
=m|y 0:k)
Where, p(xk|Hk=m, y0:k) is the posterior of the state when
Figure US12106204-20241001-P00011
=p(
Figure US12106204-20241001-P00011
=m|y0:k) is the posterior probability of selecting the m-th model.
Considering that p(
Figure US12106204-20241001-P00011
=m|y0:k) is obtained by iteration of p(
Figure US12106204-20241001-P00012
=m|y0:k-1), preferably, the Bayesian formula can be used to calculate the posterior probability of selecting the m-th model at the moment k.
p ( = m | y 0 : k ) = p ( = m | y 0 : k - 1 ) p m ( y k | y 0 : k - 1 ) Σ j = 1 M p ( = j | y 0 : k - 1 ) p j ( y k | y 0 : k - 1 )
Where, p(
Figure US12106204-20241001-P00011
=m|y0:k-1) is a priorprior probability of selecting the m-th model; pm(yk|y0:k-1) is a marginal likelihood of selecting the m-th model.
Preferably, a forgetting factor α can be determined to retain part of the historical information, that is, the prior probability of selecting the m-th model at the current moment is influenced by the posterior probability of selecting the m-th model at the previous moment:
p ( = m | y 0 : k - 1 ) = p ( = m | y 0 : k - 1 ) α Σ j = 1 M p ( = j | y 0 : k - 1 ) α
Where, p(
Figure US12106204-20241001-P00013
=m|y0:k-1) is the probability of selecting the m-th model at the k-1-th moment; α(0<α<1) is a forgetting factor.
In this example, the forgetting factor is set from 0.1 to 0.9. When the forgetting factor is set to 0.9, the influence of the previous kinematic state is considered to a great extent, so the decoded kinematic state is smoother.
Preferably, at the moment k, the calculation method of the marginal likelihood of selecting the m-th model is as follows:
p m(y k |y 0:k-1)=∫p m(y k |x k)P(x k |y 0:k-1)dx k
Where pm (yk|xk) is a likelihood function for the m-th model.
Step 3, candidate model learning and state estimation: using the training set and the validation set to obtain a set of different candidate models;
Preferably, based on the particle filter algorithm, estimating the kinematic state for each candidate model by using a weighted particle set at each time step, and calculating p(xk|
Figure US12106204-20241001-P00014
=m, y0:k) and p(
Figure US12106204-20241001-P00015
=m|y0:k) in step 2.
Assuming that this moment is the beginning of the k-th time step, and p
Figure US12106204-20241001-P00016
=m|y0:k), m=1, . . . , M and the weight of the particle set are known, where NS represents the size of the particle set and xk-1 i is the particle with the weight of ωk-1 i. Assuming p(xk-1|y0:k-1)≅Σi=1 N s ωk-1 iδ(xk-1−xk-1 i), where δ(·) represents a Dirac δ function, here is how to compute p(xk|
Figure US12106204-20241001-P00017
=m, y0:k) and p(
Figure US12106204-20241001-P00018
=m|y0:k) by using particle filtering:
3-1, p(xk|
Figure US12106204-20241001-P00019
=m, y0:k) based on the particle filtering. First, xk i, is obtained from the state transition prior p(xk|xk-1 i),i=1, . . . , NS, and then according to the principle of importance sampling, we can get:
p(x k |
Figure US12106204-20241001-P00019
=y 0:k)≈Σi=1 N s ωm,k iδ(x k −x k i)
Where, ωm,k i∝ωk-1 ipm(yk|xk i), Σi=1 N s ωm,k i=1. ωm,k i represents the normalized importance weight of the i-th particle when the m-th model is selected.
In this example, the training set is used as the particle set to avoid the unstability caused by randomly scattering the particles.
3-2, p(
Figure US12106204-20241001-P00019
=m|y0:k) based on the particle filtering. Given p(
Figure US12106204-20241001-P00020
=m|y0:k-1), first, the priorprior probability of
Figure US12106204-20241001-P00019
=m is computed by using the forgetting factor in step 2. The posterior probability pm(yk=m|y0:k) of
Figure US12106204-20241001-P00019
=m can be calculated when the likelihood pm(yk|y0:k-1), m=1, . . . , M is known. In the step 3-1, the state transition prior is used as the importance function, i.e., q(xk|xk-1, y0:k)=p(xk|xk-1). Therefore, the distribution of xk can be approximated by the particles, i.e, p(xk|y0:k-1)≈Σi=1 iδx k i. Further, we get:
p m(y k |y 0:k-1)≈Σi=1 N s ωk-1 i p m(y k |x k i)
Note that the particle filtering usually suffers from particle degradation, and after a few iterations, only a small number of particles have high weights. Therefore, we adopt a resampling method to remove particles with too small or too large weights to alleviate the particle degradation problem.
Step 4, performance evaluation of the multi-model dynamic ensemble method: using this method to compare with other methods in the test set data to evaluate the performance and robustness of this method.
This example was an online experiment with closed-loop calibration. Two phases of calibration and testing were contained in each experiment. Each session contains several blocks, and each block consists of 16 trials, where 2 trials for each target. The neural decoder was trained during the calibration phase. There were two observation blocks in the calibration phase, and the computer automatically completed the cursor control task. The neural signals of the volunteer during the observation phase were used to initialize the decoder. Subsequently, the decoder was further adjusted by using computer-assistant and ortho-impedance assistance, and the assistant ratio decreased with the increase of the number of blocks. The calibration process lasted about 10 minutes, and then the neural signals coming in the test phase were decoded by using the trained decoder.
In order to compare and illustrate the effectiveness of the neural signal decoding method proposed by the present invention, the success rate statistics were performed on the online experimental data for five consecutive days, and FIG. 2 was obtained. Compared with the Kalman filter, the online success rate of the multi-model dynamic ensemble method was increased by 2.30% (the big ball), 5.68% (the small ball), 13.85% (tracing), and the arrival time was shortened by 3.52% (the big ball), 8.81% (the small ball), 13.46% (tracing), and the performance of different days was more stable, the variance was 47.50% smaller than the Kalman filter (the big ball), 70.42% (the small ball), 64.51% (tracing). FIG. 3 is the trajectory diagram of the ball test phase on Dec. 14, 2020. It can be seen that the trajectory solved by the Kalman filter has a strong bias and cannot reach each target point on time.
However, the trajectory distribution solved by the multi-model dynamic ensemble method is relatively uniform and successfully reaches each target point. It can be seen from the above results that the performance of the multi-model dynamic ensemble method is more stable.
The above-mentioned embodiments describe the technical solutions and beneficial effects of the present invention in detail. It should be understood that the above-mentioned embodiments are only specific embodiments of the present invention, and are not intended to limit the present invention. Any modifications, additions and equivalent replacements made within the scope of the principles of the present invention shall be included within the protection scope of the present invention.

Claims (8)

The invention claimed is:
1. A computer-implemented adaptive brain-computer interface decoding method based on multi-model dynamic ensemble and executed by a processor, where the method comprises the following steps:
(1) obtaining an original motor and neural signal collected by electrodes implanted in a brain of a subject, specifically comprising: planting the electrodes in a primary motor cortex of the subject for recording the original motor and neural signal; the subject being asked to imagine performing a two dimensional (2D) cursor control task on a computer monitor to move a cursor from a center of the computer monitor through a trajectory to a target; recording the subject's neural signals simultaneously during the cursor control task to obtain the original motor and neural signal; and after pre-processing the original motor and neural signal, dividing the signal into a training set, a validation set and a test set;
(2) building a state-space decoder based on the multi-model dynamic ensemble, which is specified as follows:
(2-1) with a state-space model, using a pool of candidate models to dynamically characterize a relationship between observation variables and state variables, where the pool of candidate models comprise a linear function H1(x)=Wx+b, wherein Hl(x) represents a linear mapping function from movement states x to neural signals, W and b represent linear regression parameters; a second-order polynomial function Hp(x)=a0+a1x+a2x2, wherein Hp(x) represents a polynomial mapping function from movement states x to neural signals, where a0, a1 and a2 represent the polynomial regression parameters for the term of bias, x and x2, respectively, and two neural networks Hnn1(·) and Hnn2(·), and, the observation variables are neural signals, and the state variables are kinematic signals;
(2-2) dynamically combining the candidate models according to a Bayesian update mechanism as an observation function of the state-space model;
(3) using the training set and the validation set to train and evaluate the state-space model, and obtaining parameters of different candidate models by training;
(4) using the test set to test a performance and robustness of the state-space model, applying a neural signal decoding, and obtaining a state estimation after inputting the pre-trained neural signal to be decoded;
where in the step (2-1), an expression of the state-space model is as follows:

X k =f(x k-1)+V k-1

y k =h H k (x k)+n k
where, k represents a discrete time step; xk∈Rd x represents a state of interest; ƒ(·) represents a state transition function; yk ∈Rd y represents an observation or measurement variable; nk represents an independent and identically distributed observation noise; Hk ∈{1, 2, . . . , M} represents a model index in the observation function h(·), Hk=m represents that the model used at the moment k is hm;
for Hl(·) and Hp(·), the parameters are estimated by using the least square method, for Hnn1(·) and Hnn2(·), the parameters are optimized by using an Adam algorithm, a learning rate is set to 0.01, and a weight decay is set to 1e-4; at a same time, an early-stopping method is adopted to alleviate an overfitting problem;
where in the step (2-2), the Bayesian update mechanism is as follows:
dynamically selecting and switching among a pool of candidate models in a data-driven manner, where given an observation sequence y0:k, a posterior probability of a state at the moment k is:

p(x k |y 0:k)=Σm=1 M p(x k |
Figure US12106204-20241001-P00002
=m,y 0:k)p(
Figure US12106204-20241001-P00002
=m|y 0:k)
where, p(xk|Hk=m, y0:k) is the posterior of the state when Hk=m; p(Hk=m|y0:k) is the posterior probability of selecting the m-th model; M represents the number of the candidate models;
wherein, the step (3) specifically comprises the following steps:
step (3.1): based on a particle filter algorithm, estimating motion state for each candidate model by using a weighted particle set at each time step, and calculating p(xk|Hk=m, y0:k) and p(Hk=m|y0:k) in the step (2);
step (3.2): assuming that this moment is beginning of the k-th time step, and p(Hk-1=m|y0:k), m=1, . . . , M and a particle set with a weight are known, wherein Ns represents the size of the particle set and xk-1 i is the particle with the weight of ωk-1 i, and assuming p(xk-1|y0:k-1)≅Σi=1 N s ωk-1 iδ(xk-1−xk-1 i), wherein δ(·) represents a Dirac δ function, calculating p(xk|
Figure US12106204-20241001-P00017
=m, y0:k) and p(
Figure US12106204-20241001-P00018
=m|y0:k) by using particle filtering as follows:
step (3.2.1): obtaining xk i from the state transition priori p(xk|Xk-1 i), i=1, . . . , Ns, and then according to a principle of importance sampling, obtaining:

p(x k |H k=m,y0:k)≈Σi=1 N s ωm,k iδ(x k −x k i)
wherein, ωm,k i∝ωk-1 ipm(yk|xk i), Σi=1 N s ωm,k i=1·ωm,k i represents the normalized importance weight of the i-th particle when the m-th model is selected;
step (3.2.2): assuming p(Hk-1=m|y0:k-1), first, computing the priori probability of Hk=m by using the forgetting factor in the step (2), calculating posterior probability p(Hk=m|y0:k) of Hk=m when the likelihood Pm(yk|y0:k-1), m=1, . . . , M is known; in the step (3.2.1), using state transition priori is the function, q(xk|xk-1, y0:k)=p(xk|xk-1), approximating the distribution of xk by the particles, p(xk-1|y0:k-1)≅Σi=1 N s ωk-1 iδxk i, and obtaining:

p m(y k |y 0:k-1)≈Σi=1 N s ωk-1 i p m(y k |x k i)
step (5): outputting decoded neural signal to an exoskeleton device or a computer cursor control device.
2. The computer-implemented adaptive brain-computer interface decoding method based on multi-model dynamic ensemble and executed by a processor according to claim 1, where in step (1), the specific process of the pre-processing is as follows:
selecting an appropriate window size to calculate a firing rate of the neural signal, intercepting a valid data segment according to a state label, normalizing and smoothing the data respectively, and obtaining the pre-processed motor and neural signal.
3. The computer-implemented adaptive brain-computer interface decoding method based on multi-model dynamic ensemble and executed by a processor according to claim 1, where the posterior probability of selecting the m-th model is calculated as follows:
p ( = m | y 0 : k ) = p ( = m | y 0 : k - 1 ) p m ( y k | y 0 : k - 1 ) Σ j = 1 M p ( = j | y 0 : k - 1 ) p j ( y k | y 0 : k - 1 )
where, p(Hk=m|y0:k-1) is a prior probability of selecting the m-th model; Pm (yk|y0:k-1) is a marginal likelihood of selecting the m-th model.
4. The computer-implemented adaptive brain-computer interface decoding method based on multi-model dynamic ensemble and executed by a processor according to claim 3, where the prior probability of selecting the m-th model is calculated as follows:
p ( = m | y 0 : k - 1 ) = p ( = m | y 0 : k - 1 ) α Σ j = 1 M p ( = j | y 0 : k - 1 ) α
where, p(Hk-1=m|y0:k-1) is the probability of selecting the m-th model at the k-1-th moment; α is a forgetting factor, 0<α<1.
5. The computer-implemented adaptive brain-computer interface decoding method based on multi-model dynamic ensemble and executed by a processor according to claim 3, where the marginal likelihood of selecting the m-th model is calculated as follows:

p m(y k |y 0:k-1)=∫p m(y k |x k)P(x k |y 0:k-1)dx k
where pm(yk|xk) is a likelihood function for the m-th model.
6. The computer-implemented adaptive brain-computer interface decoding method based on multi-model dynamic ensemble and executed by a processor according to claim 3, where in the step (4), adopting a particle filter algorithm for the state estimation, and calculating p(xk|Hk=m,y0:k) and p(Hk=m|y0:k) based on particles.
7. A computer-implemented adaptive brain-computer interface decoding method based on multi-model dynamic ensemble and executed by a processor, where the method comprises the following steps:
(1) obtaining an original motor and neural signal collected by electrodes implanted in a brain of a subject, specifically comprising: planting the electrodes in a primary motor cortex of the subject for recording the original motor and neural signal; the subject being asked to imagine performing a two dimensional (2D) cursor control task on a computer monitor to move a cursor from a center of the computer monitor through a trajectory to a target; recording the subject's neural signals simultaneously during the cursor control task to obtain the original motor and neural signal; and after pre-processing the original motor and neural signal, dividing the signal into a training set, a validation set and a test set;
(2) building a state-space decoder based on the multi-model dynamic ensemble, which is specified as follows:
(2-1) with a state-space model, using a pool of candidate models to dynamically characterize a relationship between observation variables and state variables, where the pool of candidate models comprise a linear function, a second-order polynomial function, and two neural networks, and, the observation variables are neural signals, and the state variables are kinematic signals;
(2-2) dynamically combining the candidate models according to a Bayesian update mechanism as an observation function of the state-space model;
(3) using the training set and the validation set to train and evaluate the state-space model, and obtaining parameters of different candidate models by training;
(4) using the test set to test a performance and robustness of the state-space model, applying a neural signal decoding, and obtaining a state estimation after inputting the pre-trained neural signal to be decoded;
step (5): outputting decoded neural signal to an exoskeleton device or a computer cursor control device;
where in step (1), the specific process of the pre-processing is as follows:
selecting an appropriate window size to calculate a firing rate of the neural signal, intercepting a valid data segment according to a state label, normalizing and smoothing the data respectively, and obtaining the pre-processed motor and neural signal;
where in the step (2), an expression of the state-space model is as follows:

X k =f(x k-1)+V k-1

y k =h H k (x k)+n k
where, k represents a discrete time step; xk∈ Rd x represents a state of interest; f(·) represents a state transition function; yk∈Rd y represents a state of observation or measurement variable; nk represents an independent and identically distributed observation noise; Hk ∈{1,2, . . . , M} represents a model index in the observation function h(·), Hk=m represents that the model used at the moment k is hm;
where in the step (2), the Bayesian update mechanism is as follows:
dynamically selecting and switching among a pool of candidate models in a data-driven manner, where given an observation sequence y0:k, a posterior probability of a state at the moment k is:

p(x k |y 0:k)=Σm=1 M p(x k |H k =m,y 0:k)p(H k =m|y 0:k)
where, p(xk|Hk=m,y0:k) is the posterior of the state when Hk=m; p(Hk=m|y0:k)is the posterior probability of selecting the m-th model; M represents the number of the candidate models;
where the posterior probability of selecting the m-th model is calculated as follows:
p ( k = m "\[LeftBracketingBar]" y 0 : k ) = p ( k = m "\[LeftBracketingBar]" y 0 : k - 1 ) p m ( y k "\[LeftBracketingBar]" y 0 : k - 1 ) j = 1 M p ( k = j "\[LeftBracketingBar]" y 0 : k - 1 ) p j ( y k "\[LeftBracketingBar]" y 0 : k - 1 )
where, p(Hk=m|y0:k-1) is a prior probability of selecting the m-th model; Pm (yk|y0:k-1) is a marginal likelihood of selecting the m-th model;
where the prior probability of selecting the m-th model is calculated as follows:
p ( k = m "\[LeftBracketingBar]" y 0 : k - 1 ) = p ( k - 1 = m "\[LeftBracketingBar]" y 0 : k - 1 ) α j = 1 M p ( k - 1 = j "\[LeftBracketingBar]" y 0 : k - 1 ) α
where, p(Hk-1=m|y0:k-1) is the probability of selecting the m-th model at the k-1-th moment; α is a forgetting factor, 0<α<1;
where the marginal likelihood of selecting the m-th model is calculated as follows:

p m(y k |y 0:k-1)=∫p m(y k |x k)P(x k |y 0:k-1)dx k
where pm (yk|xk) is a likelihood function for the m-th model;
wherein, the step (3) specifically comprises the following steps:
step (3.1): based on a particle filter algorithm, estimating motion state for each candidate model by using a weighted particle set at each time step, and calculating p(xk|Hk=m, y0:k) and p(Hk=m|y0:k) in the step (2);
step (3.2): assuming that this moment is beginning of the k-th time step, and p(Hk-1=m|y0:k), m=1, . . . , M and a particle set with a weight are known, wherein Ns represents the size of the particle set and xk-1 i is the particle with the weight of ωk-1 i, and assuming p(xk-1|y0:k-1)≅Σi=1 N s ωk-1 iδ(xk-1−xk-1 i), wherein δ(·) represents a Dirac δ function, calculating p(xk|
Figure US12106204-20241001-P00017
=m, y0:k) and p(Hk=m|y0:k) by using particle filtering as follows:
step (3.2.1): obtaining xk i from the state transition priori p(xk|Xk-1 i), i=1, . . . , Ns, and then according to the principle of importance sampling, obtaining:

p(x k |H k=m,y0:k)≈Σi=1 N s ωm,k iδ(x k −x k i)
wherein, ωm,k i∝ωk-1 ipm(yk|xk i), Σi=1 N s ωm,k i=1·ωm,k i represents the normalized importance weight of the i-th particle when the m-th model is selected;
step (3.2.2): assuming p(Hk-1=m|y0:k-1), first, computing the priori probability of Hk=m by using the forgetting factor in the step (2), calculating posterior probability p(Hk=m|y0:k) of Hk=m when the likelihood Pm(yk|y0:k-1), m=1, . . . , M is known; in the step (3.2.1), using state transition priori is the function, q(xk|xk-1, y0:k)=p(xk|xk-1), approximating the distribution of xk by the particles, p(xk-1|y0:k-1)≅Σi=1 N s ωk-1 iδxk i, and obtaining:

p m(y k |y 0:k-1)≈Σi=1 N s ωk-1 i p m(y k |x k i)
where in the step (4), adopting a particle filter algorithm for the state estimation, and calculating p(xk|Hk=m, y0:k) and p(Hk=m|y0:k) based on particles.
8. The computer-implemented adaptive brain-computer interface decoding method based on multi-model dynamic ensemble and executed by a processor according to claim 4, wherein α=0.9.
US17/777,956 2020-12-29 2021-10-27 Adaptive brain-computer interface decoding method based on multi-model dynamic integration Active 2041-10-27 US12106204B2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
CN202011593629.3 2020-12-29
CN202011593629.3A CN112764526B (en) 2020-12-29 2020-12-29 Self-adaptive brain-computer interface decoding method based on multi-model dynamic integration
PCT/CN2021/126580 WO2022142653A1 (en) 2020-12-29 2021-10-27 Adaptive brain-computer interface decoding method based on multi-model dynamic integration

Publications (2)

Publication Number Publication Date
US20230244909A1 US20230244909A1 (en) 2023-08-03
US12106204B2 true US12106204B2 (en) 2024-10-01

Family

ID=75696097

Family Applications (1)

Application Number Title Priority Date Filing Date
US17/777,956 Active 2041-10-27 US12106204B2 (en) 2020-12-29 2021-10-27 Adaptive brain-computer interface decoding method based on multi-model dynamic integration

Country Status (3)

Country Link
US (1) US12106204B2 (en)
CN (1) CN112764526B (en)
WO (1) WO2022142653A1 (en)

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112764526B (en) * 2020-12-29 2022-10-21 浙江大学 Self-adaptive brain-computer interface decoding method based on multi-model dynamic integration
CN114925734B (en) * 2022-07-20 2022-11-25 浙江大学 Online neuron classification method based on neural mimicry calculation
CN115358367B (en) * 2022-08-09 2023-04-18 浙江大学 Dynamic self-adaptive brain-computer interface decoding method based on multi-model learning integration
CN116187152B (en) * 2022-10-24 2023-08-25 浙江大学 An Adaptive Decoding Method for Invasive Brain-Computer Interface Based on Dynamic Evolutionary Computation
CN115617180B (en) * 2022-12-02 2023-04-07 浙江大学 Smart hand motion decoding method based on invasive brain-computer interface
CN116432083A (en) * 2023-03-30 2023-07-14 中国科学院深圳先进技术研究院 Intention prediction method, device, equipment and storage medium based on brain-computer interface
WO2025043050A1 (en) * 2023-08-24 2025-02-27 Carnegie Mellon University Training and use of a posture invariant brain-computer interface
US12386424B2 (en) * 2023-11-30 2025-08-12 Zhejiang University Chinese character writing and decoding method for invasive brain-computer interface
CN118252513B (en) * 2024-05-29 2024-09-20 浙江大学 A positioning method and system based on electroencephalogram signal

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150245928A1 (en) * 2010-02-18 2015-09-03 The Board Of Trustees Of The Leland Stanford Junior University Brain-Machine Interface Utilizing Interventions to Emphasize Aspects of Neural Variance and Decode Speed and Angle
US20160048753A1 (en) 2014-08-14 2016-02-18 The Board Of Trustees Of The Leland Stanford Junior University Multiplicative recurrent neural network for fast and robust intracortical brain machine interface decoders
CN109770900A (en) 2019-01-08 2019-05-21 中国科学院自动化研究所 Brain-computer interface instruction delivery method, system and device based on convolutional neural network
US20200175354A1 (en) * 2018-12-03 2020-06-04 Deep Learn, Inc. Time and accuracy estimate-based selection of machine-learning predictive models
CN112001306A (en) 2020-08-21 2020-11-27 西安交通大学 Electroencephalogram signal decoding method for generating neural network based on deep convolution countermeasure

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101515200B (en) * 2009-04-03 2010-12-01 北京工业大学 Target selection method based on transient visually evoked EEG
CN104182042B (en) * 2014-08-14 2017-07-11 华中科技大学 A kind of brain-machine interface method of multi-modal signal
CN110090017B (en) * 2019-03-11 2021-09-14 北京工业大学 Electroencephalogram signal source positioning method based on LSTM
CN110751032B (en) * 2019-09-20 2022-08-02 华中科技大学 Training method of brain-computer interface model without calibration
CN111297379A (en) * 2020-02-10 2020-06-19 中国科学院深圳先进技术研究院 A brain-computer integration system and method based on sensory transmission
CN111783857A (en) * 2020-06-18 2020-10-16 内蒙古工业大学 Brain-computer interface for motor imagery based on nonlinear network infographics
CN112764526B (en) * 2020-12-29 2022-10-21 浙江大学 Self-adaptive brain-computer interface decoding method based on multi-model dynamic integration

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150245928A1 (en) * 2010-02-18 2015-09-03 The Board Of Trustees Of The Leland Stanford Junior University Brain-Machine Interface Utilizing Interventions to Emphasize Aspects of Neural Variance and Decode Speed and Angle
US20160048753A1 (en) 2014-08-14 2016-02-18 The Board Of Trustees Of The Leland Stanford Junior University Multiplicative recurrent neural network for fast and robust intracortical brain machine interface decoders
US20200175354A1 (en) * 2018-12-03 2020-06-04 Deep Learn, Inc. Time and accuracy estimate-based selection of machine-learning predictive models
CN109770900A (en) 2019-01-08 2019-05-21 中国科学院自动化研究所 Brain-computer interface instruction delivery method, system and device based on convolutional neural network
CN112001306A (en) 2020-08-21 2020-11-27 西安交通大学 Electroencephalogram signal decoding method for generating neural network based on deep convolution countermeasure

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Dantas et al., "Neural Decoding Using a Nonlinear Generative Model for BrainComputer Interface", 2014, 2014 IEEE International Conference on Acoustic, Speech and Signal Processing (ICASSP), pp. 4683-4687. (Year: 2014). *
Qi et al., "Dynamic Ensemble Modeling Approach to Nonstationary Neural Decoding in Brain-Computer Interfaces", Nov. 2, 2019, arXiv:1911.00714v1, pp. 1-10. (Year: 2019). *
Yu Qi et al., "Dynamic Ensemblle Modeling Approach to Nonstationary Neural Decoding in Brain-Computer Interfaces", 33rd Conference on Neural Information Processing Systems, (NeurIPS 2019), pp. Nos. 1-10, Nov. 2019.

Also Published As

Publication number Publication date
WO2022142653A1 (en) 2022-07-07
CN112764526B (en) 2022-10-21
US20230244909A1 (en) 2023-08-03
CN112764526A (en) 2021-05-07

Similar Documents

Publication Publication Date Title
US12106204B2 (en) Adaptive brain-computer interface decoding method based on multi-model dynamic integration
US12422935B2 (en) Gesture information processing method and apparatus, electronic device, and storage medium
US7826894B2 (en) Cognitive control signals for neural prosthetics
US20030023319A1 (en) Cognitive state machine for prosthetic systems
US8463371B2 (en) System and method for processing brain signals in a BCI system
US20040073414A1 (en) Method and system for inferring hand motion from multi-cell recordings in the motor cortex using a kalman filter or a bayesian model
CN111184512A (en) An action recognition method for upper limb and hand rehabilitation training in stroke patients
CN114099234B (en) An intelligent rehabilitation robot data processing method and system for assisting rehabilitation training
Li et al. Optimized artificial neural network based performance analysis of wheelchair movement for ALS patients
Jaber et al. HD-sEMG gestures recognition by SVM classifier for controlling prosthesis
CN114098768A (en) Cross-individual surface electromyographic signal gesture recognition method based on dynamic threshold and easy TL
Li et al. Cross-user gesture recognition from sEMG signals using an optimal transport assisted student-teacher framework
Kulwa et al. A multidataset characterization of window-based hyperparameters for deep CNN-driven sEMG pattern recognition
Ubeda et al. Classification method for BCIs based on the correlation of EEG maps
CN112998725A (en) Rehabilitation method and system of brain-computer interface technology based on motion observation
Aung et al. A real-time framework for EEG signal decoding with graph neural networks and reinforcement learning
Kumar et al. Analysis of CNN model based classification of diabetic retinopathy diagnosis
Baracat et al. Decoding gestures from intraneural recordings of a transradial amputee using event-based processing
Ho et al. Brain-wave bio potentials based mobile robot control: wavelet-neural network pattern recognition approach
US20230201586A1 (en) Computer vision enhanced electromyography training systems and methods thereof
Mukherjee et al. EEG and EMG Induced Pain-Sensitive Learning Controller forRobotic Knee Rehabilitation Using Deep Learning
US12277272B2 (en) Systems and methods for electromyogram-based control using neural networks
Meng et al. Direct Learning of Neuronal Firing Representations for Long-Term Motor Intent Predictions
Baracat et al. Neuromorphic Event-based Processing of Transradial Intraneural Recording for Online Gesture Recognition
Aviles et al. A novel methodology to classify myoelectric signals using genetic algorithms and support vector machines

Legal Events

Date Code Title Description
AS Assignment

Owner name: ZHEJIANG UNIVERSITY, CHINA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:QI, YU;WANG, YUEMING;ZHU, XINYUN;AND OTHERS;REEL/FRAME:059950/0986

Effective date: 20220428

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: ADVISORY ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE