EP2385789B1 - Device and method for generating a representation of a subject's attention level - Google Patents
Device and method for generating a representation of a subject's attention level Download PDFInfo
- Publication number
- EP2385789B1 EP2385789B1 EP09833759.5A EP09833759A EP2385789B1 EP 2385789 B1 EP2385789 B1 EP 2385789B1 EP 09833759 A EP09833759 A EP 09833759A EP 2385789 B1 EP2385789 B1 EP 2385789B1
- Authority
- EP
- European Patent Office
- Prior art keywords
- subject
- spectral
- score
- attention
- features
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims description 100
- 210000004556 brain Anatomy 0.000 claims description 69
- 238000012549 training Methods 0.000 claims description 51
- 230000002123 temporal effect Effects 0.000 claims description 41
- 239000013598 vector Substances 0.000 claims description 20
- 230000001419 dependent effect Effects 0.000 claims description 18
- 238000004422 calculation algorithm Methods 0.000 claims description 15
- 230000003595 spectral effect Effects 0.000 claims description 15
- 238000003860 storage Methods 0.000 claims description 7
- 230000003935 attention Effects 0.000 description 65
- 238000001514 detection method Methods 0.000 description 27
- 238000013459 approach Methods 0.000 description 22
- 238000000605 extraction Methods 0.000 description 22
- 239000011159 matrix material Substances 0.000 description 12
- 230000008569 process Effects 0.000 description 12
- 208000006096 Attention Deficit Disorder with Hyperactivity Diseases 0.000 description 11
- 208000036864 Attention deficit/hyperactivity disease Diseases 0.000 description 11
- 230000003044 adaptive effect Effects 0.000 description 11
- 208000015802 attention deficit-hyperactivity disease Diseases 0.000 description 11
- 238000004590 computer program Methods 0.000 description 10
- 230000000694 effects Effects 0.000 description 9
- 238000010801 machine learning Methods 0.000 description 8
- 238000011282 treatment Methods 0.000 description 8
- 238000010586 diagram Methods 0.000 description 7
- 238000005259 measurement Methods 0.000 description 7
- 238000012545 processing Methods 0.000 description 7
- 238000004458 analytical method Methods 0.000 description 6
- 238000013528 artificial neural network Methods 0.000 description 5
- 238000000537 electroencephalography Methods 0.000 description 5
- 238000013480 data collection Methods 0.000 description 4
- 238000009826 distribution Methods 0.000 description 4
- 238000010237 hybrid technique Methods 0.000 description 4
- 230000009467 reduction Effects 0.000 description 4
- 230000004044 response Effects 0.000 description 4
- 238000001228 spectrum Methods 0.000 description 4
- 238000012706 support-vector machine Methods 0.000 description 4
- 230000000007 visual effect Effects 0.000 description 4
- 238000002507 cathodic stripping potentiometry Methods 0.000 description 3
- 230000001276 controlling effect Effects 0.000 description 3
- 238000013500 data storage Methods 0.000 description 3
- 238000001914 filtration Methods 0.000 description 3
- 230000006872 improvement Effects 0.000 description 3
- 230000007246 mechanism Effects 0.000 description 3
- 230000015654 memory Effects 0.000 description 3
- 208000024827 Alzheimer disease Diseases 0.000 description 2
- 230000008901 benefit Effects 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 2
- 239000012141 concentrate Substances 0.000 description 2
- 238000002790 cross-validation Methods 0.000 description 2
- 229940079593 drug Drugs 0.000 description 2
- 239000003814 drug Substances 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 210000001061 forehead Anatomy 0.000 description 2
- 230000004927 fusion Effects 0.000 description 2
- 210000003128 head Anatomy 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 238000012805 post-processing Methods 0.000 description 2
- 238000005070 sampling Methods 0.000 description 2
- 210000004761 scalp Anatomy 0.000 description 2
- 230000035945 sensitivity Effects 0.000 description 2
- 230000001755 vocal effect Effects 0.000 description 2
- 206010000087 Abdominal pain upper Diseases 0.000 description 1
- 206010000117 Abnormal behaviour Diseases 0.000 description 1
- 241000251468 Actinopterygii Species 0.000 description 1
- 206010053759 Growth retardation Diseases 0.000 description 1
- 206010019233 Headaches Diseases 0.000 description 1
- 229910021607 Silver chloride Inorganic materials 0.000 description 1
- 208000013738 Sleep Initiation and Maintenance disease Diseases 0.000 description 1
- 206010047571 Visual impairment Diseases 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 230000006978 adaptation Effects 0.000 description 1
- 230000000454 anti-cipatory effect Effects 0.000 description 1
- 230000036528 appetite Effects 0.000 description 1
- 235000019789 appetite Nutrition 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 208000010877 cognitive disease Diseases 0.000 description 1
- 230000008867 communication pathway Effects 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 238000005094 computer simulation Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 230000002708 enhancing effect Effects 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 231100000001 growth retardation Toxicity 0.000 description 1
- 231100000869 headache Toxicity 0.000 description 1
- 208000013403 hyperactivity Diseases 0.000 description 1
- 208000035231 inattentive type attention deficit hyperactivity disease Diseases 0.000 description 1
- 206010022437 insomnia Diseases 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000012432 intermediate storage Methods 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 238000007726 management method Methods 0.000 description 1
- 238000002483 medication Methods 0.000 description 1
- 230000003340 mental effect Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 238000012567 pattern recognition method Methods 0.000 description 1
- 230000000144 pharmacologic effect Effects 0.000 description 1
- 238000011458 pharmacological treatment Methods 0.000 description 1
- 230000036417 physical growth Effects 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 230000002360 prefrontal effect Effects 0.000 description 1
- 208000020016 psychiatric disease Diseases 0.000 description 1
- 230000008430 psychophysiology Effects 0.000 description 1
- 230000001105 regulatory effect Effects 0.000 description 1
- 230000000284 resting effect Effects 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 238000010187 selection method Methods 0.000 description 1
- 230000010332 selective attention Effects 0.000 description 1
- 230000000638 stimulation Effects 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
- 230000002459 sustained effect Effects 0.000 description 1
- 208000024891 symptom Diseases 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 230000010389 voluntary attention Effects 0.000 description 1
- 230000002087 whitening effect Effects 0.000 description 1
- 230000003936 working memory Effects 0.000 description 1
Images
Classifications
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/20—Input arrangements for video game devices
- A63F13/21—Input arrangements for video game devices characterised by their sensors, purposes or types
- A63F13/212—Input arrangements for video game devices characterised by their sensors, purposes or types using sensors worn by the player, e.g. for measuring heart beat or leg activity
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/24—Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
- A61B5/316—Modalities, i.e. specific diagnostic methods
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/24—Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
- A61B5/316—Modalities, i.e. specific diagnostic methods
- A61B5/369—Electroencephalography [EEG]
- A61B5/372—Analysis of electroencephalograms
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/24—Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
- A61B5/316—Modalities, i.e. specific diagnostic methods
- A61B5/369—Electroencephalography [EEG]
- A61B5/372—Analysis of electroencephalograms
- A61B5/374—Detecting the frequency distribution of signals, e.g. detecting delta, theta, alpha, beta or gamma waves
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/24—Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
- A61B5/316—Modalities, i.e. specific diagnostic methods
- A61B5/369—Electroencephalography [EEG]
- A61B5/375—Electroencephalography [EEG] using biofeedback
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7235—Details of waveform analysis
- A61B5/7264—Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems
- A61B5/7267—Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems involving training the classification device
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/45—Controlling the progress of the video game
- A63F13/46—Computing the game score
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/60—Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor
- A63F13/67—Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor adaptively or by learning from player actions, e.g. skill level adjustment or by storing successful combat sequences for re-use
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/015—Input arrangements based on nervous system activity detection, e.g. brain waves [EEG] detection, electromyograms [EMG] detection, electrodermal response detection
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/24—Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
- A61B5/316—Modalities, i.e. specific diagnostic methods
- A61B5/369—Electroencephalography [EEG]
- A61B5/377—Electroencephalography [EEG] using evoked responses
- A61B5/378—Visual stimuli
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/24—Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
- A61B5/316—Modalities, i.e. specific diagnostic methods
- A61B5/369—Electroencephalography [EEG]
- A61B5/377—Electroencephalography [EEG] using evoked responses
- A61B5/38—Acoustic or auditory stimuli
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/40—Detecting, measuring or recording for evaluating the nervous system
- A61B5/4058—Detecting, measuring or recording for evaluating the nervous system for evaluating the central nervous system
- A61B5/4064—Evaluating the brain
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/10—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals
- A63F2300/1012—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals involving biosensors worn by the player, e.g. for measuring heart beat, limb activity
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/60—Methods for processing data by generating or executing the game program
- A63F2300/6027—Methods for processing data by generating or executing the game program using adaptive systems learning from user actions, e.g. for skill level adjustment
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/60—Methods for processing data by generating or executing the game program
- A63F2300/61—Score computation
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/70—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for mining of medical data, e.g. analysing previous cases of other patients
Definitions
- the present invention relates broadly to a device and method for generating a representation of a subject's attention level, and to a computer storage medium having stored thereon computer code means for instructing a computing device to execute a method of generating a representation of a subject's attention level.
- US 2002/128540 A1 discloses an approach for the enhancement of attention and further the treatment of attention deficit disorder.
- the approach includes presentation of a virtual reality environment for enhancing attention shift, selective attention, and sustained attention by correlating a biofeedback device that estimates the current attention level of the patient.
- KAI KENG ANG ET AL "Filter Bank Common Spatial Pattern (FBCSP) in Brain-Computer Interface", NEURAL NETWORKS, 2008, IJCNN 2008, (IEEE WORLD CONGRESS ON COMPUTATIONAL INTELLIGENCE), IEEE INTERNATIONAL JOINT CONFERENCE ON, IEEE, PISCATAWAY, NJ, USA, 1 June 2008 (2008-06-01), pages 2390-2397, XP031395786, ISBN: 978-1-4244-1820-6 , teach extracting discriminative patterns from an EEG using the Common Spatial Pattern algorithm. An autonomous selection of key temporal-spatial discriminative EEG characteristics is performed.
- United States Patent No. 6402520 describes regulating theta and beta wave activity (more specifically, to decrease theta wave activity and increase beta wave activity) as measured based on respective average millivolt activity.
- a clear and direct correlation between the measured average millivolt theta and beta activities and attention has not been established.
- the users involved in the training approaches in that document only leam to control their measured average millivolt theta and beta activities, which are not a direct measure of attention.
- the users may develop a mechanism for controlling the measured average millivolt theta and beta activities, this does not directly correlate with achieving higher attention levels.
- a device for generating a representation of a subject's attention level comprising:
- the means for presenting may present said score in the form of a game.
- the means for presenting said score may adaptively adjust at least one control parameter of the game based on said score.
- the adjusting of said one parameter may comprise:
- the extracting of the temporal features from the brain signals may comprise:
- the statistics of the brain waveforms may be standard deviations.
- the selecting of the spectral-spatial features containing discriminative information between concentration and non-concentration states from the set of extracted spectral-spatial features may comprise selecting spectral-spatial features based on the mutual dependence of the features with respect to the concentration and non-concentration states.
- the combining of the scores x 1 and x 2 to give a single score may comprise:
- the classifier may comprise one or more of a group consisting of a Linear Discriminant Analysis classifier, Neural Networks, Support Vector Machines, Fuzzy Inference System, Tree-based classifiers, Fuzzy Type 2 and Relevance Vector Machine.
- the device may use training data to generate parameters for classifying the extracted temporal features using a classifier, for extracting spectral-spatial features from brain signals, for selecting spectral-spatial features containing discriminative information between the concentration and non-concentration states from the set of extracted spectral-spatial features and for classifying the selected spectral-spatial features using a classifier.
- the parameters may comprise one or more of a group consisting of projection matrices of CSPs for the CSP algorithm, parameters for selecting spectral-spatial features based on mutual information and a model for the classifiers.
- the use of training data to generate parameters may comprise:
- the presenting of said score to the subject may comprise presenting said score in the form of a game.
- the presenting of said score to the subject may comprise adaptively adjusting at least one control parameter of the game based on said score.
- the adjusting of said one parameter may comprise:
- the statistics of the brain waveforms may be standard deviations.
- the parameters may comprise one or more of a group consisting of projection matrices of CSPs for the CSP algorithm, parameters for selecting spectral-spatial features based on mutual information and a model for the classifiers.
- the use of training data to generate parameters may comprise:
- a computer storage medium having stored thereon computer code means for instructing a computing device to execute a method of generating a representation of a subject's attention level, the method comprising the steps of:
- Example embodiments of the present invention provide a BCI-based system that seeks to improve ADHD treatment, e.g. improving attention of a user.
- the BCI technology of the example embodiments has been implemented in the form of computer games.
- Players can control, using their degree/level of attention and the BCI setup, various parameters of games, e.g. fish jump, racing car with speed, puzzle parts, etc.
- FIG. 1 shows a schematic diagram illustrating an overview of device 100 for generating a representation of a subject's attention level according to an example embodiment.
- the device 100 comprises a plurality of clients 102a, 102b... 102n connected to a server 106 via a network connection 104, e.g. an intranet or the Internet.
- the server 106 is also connected to a player manager 112 which preferably manages the interconnection and data flow for each client 102.
- the client 102 initiates the start of a session, e.g. by requesting the server 106 to provide a game.
- the client 102 then downloads the game from the server 106.
- a game program 110 provided by the server 106 to the client 102 comprises, inter alia, algorithm 114 used for the EEG processing, analysis and classification, attention model 116 and the game 118 embedded therewithin.
- Various types of game can be developed for use in the example embodiment.
- the client 102 can save and send information (i.e. data 108) back to the server 106 to update the respective player's profile stored on the player manager 112.
- the system 100 of the example embodiment is advantageously capable of multi-user operation.
- the user i.e. player
- the present specification also discloses apparatus for performing the operations of the methods.
- Such apparatus may be specially constructed for the required purposes, or may comprise a general purpose computer or other device selectively activated or reconfigured by a computer program stored in the computer.
- the algorithms and displays presented herein are not inherently related to any particular computer or other apparatus.
- Various general purpose machines may be used with programs in accordance with the teachings herein.
- the construction of more specialized apparatus to perform the required method steps may be appropriate.
- the structure of a conventional general purpose computer will appear from the description below.
- the present specification also implicitly discloses a computer program, in that it would be apparent to the person skilled in the art that the individual steps of the method described herein may be put into effect by computer code.
- the computer program is not intended to be limited to any particular programming language and implementation thereof. It will be appreciated that a variety of programming languages and coding thereof may be used to implement the teachings of the disclosure contained herein.
- the computer program is not intended to be limited to any particular control flow. There are many other variants of the computer program, which can use different control flows without departing from the spirit or scope of the invention.
- Such a computer program may be stored on any computer readable medium.
- the computer readable medium may include storage devices such as magnetic or optical disks, memory chips, or other storage devices suitable for interfacing with a general purpose computer.
- the computer readable medium may also include a hard-wired medium such as exemplified in the Internet system, or wireless medium such as exemplified in the GSM mobile telephone system.
- the computer program when loaded and executed on such a general-purpose computer effectively results in an apparatus that implements the steps of the preferred method.
- FIG 2 is a detailed schematic diagram of an implementation of the device of Figure 1 to one client according to an example embodiment.
- the system comprises a physiological data acquisition module 204, an attention detection module 206, an adaptive control module 210, an interface control module 212, and a presentation module 214 with stimulus generator (for providing feedback to a subject (i.e. player) 202).
- the physiological data acquisition module 204 obtains bio-signals from the player 202 and provides the results to the attention detection module 206.
- the attention detection results are in turn provided to the adaptive control module 210.
- the interface control module 210 provides a link between the adaptive control module 210 and the presentation module 214.
- the physiological data acquisition module 204 is used to record physiological bio-signals from the subject 202.
- the bio-signals include the electroencephalogram (EEG), electrooculogram (EoG), electromyogram (EMG).
- EEG electroencephalogram
- EoG electrooculogram
- EMG electromyogram
- Different sensor montages can be used for data acquisition in the example embodiment.
- EEG signals are collected from a central-parietal region of the scalp.
- EEG signals are collected from a pre-frontal area.
- the filters in the example embodiment can be Infinite Impulse Response (IIR) or Finite Impulse Response (FIR) filters.
- IIR Infinite Impulse Response
- FIR Finite Impulse Response
- artefacts are advantageously removed and detection accuracy improved in the later stage.
- the filtered signals are then sent to spatial filters corresponding to each frequency band.
- the spatial filters in the example embodiment are based on Common Spatial Pattern (CSP).
- CSP Common Spatial Pattern
- the CSP filters are trained by joint maximization/minimization of the variances for two classes involved.
- the rows of W are the stationary spatial filters and the columns of W -1 are the common spatial patterns.
- the spatially filtered signal given in Equation (2) maximizes the differences in the variance of the two classes of EEG measurements. However, the variances of only a small number m of the spatial filtered signal are used in the example embodiment as features for classification.
- these feature vectors are used for attention detection. Based on specific tasks, different features are selected and finally a classification is applied.
- the attention detection module 206 advanced pattern recognition methods are used in the example embodiment to process the incoming EEG signal and classify it into attention/non-attention states with a quantifiable score to indicate the level of attention of the subject 202.
- the attention detection of the example embodiment is capable of dealing with two types of scenarios.
- the subject 202 is presented with stimuli and the EEG is recorded so as to monitor the anticipatory aspect of attention, to classify signals which relate to the task presented. This is named reactive / dependent attention in the example embodiment.
- the subject 202 voluntarily directs his/her attention to the auditory and/or visual stimuli, and this is named self-paced attention.
- the attention detection of the example embodiment preferably provides a useful means to tackle specific aspects of attention such as spatial, verbal, and object orientation.
- FIG. 3 shows a flowchart illustrating a method 300 for concentration detection according to an embodiment of the present invention.
- step 302 multi-channel EEG acquisition is performed using a real-time data acquisition and processing platform.
- the data acquisition and processing platform implements the following steps.
- a NuAmps device from Neuroscan, Inc. is first used to measure the scalp brain signals.
- the brain signals are then recorded from Ag-AgCl electrodes placed on the surface of the user's head.
- the digitizer device for the recording of the brain signals works at a sampling rate of 250Hz.
- the recorded brain signals are then filtered via temporal filtering to remove high frequency noises and very slow waves using for example, a 5 th -order digital Butterworth filter with a passband of [0.5Hz 40Hz].
- the filtered brain signals are next downsampled by a factor of 4 in order to reduce the computational complexity.
- step 306 temporal feature extraction is performed.
- Step 306 computes statistics such as the standard deviation of the windowed and pre-processed EEG waveforms in each channel. The statistics are then concatenated into a joint feature vector. The feature vector is then input to step 308.
- a classifier such as the Linear Discriminant Analysis (LDA) is implemented to produce a score, for example x 1 , indicating the likelihood of the hypothesis whereby the hypothesis is that the subject is in a state of concentration i.e. with focused attention.
- LDA Linear Discriminant Analysis
- Steps 306 and 308 form the temporal feature extraction module in the method 300.
- a common spatial pattern (CSP) array is implemented.
- Step 312 applies the CSP algorithm to each EEG component obtained in step 310 to emphasize the difference in spatial distributions of the energy between the two classes, the concentration and the non-concentration classes corresponding to the brain states during which the subject is concentrating and not concentrating respectively.
- the CSP algorithm is detailed in Equation (7) whereby for the j th EEG component, a CSP feature cf(j) is extracted according to Equation (7).
- W l is a matrix comprising of the first I 1 and the last I 2 rows of W, whereby I 1 and I 2 are normalized for data processing efficiency and the ratio between I 1 and I 2 is kept constant.
- the spatial filtering parameters i.e. spatial patterns such as the matrix W are learnt from the examples of the two classes via a subject dependent model training approach which would be elaborated later.
- the CSP array produces an array of spectral-spatial features, each representing the energy of the EEG component projected onto a particular spatial pattern.
- Such an array of features is usually over-redundant since not every spectral-spatial feature is associated with the concentration or non-concentration state in the brain.
- the unnecessary (i.e. redundant) features are removed.
- step 314 a mutual information feature selection is implemented to remove the unnecessary features.
- Step 314 selects a set of features that contains the discriminative information between the concentration and the non-concentration states. This set is determined through a model training procedure via a subject dependent model training approach which would be elaborated later.
- a feature vector is obtained and is input into step 316.
- step 316 a classifier such as the LDA is implemented. Using the feature vector input from step 314, a score, for example x 2 , is produced by the classifier. This score indicates the likelihood of the hypothesis whereby the hypothesis is that the subject is in a state of concentration i.e. with focused attention. Steps 310 - 316 form the spectral-spatial feature extraction module of the method 300.
- Step 318 implements the fusion of the results from the temporal feature extraction module and the spectral-spatial feature extraction module to obtain a single output.
- the continuous outputs of the classifiers in the temporal feature extraction module and the spectral-spatial feature extraction module are normalized.
- the normalized output x n will be (x-m x )/s x whereby m x and s x are respectively the mean and standard deviation of the outputs obtained using the training samples.
- Two normalized outputs x 1n and x 2n from the temporal feature module and the spectral-spatial module respectively are hence obtained.
- these two normalized outputs x 1n and x 2n are combined according to Equation (9) using weights w 1 and w 2 whereby weights w 1 and w 2 correspond to x 1n and x 2n respectively and reflect the individual performance of each of the modules.
- the normalized outputs x 1n and x 2n can also be combined using non-linear methods such as a non-linear weighted regression.
- S a t x 1 ⁇ n * w 1 + x 2 ⁇ n * w 2
- a subject-dependent model training approach is used in the embodiments of the present invention to obtain the parameters and models for the method 300.
- FIG. 4 illustrates a data collection protocol 400 for the subject-dependent model training approach according to an embodiment of the present invention.
- the protocol consists of 4 different tasks to be performed by the subject.
- a subject is required to read a technical paper hence, in this task, the subject is in a state of concentration with his or her eyes opened.
- task 404 the subject is required to relax and look around hence, in this task, the subject is not in a state of concentration and has his or her eyes opened.
- the subject is required to perform mental arithmetic for example, taking 400 minus 7 repeatedly, hence, in this task, the subject is in a state of concentration with his or her eyes closed.
- the subject is required to have his or her body and mind in a resting state with his or her eyes closed, hence in this task, the subject is not in a state of concentration with his or her eyes closed.
- the ideal level of attention for each of these tasks is plotted in Figure 4 as line 410 whereby the ideal level of attention is high when the subject is required to be in a state of concentration and is low when the subject is required to be not in a state of concentration.
- the subject is required to take part in a few sessions, each session involving an array of alternate tasks.
- projection matrices of CSPs for the CSP algorithm in the spectral-spatial feature extraction module are obtained.
- the learning of these projection matrices are carried out using the CSP method that jointly diagonalizes the two covariance matrices of the two classes i.e. the concentration class and the non-concentration class.
- the CSP method includes the following steps.
- step 2 the composite spatial covariance ⁇ c is computed according to Equation (11).
- Equation (11) the spatial covariance of one distribution ⁇ d is taken to be the average over the trials of each class and d ⁇ 1, 2 ⁇ is the class index.
- ⁇ c ⁇ ⁇ 1 + ⁇ ⁇ 2
- Equation (16) is equivalent to equation (2).
- the spatial filtered signal Z given in Equation (16) maximizes the difference in the variance of the two classes of EEG measurements.
- the variances of only a small number m of the spatial filtered signals are used as features for classification
- the signals Z p , p ⁇ 1 .. 2 m ⁇ that maximize the difference in the variance of the two classes of EEG are associated with the largest eigenvalues ⁇ and ( I- ⁇ ).
- these signals are used to form the feature vector X p given in Equation (17) whereby feature vectors X p are inputs to the classifier.
- Equation (17) is equivalent to Equation (3).
- the mutual information feature selection method is based on mutual information which indicates the mutual dependence of the features with respect to the classes. Further details of the mutual information feature selection process are as follows.
- Equation (18) H(X) denotes the entropy of the feature variable X and H(Y
- the entropy and the conditional entropy are given respectively in Equation (19) and Equation (20).
- the mutual information feature selection process includes the following steps.
- F k and the Na ⁇ ve Bayesian Parzen Window are used to predict the class label Y k .
- the mutual information of the predicted class label and the true label i.e. I(Y k ; Y) is then computed.
- step 3 the feature f k which maximizes I(Y k ; Y) is then selected.
- step 6 the candidate set is updated by F ⁇ F ⁇ f k ⁇ whereas the select feature set is updated by Fopt ⁇ Fopt ⁇ S ⁇ f k ⁇ .
- a feature refers to a CSP feature from a filter bank and can take on different values at different instances.
- the mutual information feature selection process in the example embodiments as described above is applied to the training set with labelled samples. After the feature selection process is completed, the select set of features includes the CSP features determined as "important" or characteristic for concentration detection based on their mutuality amongst the labeled samples. This set of features is used during the feature selection process when processing unlabelled data for concentration detection.
- models for the classifiers in the method 300 are obtained by the traditional Fisher linear discriminant method, using labelled training data samples.
- the labelled training data samples have positive labels if they are recorded from the concentration tasks and negative labels if they are recorded from the non-concentration tasks.
- the set of parameters obtained from the subject dependent training approach can be used to recreate a model for concentration detection using a computer program.
- a setup/configuration file is created whereby this file includes the projection vector and the bias of the classifiers, projection matrices of each CSP filter, the bands to be selected for the filter banks, and the weights to be used for combining the outputs from the temporal feature extraction module and the spectral-spatial feature extraction module.
- Figure 5 shows a schematic block diagram illustrating the connection between a method for concentration detection and a subject-dependent training approach according to an embodiment of the present invention.
- units 502, 504 and 506 correspond to the subject-dependent training approach
- units 508, 510 and 512 correspond to the spectral-spatial feature extraction module in the method 300 in Figure 3
- units 514 and 516 correspond to the temporal feature extraction module in the method 300 in Figure 3 .
- the feature extraction unit 508 implements steps 310 and 312 in Figure 3 whereas the feature extraction unit 514 implements the step 306.
- the feature selection unit 510 implements the step 314.
- the classification units, 516 and 512 implement steps 308 and 316 in Figure 3 respectively whereas the post-processing unit 520 implements steps 318 and 320 in Figure 3 .
- the method in the example embodiments can be implemented in the form of a software tool for example, as add-ons to EEG systems or as internet-based web services.
- the method can also be embedded into a PDA-like medical device. Even with only a low-cost EEG acquired at a low sampling rate and from a few EEG sensors on the forehead, the method in the example embodiments is still able to provide robust attention or concentration detection and scoring.
- the method in the example embodiments can be implemented in a simple and handy system with only forehead sensors.
- the example embodiments of the present invention can provide a continuous, quantitative, accurate and robust scoring mechanism for subject attention or concentration level since the example embodiments are based on features extracted and further selected using a multi-domain (spatial, spectral and temporal) analysis of the EEG and classified using machine learning.
- the example embodiments of the present invention provide a system to capture subject-specific EEG characteristics into a computational model and an automated parameter selection process that can find the best parameters and model.
- the example embodiments of the present invention provide a post-processing fusion scheme that improves performance by a multi-scale approach.
- Table 1 shows the results achieved by a method for concentration detection according to an embodiment of the present invention and by the prior art method in Monastra and Lubar [ Monastra and Lubar, 2000 - US06097980 - Quantitative electroencephalographic (QEEG) process and apparatus for assessing attention deficit hyperactivity disorder; V. J. Monastra, S. Lynn, M. Linden, J. F. Lubar, J. Gruzelier, and T. J. LaVaque, "Electroencephalographic Biofeedback in the Treatment of Attention-Deficit/Hyperactivity Disorder," Applied Psychophysiology and Biofeedback, vol. 30, no. 2, pp.
- QEEG Quantitative electroencephalographic
- Table 1 the row corresponding to "Theta/beta (prior-art)" shows the mean accuracy obtained by the method according to the prior art
- the row corresponding to "Waveform only” shows the mean accuracy obtained from the temporal feature extraction module alone
- the row corresponding to "Spectrum only” shows the mean accuracy obtained from the spectral-spatial feature extraction module alone
- the row corresponding to "Hybrid technique” shows the mean accuracy obtained from the method in the example embodiments.
- the results in Table 1 are in percentage, expressed in the form "mean ⁇ standard deviation” and are obtained via a 2x2 fold cross-validation method.
- Table 2 shows further results achieved by a method for concentration detection according to an embodiment of the present invention and by the prior art method in Monastra and Lubar.
- EER error rate
- Table 2 shows further results achieved by a method for concentration detection according to an embodiment of the present invention and by the prior art method in Monastra and Lubar.
- the row corresponding to "Theta/beta (prior-art)" shows the equal error rate (EER) obtained by the method according to the prior art
- the row corresponding to "Waveform only” shows the EER obtained from the temporal feature extraction module alone
- the row corresponding to "Spectrum only” shows the EER obtained from the spectral-spatial feature extraction module alone
- the row corresponding to "Hybrid technique” shows the EER obtained from the method in the example embodiments.
- the EER is the rate at which the false positive rate and the false negative rate are equal.
- Table 2 results in Table 2 are in percentage, expressed in the form "mean ⁇ standard deviation" and are obtained via a 2x2 fold cross-validation method.
- the relative error reduction rate is calculated according to Equation (21). It can be seen from Table 2 that the overall error rate reduction is 42.5% indicating that the method in the example embodiments performs significantly better than the prior art method.
- Table 2 also shows that even the performance of the temporal feature extraction module alone ("Waveform only”) or the spectral-spatial feature extraction module alone (“Spectral only”) in the example embodiments is better than the prior art method. This illustrates that the subject dependent training approach can significantly improve the performance of the methods.
- the adaptive control module 210 of the example embodiment receives the results of the attention detection module 206 together with the information from the presentation module 214, and adjusts automatically and adaptively the level of difficulty of the attention training games.
- the rules embedded in the adaptive control module 210 are programmable.
- Hierarchical Temporal Control is used.
- the speed of the car is proportional to ⁇ .
- the adaptive control module 210 of the example embodiment is able to update the threshold R based on a performance indicator obtained from a specific subject 202 when he plays the game in his first trial.
- the time for a subject 202 to perform a specific task e.g. controlling a racing car around a racing course on the screen
- T g the time for a subject 202 to perform a specific task
- D' the total distance of the course
- the system of the example embodiment adjusts the level of difficulty (to make it more difficult or easier) from a temporal perspective. It should also be appreciated that the system of the example embodiment is preferably based on an asynchronous BCI mode, i.e. the subject 202 does not need to follow any cue in order to control the game.
- the adaptive control module 210 works by first setting a default level of difficulty.
- the subject 202 then plays the game at the default level of difficulty over a first period of time.
- the relevant parameters such as Q t and T are estimated. These provide a direct indication of the subject's concentration level.
- the adaptive control module 210 then changes the level of difficulty based on the estimated parameters.
- the delay time in the example embodiment ranges from about 1 second to 19 seconds depending on the type of game and/or presentation.
- the time delay can also be automatically and adaptively adjusted by the system based on statistics of the subject's respective game profile.
- the parameters are continuously calculated and the difficulty level is adjusted accordingly.
- different thresholds R may be set for respective difficulty levels. For example, the level of difficulty is increased if the player's attention level is high (e.g. time taken to complete the racing course is below a lower limit). Similarly, the level of difficulty is decreased if the player's attention level is low (e.g. time taken to complete the racing course is above an upper limit).
- Various other schemes for adjusting the level of difficulty can be implemented depending on e.g. whether the game is intended to be more or less challenging.
- the session ends after an allocated period has passed, and the relevant session information such as highest score/level, duration at the highest level, distribution of score, etc. is stored in the system.
- the interface control module 212 of the example embodiment comprises a typical BCI setup which comprises a multi-channel EEG cap, amplifiers and application programming interface (API) code for allowing the client 102 to record and store the physiological bio-signals (EEG, EoG, EMG, etc.).
- EEG EEG
- EoG EMG
- EMG EMG
- the presentation module 214 comprises the game part, which is the implementation of the game on the client 102 ( Figure 1 ), and allows interaction with the subject.
- Various game examples have been developed for use with the system of the example embodiment. It will be appreciated that the games are run on the client while the client is being connected to the interface 212 to record the subject's physiological bio-signals.
- the interface and game parameters are fully configurable using e.g. a graphic user interface (GUI).
- GUI graphic user interface
- a Stimulus generator is also provided in the presentation module 214, allowing the creation of additional auditory and visual disturbances during the training phase. Such disturbances may advantageously influence the subject's performance and thus his/her EEG/EoG/EMG.
- Figure 7 shows a flow chart 700 illustrating a method 700 for generating a representation of a subject's attention level according to an example embodiment.
- brain signals from the subject are measured.
- temporal features are extracted from the brain signals.
- the extracted temporal features are classified using a classifier to give a score x 1 .
- spectral-spatial features are from the brain signals.
- spectral-spatial features containing discriminative information between concentration and non-concentration states are selected from the set of extracted spectral-spatial features.
- the selected spectral-spatial features are classified using a classifier to give a score x 2 .
- the scores x 1 and x 2 are combined to give a single score.
- said score is presented to the subject.
- the method and system of the example embodiment described above advantageously provide an attention feedforward training protocol for attention training.
- the attention training game has an adjustable level of difficulty under time constraints.
- the training system is preferably asynchronous, i.e. initiated by the user, not the computer.
- the degree/level of difficulty is adaptively and automatically adjusted by information derived from physiological bio-signals (e.g. EEG, EOG, EMG, etc.).
- the method and system of the example embodiment also provide rules/strategy for programmable and automated adaptation of level of difficulty (e.g. by the Graphical User Interface (GUI) and Game).
- GUI Graphical User Interface
- the parameters of the GUI/Game preferably allow the creation of a learning curve (e.g. level of difficulty over time related to training).
- the presentation of the stimuli in a multiple sequence with additional disturbance can be created during the training.
- These disturbances can be auditory or/and visual, and advantageously allowing active control of the level of difficulty.
- the training game can also be tailored to cater for various aspects of attention (e.g. spatial, verbal, object orientation) and to provide measures of adaptive learning.
- the method and system of the example embodiment can be implemented on a computer system 800, schematically shown in Figure 8 . It may be implemented as software, such as a computer program being executed within the computer system 800, and instructing the computer system 800 to conduct the method of the example embodiment.
- the computer system 800 comprises a computer module 802, input modules such as a keyboard 804 and mouse 806 and a plurality of output devices such as a display 808, and printer 810.
- the computer module 802 is connected to a computer network 812 via a suitable transceiver device 814, to enable access to e.g. the Internet or other network systems such as Local Area Network (LAN) or Wide Area Network (WAN).
- LAN Local Area Network
- WAN Wide Area Network
- the computer module 802 in the example includes a processor 818, a Random Access Memory (RAM) 820 and a Read Only Memory (ROM) 822.
- the computer module 802 also includes a number of Input/Output (I/O) interfaces, for example I/O interface 824 to the display 808, and I/O interface 826 to the keyboard 804.
- I/O Input/Output
- the components of the computer module 802 typically communicate via an interconnected bus 828 and in a manner known to the person skilled in the relevant art.
- the application program is typically supplied to the user of the computer system 800 encoded on a data storage medium such as a CD-ROM or flash memory carrier and read utilising a corresponding data storage medium drive of a data storage device 830.
- the application program is read and controlled in its execution by the processor 818.
- Intermediate storage of program data maybe accomplished using RAM 820.
- the present invention has been mainly described with respect to ADHD treatment, it can also find application for other mental disorders, such as depression, degraded working memory, Mild Cognitive Disorder (MCI), Alzheimer Disease (AD), etc.
- MCI Mild Cognitive Disorder
- AD Alzheimer Disease
- the implementation of the present invention has been described as a system, it can also take the form of a software program with CDROM/DVD on a computer, a web-service over the internet, or a cartridge for a console, etc.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Multimedia (AREA)
- Heart & Thoracic Surgery (AREA)
- Biophysics (AREA)
- Public Health (AREA)
- Molecular Biology (AREA)
- Surgery (AREA)
- Animal Behavior & Ethology (AREA)
- Pathology (AREA)
- Medical Informatics (AREA)
- Veterinary Medicine (AREA)
- Psychiatry (AREA)
- Artificial Intelligence (AREA)
- Theoretical Computer Science (AREA)
- Psychology (AREA)
- General Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Fuzzy Systems (AREA)
- Mathematical Physics (AREA)
- Neurology (AREA)
- Neurosurgery (AREA)
- Evolutionary Computation (AREA)
- Signal Processing (AREA)
- General Physics & Mathematics (AREA)
- Dermatology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physiology (AREA)
- Cardiology (AREA)
- Measurement And Recording Of Electrical Phenomena And Electrical Characteristics Of The Living Body (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
- Magnetic Resonance Imaging Apparatus (AREA)
Description
- The present invention relates broadly to a device and method for generating a representation of a subject's attention level, and to a computer storage medium having stored thereon computer code means for instructing a computing device to execute a method of generating a representation of a subject's attention level.
- Attention Deficit Hyperactivity Disorder (ADHD) is a common behavioural disorder in children, characterised by symptoms such as inattention and hyperactivity. Today's management of ADHD often leads to the use of pharmacological stimulant medication. Parents, however, may be concerned about potential unknown side effects of these medications which include headache, stomach pain, sleeplessness, poor appetite, physical growth retardation, etc.
- Recently, there has been a growing interest in treatment of ADHD based on psychosocial aspects. Although important, such treatment has, to date, been shown to be less superior compared to pharmacological treatments.
- At the same time, it has been noted that advanced technologies such as brain-computer interface (BCI) can be used to improve the treatment of ADHD based on psychological aspects. BCI provides a direct communication pathway between a human brain and an external device. It relies on bio-signals such as electroencephalogram (EEG) and thus is a low cost and non-invasive interface. Various studies have demonstrated the efficacy of neurofeedback (NFB) or EEG biofeedback in the treatment of ADHD.
-
US 2002/128540 A1 discloses an approach for the enhancement of attention and further the treatment of attention deficit disorder. The approach includes presentation of a virtual reality environment for enhancing attention shift, selective attention, and sustained attention by correlating a biofeedback device that estimates the current attention level of the patient. - HAMADICHAREF B ET AL: "Learning EEG-based spectral-spatial patterns for attention level measurement", CIRCUITS AND SYSTEMS, 2009, ISCAS 2009, IEEE INTERNATIONAL SYMPOSIUM ON, IEEE, PISCATAWAY, NJ, USA, 24 May 2009 (2009-05-24), pages 1465-1468, XP031479485, ISBN: 978-1-4244-3827-3, teach an approach an EEG-based attention level measurement system, wherein best features are selected using a mutual information procedure and classified using Fisher linear discriminant.
- KAI KENG ANG ET AL: "Filter Bank Common Spatial Pattern (FBCSP) in Brain-Computer Interface", NEURAL NETWORKS, 2008, IJCNN 2008, (IEEE WORLD CONGRESS ON COMPUTATIONAL INTELLIGENCE), IEEE INTERNATIONAL JOINT CONFERENCE ON, IEEE, PISCATAWAY, NJ, USA, 1 June 2008 (2008-06-01), pages 2390-2397, XP031395786, ISBN: 978-1-4244-1820-6, teach extracting discriminative patterns from an EEG using the Common Spatial Pattern algorithm. An autonomous selection of key temporal-spatial discriminative EEG characteristics is performed.
- United States Patent No.
6402520 describes regulating theta and beta wave activity (more specifically, to decrease theta wave activity and increase beta wave activity) as measured based on respective average millivolt activity. However, a clear and direct correlation between the measured average millivolt theta and beta activities and attention has not been established. Thus, the users involved in the training approaches in that document only leam to control their measured average millivolt theta and beta activities, which are not a direct measure of attention. As a result, while the users may develop a mechanism for controlling the measured average millivolt theta and beta activities, this does not directly correlate with achieving higher attention levels. - A need therefore exists to provide device and method for generating a representation of a subject's attention level that seek to address at least one of the above problems.
- This need is solved by a device according to
claim 1, a method according to claim 7 and a computer storage medium according to claim 13. - In accordance with a first aspect of the present invention, there is provided a device for generating a representation of a subject's attention level comprising:
- means for measuring brain signals from the subject;
- means for extracting temporal features from the brain signals;
- means for classifying the extracted temporal features using a classifier to give a score x1;
- means for extracting spectral-spatial features from the brain signals;
- means for selecting spectral-spatial features containing discriminative information between concentration and non-concentration states from the set of extracted spectral-spatial features;
- means for classifying the selected spectral-spatial features using a classifier to give a score x2;
- means for combining the scores x1 and x2 to give a single score; and
- means for presenting said score to the subject.
- The means for presenting may present said score in the form of a game.
- The means for presenting said score may adaptively adjust at least one control parameter of the game based on said score.
- The adjusting of said one parameter may comprise:
- presenting said game using a relationship between said score and said one control parameter over a first period of time;
- determining a representative value for the score of the subject over the first period of time;
- adjusting the relationship based on said representative value; and
- presenting said game using the modified relationship between said score and said one control parameter over a second period of time.
- The adjusting of the relationship may be such that a level of difficulty of the game in the second period of time is proportional to the representative value over the first period.
- The extracting of the temporal features from the brain signals may comprise:
- computing statistics of brain waveforms in each of a plurality of electrode channels; and
- concatenating the statistics into a joint feature vector.
- The statistics of the brain waveforms may be standard deviations.
- The extracting of the spectral-spatial features of the brain signals may comprise:
- extracting respective brain signal components in discrete frequency windows using filter banks to obtain spectral features of brain signals; and
- applying a CSP algorithm to each of the spectral features using a CSP array to obtain the spectral-spatial features of brain signals.
- The selecting of the spectral-spatial features containing discriminative information between concentration and non-concentration states from the set of extracted spectral-spatial features may comprise selecting spectral-spatial features based on the mutual dependence of the features with respect to the concentration and non-concentration states.
- The combining of the scores x1 and x2 to give a single score may comprise:
- normalizing the scores x1 and x2 according to an equation (x-mx)/sx, wherein mx and sx are the mean and standard deviation of outputs from the classifiers using training samples to give x1n and x2n respectively;
- assigning weights w1 and w2 to normalized scores x1n and x2n respectively; and
- combining the scores x1n and x2n according to an equation x1n*w1+x2n*w2 to give a single score.
- The weights w1 and w2 may be calculated according to the equation wi=(yi)p where yi is a classification accuracy in classifying the extracted temporal features if i = 1 and in classifying the extracted spectral-spatial features if i = 2 and p (p>0) controls the power of wi in the calculation of the single score.
- The classifier may comprise one or more of a group consisting of a Linear Discriminant Analysis classifier, Neural Networks, Support Vector Machines, Fuzzy Inference System, Tree-based classifiers,
Fuzzy Type 2 and Relevance Vector Machine. - The device may use training data to generate parameters for classifying the extracted temporal features using a classifier, for extracting spectral-spatial features from brain signals, for selecting spectral-spatial features containing discriminative information between the concentration and non-concentration states from the set of extracted spectral-spatial features and for classifying the selected spectral-spatial features using a classifier.
- The parameters may comprise one or more of a group consisting of projection matrices of CSPs for the CSP algorithm, parameters for selecting spectral-spatial features based on mutual information and a model for the classifiers.
- The use of training data to generate parameters may comprise:
- collecting training data from subjects performing a set of tasks; and
- determining said parameters via machine learning methods.
- In accordance with a second aspect of the present invention, there is provided a method for generating a representation of a subject's attention level, the method comprising the steps of:
- measuring brain signals from the subject;
- extracting temporal features from the brain signals;
- classifying the extracted temporal features using a classifier to give a score x1;
- extracting spectral-spatial features from the brain signals;
- selecting spectral-spatial features containing discriminative information between concentration and non-concentration states from the set of extracted spectral-spatial features;
- classifying the selected spectral-spatial features using a classifier to give a score x2;
- combining the scores x1 and x2 to give a single score; and
- presenting said score to the subject.
- The presenting of said score to the subject may comprise presenting said score in the form of a game.
- The presenting of said score to the subject may comprise adaptively adjusting at least one control parameter of the game based on said score.
- The adjusting of said one parameter may comprise:
- presenting said game using a relationship between said score and said one control parameter over a first period of time;
- determining a representative value for the score of the subject over the first period of time;
- adjusting the relationship based on said representative value; and
- presenting said game using the modified relationship between said score and said one control parameter over a second period of time.
- The adjusting of the relationship may be such that a level of difficulty of the game in the second period of time is proportional to the representative value over the first period.
- The extracting of the temporal features from the brain signals may comprise:
- computing statistics of brain waveforms in each of a plurality of electrode channels; and
- concatenating the statistics into a joint feature vector.
- The statistics of the brain waveforms may be standard deviations.
- The extracting of the spectral-spatial features of the brain signals may comprise:
- extracting respective brain signal components in discrete frequency windows using filter banks to obtain spectral features of brain signals; and
- applying a CSP algorithm to each of the spectral features using a CSP array to obtain the spectral-spatial features of brain signals.
- The selecting of the spectral-spatial features containing discriminative information between concentration and non-concentration states from the set of extracted spectral-spatial features may comprise selecting spectral-spatial features based on the mutual dependence of the features with respect to the concentration and non-concentration states.
- The combining of the scores x1 and x2 to give a single score may comprise:
- normalizing the scores x1 and x2 according to an equation (x-mx)/sx, wherein mx and sx are the mean and standard deviation of outputs from the classifiers using training samples to give x1n and x2n respectively;
- assigning weights w1 and w2 to normalized scores x1n and x2n respectively; and
- combining the scores x1n and x2n according to an equation x1n*w1+x2n*w2 to give a single score.
- The weights w1 and w2 may be calculated according to the equation wi=(yi)p where yi is a classification accuracy in classifying the extracted temporal features if i = 1 and in classifying the extracted spectral-spatial features if i = 2 and p (p>0) controls the power of wi in the calculation of the single score.
- The classifier may comprise one or more of a group consisting of a Linear Discriminant Analysis classifier, Neural Networks, Support Vector Machines, Fuzzy Inference System, Tree-based classifiers,
Fuzzy Type 2 and Relevance Vector Machine. - The method may further comprise using training data to generate parameters for classifying the extracted temporal features using a classifier, for extracting spectral-spatial features from brain signals, for selecting spectral-spatial features containing discriminative information between the concentration and non-concentration states from the set of extracted spectral-spatial features and for classifying the selected spectral-spatial features using a classifier.
- The parameters may comprise one or more of a group consisting of projection matrices of CSPs for the CSP algorithm, parameters for selecting spectral-spatial features based on mutual information and a model for the classifiers.
- The use of training data to generate parameters may comprise:
- collecting training data from subjects performing a set of tasks; and
- determining said parameters via machine learning methods.
- In accordance with a third aspect of the present invention, there is provided a computer storage medium having stored thereon computer code means for instructing a computing device to execute a method of generating a representation of a subject's attention level, the method comprising the steps of:
- measuring brain signals from the subject;
- extracting temporal features from the brain signals;
- classifying the extracted temporal features using a classifier to give a score x1;
- extracting spectral-spatial features from the brain signals;
- selecting spectral-spatial features containing discriminative information between concentration and non-concentration states from the set of extracted spectral-spatial features;
- classifying the selected spectral-spatial features using a classifier to give a score x2;
- combining the scores x1 and x2 to give a single score; and
- presenting said score to the subject.
- Embodiments of the invention will be better understood and readily apparent to one of ordinary skill in the art from the following written description, by way of example only, and in conjunction with the drawings, in which:
-
Figure 1 shows a schematic diagram illustrating an overview of device for generating a representation of a subject's attention level according to an example embodiment. -
Figure 2 is a detailed schematic diagram of an implementation of the device ofFigure 1 to one client according to an example embodiment. -
Figure 3 shows a flowchart illustrating a method for concentration detection according to an embodiment of the present invention. -
Figure 4 illustrates a data collection protocol for a subject-dependent model training approach according to an embodiment of the present invention. -
Figure 5 shows a schematic block diagram illustrating the connection between a method for concentration detection and a subject-dependent training approach according to an embodiment of the present invention. -
Figure 6 shows an image of an experimental set-up for implementing the method and system of the example embodiment. -
Figure 7 shows a flow chart illustrating a method for generating a representation of a subject's attention level according to an example embodiment. -
Figure 8 shows a schematic diagram illustrating a computer system for implementing the method and system of the example embodiment. - Example embodiments of the present invention provide a BCI-based system that seeks to improve ADHD treatment, e.g. improving attention of a user. The BCI technology of the example embodiments has been implemented in the form of computer games. Players can control, using their degree/level of attention and the BCI setup, various parameters of games, e.g. fish jump, racing car with speed, puzzle parts, etc.
-
Figure 1 shows a schematic diagram illustrating an overview ofdevice 100 for generating a representation of a subject's attention level according to an example embodiment. Thedevice 100 comprises a plurality ofclients server 106 via anetwork connection 104, e.g. an intranet or the Internet. Theserver 106 is also connected to aplayer manager 112 which preferably manages the interconnection and data flow for each client 102. - In the example embodiment, the client 102 initiates the start of a session, e.g. by requesting the
server 106 to provide a game. The client 102 then downloads the game from theserver 106. Agame program 110 provided by theserver 106 to the client 102 comprises, inter alia,algorithm 114 used for the EEG processing, analysis and classification,attention model 116 and thegame 118 embedded therewithin. Various types of game can be developed for use in the example embodiment. While playing the game the client 102 can save and send information (i.e. data 108) back to theserver 106 to update the respective player's profile stored on theplayer manager 112. In addition, thesystem 100 of the example embodiment is advantageously capable of multi-user operation. Furthermore, while playing the game, the user (i.e. player) can be shown his/her level of attention, e.g. as visual feedback. This can advantageously help the user perform better. - Some portions of the description which follows are explicitly or implicitly presented in terms of algorithms and functional or symbolic representations of operations on data within a computer memory. These algorithmic descriptions and functional or symbolic representations are the means used by those skilled in the data processing arts to convey most effectively the substance of their work to others skilled in the art. An algorithm is here, and generally, conceived to be a self-consistent sequence of steps leading to a desired result. The steps are those requiring physical manipulations of physical quantities, such as electrical, magnetic or optical signals capable of being stored, transferred, combined, compared, and otherwise manipulated.
- Unless specifically stated otherwise, and as apparent from the following, it will be appreciated that throughout the present specification, discussions utilizing terms such as "scanning", "calculating", "determining", "replacing", "generating", "initializing", "outputting", "concatenating", "extracting", "classifying", "adjusting" or the like, refer to the action and processes of a computer system, or similar electronic device, that manipulates and transforms data represented as physical quantities within the computer system into other data similarly represented as physical quantities within the computer system or other information storage, transmission or display devices.
- The present specification also discloses apparatus for performing the operations of the methods. Such apparatus may be specially constructed for the required purposes, or may comprise a general purpose computer or other device selectively activated or reconfigured by a computer program stored in the computer. The algorithms and displays presented herein are not inherently related to any particular computer or other apparatus. Various general purpose machines may be used with programs in accordance with the teachings herein. Alternatively, the construction of more specialized apparatus to perform the required method steps may be appropriate. The structure of a conventional general purpose computer will appear from the description below.
- In addition, the present specification also implicitly discloses a computer program, in that it would be apparent to the person skilled in the art that the individual steps of the method described herein may be put into effect by computer code. The computer program is not intended to be limited to any particular programming language and implementation thereof. It will be appreciated that a variety of programming languages and coding thereof may be used to implement the teachings of the disclosure contained herein. Moreover, the computer program is not intended to be limited to any particular control flow. There are many other variants of the computer program, which can use different control flows without departing from the spirit or scope of the invention.
- Furthermore, one or more of the steps of the computer program may be performed in parallel rather than sequentially. Such a computer program may be stored on any computer readable medium. The computer readable medium may include storage devices such as magnetic or optical disks, memory chips, or other storage devices suitable for interfacing with a general purpose computer. The computer readable medium may also include a hard-wired medium such as exemplified in the Internet system, or wireless medium such as exemplified in the GSM mobile telephone system. The computer program when loaded and executed on such a general-purpose computer effectively results in an apparatus that implements the steps of the preferred method.
-
Figure 2 is a detailed schematic diagram of an implementation of the device ofFigure 1 to one client according to an example embodiment. The system comprises a physiologicaldata acquisition module 204, anattention detection module 206, anadaptive control module 210, aninterface control module 212, and apresentation module 214 with stimulus generator (for providing feedback to a subject (i.e. player) 202). - In the example embodiment, the physiological
data acquisition module 204 obtains bio-signals from theplayer 202 and provides the results to theattention detection module 206. The attention detection results are in turn provided to theadaptive control module 210. Theinterface control module 210 provides a link between theadaptive control module 210 and thepresentation module 214. - In the example embodiment, the physiological
data acquisition module 204 is used to record physiological bio-signals from the subject 202. The bio-signals include the electroencephalogram (EEG), electrooculogram (EoG), electromyogram (EMG). Different sensor montages can be used for data acquisition in the example embodiment. For example, for stimulation-dependent applications (e.g. P300 based games), EEG signals are collected from a central-parietal region of the scalp. For voluntary attention control, EEG signals are collected from a pre-frontal area. - In the example embodiment, the EEG signals are first passed through a filter bank to be broken down to sub-bands, where the number of bands N is variable depending on the tasks.
-
- The filters in the example embodiment can be Infinite Impulse Response (IIR) or Finite Impulse Response (FIR) filters. In addition, by excluding the lower frequency filter (0-4Hz) in the example embodiment, artefacts are advantageously removed and detection accuracy improved in the later stage.
- The filtered signals are then sent to spatial filters corresponding to each frequency band. The spatial filters in the example embodiment are based on Common Spatial Pattern (CSP). The CSP filters are trained by joint maximization/minimization of the variances for two classes involved. The spatially filtered signal for ith band, Yi (t), is given as
- The rows of W are the stationary spatial filters and the columns of W -1 are the common spatial patterns. The spatially filtered signal given in Equation (2) maximizes the differences in the variance of the two classes of EEG measurements. However, the variances of only a small number m of the spatial filtered signal are used in the example embodiment as features for classification. The m first and last rows of Yi (t) form a feature vector
-
- In the method and system of the example embodiment, these feature vectors are used for attention detection. Based on specific tasks, different features are selected and finally a classification is applied.
- Generally, in the
attention detection module 206, advanced pattern recognition methods are used in the example embodiment to process the incoming EEG signal and classify it into attention/non-attention states with a quantifiable score to indicate the level of attention of the subject 202. - In addition, the attention detection of the example embodiment is capable of dealing with two types of scenarios. In one scenario, the subject 202 is presented with stimuli and the EEG is recorded so as to monitor the anticipatory aspect of attention, to classify signals which relate to the task presented. This is named reactive / dependent attention in the example embodiment. In another scenario, the subject 202 voluntarily directs his/her attention to the auditory and/or visual stimuli, and this is named self-paced attention. Thus, the attention detection of the example embodiment preferably provides a useful means to tackle specific aspects of attention such as spatial, verbal, and object orientation.
- In the example embodiment, to use attention detection in ADHD training session, a score
attention detection module 206, i.e.: -
-
Figure 3 shows a flowchart illustrating amethod 300 for concentration detection according to an embodiment of the present invention. Instep 302, multi-channel EEG acquisition is performed using a real-time data acquisition and processing platform. In one example, the data acquisition and processing platform implements the following steps. A NuAmps device from Neuroscan, Inc. is first used to measure the scalp brain signals. The brain signals are then recorded from Ag-AgCl electrodes placed on the surface of the user's head. The digitizer device for the recording of the brain signals works at a sampling rate of 250Hz. The recorded brain signals are then filtered via temporal filtering to remove high frequency noises and very slow waves using for example, a 5th-order digital Butterworth filter with a passband of [0.5Hz 40Hz]. The filtered brain signals are next downsampled by a factor of 4 in order to reduce the computational complexity. - In
step 304, windowing and pre-processing are performed. Step 304 selects electrode channels of interest and segments the incoming data stream into chunks using a running windowing mechanism. The window size and shift step are determined using training data. Step 304 also removes noise and artefacts through filtering. - In
step 306, temporal feature extraction is performed. Step 306 computes statistics such as the standard deviation of the windowed and pre-processed EEG waveforms in each channel. The statistics are then concatenated into a joint feature vector. The feature vector is then input to step 308. Instep 308, a classifier, such as the Linear Discriminant Analysis (LDA), is implemented to produce a score, for example x1, indicating the likelihood of the hypothesis whereby the hypothesis is that the subject is in a state of concentration i.e. with focused attention. Other classifiers that can be used include Neural Networks (NNs), Support Vector Machines (SVM), Fuzzy Inference System (FIS), Tree-based classifiers etc., and their variants such as theFuzzy Type 2 and the Relevance Vector Machine (RVM).Steps method 300. - In
step 310, an array of band pass filters i.e. filter banks is implemented on the windowed and pre-processed EEG. Each filter bank is centred at a particular frequency, sampled at a fixed interval and is used to extract the EEG component in each discrete frequency window. For example, the fixed interval may be 4Hz for the frequency range of the EEG from 4Hz to 36Hz. In one example, the filter bank is a digital filter with a low order and a linear phase. Such a filter bank can be a Finite Impulse Response (FIR) filter or an Infinite Impulse Response (IIR) filter. In a preferred embodiment, the filter bank is a low-order bandpass Chebyshev Type II filter with a pass-band width of 4Hz. MATLAB (MathWorks Inc.) tools can be used to design and implement the filter banks. At the output of the filter banks, an EEG component is obtained for each filter bank with each component further containing separate components from each of the selected electrode channels. - In
step 312, a common spatial pattern (CSP) array is implemented. Step 312 applies the CSP algorithm to each EEG component obtained instep 310 to emphasize the difference in spatial distributions of the energy between the two classes, the concentration and the non-concentration classes corresponding to the brain states during which the subject is concentrating and not concentrating respectively. The CSP algorithm is detailed in Equation (7) whereby for the jth EEG component, a CSP feature cf(j) is extracted according to Equation (7). In Equation (7), Wl is a matrix comprising of the first I1 and the last I2 rows of W, whereby I1 and I2 are normalized for data processing efficiency and the ratio between I1 and I2 is kept constant. Furthermore, Ej is a m x n data matrix of the jth EEG component whereby m is the number of selected electrode channels and n is the number of samples in the EEG component in one channel. The relationship between W and the covariance matrices of the EEG components is given by Equation (8) in which ∑(1) and ∑(2) are the covariance matrices of the EEG components corresponding to two different classes of brain signals (i.e. different brain states), I is the identity matrix and D is a diagonal matrix. - The spatial filtering parameters i.e. spatial patterns such as the matrix W are learnt from the examples of the two classes via a subject dependent model training approach which would be elaborated later. The CSP array produces an array of spectral-spatial features, each representing the energy of the EEG component projected onto a particular spatial pattern. Such an array of features is usually over-redundant since not every spectral-spatial feature is associated with the concentration or non-concentration state in the brain. Preferably, the unnecessary (i.e. redundant) features are removed.
- In
step 314, a mutual information feature selection is implemented to remove the unnecessary features. Step 314 selects a set of features that contains the discriminative information between the concentration and the non-concentration states. This set is determined through a model training procedure via a subject dependent model training approach which would be elaborated later. At the end ofstep 314, a feature vector is obtained and is input intostep 316. - In
step 316, a classifier such as the LDA is implemented. Using the feature vector input fromstep 314, a score, for example x2, is produced by the classifier. This score indicates the likelihood of the hypothesis whereby the hypothesis is that the subject is in a state of concentration i.e. with focused attention. Steps 310 - 316 form the spectral-spatial feature extraction module of themethod 300. - Step 318 implements the fusion of the results from the temporal feature extraction module and the spectral-spatial feature extraction module to obtain a single output. In
step 318, the continuous outputs of the classifiers in the temporal feature extraction module and the spectral-spatial feature extraction module are normalized. In one example, if an output is the score x, the normalized output xn will be (x-mx)/sx whereby mx and sx are respectively the mean and standard deviation of the outputs obtained using the training samples. Two normalized outputs x1n and x2n from the temporal feature module and the spectral-spatial module respectively are hence obtained. In one example, these two normalized outputs x1n and x2n are combined according to Equation (9) using weights w1 and w2 whereby weights w1 and w2 correspond to x1n and x2n respectively and reflect the individual performance of each of the modules. However, the normalized outputs x1n and x2n can also be combined using non-linear methods such as a non-linear weighted regression. Weights w1 and w2 are calculated according to the formula wi = (yi)p where yi is the classification accuracy of the module alone and is obtained via training samples, and p (p>0) controls the power of the accuracy's weight in the combination. In one example, p is set to 1. -
- Because of the large cross-subject variances in EEG patterns, a subject-dependent model training approach is used in the embodiments of the present invention to obtain the parameters and models for the
method 300. - In the subject-dependent model training approach in the example embodiments, training data collection sessions are implemented to collect a subject's EEGs during navigated sessions.
Figure 4 illustrates adata collection protocol 400 for the subject-dependent model training approach according to an embodiment of the present invention. The protocol consists of 4 different tasks to be performed by the subject. Intask 402, a subject is required to read a technical paper hence, in this task, the subject is in a state of concentration with his or her eyes opened. Intask 404, the subject is required to relax and look around hence, in this task, the subject is not in a state of concentration and has his or her eyes opened. Intask 406, the subject is required to perform mental arithmetic for example, taking 400 minus 7 repeatedly, hence, in this task, the subject is in a state of concentration with his or her eyes closed. Intask 408, the subject is required to have his or her body and mind in a resting state with his or her eyes closed, hence in this task, the subject is not in a state of concentration with his or her eyes closed. The ideal level of attention for each of these tasks is plotted inFigure 4 asline 410 whereby the ideal level of attention is high when the subject is required to be in a state of concentration and is low when the subject is required to be not in a state of concentration. In one example, the subject is required to take part in a few sessions, each session involving an array of alternate tasks. - Furthermore, in the subject-dependent training approach in the example embodiments, groups of parameters are determined via machine learning methods. An example of a machine learning method is the automation parameter optimization which is an iterative approach. Further details of the machine learning methods are given below. In one example, three groups of parameters are generated.
- Firstly, projection matrices of CSPs for the CSP algorithm in the spectral-spatial feature extraction module (See
Figure 3 ) are obtained. The learning of these projection matrices are carried out using the CSP method that jointly diagonalizes the two covariance matrices of the two classes i.e. the concentration class and the non-concentration class. - In one example, the CSP method includes the following steps.
- In
step 1, the normalized spatial covariance ∑ of the EEG measurements is computed according to Equation (10). In Equation (10), E is an NxT matrix representing the raw EEG measurement data of a single trial, N is the number of channels, T is the number of measurement samples per channel, ' denotes the transpose operator and trace(·) denotes the operation that sums the diagonal elements. -
-
-
-
-
- Equation (16) is equivalent to equation (2).
- The spatial filtered signal Z given in Equation (16) maximizes the difference in the variance of the two classes of EEG measurements. In general, the variances of only a small number m of the spatial filtered signals are used as features for classification The signals Zp , p∈{1..2m} that maximize the difference in the variance of the two classes of EEG are associated with the largest eigenvalues λ and (I-λ). In one example, these signals are used to form the feature vector Xp given in Equation (17) whereby feature vectors Xp are inputs to the classifier.
- Equation (17) is equivalent to Equation (3).
- Secondly, a set of parameters for mutual information feature selection in the spectral-spatial feature selection module is determined. The mutual information feature selection method is based on mutual information which indicates the mutual dependence of the features with respect to the classes. Further details of the mutual information feature selection process are as follows.
- Taking into consideration a vector variable X for example, CSP features as obtained in Equation (17) and its corresponding class label Y, the mutual information between the two random variables X and Y is given by Equation (18). In Equation (18), H(X) denotes the entropy of the feature variable X and H(Y|X) represents the conditional entropy of class label variable Y given feature variable X. The entropy and the conditional entropy are given respectively in Equation (19) and Equation (20).
- In one example, the mutual information feature selection process includes the following steps.
- In
step 1, a candidate set of d features is initialized as F ={f1, f2, ... , fd} and a select feature set is initialized as a null set Fopt = ∅. - In
step 2, for each feature fk in the candidate set, a tentative feature vector Fk = Fopt ∪ {fk} is formed. Next, Fk and the Naïve Bayesian Parzen Window are used to predict the class label Yk. The mutual information of the predicted class label and the true label i.e. I(Yk; Y) is then computed. - In
step 3, the feature fk which maximizes I(Yk; Y) is then selected. - In step 4, if F = ∅ and the gain in the mutual information is less than a preset threshold δ i.e. I(Yk;Y)- I0< δ, the process is terminated. Otherwise, in step 5, I0 = I(Yk;Y).
- In step 6, the candidate set is updated by F→F\{fk} whereas the select feature set is updated by Fopt → Fopt ∪ S{fk}.
- In step 7, if the candidate set is empty, the process is terminated. Otherwise, the process is repeated from
step 2. - In the example embodiments, a feature refers to a CSP feature from a filter bank and can take on different values at different instances. The mutual information feature selection process in the example embodiments as described above is applied to the training set with labelled samples. After the feature selection process is completed, the select set of features includes the CSP features determined as "important" or characteristic for concentration detection based on their mutuality amongst the labeled samples. This set of features is used during the feature selection process when processing unlabelled data for concentration detection.
- Thirdly, models for the classifiers in the
method 300 are obtained by the traditional Fisher linear discriminant method, using labelled training data samples. In one example, the labelled training data samples have positive labels if they are recorded from the concentration tasks and negative labels if they are recorded from the non-concentration tasks. - In the example embodiments, the set of parameters obtained from the subject dependent training approach can be used to recreate a model for concentration detection using a computer program. In one example, a setup/configuration file is created whereby this file includes the projection vector and the bias of the classifiers, projection matrices of each CSP filter, the bands to be selected for the filter banks, and the weights to be used for combining the outputs from the temporal feature extraction module and the spectral-spatial feature extraction module.
-
Figure 5 shows a schematic block diagram illustrating the connection between a method for concentration detection and a subject-dependent training approach according to an embodiment of the present invention. In one example,units units method 300 inFigure 3 andunits method 300 inFigure 3 . - In
Figure 5 , training EEGs are acquired from the subjects when they are performing the required tasks during the training data collection sessions implemented in the subject-dependent training approach in the example embodiments. Machine learning techniques are then implemented in using the training EEGs in the featureextraction training unit 502, featureselection training unit 504 and themodelling unit 506 inFigure 5 . This would obtain the required parameters and model for thefeature extraction unit 508,feature selection unit 510 and theclassification units - In
Figure 5 , in one example, thefeature extraction unit 508 implementssteps Figure 3 whereas thefeature extraction unit 514 implements thestep 306. In addition, thefeature selection unit 510 implements thestep 314. Furthermore, the classification units, 516 and 512, implementsteps Figure 3 respectively whereas thepost-processing unit 520 implementssteps Figure 3 . - The advantages conferred by the embodiments of the present invention can include:
- Firstly, the method for concentration detection in the example embodiments provides an accurate quantitative measure of the subject's attention or concentration level that is not provided by any of the prior arts. The method in the example embodiments is subject-specific and uses optimized parameters. On the other hand, the prior art methods are based on spectral features alone, with their output typically based on the average of a large set of results and a comparison performed within a narrow range to detect concentration. For example, the range can be extending from the mean minus the standard deviation to the mean plus the standard deviation of the results. Hence, the method in the example embodiments is more accurate. Furthermore, in the example embodiments of the present invention, an accurate score can be obtained continuously and this is important in (near) real-time situations when a fast and accurate score is necessary.
- Secondly, the hybrid model approach implemented in the example embodiments of the present invention takes all dimensions of the EEG into consideration. Specifically, these dimensions are the temporal, spatial and spectral information of the EEG which are then combined to give a single result. On the other hand, prior arts only concentrate on the spectral information of the EEG and hence provide a less detailed picture of the subject's EEG characteristics as compared to the embodiments of the present invention. In addition, in the example embodiments, the windowing approach allows the method of concentration detection to adjust the time resolution by changing the time segmentation window size to the best window size. This allows different window sizes to be chosen under different circumstances. For example, when a long term score is desired, the EEG recording session is preferably long whereas in a real-time situation, the EEG recording segment is preferably short.
- Thirdly, the method in the example embodiments of the present invention allows the creation of the best model for each subject. The method can also be used to create models based on a small cohort and thus, investigate group-specific issues for example, a group of ADHD boys. Furthermore, using a large database, the method can also be useful in investigating generalization issues for example population based medical studies.
- Fourthly, in the example embodiments, automatic selection and combination of features is achieved as the parameters and models for the method are automatically obtained from subject-specific modelling. This can improve the performance of the concentration detection method in the example embodiments. The mutual information feature selection in the example embodiments provides a novel way to create subject-specific modelling for example, for individualized healthcare. Furthermore, the use of the subject-specific model in the example embodiments achieves a higher accuracy and the machine learning methods used to create the subject-specific models allow the method in the example embodiments to be more flexible.
- Fifthly, in the example embodiments, the metric used in the overall performance evaluation is based on receiver operating characteristics (ROC) analysis. In the example embodiments, performance curves plotting the False Positive Rate (FPR) against the False Negative Rate are used to analyze the ROC. This metric (ROC) shows objectively the true performance of the method in the example embodiments using a simple curve. It will also allow one to determine the best model to be used for each subject and also to choose a model that will fit the sensitivity and specificity requirements along the ROC curve, while taking note of the trade-off between the sensitivity and specificity.
- Furthermore, the method in the example embodiments can be implemented in the form of a software tool for example, as add-ons to EEG systems or as internet-based web services. The method can also be embedded into a PDA-like medical device. Even with only a low-cost EEG acquired at a low sampling rate and from a few EEG sensors on the forehead, the method in the example embodiments is still able to provide robust attention or concentration detection and scoring. Thus, the method in the example embodiments can be implemented in a simple and handy system with only forehead sensors.
- Hence, the example embodiments of the present invention can provide a continuous, quantitative, accurate and robust scoring mechanism for subject attention or concentration level since the example embodiments are based on features extracted and further selected using a multi-domain (spatial, spectral and temporal) analysis of the EEG and classified using machine learning. In addition, the example embodiments of the present invention provide a system to capture subject-specific EEG characteristics into a computational model and an automated parameter selection process that can find the best parameters and model. Furthermore, the example embodiments of the present invention provide a post-processing fusion scheme that improves performance by a multi-scale approach.
- To further illustrate the advantages of the example embodiments of the present invention, an experimental study involving 5 participating subjects (all male and healthy) was carried out. The EEGs from these subjects are recorded from a standard 10/20 EEG system (NeuroScan NuAmps) with 15 channels and from frontal channels (Fp1/Fp2).
- Table 1 shows the results achieved by a method for concentration detection according to an embodiment of the present invention and by the prior art method in Monastra and Lubar [
Monastra and Lubar, 2000 - US06097980 - Quantitative electroencephalographic (QEEG) process and apparatus for assessing attention deficit hyperactivity disorder; V. J. Monastra, S. Lynn, M. Linden, J. F. Lubar, J. Gruzelier, and T. J. LaVaque, "Electroencephalographic Biofeedback in the Treatment of Attention-Deficit/Hyperactivity Disorder," Applied Psychophysiology and Biofeedback, vol. 30, no. 2, pp. 95-114, June 2005.] In Table 1, the row corresponding to "Theta/beta (prior-art)" shows the mean accuracy obtained by the method according to the prior art, the row corresponding to "Waveform only" shows the mean accuracy obtained from the temporal feature extraction module alone the row corresponding to "Spectrum only" shows the mean accuracy obtained from the spectral-spatial feature extraction module alone and the row corresponding to "Hybrid technique" shows the mean accuracy obtained from the method in the example embodiments. Furthermore, the results in Table 1 are in percentage, expressed in the form "mean ± standard deviation" and are obtained via a 2x2 fold cross-validation method. From Table 1, it can be seen that the mean accuracy of the method in the example embodiments is significantly better than that of the prior art method. More specifically, the overall performance improvement (absolute value) of the method in the example embodiments over the prior art method is 14.8%. Thus, these results demonstrate the ability of the method in the example embodiments to create an optimized subject-specific model that outperforms the prior art method.Table 1 Subject 1Subject 2Subject 3Subject 4 Subject 5 Average Theta/beta (prior-art) 57.5±2.7 57.5±3.5 66.7±10.9 56.9±9.7 57.5±2.2 59.2 Waveform only 60.2±3.8 78.8±5.3 69.8±4.7 76.3±5.3 72.8±6.2 71.6 Spectrum only 64.4±4.0 87.9±6.2 72.8±3.2 76.3±0.0 59.6±8.9 72.2 Hybrid technique 62.8±4.4 83.8±3.5 76.0±1.0 76.3±1.7 71.3±5.3 74.0 Improvement 5.3 26.3 9.3 19.4 13.8 14.8 - Table 2 shows further results achieved by a method for concentration detection according to an embodiment of the present invention and by the prior art method in Monastra and Lubar. In Table 2, for each subject, the row corresponding to "Theta/beta (prior-art)" shows the equal error rate (EER) obtained by the method according to the prior art, the row corresponding to "Waveform only" shows the EER obtained from the temporal feature extraction module alone, the row corresponding to "Spectrum only" shows the EER obtained from the spectral-spatial feature extraction module alone and the row corresponding to "Hybrid technique" shows the EER obtained from the method in the example embodiments. The EER is the rate at which the false positive rate and the false negative rate are equal. Furthermore, the results in Table 2 are in percentage, expressed in the form "mean ± standard deviation" and are obtained via a 2x2 fold cross-validation method. For each subject, the best performance by each of the methods is tabulated in Table 2. The relative error reduction rate is calculated according to Equation (21). It can be seen from Table 2 that the overall error rate reduction is 42.5% indicating that the method in the example embodiments performs significantly better than the prior art method. Furthermore, Table 2 also shows that even the performance of the temporal feature extraction module alone ("Waveform only") or the spectral-spatial feature extraction module alone ("Spectral only") in the example embodiments is better than the prior art method. This illustrates that the subject dependent training approach can significantly improve the performance of the methods.
Table 2 Subject 1Subject 2Subject 3Subject 4 Subject 5 Average Theta/beta (prior-art) 42.7 44.1 30.6 39.3 38.7 39.1 Waveform only 39.2 17.9 27.5 17.8 33.9 27.3 Spectrum only 37.9 8.2 21.9 25.1 30.6 24.7 Hybrid technique 35.0 7.3 21.9 20.8 27.7 22.5 Improvement (Relative Error Reduction Rate) 18 83.4 28.4 47.0 29.7 42.5 - It should be appreciated that, as the system of the example embodiment makes use of direct attention detection to control e.g. a game which then provides feedback cues to the subject 202, it is actually a feedforward instead of feedback system. In other words, the game only provides a visual representation of the concentration level. Other forms of representation can also be used, e.g. a simple bar indicator.
- This is a major distinction between the system of the example embodiment and prior art approaches which purely rely on feedback signals and in which the connection between attention and feedback training program is implicit. In the system of the example embodiment, the control of the game is advantageously through direct attention control. That is, the subject 202 only needs to concentrate and does not need to follow any feedback cues, and the result provided (i.e. representation of concentration level) is direct. Thus, the subject 202 knows explicitly that he is controlling if he so wishes.
- The
adaptive control module 210 of the example embodiment receives the results of theattention detection module 206 together with the information from thepresentation module 214, and adjusts automatically and adaptively the level of difficulty of the attention training games. In addition, the rules embedded in theadaptive control module 210 are programmable. -
-
-
- For example, in the case of racing car game, the speed of the car is proportional to Δ. The speed of the car is set as:
Q is an average value of Qt over time T. - With the quantifiable attention score and the speed control, the
adaptive control module 210 of the example embodiment is able to update the threshold R based on a performance indicator obtained from aspecific subject 202 when he plays the game in his first trial. In the example embodiment, the time for a subject 202 to perform a specific task (e.g. controlling a racing car around a racing course on the screen) is defined as Tg , and the total distance of the course is defined as D'. Theadaptive control module 210 is advantageously able to adjust the difficulty level based on the following equation: - Thus, the system of the example embodiment adjusts the level of difficulty (to make it more difficult or easier) from a temporal perspective. It should also be appreciated that the system of the example embodiment is preferably based on an asynchronous BCI mode, i.e. the subject 202 does not need to follow any cue in order to control the game.
- Based on the above, the
adaptive control module 210 works by first setting a default level of difficulty. The subject 202 then plays the game at the default level of difficulty over a first period of time. During a subsequent delay time (i.e. learning phase), the relevant parameters such as Qt and T are estimated. These provide a direct indication of the subject's concentration level. Theadaptive control module 210 then changes the level of difficulty based on the estimated parameters. - The delay time in the example embodiment ranges from about 1 second to 19 seconds depending on the type of game and/or presentation. The time delay can also be automatically and adaptively adjusted by the system based on statistics of the subject's respective game profile.
- Subsequently, during the training session, the parameters are continuously calculated and the difficulty level is adjusted accordingly. In the example embodiment, different thresholds R may be set for respective difficulty levels. For example, the level of difficulty is increased if the player's attention level is high (e.g. time taken to complete the racing course is below a lower limit). Similarly, the level of difficulty is decreased if the player's attention level is low (e.g. time taken to complete the racing course is above an upper limit). Various other schemes for adjusting the level of difficulty can be implemented depending on e.g. whether the game is intended to be more or less challenging. The session ends after an allocated period has passed, and the relevant session information such as highest score/level, duration at the highest level, distribution of score, etc. is stored in the system.
- The
interface control module 212 of the example embodiment comprises a typical BCI setup which comprises a multi-channel EEG cap, amplifiers and application programming interface (API) code for allowing the client 102 to record and store the physiological bio-signals (EEG, EoG, EMG, etc.). - The
presentation module 214 comprises the game part, which is the implementation of the game on the client 102 (Figure 1 ), and allows interaction with the subject. Various game examples have been developed for use with the system of the example embodiment. It will be appreciated that the games are run on the client while the client is being connected to theinterface 212 to record the subject's physiological bio-signals. In the example embodiment, the interface and game parameters are fully configurable using e.g. a graphic user interface (GUI). A Stimulus generator is also provided in thepresentation module 214, allowing the creation of additional auditory and visual disturbances during the training phase. Such disturbances may advantageously influence the subject's performance and thus his/her EEG/EoG/EMG. -
Figure 6 shows an image of an experimental set-up for implementing the method and system of the example embodiment. As can be seen inFigure 6 , the client 102 (Figure 1 ) takes the form of a standard personal computer having adisplay 602 in the example embodiment. Thegame 110 and other graphic user interface (GUI)elements 604, e.g. score, are provided on saiddisplay 602.Player 202 plays thegame 110 while wearing a cap 606 having various sensors for obtaining bio-signals generated from different areas of his/her head. -
Figure 7 shows aflow chart 700 illustrating amethod 700 for generating a representation of a subject's attention level according to an example embodiment. Atstep 702, brain signals from the subject are measured. Atstep 704, temporal features are extracted from the brain signals. Atstep 706, the extracted temporal features are classified using a classifier to give a score x1. Atstep 708, spectral-spatial features are from the brain signals. Atstep 710, spectral-spatial features containing discriminative information between concentration and non-concentration states are selected from the set of extracted spectral-spatial features. Atstep 712, the selected spectral-spatial features are classified using a classifier to give a score x2. Atstep 714, the scores x1 and x2 are combined to give a single score. Atstep 716, said score is presented to the subject. - The method and system of the example embodiment described above advantageously provide an attention feedforward training protocol for attention training. In addition, the attention training game has an adjustable level of difficulty under time constraints. Further, the training system is preferably asynchronous, i.e. initiated by the user, not the computer. Advantageously, the degree/level of difficulty is adaptively and automatically adjusted by information derived from physiological bio-signals (e.g. EEG, EOG, EMG, etc.).
- The method and system of the example embodiment also provide rules/strategy for programmable and automated adaptation of level of difficulty (e.g. by the Graphical User Interface (GUI) and Game). Furthermore, the parameters of the GUI/Game preferably allow the creation of a learning curve (e.g. level of difficulty over time related to training). Also, the presentation of the stimuli in a multiple sequence with additional disturbance can be created during the training. These disturbances can be auditory or/and visual, and advantageously allowing active control of the level of difficulty. The training game can also be tailored to cater for various aspects of attention (e.g. spatial, verbal, object orientation) and to provide measures of adaptive learning.
- The method and system of the example embodiment can be implemented on a
computer system 800, schematically shown inFigure 8 . It may be implemented as software, such as a computer program being executed within thecomputer system 800, and instructing thecomputer system 800 to conduct the method of the example embodiment. - The
computer system 800 comprises acomputer module 802, input modules such as akeyboard 804 andmouse 806 and a plurality of output devices such as adisplay 808, andprinter 810. - The
computer module 802 is connected to acomputer network 812 via asuitable transceiver device 814, to enable access to e.g. the Internet or other network systems such as Local Area Network (LAN) or Wide Area Network (WAN). - The
computer module 802 in the example includes aprocessor 818, a Random Access Memory (RAM) 820 and a Read Only Memory (ROM) 822. Thecomputer module 802 also includes a number of Input/Output (I/O) interfaces, for example I/O interface 824 to thedisplay 808, and I/O interface 826 to thekeyboard 804. - The components of the
computer module 802 typically communicate via aninterconnected bus 828 and in a manner known to the person skilled in the relevant art. - The application program is typically supplied to the user of the
computer system 800 encoded on a data storage medium such as a CD-ROM or flash memory carrier and read utilising a corresponding data storage medium drive of adata storage device 830. The application program is read and controlled in its execution by theprocessor 818. Intermediate storage of program data maybe accomplished usingRAM 820. - Although the present invention has been mainly described with respect to ADHD treatment, it can also find application for other mental disorders, such as depression, degraded working memory, Mild Cognitive Disorder (MCI), Alzheimer Disease (AD), etc. In addition, while the implementation of the present invention has been described as a system, it can also take the form of a software program with CDROM/DVD on a computer, a web-service over the internet, or a cartridge for a console, etc.
- It will be appreciated by a person skilled in the art that numerous variations and/or modifications may be made to the present invention as shown in the specific embodiments without departing from the scope of the invention as defined in the claims.
- The present embodiments are, therefore, to be considered in all respects to be illustrative and not restrictive.
Claims (13)
- A device for generating a representation of a subject's attention level comprising:means for measuring brain bio-electrical signals from the subject;means for extracting temporal features from the brain bio-electrical signals;a first classifier operable to determine a first score indicative of the subject's attention level based on the extracted temporal features;means for extracting spectral-spatial features from the brain bio-electrical signals;means for selecting spectral-spatial features containing discriminative information between concentration and non-concentration states from the set of extracted spectral-spatial features;a second classifier operable to determine a second score indicative of the subject's attention level based on the selected spectral-spatial features;means for generating an attention score indicative of the subject's attention level based on a combination of the first and second scores; andmeans for presenting said attention score to the subject, characterized in that:the extracting of the temporal features from the brain bio-electrical signals comprises:computing statistics of brain waveforms in each of a plurality of electrode channels; andconcatenating the statistics into a joint feature vector for input to the first classifier to determine the first score,the extracting of the spectral-spatial features of the brain bio-electrical signals comprises:extracting respective brain bio-electrical signal components in discrete frequency windows using filter banks to obtain spectral features of brain bio-electrical signals; andapplying a CSP algorithm to each of the spectral features using a CSP array to obtain the spectral-spatial features of brain bio-electrical signals, andwherein the device is configured to use subject-dependent training data collected when the subject performs a plurality of predetermined tasks over a trial period, each task corresponding to a state of the subject's attention level, to generate parameters for determining the first score based on the extracted temporal features using the first classifier, for extracting the spectral-spatial features from brain bio-electrical signals, for selecting spectral-spatial features containing discriminative information between the concentration and non-concentration states from the set of extracted spectral-spatial features, and for determining the second score based on the selected spectral-spatial features using the second classifier.
- The device as claimed in claim 1, wherein means for presenting said attention score in the form of a game.
- The device as claimed in claim 2, wherein the means for presenting said attention score adaptively adjusts at least one control parameter of the game based on said attention score.
- The device as claimed in claim 3, wherein the adjusting of said one parameter comprises:presenting said game using a relationship between said attention score and said one control parameter over a first period of time;determining a representative value for the attention score of the subject over the first period of time;adjusting the relationship based on said representative value; andpresenting said game using the modified relationship between said attention score and said one control parameter over a second period of time.
- The device as claimed in claim 4, wherein the adjusting of the relationship is such that a level of difficulty of the game in the second period of time is proportional to the representative value over the first period.
- The device as claimed in any one of the preceding claims, wherein the combining of the first and second scores to give the attention score comprises:normalizing the first and second scores according to an equation (x-mx)/sx, wherein x is one of the first score or the second score and mx and sx are the mean and standard deviation of outputs from the classifiers using training samples, to give normalized scores x1n and x2n respectively;assigning weights w1 and w2 to the normalized scores x1n and x2n respectively; andcombining the normalized scores x1n and x2n according to an equation x1n*w1+x2n*w2 to give the attention score.
- A computer-implemented method for generating a representation of a subject's attention level, the method comprising the steps of:measuring brain bio-electrical signals from the subject;extracting temporal features from the brain bio-electrical signals;determining a first score indicative of the subject's attention level based on the extracted temporal features using a first classifier;extracting spectral-spatial features from the brain bio-electrical signals;selecting spectral-spatial features containing discriminative information between the concentration and non-concentration states from the set of extracted spectral-spatial features;determining a second score indicative of the subject's attention level based on the selected spectral-spatial features using a second classifier;generating an attention score indicative of the subject's attention level based on a combination of the first and second scores; andpresenting said attention score to the subject, characterized in that:the extracting of the temporal features from the brain bio-electrical signals comprises:computing statistics of brain waveforms in each of a plurality of electrode channels; andconcatenating the statistics into a joint feature vector for input to the first classifier to determine the first score,the extracting of the spectral-spatial features of the brain bio-electrical signals comprises:extracting respective brain bio-electrical signal components in discrete frequency windows using filter banks to obtain spectral features of brain bio-electrical signals; andapplying a CSP algorithm to each of the spectral features using a CSP array to obtain the spectral-spatial features of brain bio-electrical signals, andthe method further comprises using subject-dependent training data collected when the subject performs a plurality of predetermined tasks over a trial period, each task corresponding to a state of the subject's attention level, to generate parameters for determining the first score based on the extracted temporal features using the first classifier, for extracting the spectral-spatial features from brain bio-electrical signals, for selecting spectral-spatial features containing discriminative information between the concentration and non-concentration states from the set of extracted spectral-spatial features, and for determining the second score based on the selected spectral-spatial features using the second classifier.
- The method as claimed in claim 7, wherein presenting said attention score to the subject comprises presenting said attention score in the form of a game.
- The method as claimed in claim 8, wherein presenting said attention score to the subject comprises adaptively adjusting at least one control parameter of the game based on said attention score.
- The method as claimed in claim 9, wherein the adjusting of said one parameter comprises:presenting said game using a relationship between said attention score and said one control parameter over a first period of time;determining a representative value for the attention score of the subject over the first period of time;adjusting the relationship based on said representative value; andpresenting said game using the modified relationship between said attention score and said one control parameter over a second period of time.
- The method as claimed in claim 10, wherein the adjusting of the relationship is such that a level of difficulty of the game in the second period of time is proportional to the representative value over the first period.
- The method as claimed in any one of claims 7 to 11, wherein the combining of the first and second scores to give the attention score comprises:normalizing the first and second scores according to an equation (x-mx)/sx, wherein x is one of the first score or the second score and mx and sx are the mean and standard deviation of outputs from the classifiers using training samples, to give normalized scores x1n and x2n respectively;assigning weights w1 and w2 to the normalized scores x1n and x2n respectively; andcombining the normalized scores x1n and x2n according to an equation x1n*w1+x2n*w2 to give the attention score.
- A computer storage medium having stored thereon computer code means for instructing a computing device to execute a method of generating a representation of a subject's attention level, the method comprising the steps of:measuring brain bio-electrical signals from the subject;extracting temporal features from the brain bio-electrical signals;determining a first score indicative of the subject's attention level based on the extracted temporal features using a first classifier;extracting spectral-spatial features from the brain bio-electrical signals;selecting spectral-spatial features containing discriminative information between the concentration and non-concentration states from the set of extracted spectral-spatial features;determining a second score indicative of the subject's attention level based on the selected spectral-spatial features using a second classifier;generating an attention score indicative of the subject's attention level based on a combination of the first and second scores; andpresenting said attention score to the subject, characterized in thatthe extracting of the temporal features from the brain bio-electrical signals comprises:computing statistics of brain waveforms in each of a plurality of electrode channels; andconcatenating the statistics into a joint feature vector for input to the first classifier to determine the first score,the extracting of the spectral-spatial features of the brain bio-electrical signals comprises:extracting respective brain bio-electrical signal components in discrete frequency windows using filter banks to obtain spectral features of brain bio-electrical signals; andapplying a CSP algorithm to each of the spectral features using a CSP array to obtain the spectral-spatial features of brain bio-electrical signals, andthe method further comprises using subject-dependent training data collected when the subject performs a plurality of predetermined tasks over a trial period, each task corresponding to a state of the subject's attention level, to generate parameters for determining the first score based on the extracted temporal features using the first classifier, for extracting the spectral-spatial features from brain bio-electrical signals, for selecting spectral-spatial features containing discriminative information between the concentration and non-concentration states from the set of extracted spectral-spatial features, and for determining the second score based on the selected spectral-spatial features using the second classifier.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13929108P | 2008-12-19 | 2008-12-19 | |
PCT/SG2009/000337 WO2010071604A1 (en) | 2008-12-19 | 2009-09-14 | Device and method for generating a representation of a subject's attention level |
Publications (3)
Publication Number | Publication Date |
---|---|
EP2385789A1 EP2385789A1 (en) | 2011-11-16 |
EP2385789A4 EP2385789A4 (en) | 2012-08-01 |
EP2385789B1 true EP2385789B1 (en) | 2019-03-13 |
Family
ID=42269057
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP09833759.5A Active EP2385789B1 (en) | 2008-12-19 | 2009-09-14 | Device and method for generating a representation of a subject's attention level |
Country Status (7)
Country | Link |
---|---|
US (1) | US9149719B2 (en) |
EP (1) | EP2385789B1 (en) |
CN (1) | CN102573619B (en) |
AU (1) | AU2009327552A1 (en) |
CA (1) | CA2747631A1 (en) |
SG (1) | SG172262A1 (en) |
WO (1) | WO2010071604A1 (en) |
Families Citing this family (56)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20110072730A (en) * | 2009-12-23 | 2011-06-29 | 한국과학기술원 | Adaptive brain-computer interface device |
US20120203725A1 (en) * | 2011-01-19 | 2012-08-09 | California Institute Of Technology | Aggregation of bio-signals from multiple individuals to achieve a collective outcome |
US9443205B2 (en) * | 2011-10-24 | 2016-09-13 | President And Fellows Of Harvard College | Enhancing diagnosis of disorder through artificial intelligence and mobile health technologies without compromising accuracy |
GB2514726B (en) * | 2012-03-27 | 2017-05-03 | Myndplay Ltd | Improvements relating to control mechanisms |
WO2014031083A1 (en) | 2012-08-24 | 2014-02-27 | Agency For Science, Technology And Research | Autodidactic cognitive training device and method thereof |
EP3441896B1 (en) * | 2012-09-14 | 2021-04-21 | InteraXon Inc. | Systems and methods for collecting, analyzing, and sharing bio-signal and non-bio-signal data |
US10606353B2 (en) | 2012-09-14 | 2020-03-31 | Interaxon Inc. | Systems and methods for collecting, analyzing, and sharing bio-signal and non-bio-signal data |
IN2013MU03025A (en) * | 2013-09-19 | 2015-07-03 | Tata Consultancy Services Ltd | |
US20150230020A1 (en) * | 2014-02-12 | 2015-08-13 | Kee Sook Jeon | Headset |
CN103971124B (en) * | 2014-05-04 | 2017-02-15 | 杭州电子科技大学 | Multi-class motor imagery brain electrical signal classification method based on phase synchronization |
CA2949431C (en) | 2014-05-21 | 2023-09-26 | Akili Interactive Labs, Inc. | Processor-implemented systems and methods for enhancing cognitive abilities by personalizing cognitive training regimens |
CA2979390A1 (en) * | 2015-03-12 | 2016-09-15 | Akili Interactive Labs, Inc. | Processor implemented systems and methods for measuring cognitive abilities |
IL239191A0 (en) * | 2015-06-03 | 2015-11-30 | Amir B Geva | Image classification system |
EP3302259B1 (en) | 2015-06-08 | 2022-02-16 | Atentiv LLC | Functional eeg montage for cognitive skill development |
JP6018266B1 (en) * | 2015-07-17 | 2016-11-02 | 株式会社スクウェア・エニックス | Video game processing program and video game processing system |
JP6450350B2 (en) * | 2015-07-20 | 2019-01-09 | ネオウィズ コーポレーション | GAME CONTROL METHOD, GAME CONTROL DEVICE, AND ITS RECORDING MEDIUM |
US20190008436A1 (en) * | 2015-07-31 | 2019-01-10 | Atentiv Llc | Method and system for monitoring and improving attention |
US20180286272A1 (en) | 2015-08-28 | 2018-10-04 | Atentiv Llc | System and program for cognitive skill training |
CN106128201B (en) * | 2016-06-14 | 2018-12-21 | 北京航空航天大学 | A kind of attention training system of immersion vision and the combination of discrete force control task |
CN106108893B (en) * | 2016-06-20 | 2019-03-01 | 杭州电子科技大学 | Mental imagery training Design of man-machine Conversation method based on eye electricity, brain electricity |
US11839472B2 (en) | 2016-07-19 | 2023-12-12 | Akili Interactive Labs, Inc. | Platforms to implement signal detection metrics in adaptive response-deadline procedures |
AU2017314831C1 (en) | 2016-08-26 | 2023-01-05 | Akili Interactive Labs, Inc. | Cognitive platform coupled with a physiological component |
US10490092B2 (en) | 2017-03-17 | 2019-11-26 | Age Of Learning, Inc. | Personalized mastery learning platforms, systems, media, and methods |
US11402909B2 (en) * | 2017-04-26 | 2022-08-02 | Cognixion | Brain computer interface for augmented reality |
FR3070852B1 (en) | 2017-09-08 | 2019-09-20 | Centre National De La Recherche Scientifique | DECODING THE VISUAL ATTENTION OF AN INDIVIDUAL FROM ELECTROENCEPHALOGRAPHIC SIGNALS |
US11723579B2 (en) | 2017-09-19 | 2023-08-15 | Neuroenhancement Lab, LLC | Method and apparatus for neuroenhancement |
US11717686B2 (en) | 2017-12-04 | 2023-08-08 | Neuroenhancement Lab, LLC | Method and apparatus for neuroenhancement to facilitate learning and performance |
CN108108763B (en) * | 2017-12-25 | 2021-07-23 | 中国科学院深圳先进技术研究院 | Electroencephalogram classification model generation method and device and electronic equipment |
CN107961006B (en) * | 2017-12-28 | 2023-12-05 | 出门问问信息科技有限公司 | Wrist type intelligent wearable device and information prompting method |
US11478603B2 (en) | 2017-12-31 | 2022-10-25 | Neuroenhancement Lab, LLC | Method and apparatus for neuroenhancement to enhance emotional response |
CN108346192B (en) * | 2018-02-13 | 2020-12-15 | 福州外语外贸学院 | Attendance checking method and terminal for computer room course student concentration degree |
US11364361B2 (en) | 2018-04-20 | 2022-06-21 | Neuroenhancement Lab, LLC | System and method for inducing sleep by transplanting mental states |
US10909157B2 (en) * | 2018-05-22 | 2021-02-02 | Salesforce.Com, Inc. | Abstraction of text summarization |
WO2019226947A1 (en) * | 2018-05-25 | 2019-11-28 | The Regents Of The University Of California | Systems and methods for enhanced wearable attention monitoring |
EP3594861B1 (en) | 2018-07-09 | 2024-04-03 | Tata Consultancy Services Limited | Systems and methods for classification of multi-dimensional time series of parameters |
US11452839B2 (en) | 2018-09-14 | 2022-09-27 | Neuroenhancement Lab, LLC | System and method of improving sleep |
US11380211B2 (en) | 2018-09-18 | 2022-07-05 | Age Of Learning, Inc. | Personalized mastery learning platforms, systems, media, and methods |
CN109864749A (en) * | 2019-01-30 | 2019-06-11 | 北京津发科技股份有限公司 | Pay attention to training method, assessment method and its device of control ability |
US11786694B2 (en) | 2019-05-24 | 2023-10-17 | NeuroLight, Inc. | Device, method, and app for facilitating sleep |
CN110400619B (en) * | 2019-08-30 | 2023-07-21 | 上海大学 | Hand function rehabilitation training method based on surface electromyographic signals |
CN111563520B (en) * | 2020-01-16 | 2023-01-13 | 北京航空航天大学 | Hyperspectral image classification method based on space-spectrum combined attention mechanism |
US11645604B2 (en) * | 2020-01-31 | 2023-05-09 | International Business Machines Corporation | Methods and systems for managing concentration in work environments |
US20210386351A1 (en) * | 2020-06-12 | 2021-12-16 | Korea University Research And Business Foundation | Brain-computer interface apparatus for minimizing signal correction process between users using clustering technology based on brain activity and method of operating the same |
CN112244772A (en) * | 2020-10-15 | 2021-01-22 | 王映姗 | Sleep stage prediction method based on deep learning, storage medium and terminal equipment |
CN112494774A (en) * | 2020-10-16 | 2021-03-16 | 杭州师范大学 | Continuous attention regulation and control system based on delayed electroencephalogram feedback |
CN112515685B (en) * | 2020-11-10 | 2023-03-24 | 上海大学 | Multi-channel electroencephalogram signal channel selection method based on time-frequency co-fusion |
CN112932509B (en) * | 2021-01-28 | 2022-10-28 | 福建脉行人生医疗科技有限公司 | Method and device for picking up and optimizing analysis of ocular signals |
CN113143273B (en) * | 2021-03-23 | 2024-06-28 | 陕西师范大学 | Intelligent detection system and method for attention state of learner in online video learning |
CN113143291B (en) * | 2021-05-11 | 2023-02-03 | 燕山大学 | Electroencephalogram feature extraction method under rapid sequence visual presentation |
CN113157101B (en) * | 2021-06-07 | 2022-08-19 | 成都华脑科技有限公司 | Fragmentation reading habit identification method and device, readable medium and electronic equipment |
CN113577495A (en) * | 2021-07-21 | 2021-11-02 | 大连民族大学 | Children attention deficit hyperactivity disorder auxiliary treatment system based on BCI-VR |
WO2023069968A1 (en) * | 2021-10-18 | 2023-04-27 | The Regents Of The University Of California | Multi channel data aggregation based data compression |
CN114209283B (en) * | 2021-12-27 | 2024-07-26 | 燕山大学 | Intraoperative pain detection method based on brain-computer space complexity |
CN114209323B (en) * | 2022-01-21 | 2024-05-10 | 中国科学院计算技术研究所 | Method for identifying emotion and emotion identification model based on electroencephalogram data |
CN114391854A (en) * | 2022-02-16 | 2022-04-26 | 北京脑陆科技有限公司 | Electroencephalogram signal processing device, method and medium based on lateralization index |
CN116807479B (en) * | 2023-08-28 | 2023-11-10 | 成都信息工程大学 | Driving attention detection method based on multi-mode deep neural network |
Family Cites Families (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5724987A (en) | 1991-09-26 | 1998-03-10 | Sam Technology, Inc. | Neurocognitive adaptive computer-aided training method and system |
US5377100A (en) | 1993-03-08 | 1994-12-27 | The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration | Method of encouraging attention by correlating video game difficulty with attention level |
US6402520B1 (en) * | 1997-04-30 | 2002-06-11 | Unique Logic And Technology, Inc. | Electroencephalograph based biofeedback system for improving learning skills |
KR20020068825A (en) | 2001-02-23 | 2002-08-28 | 학교법인 한양학원 | System and method of enhancing attention by correlating virtual reality with biofeedback |
US7403814B2 (en) * | 2001-05-04 | 2008-07-22 | University Of Virginia Patent Foundation | Method, apparatus, and computer program product for assessment of attentional impairments |
AU2005223637A1 (en) * | 2004-03-18 | 2005-09-29 | University Of Virginia Patent Foundation | Method, apparatus, and computer program product for stochastic psycho-physiological assessment of attentional impairments |
US7826894B2 (en) * | 2004-03-22 | 2010-11-02 | California Institute Of Technology | Cognitive control signals for neural prosthetics |
US7520848B2 (en) * | 2004-04-09 | 2009-04-21 | The Board Of Trustees Of The Leland Stanford Junior University | Robotic apparatus for targeting and producing deep, focused transcranial magnetic stimulation |
US20180146879A9 (en) * | 2004-08-30 | 2018-05-31 | Kalford C. Fadem | Biopotential Waveform Data Fusion Analysis and Classification Method |
WO2007007321A2 (en) * | 2005-07-07 | 2007-01-18 | Bar-Ilan University | Method and system for processing an electroencephalograph (eeg) signal |
US20100217145A1 (en) * | 2006-06-09 | 2010-08-26 | Bracco Spa | Method of processing multichannel and multivariate signals and method of classifying sources of multichannel and multivariate signals operating according to such processing method |
US20090306531A1 (en) * | 2008-06-05 | 2009-12-10 | Eric Claude Leuthardt | Methods and systems for controlling body parts and devices using ipsilateral motor cortex and motor related cortex |
US8244475B2 (en) * | 2007-12-27 | 2012-08-14 | Teledyne Scientific & Imaging, Llc | Coupling human neural response with computer pattern analysis for single-event detection of significant brain responses for task-relevant stimuli |
EP2231007A1 (en) * | 2008-01-11 | 2010-09-29 | Oregon Health and Science University | Rapid serial presentation communication systems and methods |
CN102149320A (en) | 2008-04-28 | 2011-08-10 | 科技研究局(A*Star) | Method and system for concentration detection |
US20100249636A1 (en) * | 2009-03-27 | 2010-09-30 | Neurofocus, Inc. | Personalized stimulus placement in video games |
US8442626B2 (en) * | 2010-06-21 | 2013-05-14 | Aleksandrs Zavoronkovs | Systems and methods for communicating with a computer using brain activity patterns |
-
2009
- 2009-09-14 EP EP09833759.5A patent/EP2385789B1/en active Active
- 2009-09-14 CN CN200980157260.7A patent/CN102573619B/en active Active
- 2009-09-14 SG SG2011044880A patent/SG172262A1/en unknown
- 2009-09-14 US US13/140,599 patent/US9149719B2/en active Active
- 2009-09-14 CA CA2747631A patent/CA2747631A1/en not_active Abandoned
- 2009-09-14 AU AU2009327552A patent/AU2009327552A1/en not_active Abandoned
- 2009-09-14 WO PCT/SG2009/000337 patent/WO2010071604A1/en active Application Filing
Non-Patent Citations (1)
Title |
---|
None * |
Also Published As
Publication number | Publication date |
---|---|
AU2009327552A1 (en) | 2011-08-11 |
US9149719B2 (en) | 2015-10-06 |
EP2385789A1 (en) | 2011-11-16 |
CN102573619A (en) | 2012-07-11 |
SG172262A1 (en) | 2011-07-28 |
EP2385789A4 (en) | 2012-08-01 |
US20120108997A1 (en) | 2012-05-03 |
WO2010071604A1 (en) | 2010-06-24 |
CN102573619B (en) | 2015-04-22 |
CA2747631A1 (en) | 2010-06-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2385789B1 (en) | Device and method for generating a representation of a subject's attention level | |
US8862581B2 (en) | Method and system for concentration detection | |
Sawangjai et al. | EEGANet: Removal of ocular artifacts from the EEG signal using generative adversarial networks | |
US11890246B2 (en) | Motion teaching system and motion teaching method | |
Nezam et al. | A novel classification strategy to distinguish five levels of pain using the EEG signal features | |
WO2012102675A1 (en) | A method and system for detecting attention | |
Piciucco et al. | Biometric recognition using wearable devices in real-life settings | |
Lee et al. | A comparative analysis of multi-class EEG classification for brain computer interface | |
Szachewicz | Classification of motor imagery for brain-computer interfaces | |
Ogino et al. | An EEG-based robot arm control to express human emotions | |
Sannelli et al. | Estimating noise and dimensionality in BCI data sets: Towards illiteracy comprehension | |
Sorinas et al. | Unraveling the development of an algorithm for recognizing primary emotions through electroencephalography | |
Iáñez et al. | Mental tasks selection method for a SVM-based BCI system | |
Wessel | Pioneering research into brain computer interfaces | |
Khalaf et al. | Hybrid EEG–fTCD Brain–computer interfaces | |
Karam et al. | Neural network for recognition of brain wave signals | |
Chowdhury | EEG based assessment of emotional wellbeing in smart environment | |
Henshaw | Improved brain-computer interface methods with application to gaming | |
Alzoubi | Automatic affect detection from physiological signals: Practical issues | |
Ramar et al. | Classification of EEG Signals on SEED Dataset Using Improved CNN | |
Valderrama et al. | Combining spectral and fractal features for emotion recognition on electroencephalographic signals | |
Kim et al. | Passive EEG-based Emotional state Cognition Using Subject-independent Feature Extraction | |
van Suijdam BSc et al. | EEG Data Classification | |
Hossain | Signal Processing Combined with Machine Learning for Biomedical Applications | |
Chatzilari et al. | Multimodal BCIs–the hands-free Tetris paradigm |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
17P | Request for examination filed |
Effective date: 20110719 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO SE SI SK SM TR |
|
DAX | Request for extension of the european patent (deleted) | ||
A4 | Supplementary search report drawn up and despatched |
Effective date: 20120628 |
|
RIC1 | Information provided on ipc code assigned before grant |
Ipc: G06F 3/05 20060101ALI20120622BHEP Ipc: A61B 5/0482 20060101ALI20120622BHEP Ipc: A63F 13/06 20060101AFI20120622BHEP |
|
17Q | First examination report despatched |
Effective date: 20140226 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R079 Ref document number: 602009057474 Country of ref document: DE Free format text: PREVIOUS MAIN CLASS: A61B0005048200 Ipc: A63F0013212000 |
|
GRAP | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOSNIGR1 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: GRANT OF PATENT IS INTENDED |
|
RIC1 | Information provided on ipc code assigned before grant |
Ipc: G06F 3/05 20060101ALI20180817BHEP Ipc: A63F 13/212 20140101AFI20180817BHEP Ipc: A63F 13/67 20140101ALI20180817BHEP Ipc: A61B 5/0482 20060101ALI20180817BHEP |
|
INTG | Intention to grant announced |
Effective date: 20180921 |
|
RAP1 | Party data changed (applicant data changed or rights of an application transferred) |
Owner name: NATIONAL UNIVERSITY OF SINGAPORE Owner name: INSTITUTE OF MENTAL HEALTH Owner name: AGENCY FOR SCIENCE, TECHNOLOGY AND RESEARCH |
|
GRAS | Grant fee paid |
Free format text: ORIGINAL CODE: EPIDOSNIGR3 |
|
GRAA | (expected) grant |
Free format text: ORIGINAL CODE: 0009210 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE PATENT HAS BEEN GRANTED |
|
AK | Designated contracting states |
Kind code of ref document: B1 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO SE SI SK SM TR |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: FG4D |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: EP Ref country code: AT Ref legal event code: REF Ref document number: 1106924 Country of ref document: AT Kind code of ref document: T Effective date: 20190315 |
|
REG | Reference to a national code |
Ref country code: IE Ref legal event code: FG4D |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R096 Ref document number: 602009057474 Country of ref document: DE |
|
REG | Reference to a national code |
Ref country code: NL Ref legal event code: MP Effective date: 20190313 |
|
REG | Reference to a national code |
Ref country code: LT Ref legal event code: MG4D |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 Ref country code: SE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 Ref country code: NO Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190613 Ref country code: FI Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LV Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 Ref country code: BG Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190613 Ref country code: GR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190614 Ref country code: HR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 Ref country code: NL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 |
|
REG | Reference to a national code |
Ref country code: AT Ref legal event code: MK05 Ref document number: 1106924 Country of ref document: AT Kind code of ref document: T Effective date: 20190313 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: ES Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 Ref country code: PT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190713 Ref country code: SK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 Ref country code: RO Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 Ref country code: CZ Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 Ref country code: IT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 Ref country code: EE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: PL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 Ref country code: SM Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R097 Ref document number: 602009057474 Country of ref document: DE |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: IS Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190713 Ref country code: AT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 |
|
PLBE | No opposition filed within time limit |
Free format text: ORIGINAL CODE: 0009261 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: DK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 |
|
26N | No opposition filed |
Effective date: 20191216 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: SI Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: TR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: MC Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: PL |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LI Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20190930 Ref country code: LU Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20190914 Ref country code: IE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20190914 Ref country code: CH Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20190930 |
|
REG | Reference to a national code |
Ref country code: BE Ref legal event code: MM Effective date: 20190930 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: BE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20190930 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: FR Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20190930 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: CY Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: HU Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT; INVALID AB INITIO Effective date: 20090914 Ref country code: MT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: MK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20190313 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: DE Payment date: 20240906 Year of fee payment: 16 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: GB Payment date: 20240922 Year of fee payment: 16 |