JPH0485680A - Head skin electric potential distribution pattern recognizing type syllable input device - Google Patents

Head skin electric potential distribution pattern recognizing type syllable input device

Info

Publication number
JPH0485680A
JPH0485680A JP2201647A JP20164790A JPH0485680A JP H0485680 A JPH0485680 A JP H0485680A JP 2201647 A JP2201647 A JP 2201647A JP 20164790 A JP20164790 A JP 20164790A JP H0485680 A JPH0485680 A JP H0485680A
Authority
JP
Japan
Prior art keywords
syllable
distribution pattern
user
recognition
input device
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
JP2201647A
Other languages
Japanese (ja)
Other versions
JPH07118015B2 (en
Inventor
Akira Hiraiwa
明 平岩
Katsunori Shimohara
勝憲 下原
Tadashi Uchiyama
匡 内山
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nippon Telegraph and Telephone Corp
Original Assignee
Nippon Telegraph and Telephone Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nippon Telegraph and Telephone Corp filed Critical Nippon Telegraph and Telephone Corp
Priority to JP2201647A priority Critical patent/JPH07118015B2/en
Publication of JPH0485680A publication Critical patent/JPH0485680A/en
Publication of JPH07118015B2 publication Critical patent/JPH07118015B2/en
Anticipated expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Abstract

PURPOSE:To use the title device even by a user unable to execute the voluntary movement of actual vocalization and so on by providing a presenting part to present the king of a syllable that the vocalization is to be intended by the user and the timing of the reading by heart of it in learning the pattern of a recognizing part and constituting the recognizing part with a neuronetwork. CONSTITUTION:The title device possesses a skin surface electrode 1, a detecting part 2 to detect head skin electric potential based on electric potential detected by the electrode 1, a storage part 3 to store the multichannel time series distribution pattern of the head skin electric potential and a recognizing pat 4 to identify and to output a corresponding vocalized syllable to the stored multichannel time series distribution pattern. Then, a presenting part 7 to present the kind of the syllable that vocalization is intended by a user and the timing of the reading by heart of it in learning the pattern of the recognizing part 4 is provided and the recognizing part 4 is constituted with a neuronetwork. Thus, the input device, for which the actual vocalization is unnecessary, applicable even to a person unable to execute any voluntary movement can be realized.

Description

【発明の詳細な説明】 〔産業上の利用分野〕 本発明は、側索性硬化症等の疾病によって1発声や、眼
球動作、キーボード操作等の随意運動を全く行うことが
できない障害者や、過大な重力加速度にさらされるため
に腕や指1口の随意運動ができない状況下のユーザ等が
使用する頭皮電位分布パターン!!!!識型音節入力装
置(以下、単に「入力装置」ともいう)に関し、特に、
上述の如きユーザの頭皮から検出される準備、電位分布
パターンから発声を意図する音節を認識するようにした
入力装置に関する。
[Detailed Description of the Invention] [Industrial Application Field] The present invention is applicable to people with disabilities who are unable to make any voluntary movements such as vocalization, eye movement, or keyboard operation due to diseases such as lateral sclerosis, and those with excessive Scalp potential distribution pattern used by users who are unable to make voluntary movements of their arms or fingers due to exposure to gravitational acceleration! ! ! ! Regarding the cognition syllable input device (hereinafter also simply referred to as "input device"), in particular,
The present invention relates to an input device that recognizes a syllable intended to be uttered from a potential distribution pattern detected from a user's scalp as described above.

〔従来の技術〕[Conventional technology]

従来、上述の如き随意運動が充分にできない障害者や、
過大な重力加速度にさらされるために腕や指1口の随意
運動ができない状況下のユーザ等を対象とした入力装置
として、例えば、脳波トポグラフィを応用した準備電位
分布パターンを認識・解析する方式に基づくものが研究
されている。
Traditionally, people with disabilities who are unable to make sufficient voluntary movements as described above,
As an input device for users who are unable to make voluntary movements of their arms or fingers due to exposure to excessive gravitational acceleration, for example, it can be used as an input device that recognizes and analyzes readiness potential distribution patterns using electroencephalogram topography. What is based on this is being researched.

なお、上述の脳波トポグラフィを応用した準備電位分布
パターンをコ識・解析する方式に関しては、塚原による
「脳の情報処理J(1964年)、180頁、金子によ
る1最近脳科学J(1988年10月)、+79頁の記
載が参考になるわ また、L述の脳波トポグラフィを応用した準備電位分布
パターンを認識・解析する方式を応用した入力装置には
、本発明者等が特願平1−52872号「脳波トポグラ
フィによる音節認識装置」に提案り。
Regarding the method of recognizing and analyzing the preparedness potential distribution pattern applying the above-mentioned electroencephalogram topography, please refer to Tsukahara's ``Brain Information Processing J (1964), p. 180, and Kaneko's 1 Recent Brain Science J (1988, The description on page 79 of the Japanese Patent Application No. 1999-1-1 (Japanese Patent Application No. No. 52872 "Syllable recognition device using electroencephalogram topography".

た装置、同じく、本発明者等が特願平2−37456号
[頭皮P$備電位パターンによる操作装置」に提案し、
た装置がある。
The device was also proposed by the present inventors in Japanese Patent Application No. 2-37456 [Operation device using scalp P$ potential pattern],
There is a device.

[発明が解決しようとする護摩] 上述の従来技術の項に記載した装置は、いずれも、人が
随意運動とし、ての発声や手の動作を行う直曲に、頭皮
から電位分布変化と(7,で検出される準備電位パター
ンを、パターンの学習・認識能力に優れた神経回路網に
ューラル・ネットワーク)を用いて、ユーザが意図した
発声音節や手の動作を認識・識別するようにしたもので
ある。これらの装置においては、認識部を構成する神経
回路網が、頭皮準備電位パターンと意図する随意運動の
関係をq習するどきに用いた頭皮ff;備電位パタンは
、発声もしくは手等の動作を実際にコ〜ザ(、行わせる
こと1:よって随意運動の開始時点を特定し、その開始
時点を基準どして、それより遡って設定した観測窓によ
り得られる時系列のパターンであった。
[Goma to be Solved by the Invention] All of the devices described in the above-mentioned prior art section detect changes in electric potential distribution from the scalp and ( The preparedness potential pattern detected in step 7 is used to recognize and identify the vocal syllables and hand movements intended by the user using a neural network (neural network) with excellent pattern learning and recognition abilities. It is something. In these devices, the neural network constituting the recognition unit learns the relationship between the scalp readiness potential pattern and the intended voluntary movement; In fact, it was a time-series pattern obtained by identifying the starting point of the voluntary movement, and using the observation window set from that starting point as a reference point.

従って、1・、述の認識部の神経回路網の)f習をPi
うたぬには1、ユーザが何等かの随意運動を行う体力を
有することが不可欠であり、疾病等のためにあらゆる随
意運動を行うことができない人に対しては、この装置を
用いての人力を行うことができないどいつ問題があった
4、 本発明は−J−記事情に鑑みてなされたもので、子の目
的どするところは、従来の技術におけるドア述の如き問
題を解消し、疾病等のためにあらゆる随意運動を行うこ
とができない人に対しても適用可能な、実際の発声等の
随意運動を必要どしない入力装置を提供することにある
Therefore, 1., the neural network of the recognition part mentioned above) is Pi
1. It is essential for the user to have the physical strength to perform some voluntary movements, and for people who are unable to perform any voluntary movements due to illness etc., this device can be used without manual effort. The present invention has been made in view of the above circumstances, and the purpose of the present invention is to solve the problems such as the door description in the conventional technology, It is an object of the present invention to provide an input device that does not require voluntary movements such as actual vocalization and is applicable even to people who cannot perform any voluntary movements due to illness or the like.

[謀顧を解決するための手段] 本発明の上述の目的は、複数の皮膚表1fi電極ど、該
電極により検出される電位を基に頭皮電位を検出する検
出部と、前記頭皮電位の多チャンネル時系列分布パター
ンを記憶するとともに必要な前処理を行う記憶部と、該
記憶部により記憶・前処理された多チャンネル時系列分
布パターンに対して、対応する発声音節を識別・出力す
る認識部とを壱する入力装置において、前記認識部のパ
ターン学習時に、ユーザが発声を意図すべき音節の種類
とその必読のタイミングを呈示する呈示部を設けるとと
もに、前記認識部を、m数のユニットとそれらを結ぶ重
み付きのリンクから成る神経回路網で構成することを特
徴とする入力装置によっで達成される。
[Means for solving the problem] The above-mentioned object of the present invention is to provide a detection unit that detects a scalp potential based on the potential detected by the electrodes, such as a plurality of skin surface 1fi electrodes; A storage unit that stores channel time-series distribution patterns and performs necessary preprocessing; and a recognition unit that identifies and outputs uttered syllables corresponding to the multi-channel time-series distribution patterns stored and preprocessed by the storage unit. In the input device, the recognition unit is provided with a presentation unit that presents the type of syllable that the user intends to utter and the timing at which it must be read when the recognition unit learns the pattern, and the recognition unit is formed of m units. This is achieved by an input device characterized by being constructed of a neural network consisting of weighted links connecting them.

〔作用] 本発明に係る入力装置においては、ユーザの頭皮から検
出される電位分布パターンからユーザが意図する発声音
節を認識する入力装置において、ユーザに、発声を意図
すべき音節の種類と、その必読のタイミングを呈示部に
呈示して、ユーザに発声等を意図させることにより、認
識、部のバタ・−ン学習時においても実際の発声等の随
意運動の必要性をなくシ2、あらゆる随意運動ができな
いユーザも使用することを可能にしたものである。
[Function] In the input device according to the present invention, the input device recognizes the syllable that the user intends to pronounce from the potential distribution pattern detected from the user's scalp, and the input device allows the user to know the type of syllable that the user intends to pronounce and the type of syllable that the user intends to pronounce. By presenting the timing of the must-read on the presentation unit and making the user intend to utter, etc., the need for voluntary movements such as actual utterance is eliminated even when learning recognition and part bata.2. This allows even users who are unable to exercise to use it.

なお、本明細書中において、「必読」とは、実際に発声
することなく文字を音節単位で発声を意図しながら読む
ことを言い、[発声を意図するjまたは[発声をイメー
ジするJとは、音節単位で発声を意識しながら実際には
発声しないことを言う。また、本発明において「音節」
とaう語は、いわゆるモノシラブルの発声単位から、こ
れらが複数個連続して(組み合わされて)発声されるシ
ラブルまでを含み、例えば、アルファベットまたは五4
−音から一文字漢字までをも含むものとする4、[実施
例] 以下、本発明の実施例を図面に基づいて詳細1.7説明
する。
In addition, in this specification, "must read" refers to reading the letters syllable by syllable without actually saying them. , to say that you are conscious of uttering each syllable but do not actually utter it. In addition, in the present invention, "syllable"
The word a includes everything from so-called monosyllable units to syllables in which a plurality of monosyllables are uttered consecutively (combined), for example, the alphabet or
4. [Embodiments] Hereinafter, embodiments of the present invention will be described in detail in 1.7 based on the drawings.

第1図は、本発明の一実施例である入力装置の全体構成
を示す図である。図において、1は多数の皮膚表面電極
、2は該皮膚表面電極1から検出される電位を基に、人
の頭皮から検出される電イΩを検出する検出部、3は上
述の人の頭皮から検出される電位の時系列パターンを記
憶する記憶部、4は該記憶部3により記憶された多チャ
ンネルの時系列電位パターンを入力として、該多チャン
ネル電位パターンを認識して、対応する音節を出力する
認識部、5は該認識部4で認識した音節を呈示する呈示
部を示している。また、6は上記認識部4のパターン学
習時に、ユーザが発声を意図すべき音節の種類を生成し
、音節の呈示信号を生成する呈示信号生成部、7は該呈
示信号生成部6で生成された音節信号を呈示する呈示部
、8は上述の各部のタイミングをとって、各部の動作を
制御する制御部である。
FIG. 1 is a diagram showing the overall configuration of an input device that is an embodiment of the present invention. In the figure, 1 is a large number of skin surface electrodes, 2 is a detection unit that detects the electric potential detected from the scalp of a person based on the potential detected from the skin surface electrode 1, and 3 is the scalp of the above-mentioned person. A storage unit 4 stores the time-series pattern of potentials detected from the storage unit 3, receives as input the multi-channel time-series potential pattern stored in the storage unit 3, recognizes the multi-channel potential pattern, and generates the corresponding syllable. The output recognition unit 5 indicates a presentation unit that presents the syllables recognized by the recognition unit 4. Further, 6 is a presentation signal generation unit that generates the type of syllable that the user intends to utter and generates a syllable presentation signal when the recognition unit 4 learns the pattern; 7 is a presentation signal generation unit that generates a syllable presentation signal; A presentation section 8 is a control section that controls the operation of each section by timing the above-mentioned sections.

なお、第1図中の各信号a−jの意味は、下記の通りで
ある。aは皮膚表面電極1で検出された頭皮電位時系列
信号、bは検出部2によって処理された頭皮電位時系列
信号、Cは記憶部3によって約処理された、数値化され
た頭皮電位時系列信号、dは認識部4によって識別され
た音節信号、e+Lgおよびbは制御部8と各部の動作
を動期させ、制御する制御信号、1は認識部4のパター
ン学習時に、ユーザが発声を意図すべき音節の種類を認
識部4に教師データとして呈示する信号、Jは呈示信号
生成部6により呈示部7に呈示すべく生成された信号を
示している。
The meanings of each signal aj in FIG. 1 are as follows. a is a scalp potential time series signal detected by the skin surface electrode 1, b is a scalp potential time series signal processed by the detection unit 2, and C is a digitized scalp potential time series processed by the storage unit 3. signal, d is a syllable signal identified by the recognition unit 4, e+Lg and b are control signals that activate and control the operation of the control unit 8 and each part, 1 is the syllable signal that the user intends to utter when the recognition unit 4 learns the pattern. A signal for presenting the type of syllable to be used as teacher data to the recognition section 4, J indicates a signal generated by the presentation signal generation section 6 to be presented to the presentation section 7.

第2図は、上述の認識部4の詳細を示す構成図であり、
4−1は神経回路網、4−2は重み更新制御部を示して
いる。ここでは、神経回路網4−1を層状に接続し、3
層で構成した場合を示しており、重み更新制御部4−2
は、認識部4の学習モード時に、制御部8により起動さ
れ、誤差検出4−28によって呈示信号生成部6からの
信号と神経回路網4−1の出力との差を検出し、これに
基づいて、荷重更新値計算4−2bによって荷重更新の
ための信号処理を行うものである。
FIG. 2 is a configuration diagram showing details of the above-mentioned recognition unit 4,
Reference numeral 4-1 indicates a neural network, and reference numeral 4-2 indicates a weight update control unit. Here, the neural network 4-1 is connected in layers, and 3
The case is shown in which the weight update control unit 4-2 is configured with layers.
is activated by the control unit 8 when the recognition unit 4 is in the learning mode, and the error detection unit 4-28 detects the difference between the signal from the presentation signal generation unit 6 and the output of the neural network 4-1, and based on this, Then, signal processing for updating the load is performed by the load update value calculation 4-2b.

以下、上述の如く構成された本実施例の動作を説明する
。なお、以下の説明においては、認識部4の神経回路網
4−1の学習モードAと、ユーザがある音節の発声を意
図して、発声をイメージしたときに、頭皮電位分布変化
に混在して出現する準備電位パターンを認識・識別して
発声を意図する音節を認識する認識モードBについて、
順次、説明する。
The operation of this embodiment configured as described above will be explained below. In addition, in the following explanation, learning mode A of the neural network 4-1 of the recognition unit 4 and a change in the scalp potential distribution when the user intends to utter a certain syllable and visualizes the utterance will be explained. Regarding recognition mode B, which recognizes and identifies the emerging readiness potential pattern and recognizes the syllable intended to be uttered,
I will explain them one by one.

まず、認識部4が頭皮電位分布パターンと発声を意図す
る音節との関係を学習する学習モードAについて説明す
る。
First, learning mode A in which the recognition unit 4 learns the relationship between the scalp potential distribution pattern and the syllable intended to be uttered will be described.

学習モードAとは、ユーザの頭皮から検出される電位分
布パターンとユーザが意図する発声音節の関係を、認識
部4の神経回路網4−1に学習させるモードである。音
節の種類と、その必読のタイミングをデイスプレィ装置
に呈示し、それに従って、ユーザに発声を意図させるよ
う指示することによって、実際の随意運動を行わせるこ
となく、学習パターンデータを特定できるようにしたも
ので、実際の随意運動の必要性をなくし、あらゆる随意
運動ができないユーザでも使用することが可能となる。
Learning mode A is a mode in which the neural network 4-1 of the recognition unit 4 learns the relationship between the potential distribution pattern detected from the user's scalp and the uttered syllable intended by the user. By presenting the type of syllable and the timing at which it must be read on the display device, and instructing the user to intend to pronounce it accordingly, learning pattern data can be identified without actually making voluntary movements. This eliminates the need for actual voluntary movements, making it possible for users who are unable to perform any voluntary movements to use it.

この動作のタイミングチャートを第4図に、また、フロ
ーチャートを第7図に示す。
A timing chart of this operation is shown in FIG. 4, and a flowchart is shown in FIG.

学習モードAは、大きく分けて、ユーザの必読トレーニ
ングを行う期間(ステップ72〜79)と、必読を意識
しない期間(ステップ80〜83)から成る。
The learning mode A is roughly divided into a period in which the user is trained to read the required reading (steps 72 to 79), and a period in which the user is not conscious of the required reading (steps 80 to 83).

学習モードAでは、まず、ユーザに、呈示される音節を
意図して必読するように指示して、ユーザが、呈示部7
に呈示される音節を必読し、必読に伴ってユーザの頭皮
上に発生する準備電位を含む頭皮電位分布を認識部4の
入力とし、呈示された音節の種類を出力として、認識部
4の神経回路網4−1が入力と出力との関係を学習する
In learning mode A, first, the user is instructed to read the presented syllables intentionally, and the user selects the presentation section 7.
The user must read the syllables that are presented, the scalp potential distribution including the readiness potential generated on the user's scalp due to the must-read is input to the recognition unit 4, and the type of the presented syllable is output, and the neural network of the recognition unit 4 is The circuit network 4-1 learns the relationship between input and output.

今、呈示部7への音節の呈示が、第3図に示す如くアル
ファベットを対象とする場合、格子状に表示されたアル
ファベットが、この順に、第4図のタイミングチャート
における呈示部7の文字のフラッシュ時に、発声を促す
音節の文字のみ背景色が反転してフラッシュするものと
する。この文字のフラッシュは、アルファベット類に行
われ、時間的にも等間隔で行われる。ユーザは、この等
間隔のタイミングを見計いながら、発声を意図してデイ
スプレィ装置に表示される音節を必読するものとする。
Now, when the presentation of syllables to the presentation section 7 targets the alphabet as shown in FIG. At the time of flashing, only the characters of the syllables that encourage utterance are flashed with the background color inverted. This flashing of characters is performed alphabetically and at equal intervals in time. It is assumed that the user must read the syllables displayed on the display device with the intention of uttering them, while checking the timing of the equal intervals.

ユーザが呈示部7のフラッシュのタイミングに慣れると
、ユーザは、フラッシュ刺激の直前に、次に必読すべき
文字の発声準備を意識し、フラッシュのタイミングに必
読できるようになる。このとき、ユーザの頭皮に装着さ
れた皮膚表面電極から得られた頭皮電位は、検出部2に
よって信号として検出され、記憶部3に記憶された呈示
部7のフラッシュの直前、1秒前からフラッシュ直前ま
での時間の頭皮電位時系列信号として処理され、第5図
に示す如く数値マトリクス化され、認識部4の神経回路
網4−1の入力層4−18に、第6図に示す如き入カバ
ターンとして与えられる。
Once the user gets used to the timing of the flash on the presentation unit 7, the user will be aware of preparing to speak the next must-read character immediately before the flash stimulation, and will be able to read the must-read character at the timing of the flash. At this time, the scalp potential obtained from the skin surface electrode attached to the user's scalp is detected as a signal by the detection unit 2, and is stored in the storage unit 3 immediately before the flash of the presentation unit 7, from 1 second before the flash. It is processed as a scalp potential time-series signal of the previous time, converted into a numerical matrix as shown in FIG. Given as a cover turn.

このとき、認識部4の神経回路網4−1の出力層4−1
0には、呈示された音節の種類が教師信号として与えら
れる。第6図では、入力層4−1aに音節“b″の必読
直前に得られた、頭皮電位時系列信号の数値マトリクス
を与えた場合を示しており、教師信号としては、図に示
す如く、出力層4−10の”b”のユニットのみに“1
″を、他のユニットには“0″を与える例を示している
At this time, the output layer 4-1 of the neural network 4-1 of the recognition unit 4
0 is given the type of syllable presented as a teacher signal. FIG. 6 shows a case where the input layer 4-1a is given a numerical matrix of scalp potential time-series signals obtained just before the must-read syllable "b", and as the teacher signal, as shown in the figure, “1” is applied only to unit “b” of output layer 4-10.
”, and “0” is given to other units.

すべての音節に対して、複数回、上述の必読を繰り返し
た後、必読による認識部4の学習は終了する。その後、
そのままの状態で、認識部4にはユーザが必読を意識し
ていない状態の頭皮電位時系列信号を与え、出力層4−
1cには入力音節無しを教示し、これを複数回繰り返す
。入力音節無しの頭皮電位のパターンの学習が終了した
ら、学習モードAは終了とする。
After repeating the above-mentioned required reading multiple times for all syllables, the learning of the recognition unit 4 through the required reading is completed. after that,
In this state, the recognition unit 4 is given a scalp potential time series signal that the user is not aware of as a must-read, and the output layer 4-
1c is taught no input syllable, and this is repeated multiple times. When learning of the scalp potential pattern without input syllables is completed, learning mode A is ended.

ここで、認識部4の神経回路網4−1の学習法には、既
に提案されている種々の方法を利用することができ、例
えば、誤差逆伝播法の場合、認識部4の学習において、
一つのパターンの学習は、神経回路網が、入力と出力に
与えられたパターン関係に誤差がなくなるように、ある
いは、誤差許容量に達するまで何度も学習するものであ
る。
Here, various methods that have already been proposed can be used for the learning method of the neural network 4-1 of the recognition unit 4. For example, in the case of the error backpropagation method, in the learning of the recognition unit 4,
In learning one pattern, the neural network repeatedly learns the pattern relationship given to the input and output until there is no error or until an error tolerance is reached.

次に、認識モードBについて、第9図に示すフローチャ
ートに基づいて説明する。認識部4の学習モードが終っ
たら、ユーザの頭皮電位の取り込みを開始する(ステッ
プ91)。以後、ユーザは、自分が入力を意図する音節
を順次、必読して、発声のイメージを行う。ここで、第
8図に示す如く、時系列の頭皮電位信号に対して、学習
モードAで開いたと同じ長さの観測窓(ここでは1秒間
)を開き、観測窓内の信号を学習モードAの場合と同様
に、数値マトリクス化し、認識部4の神経回路網4−1
の入力層4−18に与える(ステップ92〜93)。
Next, recognition mode B will be explained based on the flowchart shown in FIG. When the learning mode of the recognition unit 4 is finished, the acquisition of the user's scalp potential is started (step 91). Thereafter, the user must read the syllables that he/she intends to input one after another, and visualizes how to pronounce them. Here, as shown in Fig. 8, an observation window of the same length (here, 1 second) as opened in learning mode A is opened for the time-series scalp potential signal, and the signal within the observation window is observed in learning mode A. Similarly to the case of , the neural network 4-1 of the recognition unit 4 is converted into a numerical matrix
input layer 4-18 (steps 92-93).

与えられた入カバターンとしての数値マトリクスに対し
て、それぞれの音節に対応した出力層4−1cの各ユニ
ットが出力するが、この出力層のユニットのなかで、−
書出力値の高いユニットを、認識した音節とする(ステ
ップ94)。例えば、出力層ユニットの出力値のうち、
 e″のユニットの出力値が一番大きい値ならば、ユー
ザが発声を意図する音節は、 e”であるとする。サン
プリングレートをΔtとすると、連続した頭皮電位時系
列信号に対して、第8図に示す如く、順次、時間窓をず
らしていきくステップ95〜96)、それぞれの時間窓
内の頭皮電位時系列信号に対して、認識部4によって認
識を行っていく。上述の処理により、出力された音節列
によって、ユーザが発声を意図した文字列が形成される
Each unit of the output layer 4-1c corresponding to each syllable outputs a numerical matrix as a given input cover pattern, but among the units of this output layer, -
The unit with the highest written output value is set as the recognized syllable (step 94). For example, among the output values of the output layer unit,
If the output value of the unit e'' is the largest value, it is assumed that the syllable that the user intends to utter is e''. Assuming that the sampling rate is Δt, steps 95 to 96) in which the time windows are sequentially shifted as shown in FIG. 8 for the continuous scalp potential time series signals, calculate the scalp potential time series signals within each time window. The recognition unit 4 then performs recognition. Through the above-described processing, a string of characters intended to be uttered by the user is formed from the output syllable string.

上記実施例によれば、背景脳波に混在し、必読または発
声を意図することもしくは実際の発声に伴って出現する
準備電位変化のパターンの認識と識別を、学習能力を有
する神経回路網によって行い、かつ、神経回路網の学習
時には、ユーザが、ユーザに呈示される信号に対して、
必読のタイミングを同期して行うので、随意運動を全く
行うことができない状態においても、自分の意図する音
節列を入力することができる。
According to the above embodiment, a neural network having a learning ability recognizes and identifies patterns of change in readiness potential that are mixed in the background brain waves and appear in conjunction with must-read or intended vocalization or actual vocalization, In addition, when learning the neural network, the user responds to the signals presented to the user,
Since the required reading timings are synchronized, you can input the syllable string you intend even if you are unable to make any voluntary movements.

なお、上述の説明においては、アルファベットを対象音
節の例として説明したが、前述の如く、五十音を対象に
すること、他の言語の音節を対象とすることも可能であ
る。また、第4図に示したタイミングチャートにおいて
は、呈示部のフラッシュ刺激のインターバルを1.6〜
1.9秒とした例を示したが、フラッシュ刺激のインタ
ーバルを変えること、フラッシュの継続時間を変えるこ
とは、フラッシュ刺激に伴ってユーザの頭皮に観測され
る誘発電位P300の影響が観測窓内に入らない範囲に
おいて、かつ、ユーザが認知できる範囲内において、そ
れぞれ可能である。
In addition, in the above description, the alphabet was explained as an example of target syllables, but as mentioned above, it is also possible to target Japanese syllables or syllables of other languages. In addition, in the timing chart shown in Figure 4, the interval of flash stimulation of the presentation section is set to 1.6~
Although we have shown an example of 1.9 seconds, changing the interval of flash stimulation and changing the duration of flash will reduce the influence of the evoked potential P300 observed in the user's scalp due to flash stimulation within the observation window. Each of these is possible as long as it does not fall within the range and within the range that the user can recognize.

更に、認識部の神経回路網の充分な学習が終了している
場合、すなわち、不特定多数の充分な数のユーザの発す
る頭皮電位時系列パターンと、上記ユーザの意図する発
声音節の関係を学習した認識部の神経回路網の重み値を
用いることができる場合には、少数カテゴリ、例えば、
数個の音節の弁別に関しては、前述の頭皮電位時系列パ
ターンの個人差を、前記神経回路網が吸収できることか
ら、上記実施例に示した入力装置から、呈示信号生成部
6と呈示部1とを除去した構成として、学習モードAを
経ていない不特定のユーザに対しても、本装置を、認識
モードBの専用機として提供することも可能である。
Furthermore, if the neural network of the recognition unit has completed sufficient learning, in other words, it learns the relationship between the scalp potential time-series patterns uttered by a sufficient number of unspecified users and the syllables uttered by the users. If the weight values of the neural network of the recognized recognition unit can be used, the minority category, e.g.
Regarding the discrimination of several syllables, since the neural network can absorb the aforementioned individual differences in the scalp potential time series pattern, the presentation signal generation section 6 and the presentation section 1 can be connected to each other from the input device shown in the above embodiment. As a configuration in which the above is removed, it is also possible to provide this device as a dedicated machine for recognition mode B even to unspecified users who have not gone through learning mode A.

つまり、複数の皮膚表面電極と、該電極により検出され
る電位を基に頭皮電位を検出する検出部と、前記頭皮電
位の多チャンネル時系列分布パターンを記憶するととも
に必要な前処理を行う記憶部と、該記憶部により記憶・
前処理された多チャンネル時系列分布パターンに対して
、対応する発声音節を識別・出力する認識部とを有する
頭皮電位分布パターン認識型音節入力装置において、前
記認識部を、複数のユニットとそれらを結ぶ重み付きの
リンクから成る神経回路網で構成し、かつ、該認識部は
、充分な学習を終了したものとすることを特徴とする頭
皮電位分布パターン認識型音節入力装置も実現可能であ
る。
That is, a plurality of skin surface electrodes, a detection unit that detects scalp potential based on the potential detected by the electrodes, and a storage unit that stores the multichannel time-series distribution pattern of the scalp potential and performs necessary preprocessing. and the storage unit stores and
A scalp potential distribution pattern recognition type syllable input device comprising a recognition unit that identifies and outputs a corresponding uttered syllable with respect to a preprocessed multi-channel time-series distribution pattern. It is also possible to realize a scalp potential distribution pattern recognition type syllable input device comprising a neural network consisting of weighted links, and characterized in that the recognition section has completed sufficient learning.

〔発明の効果〕〔Effect of the invention〕

以上、詳細に説明した如く、本発明によれば、複数の皮
膚表面電極と、該電極により検出される電位を基に頭皮
電位を検出する検出部と、前記頭皮電位の多チャンネル
時系列分布パターンを記憶するとともに必要な前処理を
行う記憶部と、該記憶部により記憶・前処理された多チ
ャンネル時系列分布パターンに対して、対応する発声音
節を識別・出力する認識部とを有する頭皮電位分布パタ
ーン認識型音節入力装置において、前記認識部のパター
ン学習時に、ユーザが発声を意図すべき音節の種類とそ
の必読のタイミングを呈示する呈示部を設けるとともに
、前記認R部を、複数のユニットとそれらを結ぶ重み付
きのリンクから成る神経回路網で構成したので、疾病等
のためにあらゆる随意運動を行うことができない人に対
しても適用可能な、実際の発声等の随意運動を必要とし
ない入力装置を実現できるという顕著な効果を奏するも
のである。
As described in detail above, according to the present invention, there are provided a plurality of skin surface electrodes, a detection unit that detects a scalp potential based on the potential detected by the electrodes, and a multichannel time-series distribution pattern of the scalp potential. and a recognition section that identifies and outputs a corresponding uttered syllable with respect to the multi-channel time-series distribution pattern stored and preprocessed by the storage section. In the distribution pattern recognition type syllable input device, a presentation section is provided that presents the type of syllable that the user intends to utter and the timing at which it must be read when the recognition section learns the pattern, and the recognition R section is configured to include a plurality of units. Since it is constructed with a neural network consisting of weighted links connecting them, it can be applied to people who are unable to perform any voluntary movements due to illness, etc., and does not require voluntary movements such as actual vocalization. This has the remarkable effect that it is possible to realize an input device that does not require any input.

【図面の簡単な説明】[Brief explanation of drawings]

第1図は本発明の一実施例である入力装置の全体構成を
示す図、第2図は認識部4の詳細を示す構成図、第3図
は呈示部の呈示画面の一例を示す図、第4図は学習モー
ドAにおけるタイミングチャート、第5図は第4図のタ
イミングチャートに基づいて時系列頭皮電位信号を数値
マトリクスに変換する処理の説明図、第6図は上記数値
マトリクスを認識部に入力し、対応する音節を教師信号
として出力層へ教示する状況を説明する図、第7図は学
習モードへ時の動作フローチャート、第8図は学習モー
ドが終了した認識部が時系列頭皮電位信号を認識する状
況を示す図、第9図は認識モードBの動作フローチャー
トである。 1・皮膚表面電極、2:検出部、 :認識部、4−1・神経回路網、4 4−1b:隠れ層、4−1c:出力層、新制術部、4−
2a:誤差検出、4 値計算、5 認識内容呈示部、6 部、7:呈示部、8:制御部。 38記憶部、4 18 入力層、 4−2:重み更 2b:荷重更新 :呈示信号生成
FIG. 1 is a diagram showing the overall configuration of an input device that is an embodiment of the present invention, FIG. 2 is a configuration diagram showing details of the recognition section 4, and FIG. 3 is a diagram showing an example of the presentation screen of the presentation section. FIG. 4 is a timing chart in learning mode A, FIG. 5 is an explanatory diagram of the process of converting a time-series scalp potential signal into a numerical matrix based on the timing chart of FIG. 4, and FIG. 6 is a recognition unit , and the corresponding syllable is taught to the output layer as a teacher signal. Figure 7 is an operation flowchart when entering the learning mode. Figure 8 is a diagram explaining the situation in which the recognition unit outputs the time-series scalp potential after the learning mode is completed. FIG. 9, which is a diagram showing the situation in which signals are recognized, is an operation flowchart of recognition mode B. 1. Skin surface electrode, 2: Detection section, : Recognition section, 4-1. Neural network, 4 4-1b: Hidden layer, 4-1c: Output layer, New system section, 4-
2a: error detection, 4 value calculation, 5 recognition content presentation section, 6 section, 7: presentation section, 8: control section. 38 storage unit, 4 18 input layer, 4-2: Weight update 2b: Load update: Presentation signal generation

Claims (1)

【特許請求の範囲】[Claims] (1)複数の皮膚表面電極と、該電極により検出される
電位を基に頭皮電位を検出する検出部と、前記頭皮電位
の多チャンネル時系列分布パターンを記憶するとともに
必要な前処理を行う記憶部と、該記憶部により記憶・前
処理された多チャンネル時系列分布パターンに対して、
対応する発声音節を識別・出力する認識部とを有する頭
皮電位分布パターン認識型音節入力装置において、前記
認識部のパターン学習時に、ユーザが発声を意図すべき
音節の種類とその必読のタイミングを呈示する呈示部を
設けるとともに、前記認識部を、複数のユニットとそれ
らを結ぶ重み付きのリンクから成る神経回路網で構成す
ることを特徴とする頭皮電位分布パターン認識型音節入
力装置。
(1) A plurality of skin surface electrodes, a detection unit that detects scalp potential based on the potential detected by the electrodes, and a memory that stores the multichannel time-series distribution pattern of the scalp potential and performs necessary preprocessing. and the multi-channel time-series distribution pattern stored and preprocessed by the storage unit,
In a scalp potential distribution pattern recognition type syllable input device having a recognition unit that identifies and outputs a corresponding uttered syllable, when the recognition unit learns the pattern, the type of syllable that the user intends to utter and the timing at which it must be read are presented. 1. A scalp potential distribution pattern recognition type syllable input device, characterized in that a presentation section is provided, and the recognition section is constituted by a neural network consisting of a plurality of units and weighted links connecting them.
JP2201647A 1990-07-30 1990-07-30 Scalp potential distribution pattern recognition type syllable input device Expired - Fee Related JPH07118015B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2201647A JPH07118015B2 (en) 1990-07-30 1990-07-30 Scalp potential distribution pattern recognition type syllable input device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
JP2201647A JPH07118015B2 (en) 1990-07-30 1990-07-30 Scalp potential distribution pattern recognition type syllable input device

Publications (2)

Publication Number Publication Date
JPH0485680A true JPH0485680A (en) 1992-03-18
JPH07118015B2 JPH07118015B2 (en) 1995-12-18

Family

ID=16444557

Family Applications (1)

Application Number Title Priority Date Filing Date
JP2201647A Expired - Fee Related JPH07118015B2 (en) 1990-07-30 1990-07-30 Scalp potential distribution pattern recognition type syllable input device

Country Status (1)

Country Link
JP (1) JPH07118015B2 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0833616A (en) * 1994-07-22 1996-02-06 Nec Corp Device and method for transmitting living body internal state
JP2012183292A (en) * 2011-03-03 2012-09-27 Seoul National Univ R&Db Foundation Interface device and method between brain of preparation potential base and computer
CN111221419A (en) * 2020-01-13 2020-06-02 武汉大学 Array type flexible capacitor electronic skin for sensing human motion intention

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
JOURNAL OF NEUROPHYSIOLOGY=1980 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0833616A (en) * 1994-07-22 1996-02-06 Nec Corp Device and method for transmitting living body internal state
JP2012183292A (en) * 2011-03-03 2012-09-27 Seoul National Univ R&Db Foundation Interface device and method between brain of preparation potential base and computer
CN111221419A (en) * 2020-01-13 2020-06-02 武汉大学 Array type flexible capacitor electronic skin for sensing human motion intention

Also Published As

Publication number Publication date
JPH07118015B2 (en) 1995-12-18

Similar Documents

Publication Publication Date Title
Breznitz Asynchrony of visual-orthographic and auditory-phonological word recognition processes: An underlying factor in dyslexia
Kröger et al. Towards a neurocomputational model of speech production and perception
Vernon Reading and Its Difficulties: A Physiological Study
Van Turennout et al. Electrophysiological evidence on the time course of semantic and phonological processes in speech production.
Moriarty et al. Phonological awareness intervention for children with childhood apraxia of speech
Balota et al. Mental chronometry: beyond onset latencies in the lexical decision task.
Männel et al. Accentuate or repeat? Brain signatures of developmental periods in infant word recognition
JP2003504646A (en) Systems and methods for training phonological recognition, phonological processing and reading skills
Angrick et al. Interpretation of convolutional neural networks for speech spectrogram regression from intracranial recordings
AlSaleh et al. Brain-computer interface technology for speech recognition: A review
Williams et al. Interlanguage dynamics and lexical networks in nonnative L2 signers of ASL: Cross-modal rhyme priming
Corina On the nature of left hemisphere specialization for signed language
Kóbor et al. ERP evidence for implicit L2 word stress knowledge in listeners of a fixed-stress language
Studer-Eichenberger et al. Statistical learning, syllable processing, and speech production in healthy hearing and hearing-impaired preschool children: A mismatch negativity study
Ye et al. Attention bidirectional LSTM networks based mime speech recognition using sEMG data
JPH0485680A (en) Head skin electric potential distribution pattern recognizing type syllable input device
Wilcox The structure of signed & spoken languages
Cohen-Mimran Temporal processing deficits in Hebrew speaking children with reading disabilities
Patel et al. Teachable interfaces for individuals with dysarthric speech and severe physical disabilities
Hamidi Digital tangible games for speech intervention
Dozorska The mathematical model of electroencephalographic and electromyographic signals for the task of human communicative function restoration
Lalain et al. Dyslexia: The articulatory hypothesis revisited
US20220208173A1 (en) Methods of Generating Speech Using Articulatory Physiology and Systems for Practicing the Same
JPH02232783A (en) Syllable recognizing device by brain wave topography
Marecka et al. An ear for language: sensitivity to fast amplitude rise times predicts novel vocabulary learning

Legal Events

Date Code Title Description
LAPS Cancellation because of no payment of annual fees