JPH1031497A - Voice conversation control method and voice conversation system - Google Patents
Voice conversation control method and voice conversation systemInfo
- Publication number
- JPH1031497A JPH1031497A JP8189060A JP18906096A JPH1031497A JP H1031497 A JPH1031497 A JP H1031497A JP 8189060 A JP8189060 A JP 8189060A JP 18906096 A JP18906096 A JP 18906096A JP H1031497 A JPH1031497 A JP H1031497A
- Authority
- JP
- Japan
- Prior art keywords
- dialogue
- keyword
- response
- recognition
- word
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Abstract
Description
【0001】[0001]
【発明の属する技術分野】本発明は、情報検索などを行
なうために利用する計算機システムに係り、特に、マイ
クとスピーカあるいは電話などの音声入出力インタフェ
ースを備え、誰でも容易に利用することができる音声対
話システムに関するものである。BACKGROUND OF THE INVENTION 1. Field of the Invention The present invention relates to a computer system used for performing information retrieval and the like, and more particularly, to a computer system provided with a microphone and a voice input / output interface such as a speaker or a telephone so that anyone can use it easily. The present invention relates to a voice interaction system.
【0002】[0002]
【従来の技術】音声対話システムでは一般に、利用者の
システムに対するデータ入力において音声認識の技術を
用いている。音声によるデータ入力は、電話を通して使
う場合、あるいは車の運転中などのいわゆるハンズアイ
ビジーの状況下使う場合、非常に有効である。2. Description of the Related Art Generally, a speech dialogue system uses a speech recognition technique for data input to a user's system. Voice data input is very effective when used over a telephone or when used in a so-called hands-eye busy situation such as driving a car.
【0003】普通は複数の項目(例えば、電話番号検索
のアプリケーションでは相手の所属と名前、チケットの
予約のアプリケーションでは月日、時間、大人または子
供、枚数等)に対してデータを入力する。しかし一方、
音声認識において100%の認識率を実現することは不
可能であり、利用者の入力に対して、入力項目毎にシス
テムの認識結果を確認することが必須である。例えば、
文または文節単位の音声認識が可能な音声対話システム
での対話例は、 システム:「相手の所属と名前をどうぞ」 利用者 :「資材課の佐藤さん」 システム:「資材課の佐藤ですね」 利用者 :「はい」 となる。この例のように、文または文節単位の音声認識
が可能なシステムで認識が正しく行われた場合は、入力
時間も短く利用効率が良くなる。しかし、文または文節
単位の認識の場合、複数のキーワード(ここでの例では
相手の所属と名前)の組合せ分だけ認識対象の種類が多
くなるばかりでなく、助詞も含めた表現のバラエティが
多様になるという点で、単語認識に比較して認識性能が
低くなる。さらに、認識を間違えた場合は、次のような
問題も発生する。Normally, data is input for a plurality of items (for example, the affiliation and name of a partner in a telephone number search application, the date and time, adults or children, the number of sheets, etc. in a ticket reservation application). But on the other hand,
It is impossible to achieve a recognition rate of 100% in voice recognition, and it is essential to check the recognition result of the system for each input item for a user input. For example,
An example of dialogue in a voice dialogue system that can recognize sentences or phrases is as follows: System: "Please select the affiliation and name of the other party" User: "Mr. Sato of the Materials Division" System: "Sato of the Materials Division" User: Yes. As in this example, when recognition is correctly performed in a system capable of performing speech recognition in units of sentences or phrases, the input time is short and the use efficiency is improved. However, in the case of recognition by sentence or phrase, not only the number of types of recognition targets increases by the combination of a plurality of keywords (in this example, the affiliation and name of the partner), but also the variety of expressions, including particles, varies. , The recognition performance is lower than that of word recognition. Further, if the recognition is wrong, the following problem also occurs.
【0004】 システム:「相手の所属と名前をどうぞ」 利用者 :「資材課の佐藤さん」 システム:「資材課の加藤ですね」 利用者 :「いいえ」 システム:「相手の所属と名前をどうぞ」 最初のシステムの質問に、所属と名前という2つのキー
ワードが含まれている。この場合、システム側でこのや
りとりだけからでは、利用者が否定したことが所属と名
前のどちらか片方だけを誤認識したことを意味するの
か、あるいは両方誤認識したことを意味するのか、が特
定できない。したがって、所属と名前の両方のキーワー
ドが正しく認識されるまで、同じ質問を繰り返すことに
なる。この場合は、時間がかかるという問題がある。ま
た別な対処方法として、利用者に誤認識された方だけ再
入力してもらう方法も考えられる。この場合は、所属か
名前かわからない発声を認識しなければならないので、
音声認識の性能に対する要求が現状で対応可能なレベル
より高くなるという問題がある上、そのように利用者の
発声を誘導するにはどうすれば良いかという、解決が非
常に困難な課題が生じる。[0004] System: "Please choose the affiliation and name of the other party" User: "Mr. Sato of the Materials Division" System: "It's Kato of the Materials Division" User: "No" System: "Please choose the affiliation and name of the other party The first system question included two keywords: affiliation and name. In this case, the system alone can determine whether the user's rejection means that either the affiliation or the name has been misrecognized, or that both have been misrecognized. Can not. Thus, the same question is repeated until both the affiliation and name keywords are correctly recognized. In this case, there is a problem that it takes time. As another countermeasure, a method in which only the user who has been erroneously recognized may input again is also conceivable. In this case, you need to recognize utterances that do not know your affiliation or name,
In addition to the problem that the demand for the performance of speech recognition is higher than the level that can be currently dealt with, there is a very difficult problem of how to guide the utterance of the user.
【0005】これに対して、単語認識の場合は認識対象
の表現のバラエティが抑えられ、現状レベルでもほぼ満
足のいく認識性能が得られる。また、項目を1つずつ質
問し、確認するために、確実に1項目ずつのデータ入力
が行える。例えば、次のようになる。On the other hand, in the case of word recognition, the variety of expressions to be recognized is suppressed, and almost satisfactory recognition performance can be obtained even at the current level. In addition, in order to ask and confirm items one by one, data can be input one item at a time. For example,
【0006】 システム:「相手の所属をどうぞ」 利用者 :「資材課」 システム:「資材課ですか」 利用者 :「はい」 システム:「相手の名前をどうぞ」 利用者 :「佐藤」 システム:「佐藤ですか」 利用者 :「はい」 ただし、ここに示したようにやりとりが長くなり全体に
時間がかかる傾向になるという大きな問題が残る。[0006] System: "Please add the other party's affiliation" User: "Materials Section" System: "Is the Materials Section?" User: "Yes" System: "Please enter the name of the other party" User: "Sato" System: "Is Sato?" User: "Yes" However, as shown here, there remains a major problem that the exchange becomes longer and the whole tends to take longer.
【0007】[0007]
【発明が解決しようとする課題】上記のような従来の音
声対話システムにおいて、現状の音声認識性能と時間的
な利用効率は相反するパラメータであった。In the above-mentioned conventional speech dialogue system, the current speech recognition performance and the temporal utilization efficiency are incompatible parameters.
【0008】本発明の目的は、現状で可能な音声認識性
能の範囲で、最も効率良くデータ入力が行え、利用者と
システムとの間で円滑な対話を実現できる音声対話シス
テムを提供することにある。[0008] It is an object of the present invention to provide a speech dialogue system capable of inputting data most efficiently within the range of currently available speech recognition performance and realizing a smooth dialogue between a user and the system. is there.
【0009】[0009]
【課題を解決するための手段】本発明によれば、以下の
ような手段による対話制御方法および以下のような手段
を設けた音声対話システムが提供される。According to the present invention, there is provided a dialogue control method using the following means and a voice dialogue system provided with the following means.
【0010】タスク管理手段に要求を出しその返答結果
に応じて、応答生成手段、認識語彙補足手段、キーワー
ド判定手段、およびキーワード保持手段を制御し、また
認識語彙補足手段を介して音声認識手段を、さらに応答
生成手段を介して音声合成手段を制御し、システムと利
用者の間の対話を進行させる対話制御手段の元で、前記
キーワード保持手段は、前記対話制御手段の要求に基づ
き、キーワードの保持、削除、最新のキーワードの通知
を行い、前記対話制御手段は、タスク管理手段に対話の
進行上における次のアクションを促すガイダンス内容を
要求して該ガイダンス内容を受け取り、また前記キーワ
ード保持手段に最新のキーワードを要求して該最新のキ
ーワードを受け取り、前記ガイダンス内容と前記最新の
キーワード、およびそれらを用いて応答文を生成させる
指示を応答生成手段へ通知し、前記応答生成手段は、前
記対話制御手段から受け取った指示に従い、 対話の進
行上における次のアクションを促すガイダンス文の中
に、同時に受け取った前段階の認識結果でもある最新の
キーワードを含めた応答文を生成し、音声合成手段へ出
力し、次に前記対話制御手段は、タスク管理手段に対話
の進行上における次の認識語彙を要求して該認識語彙を
受け取り、タスクに依存したキーワードからなる該認識
語彙を認識語彙補足手段へ送り、前記認識語彙補足手段
は、前記対話制御手段より受け取った認識語彙に、「取
消」「ヘルプ」「ストップ」「もう一度」「わからな
い」「任意」などのタスクから独立なコマンドを表す単
語を補足して、音声認識手段およびキーワード判定手段
に渡し、前記キーワード判定手段は、前記認識語彙補足
手段より得た補足後の認識語彙と前記音声認識手段より
得た認識結果とを比較し、該認識結果が前記タスクから
独立なコマンドか、タスクに依存したキーワードかを判
定し、その判定結果を対話制御手段へ送り、さらに次に
前記対話制御手段は、前記キーワード判定手段の判定結
果を元に、該判定結果がキーワードである場合は、それ
をキーワード保持手段に送ると共に該キーワードをタス
ク管理手段に送り、 前記判定結果がコマンドである場
合は、各コマンドに対する処理を行う。[0010] A request is issued to the task management means, and the response generation means, the recognized vocabulary supplementing means, the keyword determination means, and the keyword holding means are controlled in accordance with the response result, and the speech recognition means is controlled via the recognized vocabulary supplementing means. Further, under the control of the speech synthesizing means via the response generating means, and the dialog control means for proceeding the dialog between the system and the user, the keyword holding means, based on the request of the dialog control means, Holding, deleting, notifying of the latest keyword, the dialog control means requests the task management means for guidance content for prompting the next action in the progress of the dialogue, receives the guidance content, and sends the guidance content to the keyword holding means. Requesting the latest keyword and receiving the latest keyword, the guidance content and the latest keyword, and An instruction to generate a response sentence using them is sent to the response generation means, and the response generation means, in accordance with the instruction received from the interaction control means, includes a guidance sentence for prompting a next action in the progress of the interaction, At the same time, a response sentence including the latest keyword, which is also the recognition result of the previous stage received, is generated and output to the speech synthesis means. Then, the dialog control means sends the task management means the next recognized vocabulary in the progress of the dialog. , Receiving the recognition vocabulary, and sending the recognition vocabulary comprising task-dependent keywords to the recognition vocabulary supplementing means. The recognition vocabulary supplementing means adds “cancel”, “cancel” to the recognition vocabulary received from the dialog control means. Supplement words that represent commands that are independent of tasks, such as help, stop, repeat, don't know, or any Handover, and the keyword determination unit compares the supplemented recognized vocabulary obtained by the recognized vocabulary supplementing unit with the recognition result obtained by the voice recognition unit, and the recognition result is a command independent of the task. Or a keyword depending on the task, and sends the result of the determination to the dialogue control means. Next, the dialogue control means determines whether the keyword is a keyword based on the determination result of the keyword determination means. Sends the keyword to the keyword holding unit and sends the keyword to the task management unit, and if the determination result is a command, performs a process for each command.
【0011】[0011]
【発明の実施の形態】以下図を用いて本発明の実施例を
説明する。DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS An embodiment of the present invention will be described below with reference to the drawings.
【0012】図1は本発明による音声対話システムの一
実施例を示すブロック図である。音声対話システムは、
あるタスクを遂行することを目的として、利用者とシス
テムが音声を使ってやりとりをするようなシステムであ
る。本発明による音声対話システムは、対話制御手段
(5)の制御の元に動作する。対話制御手段(5)の動
作については、後で詳述する。FIG. 1 is a block diagram showing one embodiment of a voice dialogue system according to the present invention. The spoken dialogue system
It is a system in which the user and the system communicate using voice for the purpose of performing a certain task. The voice dialogue system according to the present invention operates under the control of the dialogue control means (5). The operation of the dialog control means (5) will be described later in detail.
【0013】タスク管理手段(10)は、タスクに依存
した処理を全て担当する。タスクに依存した情報(タス
ク遂行の手順、場面毎の入力待ち語彙、等)を管理した
り、対話制御手段からのタスクの進行等についての問い
合わせなどに返答したりする。The task management means (10) is in charge of all tasks dependent processing. It manages task-dependent information (task execution procedure, input waiting vocabulary for each scene, etc.) and responds to inquiries about the progress of tasks from the dialog control means.
【0014】音声認識手段(2)は、与えられた認識語
彙の範囲内で、入力される音声(1)を認識し、認識結
果を1つの単語あるいは複数の単語の列として出力す
る。音声認識手段(2)の実現方法としては様々な手法
が考えられ、本発明はその方法を限定しない。例えば、
HMM(HiddenMarkovModel)という
確率モデルを用いる手法が扱いやすい。この方法を用い
れば、任意の認識語彙をテキストとして与えるだけで、
音声認識できる構成にすることができる。詳細は、“中
川聖一:”確率モデルによる音声認識、電子情報通信学
会、1988”他の文献に詳しいので、本明細書では省
略する。The speech recognition means (2) recognizes the input speech (1) within the range of the given recognition vocabulary, and outputs the recognition result as one word or a sequence of a plurality of words. Various methods are conceivable as a method of realizing the voice recognition means (2), and the present invention does not limit the method. For example,
A method using a probabilistic model called HMM (Hidden Markov Model) is easy to handle. With this method, just giving any recognized vocabulary as text,
A configuration capable of voice recognition can be provided. The details are described in "Seiichi Nakagawa: Speech Recognition by Probabilistic Model, IEICE, 1988" and other documents, and are omitted in this specification.
【0015】認識語彙補足手段(3)は、対話制御手段
(5)より受け取った認識語彙に、特にタスクから独立
なコマンドを表す単語等を補足して、音声認識手段に渡
す。ここで、コマンドとしては、 「違います」「違
う」「いいえ」「取消」「キャンセル」などの否定ある
いは取消を意味する単語、 「ヘルプ」などの助言要求
を意味する単語、 「ストップ」「停止」「中止」など
のシステムの処理の停止要求を意味する単語、「もう一
度」「リピート」などの直前のシステム応答の再出力要
求を意味する単語、などがある。さらに、「わかりませ
ん」「わからない」などの不明を意味する単語、「問わ
ない」「何でも良い」「任意」などの任意を意味する単
語なども認識語彙補足手段(3)で補足され得る単語と
する。図3に認識語彙の例を示しており、人名の認識語
彙を例に取り、補足前(a)と補足後(b)の認識語彙
を示している。The recognition vocabulary supplementing means (3) supplements the recognition vocabulary received from the dialog control means (5), especially words representing commands independent of tasks, and passes them to the speech recognition means. Here, the command is a word that means negation or cancellation, such as "no,""no,""no,""cancel," or "cancel," a word that means an advisory request, such as "help,""stop,""stop." There are words that mean a request to stop the processing of the system, such as "", "stop", and words, such as "again" and "repeat", that mean a request to re-output the immediately previous system response. Furthermore, words that can be supplemented by the recognized vocabulary supplementing means (3) include words that mean unknown, such as “I don't understand” or “I don't know”, and words that mean arbitrary, such as “I don't care”, “Anything is fine”, or “Any”. And FIG. 3 shows an example of the recognition vocabulary. The recognition vocabulary of the personal name is taken as an example, and the recognition vocabulary before (a) and after (b) the supplement is shown.
【0016】キーワード判定手段(6)は、音声認識手
段(2)より得られた認識結果がタスクから独立なコマ
ンドか、タスクに依存したキーワードかを判定し、その
結果を対話制御手段へ送る。ここで、判定結果は、例え
ば図4に示すように表現される。図4に示した例では、
1つの判定結果が、2つの値の組み合わせで表現され
る。左側が判定結果の種類を表しており、「CMD」は
コマンド、「KW」はキーワードを表す。 また、「C
MD」と組み合わされている「CANCEL」、「HE
LP」はコマンドの種類を表している。さらに「KW」
と組み合わされている“資材課”や“佐藤”は実データ
値を示している。The keyword determining means (6) determines whether the recognition result obtained from the voice recognition means (2) is a command independent of the task or a keyword dependent on the task, and sends the result to the dialog control means. Here, the determination result is expressed, for example, as shown in FIG. In the example shown in FIG.
One determination result is represented by a combination of two values. The left side shows the type of the determination result, “CMD” represents a command, and “KW” represents a keyword. Also, "C
"CANCEL", "HE" combined with "MD"
"LP" indicates the type of command. Furthermore, "KW"
“Materials section” and “Sato” combined with are showing actual data values.
【0017】キーワード保持手段(6)は、対話制御手
段の指示に従って、渡されたキーワードをスタック形式
で保持したり、スタックに積まれているキーワードを対
話制御手段に通知したりする。The keyword holding means (6) holds the passed keywords in a stack format or notifies the dialog control means of the keywords stored in the stack, according to the instruction of the dialog control means.
【0018】応答生成手段(7)は、対話制御手段
(5)からの指示に従い、タスクを遂行するために必要
な項目の内容(名前などの実データ値)を質問するよう
な応答文を生成する。The response generation means (7) generates a response sentence which inquires of the contents (actual data values such as names) of the items necessary for performing the task in accordance with the instruction from the interaction control means (5). I do.
【0019】音声合成手段(8)は、応答生成手段
(7)から得られる応答文を音声波形に変換してスピー
カなどの電気信号から音波へ変換するデバイスにより空
間中を伝播する音声(9)として出力する。The voice synthesizing means (8) converts the response sentence obtained from the response generating means (7) into a voice waveform and converts the electric signal into a sound wave by a device such as a speaker. Output as
【0020】図2は対話制御手段の処理フローを示す図
である。簡単のために、本発明による音声対話システム
でデータ入力のみが行われる場合のフローを示してい
る。実際のタスクにおいては、利用者に対する結果のみ
の提示なども含まれることになる。FIG. 2 is a diagram showing a processing flow of the dialog control means. For simplicity, the flow when only data input is performed in the voice interaction system according to the present invention is shown. In the actual task, presentation of only the result to the user is also included.
【0021】次に、図2のフローに従って、電話接続サ
ービスなどをタスクとして、所属名に‘資材課’、人名
に‘佐藤’を入力する場合を仮定して処理手順を説明す
る。Next, the processing procedure will be described in accordance with the flow of FIG. 2 on the assumption that a telephone connection service or the like is a task and "materials section" is input as the affiliation name and "Sato" is input as the personal name.
【0022】同タスク管理手段に対話の進行上における
次のアクションを促すガイダンス内容を要求し、まずサ
ービスが開始された時点では、対話制御手段(5)は、
タスク管理手段(10)へ対話の進行上における次の
アクションを促すガイダンス内容を要求する。この要求
に対する返答は、 次のガイダンス内容が「所属の質
問」である、として得られる。次に、対話制御手段
(5)は、キーワード保持手段から最新のキーワードを
受理しようとする。システム利用開始直後は、キーワー
ド保持手段はからであり、その旨が対話制御手段へ通知
される。次に、対話制御手段(5)は、次のガイダンス
内容(「所属の質問」)と、あれば最新のキーワード
(この段階では「なし」)を応答生成手段(7)に送
り、応答文を生成するように指示する。応答生成手段
(7)では、 ガイダンス内容が「所属の質問」である
ので、例えば“所属をどうぞ”というような質問文を生
成して、音声出力手段(8)に送る。続いて、対話制御
手段(5)は、「所属」を音声認識するための認識語彙
を、認識語彙補足手段(3)へ送る。ここでの認識語彙
は、先の、次のアクションを促すガイダンス内容を要求
した際に、タスク管理手段(10)より一緒に受け取る
こととする。あるいは、この段階であらためて、タスク
管理手段に問い合わせて、受理しても良い。この後、認
識語彙補足手段(3)で補足された認識語彙は、音声認
識手段(2)に送られる。そして、利用者の発声に対す
る音声認識手段(2)よる認識結果は、キーワード判定
手段(4)を介してコマンドかキーワードかの判定が付
加され、図4に示したような形式で対話制御手段(5)
に戻される。今の仮定では、ここでの受理結果は「K
W]+「資材課」である。すなわち、対話制御手段
(5)はキーワードをキーワード保持手段(6)のスタ
ックにキーワードをPOPし、さらに、タスク管理手段
(10)に対してキーワードを通知する。この時点で、
タスク管理手段(10)は、内部でタスクの進行状態が
更新される。The task management means is requested for guidance contents for prompting the next action in the progress of the dialog. First, when the service is started, the dialog control means (5)
It requests the task management means (10) for guidance contents for prompting the next action in the progress of the dialogue. The response to this request is obtained as the following guidance content is "affiliation question". Next, the dialog control means (5) tries to receive the latest keyword from the keyword holding means. Immediately after the start of use of the system, the keyword holding unit is empty, and the fact is notified to the dialog control unit. Next, the dialogue control means (5) sends the next guidance content ("Affiliation question") and the latest keyword ("None" at this stage) to the response generation means (7), and sends the response sentence. Instruct to generate. In the response generation means (7), since the guidance content is "question of affiliation", a question sentence such as "please belong" is generated and sent to the voice output means (8). Subsequently, the dialog control means (5) sends a recognized vocabulary for voice recognition of "belonging" to the recognized vocabulary supplementing means (3). The recognition vocabulary here is received together from the task management means (10) when the guidance content for prompting the next action is requested. Alternatively, an inquiry may be made to the task management means again at this stage to accept it. Thereafter, the recognized vocabulary supplemented by the recognized vocabulary supplementing means (3) is sent to the speech recognition means (2). Then, the recognition result of the user's utterance by the voice recognition means (2) is added with a determination as to whether it is a command or a keyword via the keyword determination means (4), and is performed in the form shown in FIG. 5)
Is returned to. According to the current assumption, the result of acceptance here is "K
W] + “Materials Section”. That is, the dialog control means (5) POPs the keyword on the stack of the keyword holding means (6) and further notifies the task management means (10) of the keyword. at this point,
The task management means (10) internally updates the task progress status.
【0023】ここで、図2のフローの最初に戻り、対話
制御手段(5)は、再びタスク管理手段(10)へ、対
話の進行上における次のアクションを促すガイダンス内
容を要求する。この要求に対する返答は、次のガイダン
ス内容が「名前の質問」であるである、として得られ
る。次に、対話制御手段(5)は、キーワード保持手段
から最新のキーワードを問い合わせ、先ほどPOPされ
たばかりの「資材課」が得られる。そして、対話管理手
段(5)は、次のガイダンス内容(「名前の質問」)と
最新のキーワード(「資材課」)を応答生成手段(7)
に送り、応答文を生成するように指示する。応答生成手
段(7)では、 「資材課」を含めて、かつ、「名前」
を問い合わせるような質問文、例えば“資材課で名前は
何ですか”とか“資材課の誰ですか”というような質問
文を生成して、音声出力手段(8)に送る。続いて、対
話制御手段(5)は、「名前」を音声認識するための認
識語彙を、認識語彙補足手段(3)へ送る。そしてま
た、音声認識手段(2)による認識結果は、キーワード
判定手段(4)を介してコマンドかキーワードかの判定
が付加され、図4に示したような形式で対話制御手段
(5)に戻される。今の仮定では、ここでの受理結果は
「KW]+「佐藤」となる。以上の繰り返しで、基本的
な対話が進行していく。Here, returning to the beginning of the flow of FIG. 2, the dialogue control means (5) requests the task management means (10) again for guidance contents for prompting the next action in the progress of the dialogue. The answer to this request is obtained as the next guidance content is "name question". Next, the dialogue control means (5) inquires the latest keyword from the keyword holding means, and obtains the "materials section" that has just been POP. Then, the dialogue management means (5) sends the next guidance content ("question of name") and the latest keyword ("materials section") to the response generation means (7).
And instruct it to generate a response sentence. In response generation means (7), include "material section" and "name"
, Such as "What is the name in the material section?" Or "Who is the material section?", And sends it to the voice output means (8). Subsequently, the interaction control means (5) sends a recognition vocabulary for voice recognition of the "name" to the recognition vocabulary supplementing means (3). Further, the result of recognition by the voice recognition means (2) is added with a judgment as to whether it is a command or a keyword through the keyword judgment means (4), and is returned to the dialogue control means (5) in a format as shown in FIG. It is. Under the current assumption, the result of acceptance here is “KW” + “Sato”. By repeating the above, the basic dialogue proceeds.
【0024】次に、利用者の「資材課」の発声が誤認識
されて、認識結果が「施設課」になったと仮定した場合
の例を説明する。先の“資材課の誰ですか”の代わりに
“施設課の誰ですか”というシステム出力になっている
はずであり、それに対して、利用者は“違います”と答
えたとする。キーワード判定手段(4)より受け取った
結果が「KW」+「佐藤」でなく、「CMD」+「CA
NCEL」となる。この場合、対話制御手段(5)はキ
ーワード保持手段(6)に対して、キーワードを1つ
(今の場合「資材課」)POPするように指示する。さ
らに、タスク管理手段(10)に対して、POPされた
「資材課」を取り消すように通知する。ここで、対話制
御手段(5)の処理は図2のフローの最初に戻り、タス
ク管理手段(10)、キーワード保持手段(6)への問
い合わせをし、次のアクションを促すガイダンス内容が
「所属の質問」であり、最新のキーワードは「なし」で
あることを応答生成手段(7)に送り、応答生成手段
(7)で再び、“所属をどうぞ”というような質問文が
生成される。Next, an example will be described in which it is assumed that the user's utterance of the "material section" is erroneously recognized and the recognition result is "the facility section". It is assumed that the system output is "Who is in the facility section" instead of "Who is in the material section", and the user answers "No" to the system output. The result received from the keyword determination means (4) is not "KW" + "Sato" but "CMD" + "CA
NCEL ". In this case, the dialogue control means (5) instructs the keyword holding means (6) to POP one keyword (in this case, “materials section”). Furthermore, the task management means (10) is notified to cancel the POP "material section". Here, the processing of the dialogue control means (5) returns to the beginning of the flow in FIG. 2, and inquires of the task management means (10) and the keyword holding means (6), and the guidance content for prompting the next action is "belonging Is sent to the response generating means (7), and the response generating means (7) again generates a question sentence such as "Please belong."
【0025】最後に、本発明による音声対話システムを
利用した場合の、対話例を示す。タスクを電話の接続サ
ービスとした場合の例である。Finally, an example of a dialogue when the voice dialogue system according to the present invention is used will be described. This is an example where the task is a telephone connection service.
【0026】 システム:「相手の所属をどうぞ」 利用者 :「資材課」 システム:「資材課の誰ですか」 利用者 :「佐藤」 システム:「佐藤さんに電話をつなぎます」 次に、1番目の項目で誤認識した場合の対話例を示す。System: “Please add the other party's affiliation” User: “Materials Section” System: “Who is in the Materials Section?” User: “Sato” System: “Call Sato-san” Here is an example of a dialogue in the case of misrecognition in the second item.
【0027】 システム:「相手の所属をどうぞ」 利用者 :「資材課」 システム:「施設課の誰ですか」 利用者 :「違います」 システム:「もう一度、相手の所属をどうぞ」 利用者 :「資材課」 システム:「資材課の誰ですか」 :(以下省略) 次に、2番目以降の項目で誤認識した場合の対話例を示
す。[0027] System: "Please go to the other party's affiliation" User: "Materials Section" System: "Who is the facility section?" User: "No," System: "Please try again to belong to the other party" User: "Materials Section" System: "Who is the Materials Section?": (Omitted hereafter) Next, an example of dialogue when misrecognition is made in the second and subsequent items is shown.
【0028】 :(途中まで省略) システム:「資材課の誰ですか」 利用者 :「佐藤」 システム:「加藤さんに電話をつなぎます」 利用者 :「違います」 システム:「もう一度、資材課の誰ですか」 利用者 :「佐藤」 システム:「佐藤さんに電話をつなぎます」 図1において、利用者とシステムとの間のメディアとし
て、音声のみしか描いていないが、文字、画像、など他
のメディアも含めた対話システムでも良い。また、ボタ
ン入力を用意し、取消などをボタン入力するようにして
も良い。遠隔地の電話機から利用している際には、取消
などにタッチトーン信号を用いても良い。: (Some part omitted) System: "Who is in the materials section?" User: "Sato" System: "Connects the phone to Mr. Kato" User: "Is different" System: "Again, Materials section User: "Sato" System: "Connect Sato-san" In Figure 1, only audio is drawn as media between the user and the system, but characters, images, etc. A dialog system including other media may be used. Alternatively, a button input may be prepared, and a button such as a cancel button may be input. When using from a remote telephone, a touch tone signal may be used for cancellation or the like.
【0029】図1におけるタスク管理手段はタスクに特
有の処理も含んでいる。例えば、遠隔地の電話機からシ
ステムを利用する形態の場合、電話回線の制御などが考
えられる。The task management means in FIG. 1 also includes processing specific to a task. For example, in the case of using the system from a remote telephone, control of a telephone line may be considered.
【0030】[0030]
【発明の効果】本発明によれば、複数項目のデータ入力
を音声で行う音声対話システムにおいて、音声認識を利
用する際に必須の確認に関して、確認のみのシステム出
力を省略し、認識結果を次のガイダンス文に含めて提示
することでやりとりの数を減少させることにより、 単
語音声認識程度の音声認識能力しか持たない音声対話シ
ステムでも、利用者とシステムとの間で効率良く音声対
話が進行できる効果が得られる。According to the present invention, in a voice dialogue system for inputting data of a plurality of items by voice, a system output of only verification is omitted for a required confirmation when using voice recognition, and the recognition result is set as follows. By reducing the number of conversations by presenting it in a guidance sentence, the user and the system can proceed efficiently with a dialogue even with a spoken dialogue system that has only a speech recognition capability equivalent to word speech recognition. The effect is obtained.
【図1】本発明による音声対話システムの構成の一実施
例を示すブロック図である。FIG. 1 is a block diagram showing one embodiment of the configuration of a voice dialogue system according to the present invention.
【図2】対話制御手段の処理手順を示すフローチャート
である。FIG. 2 is a flowchart illustrating a processing procedure of a dialog control unit.
【図3】認識語彙の一例を示す図である。FIG. 3 is a diagram showing an example of a recognized vocabulary.
【図4】キーワード判定手段から対話制御手段に渡され
るデータ構造の一例を示す図である。FIG. 4 is a diagram showing an example of a data structure passed from a keyword determination unit to a dialogue control unit.
【図5】キーワード保持手段に保持されるデータの一例
を示す図である。FIG. 5 is a diagram illustrating an example of data stored in a keyword storage unit.
1…利用者の音声、2…音声認識手段、3…認識語彙補
足手段、4…キーワード判定手段、5…対話制御手段、
6…キーワード保持手段、7…応答生成手段、8…音声
出力手段、9…システムの出力音声、10…タスク管理
手段。DESCRIPTION OF SYMBOLS 1 ... User's voice, 2 ... Voice recognition means, 3 ... Recognition vocabulary supplement means, 4 ... Keyword judgment means, 5 ... Dialogue control means,
6 ... Keyword holding means, 7 ... Response generation means, 8 ... Speech output means, 9 ... System output speech, 10 ... Task management means.
フロントページの続き (72)発明者 畑岡 信夫 東京都国分寺市東恋ケ窪一丁目280番地 株式会社日立製作所中央研究所内Continued on the front page (72) Inventor Nobuo Hataoka 1-280 Higashi Koigakubo, Kokubunji-shi, Tokyo Inside Central Research Laboratory, Hitachi, Ltd.
Claims (18)
上における次のアクションを促すガイダンス内容と、キ
ーワード保持手段が保持している前段階の認識結果であ
る最新のキーワードを用いて、 対話の進行上における
次のアクションを促すガイダンス文の中に、前記最新の
キーワードを含めた応答文を応答生成手段で生成し、出
力させると共に、認識語彙補足手段において、タスク管
理手段の管理している対話の進行上における次の場面で
の認識語彙に、タスクから独立なコマンドを表す単語を
補足し、その補足後の認識語彙の範囲内で、音声認識手
段において利用者の発話した音声を認識させ、キーワー
ド判定手段において、その認識結果が前記タスクから独
立なコマンドか、それ以外のタスクに依存したキーワー
ドかを判定し、その判定結果がコマンドである場合は、
各コマンドに対する処理を行い、さらに前記判定結果が
前記キーワードである場合は、該判定結果であるキーワ
ードを新たに最新のキーワードとしてキーワード保持手
段へ保持させると同時に、さらに該判定結果であるキー
ワードをタスク管理手段へ通知する、以上を繰り返すこ
とにより音声を進行させることを特徴とする音声対話制
御方法。A dialogue is provided by using guidance contents for prompting a next action in the progress of a dialogue managed by a task management means and the latest keyword which is a recognition result of a previous stage held by a keyword holding means. A response sentence including the latest keyword is generated and output by the response generating unit in the guidance sentence for prompting the next action in the progress of the process, and the task is managed by the task managing unit in the recognition vocabulary supplementing unit. A word representing a command independent from the task is supplemented to the recognition vocabulary in the next scene in the progress of the dialogue, and within the range of the recognition vocabulary after the supplementation, the speech recognition means recognizes the speech uttered by the user. The keyword determination means determines whether the recognition result is a command independent of the task or a keyword dependent on other tasks. If the judgment result is a command,
The process for each command is performed, and when the result of the determination is the keyword, the keyword that is the result of the determination is newly held in the keyword holding unit as the latest keyword. A voice dialogue control method, characterized in that the voice is advanced by notifying the management means and repeating the above.
いて、前記コマンドは、少なくとも「違います」「違
う」「いいえ」「取消」「キャンセル」などの否定ある
いは取消を意味する単語を含み、 前記判定結果が該否
定あるいは取消を意味する単語である場合に、前記キー
ワード保持手段に最新のキーワードを1つ削除するよう
に指示し、同時に前記タスク管理手段に1つ前の要求項
目の内容が取り消されたことを通知することを特徴とす
る音声対話制御方法。2. The method according to claim 1, wherein the command includes at least a word meaning negation or cancellation, such as "no,""no,""no,""cancel," or "cancel." If the result of the determination is a word meaning the negation or cancellation, the keyword holding means is instructed to delete one of the latest keywords, and at the same time, the content of the immediately preceding request item is instructed to the task management means. A voice dialogue control method, wherein a notification of cancellation is provided.
いて、前記コマンドは、少なくとも「ヘルプ」などの助
言要求を意味する単語を含み、 前記判定結果が該助言
要求を意味する単語である場合に、応答生成手段におい
て、使い方についてのメッセージ文を出力させることを
特徴とする音声対話制御方法。3. The voice dialogue control method according to claim 1, wherein the command includes at least a word meaning an advice request such as "help", and the determination result is a word meaning the advice request. And a response generation means for outputting a message about the usage.
いて、前記コマンドは、少なくとも「ストップ」「停
止」「中止」「中断」などのシステムの処理の停止要求
を意味する単語を含み、前記判定結果が該停止要求を意
味する単語である場合に、対話進行を停止することを特
徴とする音声対話制御方法。4. The voice interaction control method according to claim 1, wherein the command includes at least a word indicating a request to stop processing of the system such as “stop”, “stop”, “stop”, “stop”, and the like. A speech dialogue control method, wherein the dialogue progress is stopped when the determination result is a word meaning the stop request.
いて、 前記応答生成手段は1つ以上の応答文を蓄積
し、後で参照できる手段を持ち、前記コマンドは、少な
くとも「もう一度」「リピート」などの直前のシステム
応答の再出力要求を意味する単語を含み、前記判定結果
が該再出力要求を意味する単語である場合に、応答生成
手段において直前に出力した応答文の再出力させること
を特徴とする音声対話制御方法。5. The voice dialogue control method according to claim 1, wherein said response generation means has means for storing one or more response sentences, which can be referred to later, and wherein said command is at least “again” or “repeat”. If the determination result is a word meaning the re-output request including a word meaning a re-output request of the immediately preceding system response, the response generation means may re-output the response sent immediately before. A voice interaction control method characterized by the following.
いて、前記コマンドは、少なくとも「わかりません」
「わからない」などの不明を意味する単語を含み、前記
判定結果が該不明を意味する単語である場合に、前記タ
スク管理手段に1つ前の要求項目の内容が不明であるこ
とを通知し、同時に同タスク管理手段に対話の進行上に
おける次のアクションを促すガイダンス内容を要求し、
その返答内容に従って対話進行することを特徴とする音
声対話制御方法。6. The method according to claim 1, wherein the command is at least "I do not know."
Including a word meaning unknown such as "I don't know", and when the determination result is a word meaning unknown, notifies the task management means that the content of the immediately preceding request item is unknown, At the same time, request the task management means for guidance that prompts the next action in the progress of the dialogue,
A spoken dialogue control method characterized in that the dialogue proceeds according to the contents of the response.
いて、前記コマンドは、少なくとも「問わない」「何で
も良い」「任意」などの任意を意味する単語を含み、前
記判定結果が該任意を意味する単語である場合に、タス
ク管理手段に要求項目の内容が任意であることを通知
し、同時に同タスク管理手段に対話の進行上における次
のアクションを促すガイダンス内容を要求し、 その返
答内容に従って対話進行することを特徴とする音声対話
制御方法。7. The voice interaction control method according to claim 1, wherein said command includes at least a word meaning an arbitrary word such as "don't care", "anything", "any", and said judgment result indicates that said arbitrary word is arbitrary. If it is a meaningful word, notify the task management means that the content of the request item is optional, and at the same time, request the task management means for guidance content that prompts the next action in the progress of the dialogue, and the response content A dialogue control method characterized in that the dialogue proceeds according to the following.
し、対話の進行上における次のアクションを促すガイダ
ンス内容のように、タスクに依存した対話の進行状況に
関連した情報提供の要求を受け、その要求に対する情報
を返答するタスク管理手段と、前記タスク管理手段に要
求を出しその返答結果に応じて、認識語彙補足手段、キ
ーワード判定手段、キーワード保持手段、および応答生
成手段を制御し、システムと利用者の間の対話を進行さ
せる対話制御手段と、前記対話制御手段より受け取った
認識語彙に新たな単語を補足して、補足した結果を音声
認識手段へ渡す認識語彙補足手段と、該認識語彙補足手
段の出力により指定された認識語彙の範囲内で、利用者
の発話した音声を認識して一つまたは複数の単語系列を
出力する音声認識手段と、前記一つまたは複数の単語系
列を判定するキーワード判定手段と、前記対話制御手段
より渡されたキーワードを保持したり、保持しているキ
ーワードの中で最新のキーワードを削除、あるいは対話
制御手段に通知したりするキーワード保持手段と、前記
対話制御手段からの指示に従い、場合によっては前記対
話制御手段からのデータを用いて、応答文を生成する応
答生成手段と、該応答生成手段から得られる応答文を音
声波形に変換して出力する音声合成手段とから構成さ
れ、特に、前記対話制御手段は、タスク管理手段に対話
の進行上における次のアクションを促すガイダンス内容
を要求して該ガイダンス内容を受け取り、また前記キー
ワード保持手段に最新のキーワードを要求して該最新の
キーワードを受け取り、前記ガイダンス内容と前記最新
のキーワード、およびそれらを用いて応答文を生成させ
る指示を応答生成手段へ通知し、前記応答生成手段は、
前記対話制御手段から受け取った指示に従い、 対話の
進行上における次のアクションを促すガイダンス文の中
に、同時に受け取った前段階の認識結果でもある最新の
キーワードを含めた応答文を生成し、音声合成手段へ出
力し、次に前記対話制御手段は、タスク管理手段に対話
の進行上における次の認識語彙を要求して該認識語彙を
受け取り、タスクに依存したキーワードからなる該認識
語彙を認識語彙補足手段へ送り、前記認識語彙補足手段
は、前記対話制御手段から受け取った認識語彙に、タス
クから独立なコマンドを表す単語を補足して、前記音声
認識手段およびキーワード判定手段に渡し、前記キーワ
ード判定手段は、前記認識語彙補足手段より得た補足後
の認識語彙と前記音声認識手段より得た認識結果とを比
較し、該認識結果が前記タスクから独立なコマンドか、
タスクに依存したキーワードかを判定し、その判定結果
を対話制御手段へ送り、さらに次に前記対話制御手段
は、前記キーワード判定手段の判定結果を元に、該判定
結果がコマンドである場合は、各コマンドに対する処理
を行い、前記判定結果がキーワードである場合は、該判
定結果であるキーワードをキーワード保持手段に送ると
同時に、該キーワードをタスク管理手段に送ることを特
徴とする音声対話システム。8. A request for providing information related to the progress of a task-dependent dialogue, such as guidance content for managing the progress of a task-dependent dialogue and prompting the next action in the progress of the dialogue. A task management means for responding information to the request, a recognition vocabulary supplementing means, a keyword determination means, a keyword holding means, and a response generation means in response to a request issued to the task management means and the response result. Dialogue control means for promoting a dialogue between a user and a user; a recognition vocabulary supplementary means for supplementing a new word to the recognition vocabulary received from the dialogue control means and passing the supplemented result to a speech recognition means; A speech recognition method for recognizing a speech uttered by a user and outputting one or more word sequences within the range of the recognition vocabulary specified by the output of the vocabulary supplementing means. And a keyword determining unit for determining the one or more word sequences; holding a keyword passed from the dialog control unit; deleting a latest keyword from the held keywords; or a dialog control unit A response generating means for generating a response sentence using data from the dialog control means in accordance with an instruction from the dialog control means, and in some cases, obtained from the response generating means. And speech synthesis means for converting the response sentence into a speech waveform and outputting the speech. In particular, the dialogue control means requests the task management means for guidance content for prompting the next action in the progress of the dialogue. Requesting the latest keyword from the keyword holding means and receiving the latest keyword, Wherein the volume latest keywords, and notifies the response generation means an instruction to generate a response sentence by using them, the response generation means,
According to the instruction received from the interaction control means, a response sentence including the latest keyword, which is also a recognition result of the previous stage received at the same time, is generated in the guidance sentence for prompting the next action in the progress of the dialogue, and speech synthesis is performed. The dialogue control means requests the task management means for the next recognition vocabulary in the course of the dialogue, receives the recognition vocabulary, and supplements the recognition vocabulary consisting of task-dependent keywords with the recognition vocabulary supplement. Means, the recognition vocabulary supplementing means supplements the recognition vocabulary received from the dialog control means with a word representing a command independent of a task, passes the word to the speech recognition means and the keyword determination means, and passes the word to the keyword recognition means. Compares the supplemented recognition vocabulary obtained by the recognition vocabulary supplementing means with the recognition result obtained by the speech recognition means, and the recognition result is A command independent of the task,
It is determined whether the keyword depends on the task, the determination result is sent to the dialog control means, and further, the dialog control means, based on the determination result of the keyword determination means, if the determination result is a command, A voice dialogue system which performs a process for each command and, when the determination result is a keyword, sends the keyword as the determination result to the keyword holding means and simultaneously sends the keyword to the task management means.
いて、前記コマンドは、少なくとも「違います」「違
う」「いいえ」「取消」「キャンセル」などの否定ある
いは取消を意味する単語を含み、 前記判定結果が該否
定あるいは取消を意味する単語である場合に、前記対話
制御手段は前記キーワード保持手段に最新のキーワード
を1つ削除するように指示し、同時に前記タスク管理手
段に1つ前の要求項目の内容が取り消されたことを通知
することを特徴とする音声対話システム。9. The speech dialogue system according to claim 2, wherein the command includes at least a word meaning negation or cancellation such as "no,""no,""no,""cancel," or "cancel." If the determination result is a word meaning the negation or cancellation, the dialogue control means instructs the keyword holding means to delete one newest keyword, and at the same time, issues a request to the task management means. A voice dialogue system for notifying that the content of an item has been canceled.
おいて、前記コマンドは、少なくとも「ヘルプ」などの
助言要求を意味する単語を含み、 前記キーワード判定
手段の判定結果が該助言要求を意味する単語である場合
に、前記対話制御手段は応答生成手段にヘルプ応答を指
示し、該指示に従い前記応答生成手段は使い方などのメ
ッセージ文を出力することを特徴とする音声対話システ
ム。10. The voice interaction system according to claim 2, wherein said command includes at least a word meaning an advice request such as "help", and the result of judgment by said keyword judgment means is a word meaning the advice request. Wherein the dialogue control means instructs the response generation means to provide a help response, and the response generation means outputs a message such as usage in accordance with the instruction.
おいて、前記コマンドは、少なくとも「ストップ」「停
止」「中止」「中断」などのシステムの処理の停止要求
を意味する単語を含み、前記キーワード判定手段の判定
結果が該停止要求を意味する単語である場合に、前記対
話制御手段は対話進行を停止することを特徴とする音声
対話システム。11. The voice interaction system according to claim 2, wherein the command includes at least a word indicating a request to stop processing of the system, such as "stop", "stop", "stop", "stop", and the keyword. If the result of the determination by the determination means is a word meaning the stop request, the dialogue control means stops the dialogue progress.
おいて、前記応答生成手段は1つ以上の応答文を蓄積
し、後で参照できる手段を持ち、前記コマンドは、少な
くとも「もう一度」「リピート」などの直前のシステム
応答の再出力要求を意味する単語を含み、前記キーワー
ド判定手段の判定結果が該再出力要求を意味する単語で
ある場合に、前記対話制御手段は応答生成手段に直前に
出力した応答文の再出力する指示を出すことを特徴とす
る音声対話システム。12. The speech dialogue system according to claim 2, wherein said response generating means has means for storing one or more response sentences, which can be referred to later, and wherein said command is at least "again" or "repeat". When the result of the keyword determination means is a word meaning the re-output request, the dialog control means outputs the word immediately before to the response generation means. A spoken dialogue system that issues an instruction to re-output a response sentence.
おいて、前記コマンドは、少なくとも「わかりません」
「わからない」などの不明を意味する単語を含み、前記
キーワード判定手段の判定結果が該不明を意味する単語
である場合に、前記対話制御手段は前記タスク管理手段
に1つ前の要求項目の内容が不明であることを通知し、
同時に同タスク管理手段に対話の進行上における次のア
クションを促すガイダンス内容を要求し、その返答内容
に従って対話進行することを特徴とする音声対話システ
ム。13. The voice interaction system according to claim 2, wherein said command is at least "I do not understand".
If the keyword includes a word meaning unknown, such as "I don't know," and the result of the determination by the keyword determination means is a word meaning the unknown, the dialog control means sends the content of the immediately preceding request item to the task management means. Notify that is unknown,
At the same time, a voice dialogue system which requests the task management means for guidance content for prompting the next action in the progress of the dialogue, and proceeds with the dialogue according to the response content.
おいて、前記コマンドは、少なくとも「問わない」「何
でも良い」「任意」などの任意を意味する単語を含み、
前記キーワード判定手段の判定結果が該任意を意味する
単語である場合に、前記対話制御手段は、タスク管理手
段に要求項目の内容が任意であることを通知し、同時に
同タスク管理手段に対話の進行上における次のアクショ
ンを促すガイダンス内容を要求し、 その返答内容に従
って対話進行することを特徴とする音声対話システム。14. The voice dialogue system according to claim 2, wherein the command includes at least a word meaning an arbitrary word such as “don't care”, “anything”, “any”,
When the result of the determination by the keyword determination means is a word meaning the arbitraryness, the dialogue control means notifies the task management means that the content of the request item is optional, and at the same time, notifies the task management means of the dialogue. A spoken dialogue system that requests guidance content for prompting the next action in the process, and performs dialogue according to the response content.
3、および14記載の前記タスクの内容が電話番号の検
索サービス、あるいは電話回線の接続サービスである音
声対話システム。15. The method of claim 8, 9, 10, 11, 12, 1.
15. A voice interaction system in which the contents of the tasks described in 3 and 14 are a telephone number search service or a telephone line connection service.
3、および14記載の前記タスクの内容が交通機関、催
し物、あるいは施設の検索サービス、あるいは予約サー
ビスである音声対話システム。16. The method of claim 8, 9, 10, 11, 12, or 1.
15. A spoken dialogue system in which the contents of the tasks described in 3 and 14 are a transportation service, an entertainment or facility search service, or a reservation service.
3、および14記載の前記タスクの内容が商品の検索サ
ービス、販売サービス、あるいはサポートサービスであ
る音声対話システム。17. The method of claim 8, 9, 10, 11, 12, 1.
15. The voice interaction system according to items 3 and 14, wherein the content of the task is a product search service, a sales service, or a support service.
3、および14記載の前記タスクの内容が経路情報提供
サービス、あるいは経路誘導サービスである音声対話シ
ステム。18. The method of claim 8, 9, 10, 11, 12, 1.
15. A voice interaction system in which the contents of the tasks described in 3 and 14 are a route information providing service or a route guidance service.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP18906096A JP3700266B2 (en) | 1996-07-18 | 1996-07-18 | Spoken dialogue control method and spoken dialogue system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP18906096A JP3700266B2 (en) | 1996-07-18 | 1996-07-18 | Spoken dialogue control method and spoken dialogue system |
Publications (2)
Publication Number | Publication Date |
---|---|
JPH1031497A true JPH1031497A (en) | 1998-02-03 |
JP3700266B2 JP3700266B2 (en) | 2005-09-28 |
Family
ID=16234630
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
JP18906096A Expired - Fee Related JP3700266B2 (en) | 1996-07-18 | 1996-07-18 | Spoken dialogue control method and spoken dialogue system |
Country Status (1)
Country | Link |
---|---|
JP (1) | JP3700266B2 (en) |
Cited By (96)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH11126089A (en) * | 1997-10-24 | 1999-05-11 | Nissan Motor Co Ltd | Voice interaction device |
JP2000020089A (en) * | 1998-07-07 | 2000-01-21 | Matsushita Electric Ind Co Ltd | Speed recognition method and apparatus therefor as well as voice control system |
JP2002073080A (en) * | 2000-09-01 | 2002-03-12 | Fujitsu Ten Ltd | Voice interactive system |
US6397188B1 (en) | 1998-07-29 | 2002-05-28 | Nec Corporation | Natural language dialogue system automatically continuing conversation on behalf of a user who does not respond |
JP2003091297A (en) * | 2001-09-19 | 2003-03-28 | Matsushita Electric Ind Co Ltd | Voice interaction device |
JP2006284677A (en) * | 2005-03-31 | 2006-10-19 | Clarion Co Ltd | Voice guiding device, and control method and control program for voice guiding device |
JP2007047488A (en) * | 2005-08-10 | 2007-02-22 | Nippon Telegr & Teleph Corp <Ntt> | Interactive method, interactive device, interactive program, and recording medium |
KR101511831B1 (en) * | 2010-01-18 | 2015-04-14 | 애플 인크. | Active input elicitation by intelligent automated assistant |
US9330720B2 (en) | 2008-01-03 | 2016-05-03 | Apple Inc. | Methods and apparatus for altering audio output signals |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US9495129B2 (en) | 2012-06-29 | 2016-11-15 | Apple Inc. | Device, method, and user interface for voice-activated navigation and browsing of a document |
US9535906B2 (en) | 2008-07-31 | 2017-01-03 | Apple Inc. | Mobile device having human language translation capability with positional feedback |
US9582608B2 (en) | 2013-06-07 | 2017-02-28 | Apple Inc. | Unified ranking with entropy-weighted information for phrase-based semantic auto-completion |
JP2017058545A (en) * | 2015-09-17 | 2017-03-23 | 本田技研工業株式会社 | Voice processing device and voice processing method |
US9620104B2 (en) | 2013-06-07 | 2017-04-11 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
US9626955B2 (en) | 2008-04-05 | 2017-04-18 | Apple Inc. | Intelligent text-to-speech conversion |
US9633674B2 (en) | 2013-06-07 | 2017-04-25 | Apple Inc. | System and method for detecting errors in interactions with a voice-based digital assistant |
US9633660B2 (en) | 2010-02-25 | 2017-04-25 | Apple Inc. | User profiling for voice input processing |
US9646609B2 (en) | 2014-09-30 | 2017-05-09 | Apple Inc. | Caching apparatus for serving phonetic pronunciations |
US9646614B2 (en) | 2000-03-16 | 2017-05-09 | Apple Inc. | Fast, language-independent method for user authentication by voice |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US9697820B2 (en) | 2015-09-24 | 2017-07-04 | Apple Inc. | Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US9760559B2 (en) | 2014-05-30 | 2017-09-12 | Apple Inc. | Predictive text input |
US9785630B2 (en) | 2014-05-30 | 2017-10-10 | Apple Inc. | Text prediction using combined word N-gram and unigram language models |
US9798393B2 (en) | 2011-08-29 | 2017-10-24 | Apple Inc. | Text correction processing |
US9818400B2 (en) | 2014-09-11 | 2017-11-14 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US9842101B2 (en) | 2014-05-30 | 2017-12-12 | Apple Inc. | Predictive conversion of language input |
US9842105B2 (en) | 2015-04-16 | 2017-12-12 | Apple Inc. | Parsimonious continuous-space phrase representations for natural language processing |
US9858925B2 (en) | 2009-06-05 | 2018-01-02 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US9865280B2 (en) | 2015-03-06 | 2018-01-09 | Apple Inc. | Structured dictation using intelligent automated assistants |
US9886432B2 (en) | 2014-09-30 | 2018-02-06 | Apple Inc. | Parsimonious handling of word inflection via categorical stem + suffix N-gram language models |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US9899019B2 (en) | 2015-03-18 | 2018-02-20 | Apple Inc. | Systems and methods for structured stem and suffix language models |
US9934775B2 (en) | 2016-05-26 | 2018-04-03 | Apple Inc. | Unit-selection text-to-speech synthesis based on predicted concatenation parameters |
US9953088B2 (en) | 2012-05-14 | 2018-04-24 | Apple Inc. | Crowd sourcing information to fulfill user requests |
US9966065B2 (en) | 2014-05-30 | 2018-05-08 | Apple Inc. | Multi-command single utterance input method |
US9966068B2 (en) | 2013-06-08 | 2018-05-08 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
US9971774B2 (en) | 2012-09-19 | 2018-05-15 | Apple Inc. | Voice-based media searching |
US9972304B2 (en) | 2016-06-03 | 2018-05-15 | Apple Inc. | Privacy preserving distributed evaluation framework for embedded personalized systems |
US10049663B2 (en) | 2016-06-08 | 2018-08-14 | Apple, Inc. | Intelligent automated assistant for media exploration |
US10049668B2 (en) | 2015-12-02 | 2018-08-14 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10057736B2 (en) | 2011-06-03 | 2018-08-21 | Apple Inc. | Active transport based notifications |
US10067938B2 (en) | 2016-06-10 | 2018-09-04 | Apple Inc. | Multilingual word prediction |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US10079014B2 (en) | 2012-06-08 | 2018-09-18 | Apple Inc. | Name recognition system |
US10078631B2 (en) | 2014-05-30 | 2018-09-18 | Apple Inc. | Entropy-guided text prediction using combined word and character n-gram language models |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US10089072B2 (en) | 2016-06-11 | 2018-10-02 | Apple Inc. | Intelligent device arbitration and control |
US10102359B2 (en) | 2011-03-21 | 2018-10-16 | Apple Inc. | Device access using voice authentication |
US10101822B2 (en) | 2015-06-05 | 2018-10-16 | Apple Inc. | Language input correction |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US10127220B2 (en) | 2015-06-04 | 2018-11-13 | Apple Inc. | Language identification from short strings |
US10169329B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Exemplar-based natural language processing |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
US10186254B2 (en) | 2015-06-07 | 2019-01-22 | Apple Inc. | Context-based endpoint detection |
US10185542B2 (en) | 2013-06-09 | 2019-01-22 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US10192552B2 (en) | 2016-06-10 | 2019-01-29 | Apple Inc. | Digital assistant providing whispered speech |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10241752B2 (en) | 2011-09-30 | 2019-03-26 | Apple Inc. | Interface for a virtual digital assistant |
US10241644B2 (en) | 2011-06-03 | 2019-03-26 | Apple Inc. | Actionable reminder entries |
US10249300B2 (en) | 2016-06-06 | 2019-04-02 | Apple Inc. | Intelligent list reading |
US10255907B2 (en) | 2015-06-07 | 2019-04-09 | Apple Inc. | Automatic accent detection using acoustic models |
US10269345B2 (en) | 2016-06-11 | 2019-04-23 | Apple Inc. | Intelligent task discovery |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US10283110B2 (en) | 2009-07-02 | 2019-05-07 | Apple Inc. | Methods and apparatuses for automatic speech recognition |
US10297253B2 (en) | 2016-06-11 | 2019-05-21 | Apple Inc. | Application integration with a digital assistant |
US10318871B2 (en) | 2005-09-08 | 2019-06-11 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US10354011B2 (en) | 2016-06-09 | 2019-07-16 | Apple Inc. | Intelligent automated assistant in a home environment |
US10366158B2 (en) | 2015-09-29 | 2019-07-30 | Apple Inc. | Efficient word encoding for recurrent neural network language models |
US10446143B2 (en) | 2016-03-14 | 2019-10-15 | Apple Inc. | Identification of voice inputs providing credentials |
US10490187B2 (en) | 2016-06-10 | 2019-11-26 | Apple Inc. | Digital assistant providing automated status report |
US10496753B2 (en) | 2010-01-18 | 2019-12-03 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US10509862B2 (en) | 2016-06-10 | 2019-12-17 | Apple Inc. | Dynamic phrase expansion of language input |
US10521466B2 (en) | 2016-06-11 | 2019-12-31 | Apple Inc. | Data driven natural language event detection and classification |
US10552013B2 (en) | 2014-12-02 | 2020-02-04 | Apple Inc. | Data detection |
US10553209B2 (en) | 2010-01-18 | 2020-02-04 | Apple Inc. | Systems and methods for hands-free notification summaries |
US10567477B2 (en) | 2015-03-08 | 2020-02-18 | Apple Inc. | Virtual assistant continuity |
US10568032B2 (en) | 2007-04-03 | 2020-02-18 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US10593346B2 (en) | 2016-12-22 | 2020-03-17 | Apple Inc. | Rank-reduced token representation for automatic speech recognition |
US10607141B2 (en) | 2010-01-25 | 2020-03-31 | Newvaluexchange Ltd. | Apparatuses, methods and systems for a digital conversation management platform |
US10659851B2 (en) | 2014-06-30 | 2020-05-19 | Apple Inc. | Real-time digital assistant knowledge updates |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US10679605B2 (en) | 2010-01-18 | 2020-06-09 | Apple Inc. | Hands-free list-reading by intelligent automated assistant |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10705794B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US10706373B2 (en) | 2011-06-03 | 2020-07-07 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US10733993B2 (en) | 2016-06-10 | 2020-08-04 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US10789041B2 (en) | 2014-09-12 | 2020-09-29 | Apple Inc. | Dynamic thresholds for always listening speech trigger |
US10791176B2 (en) | 2017-05-12 | 2020-09-29 | Apple Inc. | Synchronization and task delegation of a digital assistant |
US10810274B2 (en) | 2017-05-15 | 2020-10-20 | Apple Inc. | Optimizing dialogue policy decisions for digital assistants using implicit feedback |
US11010550B2 (en) | 2015-09-29 | 2021-05-18 | Apple Inc. | Unified language modeling framework for word prediction, auto-completion and auto-correction |
US11025565B2 (en) | 2015-06-07 | 2021-06-01 | Apple Inc. | Personalized prediction of responses for instant messaging |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
-
1996
- 1996-07-18 JP JP18906096A patent/JP3700266B2/en not_active Expired - Fee Related
Cited By (133)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH11126089A (en) * | 1997-10-24 | 1999-05-11 | Nissan Motor Co Ltd | Voice interaction device |
JP2000020089A (en) * | 1998-07-07 | 2000-01-21 | Matsushita Electric Ind Co Ltd | Speed recognition method and apparatus therefor as well as voice control system |
US6397188B1 (en) | 1998-07-29 | 2002-05-28 | Nec Corporation | Natural language dialogue system automatically continuing conversation on behalf of a user who does not respond |
US9646614B2 (en) | 2000-03-16 | 2017-05-09 | Apple Inc. | Fast, language-independent method for user authentication by voice |
JP2002073080A (en) * | 2000-09-01 | 2002-03-12 | Fujitsu Ten Ltd | Voice interactive system |
JP2003091297A (en) * | 2001-09-19 | 2003-03-28 | Matsushita Electric Ind Co Ltd | Voice interaction device |
JP2006284677A (en) * | 2005-03-31 | 2006-10-19 | Clarion Co Ltd | Voice guiding device, and control method and control program for voice guiding device |
JP4705398B2 (en) * | 2005-03-31 | 2011-06-22 | クラリオン株式会社 | Voice guidance device, control method and program for voice guidance device |
JP2007047488A (en) * | 2005-08-10 | 2007-02-22 | Nippon Telegr & Teleph Corp <Ntt> | Interactive method, interactive device, interactive program, and recording medium |
JP4589843B2 (en) * | 2005-08-10 | 2010-12-01 | 日本電信電話株式会社 | Dialogue method, dialogue apparatus, dialogue program, and recording medium |
US10318871B2 (en) | 2005-09-08 | 2019-06-11 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US10568032B2 (en) | 2007-04-03 | 2020-02-18 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US9330720B2 (en) | 2008-01-03 | 2016-05-03 | Apple Inc. | Methods and apparatus for altering audio output signals |
US10381016B2 (en) | 2008-01-03 | 2019-08-13 | Apple Inc. | Methods and apparatus for altering audio output signals |
US9865248B2 (en) | 2008-04-05 | 2018-01-09 | Apple Inc. | Intelligent text-to-speech conversion |
US9626955B2 (en) | 2008-04-05 | 2017-04-18 | Apple Inc. | Intelligent text-to-speech conversion |
US9535906B2 (en) | 2008-07-31 | 2017-01-03 | Apple Inc. | Mobile device having human language translation capability with positional feedback |
US10108612B2 (en) | 2008-07-31 | 2018-10-23 | Apple Inc. | Mobile device having human language translation capability with positional feedback |
US10795541B2 (en) | 2009-06-05 | 2020-10-06 | Apple Inc. | Intelligent organization of tasks items |
US9858925B2 (en) | 2009-06-05 | 2018-01-02 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US11080012B2 (en) | 2009-06-05 | 2021-08-03 | Apple Inc. | Interface for a virtual digital assistant |
US10475446B2 (en) | 2009-06-05 | 2019-11-12 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US10283110B2 (en) | 2009-07-02 | 2019-05-07 | Apple Inc. | Methods and apparatuses for automatic speech recognition |
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
US10706841B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Task flow identification based on user intent |
US10496753B2 (en) | 2010-01-18 | 2019-12-03 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US11423886B2 (en) | 2010-01-18 | 2022-08-23 | Apple Inc. | Task flow identification based on user intent |
US10553209B2 (en) | 2010-01-18 | 2020-02-04 | Apple Inc. | Systems and methods for hands-free notification summaries |
US10705794B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US9548050B2 (en) | 2010-01-18 | 2017-01-17 | Apple Inc. | Intelligent automated assistant |
US10679605B2 (en) | 2010-01-18 | 2020-06-09 | Apple Inc. | Hands-free list-reading by intelligent automated assistant |
KR101511831B1 (en) * | 2010-01-18 | 2015-04-14 | 애플 인크. | Active input elicitation by intelligent automated assistant |
US10741185B2 (en) | 2010-01-18 | 2020-08-11 | Apple Inc. | Intelligent automated assistant |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US12087308B2 (en) | 2010-01-18 | 2024-09-10 | Apple Inc. | Intelligent automated assistant |
US10984327B2 (en) | 2010-01-25 | 2021-04-20 | New Valuexchange Ltd. | Apparatuses, methods and systems for a digital conversation management platform |
US10984326B2 (en) | 2010-01-25 | 2021-04-20 | Newvaluexchange Ltd. | Apparatuses, methods and systems for a digital conversation management platform |
US10607140B2 (en) | 2010-01-25 | 2020-03-31 | Newvaluexchange Ltd. | Apparatuses, methods and systems for a digital conversation management platform |
US10607141B2 (en) | 2010-01-25 | 2020-03-31 | Newvaluexchange Ltd. | Apparatuses, methods and systems for a digital conversation management platform |
US11410053B2 (en) | 2010-01-25 | 2022-08-09 | Newvaluexchange Ltd. | Apparatuses, methods and systems for a digital conversation management platform |
US10049675B2 (en) | 2010-02-25 | 2018-08-14 | Apple Inc. | User profiling for voice input processing |
US9633660B2 (en) | 2010-02-25 | 2017-04-25 | Apple Inc. | User profiling for voice input processing |
US10102359B2 (en) | 2011-03-21 | 2018-10-16 | Apple Inc. | Device access using voice authentication |
US10241644B2 (en) | 2011-06-03 | 2019-03-26 | Apple Inc. | Actionable reminder entries |
US10706373B2 (en) | 2011-06-03 | 2020-07-07 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US11120372B2 (en) | 2011-06-03 | 2021-09-14 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US10057736B2 (en) | 2011-06-03 | 2018-08-21 | Apple Inc. | Active transport based notifications |
US9798393B2 (en) | 2011-08-29 | 2017-10-24 | Apple Inc. | Text correction processing |
US10241752B2 (en) | 2011-09-30 | 2019-03-26 | Apple Inc. | Interface for a virtual digital assistant |
US9953088B2 (en) | 2012-05-14 | 2018-04-24 | Apple Inc. | Crowd sourcing information to fulfill user requests |
US10079014B2 (en) | 2012-06-08 | 2018-09-18 | Apple Inc. | Name recognition system |
US9495129B2 (en) | 2012-06-29 | 2016-11-15 | Apple Inc. | Device, method, and user interface for voice-activated navigation and browsing of a document |
US9971774B2 (en) | 2012-09-19 | 2018-05-15 | Apple Inc. | Voice-based media searching |
US9633674B2 (en) | 2013-06-07 | 2017-04-25 | Apple Inc. | System and method for detecting errors in interactions with a voice-based digital assistant |
US9620104B2 (en) | 2013-06-07 | 2017-04-11 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
US9966060B2 (en) | 2013-06-07 | 2018-05-08 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
US9582608B2 (en) | 2013-06-07 | 2017-02-28 | Apple Inc. | Unified ranking with entropy-weighted information for phrase-based semantic auto-completion |
US9966068B2 (en) | 2013-06-08 | 2018-05-08 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
US10657961B2 (en) | 2013-06-08 | 2020-05-19 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
US10185542B2 (en) | 2013-06-09 | 2019-01-22 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US9785630B2 (en) | 2014-05-30 | 2017-10-10 | Apple Inc. | Text prediction using combined word N-gram and unigram language models |
US9842101B2 (en) | 2014-05-30 | 2017-12-12 | Apple Inc. | Predictive conversion of language input |
US11133008B2 (en) | 2014-05-30 | 2021-09-28 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US10497365B2 (en) | 2014-05-30 | 2019-12-03 | Apple Inc. | Multi-command single utterance input method |
US10169329B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Exemplar-based natural language processing |
US10078631B2 (en) | 2014-05-30 | 2018-09-18 | Apple Inc. | Entropy-guided text prediction using combined word and character n-gram language models |
US9760559B2 (en) | 2014-05-30 | 2017-09-12 | Apple Inc. | Predictive text input |
US9966065B2 (en) | 2014-05-30 | 2018-05-08 | Apple Inc. | Multi-command single utterance input method |
US10659851B2 (en) | 2014-06-30 | 2020-05-19 | Apple Inc. | Real-time digital assistant knowledge updates |
US9668024B2 (en) | 2014-06-30 | 2017-05-30 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10904611B2 (en) | 2014-06-30 | 2021-01-26 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10431204B2 (en) | 2014-09-11 | 2019-10-01 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US9818400B2 (en) | 2014-09-11 | 2017-11-14 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US10789041B2 (en) | 2014-09-12 | 2020-09-29 | Apple Inc. | Dynamic thresholds for always listening speech trigger |
US9986419B2 (en) | 2014-09-30 | 2018-05-29 | Apple Inc. | Social reminders |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US9886432B2 (en) | 2014-09-30 | 2018-02-06 | Apple Inc. | Parsimonious handling of word inflection via categorical stem + suffix N-gram language models |
US9646609B2 (en) | 2014-09-30 | 2017-05-09 | Apple Inc. | Caching apparatus for serving phonetic pronunciations |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US11556230B2 (en) | 2014-12-02 | 2023-01-17 | Apple Inc. | Data detection |
US10552013B2 (en) | 2014-12-02 | 2020-02-04 | Apple Inc. | Data detection |
US9865280B2 (en) | 2015-03-06 | 2018-01-09 | Apple Inc. | Structured dictation using intelligent automated assistants |
US11087759B2 (en) | 2015-03-08 | 2021-08-10 | Apple Inc. | Virtual assistant activation |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US10311871B2 (en) | 2015-03-08 | 2019-06-04 | Apple Inc. | Competing devices responding to voice triggers |
US10567477B2 (en) | 2015-03-08 | 2020-02-18 | Apple Inc. | Virtual assistant continuity |
US9899019B2 (en) | 2015-03-18 | 2018-02-20 | Apple Inc. | Systems and methods for structured stem and suffix language models |
US9842105B2 (en) | 2015-04-16 | 2017-12-12 | Apple Inc. | Parsimonious continuous-space phrase representations for natural language processing |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US10127220B2 (en) | 2015-06-04 | 2018-11-13 | Apple Inc. | Language identification from short strings |
US10101822B2 (en) | 2015-06-05 | 2018-10-16 | Apple Inc. | Language input correction |
US11025565B2 (en) | 2015-06-07 | 2021-06-01 | Apple Inc. | Personalized prediction of responses for instant messaging |
US10186254B2 (en) | 2015-06-07 | 2019-01-22 | Apple Inc. | Context-based endpoint detection |
US10255907B2 (en) | 2015-06-07 | 2019-04-09 | Apple Inc. | Automatic accent detection using acoustic models |
US11500672B2 (en) | 2015-09-08 | 2022-11-15 | Apple Inc. | Distributed personal assistant |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
JP2017058545A (en) * | 2015-09-17 | 2017-03-23 | 本田技研工業株式会社 | Voice processing device and voice processing method |
US9697820B2 (en) | 2015-09-24 | 2017-07-04 | Apple Inc. | Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks |
US11010550B2 (en) | 2015-09-29 | 2021-05-18 | Apple Inc. | Unified language modeling framework for word prediction, auto-completion and auto-correction |
US10366158B2 (en) | 2015-09-29 | 2019-07-30 | Apple Inc. | Efficient word encoding for recurrent neural network language models |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US11526368B2 (en) | 2015-11-06 | 2022-12-13 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10049668B2 (en) | 2015-12-02 | 2018-08-14 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10446143B2 (en) | 2016-03-14 | 2019-10-15 | Apple Inc. | Identification of voice inputs providing credentials |
US9934775B2 (en) | 2016-05-26 | 2018-04-03 | Apple Inc. | Unit-selection text-to-speech synthesis based on predicted concatenation parameters |
US9972304B2 (en) | 2016-06-03 | 2018-05-15 | Apple Inc. | Privacy preserving distributed evaluation framework for embedded personalized systems |
US10249300B2 (en) | 2016-06-06 | 2019-04-02 | Apple Inc. | Intelligent list reading |
US10049663B2 (en) | 2016-06-08 | 2018-08-14 | Apple, Inc. | Intelligent automated assistant for media exploration |
US11069347B2 (en) | 2016-06-08 | 2021-07-20 | Apple Inc. | Intelligent automated assistant for media exploration |
US10354011B2 (en) | 2016-06-09 | 2019-07-16 | Apple Inc. | Intelligent automated assistant in a home environment |
US10733993B2 (en) | 2016-06-10 | 2020-08-04 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US10067938B2 (en) | 2016-06-10 | 2018-09-04 | Apple Inc. | Multilingual word prediction |
US11037565B2 (en) | 2016-06-10 | 2021-06-15 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US10192552B2 (en) | 2016-06-10 | 2019-01-29 | Apple Inc. | Digital assistant providing whispered speech |
US10509862B2 (en) | 2016-06-10 | 2019-12-17 | Apple Inc. | Dynamic phrase expansion of language input |
US10490187B2 (en) | 2016-06-10 | 2019-11-26 | Apple Inc. | Digital assistant providing automated status report |
US11152002B2 (en) | 2016-06-11 | 2021-10-19 | Apple Inc. | Application integration with a digital assistant |
US10089072B2 (en) | 2016-06-11 | 2018-10-02 | Apple Inc. | Intelligent device arbitration and control |
US10297253B2 (en) | 2016-06-11 | 2019-05-21 | Apple Inc. | Application integration with a digital assistant |
US10521466B2 (en) | 2016-06-11 | 2019-12-31 | Apple Inc. | Data driven natural language event detection and classification |
US10269345B2 (en) | 2016-06-11 | 2019-04-23 | Apple Inc. | Intelligent task discovery |
US10593346B2 (en) | 2016-12-22 | 2020-03-17 | Apple Inc. | Rank-reduced token representation for automatic speech recognition |
US11405466B2 (en) | 2017-05-12 | 2022-08-02 | Apple Inc. | Synchronization and task delegation of a digital assistant |
US10791176B2 (en) | 2017-05-12 | 2020-09-29 | Apple Inc. | Synchronization and task delegation of a digital assistant |
US10810274B2 (en) | 2017-05-15 | 2020-10-20 | Apple Inc. | Optimizing dialogue policy decisions for digital assistants using implicit feedback |
Also Published As
Publication number | Publication date |
---|---|
JP3700266B2 (en) | 2005-09-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JPH1031497A (en) | Voice conversation control method and voice conversation system | |
JP3454897B2 (en) | Spoken dialogue system | |
US8064573B2 (en) | Computer generated prompting | |
US7747438B2 (en) | Multi-slot dialog systems and methods | |
US9015048B2 (en) | Incremental speech recognition for dialog systems | |
US20210193116A1 (en) | Data driven dialog management | |
US20120253823A1 (en) | Hybrid Dialog Speech Recognition for In-Vehicle Automated Interaction and In-Vehicle Interfaces Requiring Minimal Driver Processing | |
WO2005064592A1 (en) | Device control device, speech recognition device, agent device, on-vehicle device control device, navigation device, audio device, device control method, speech recognition method, agent processing method, on-vehicle device control method, navigation method, and audio device control method, and program | |
JP2001005488A (en) | Voice interactive system | |
JP2007529831A (en) | Voice interactive messaging method and apparatus | |
CN109036406A (en) | A kind of processing method of voice messaging, device, equipment and storage medium | |
Boyce | Natural spoken dialogue systems for telephony applications | |
US20240087562A1 (en) | Interactive content output | |
AU2023274128A1 (en) | Hot-word free pre-emption of automated assistant response presentation | |
US10854196B1 (en) | Functional prerequisites and acknowledgments | |
JP7177348B2 (en) | Speech recognition device, speech recognition method and program | |
JP4304959B2 (en) | Voice dialogue control method, voice dialogue control apparatus, and voice dialogue control program | |
JP2001134285A (en) | Speech recognition device | |
Wang et al. | Cross Cultural Comparison of Users’ Barge-in with the In-Vehicle Speech System | |
JP2003036094A (en) | Device for speech dialogue and method for processing speech dialogue | |
JP7058305B2 (en) | Information processing device, audio output method, audio output program | |
JP2003228393A (en) | Device and method for voice interaction, voice interaction program and recording medium therefor | |
US12125477B2 (en) | Hot-word free pre-emption of automated assistant response presentation | |
CN114268694A (en) | Service request response method, device, equipment, system and medium | |
CN113885825A (en) | Method and device for intelligently creating application form |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
A131 | Notification of reasons for refusal |
Free format text: JAPANESE INTERMEDIATE CODE: A131 Effective date: 20040316 |
|
A131 | Notification of reasons for refusal |
Free format text: JAPANESE INTERMEDIATE CODE: A131 Effective date: 20050405 |
|
A521 | Request for written amendment filed |
Free format text: JAPANESE INTERMEDIATE CODE: A523 Effective date: 20050601 |
|
TRDD | Decision of grant or rejection written | ||
A01 | Written decision to grant a patent or to grant a registration (utility model) |
Free format text: JAPANESE INTERMEDIATE CODE: A01 Effective date: 20050621 |
|
A61 | First payment of annual fees (during grant procedure) |
Free format text: JAPANESE INTERMEDIATE CODE: A61 Effective date: 20050704 |
|
LAPS | Cancellation because of no payment of annual fees |