WO2008001549A1 - Dispositif audio interactif, procédé audio interactif, et programme correspondant - Google Patents
Dispositif audio interactif, procédé audio interactif, et programme correspondant Download PDFInfo
- Publication number
- WO2008001549A1 WO2008001549A1 PCT/JP2007/059586 JP2007059586W WO2008001549A1 WO 2008001549 A1 WO2008001549 A1 WO 2008001549A1 JP 2007059586 W JP2007059586 W JP 2007059586W WO 2008001549 A1 WO2008001549 A1 WO 2008001549A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- user
- dialogue
- interest
- confirmation
- voice
- Prior art date
Links
- 230000003993 interaction Effects 0.000 title claims abstract description 25
- 238000000034 method Methods 0.000 title claims description 28
- 238000012790 confirmation Methods 0.000 claims description 38
- 230000002452 interceptive effect Effects 0.000 claims description 10
- 230000004044 response Effects 0.000 abstract description 8
- 230000008569 process Effects 0.000 description 21
- 238000010586 diagram Methods 0.000 description 5
- 238000011156 evaluation Methods 0.000 description 5
- 230000008859 change Effects 0.000 description 4
- 238000001514 detection method Methods 0.000 description 4
- 238000013459 approach Methods 0.000 description 3
- 238000012545 processing Methods 0.000 description 3
- 230000000694 effects Effects 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
Definitions
- Voice dialogue apparatus voice dialogue method and program thereof
- the present invention relates to a voice dialogue apparatus such as a guidance robot, a voice dialogue method, and a dialogue program, which prevents the dialogue apparatus from speaking into the air and enables natural conversation with a person. It is in.
- Patent Document 1 Patent No. 3273620 discloses that a voice interaction device monitors human behavior, speaks when it stops in front of the interaction device for a predetermined time or more, and terminates the conversation when the user leaves. Disclosure.
- Patent Document 1 Japanese Patent No. 3273620
- An object of the present invention is to interrupt a dialogue when the user's interest is reduced during the dialogue, to resume the dialogue when the user's interest is recovered, and to terminate the dialogue when the user's interest does not return. is there.
- An additional problem in the present invention is to confirm the user's intention to interact with the natural point in the conversation scenario, and to enable the user to interrupt the conversation with an uninterested user.
- An additional problem with the present invention is to make the interaction with the user more natural.
- An additional problem with the present invention is to eliminate the situation where the dialog device advances by ignoring the user when the voice dialog device guides the user.
- the present invention is a voice interactive device including a speaker and a microphone
- And ending means for ending the dialogue when the degree of waiting exceeds a predetermined value.
- the present invention is also a voice interaction method using a speaker and a microphone
- the user's interest is confirmed by detecting the distance and direction to the user,
- While waiting utter the speaker power reminder to the user, confirm the willingness to continue the dialogue, continue the dialogue, and if the intention to continue the dialogue cannot be confirmed, wait further.
- the dialogue is terminated when the degree of waiting exceeds a predetermined value.
- the present invention is a program for a voice interaction device comprising a speaker, a microphone, and a sensor for detecting a distance and direction to a user,
- a voice command according to the confirmation result of the confirmation command is used so that the confirmation command for confirming the user's interest by the sensor and the conversation is continued if the user's interest is high, and waits if the user's interest is low.
- a command for switching the operation of the interactive device is used so that the confirmation command for confirming the user's interest by the sensor and the conversation is continued if the user's interest is high, and waits if the user's interest is low.
- a scenario of a dialog with the user describing the confirmation point of the user's interest is stored, and when the user's interest is low at the stored confirmation point, the dialog is interrupted and waited.
- the reminder means utter the reminder when the distance to the user decreases. Do.
- a moving means is provided so that the user can move with the user, and when the user is away from the user, the user is called from the speaker or the voice interactive device looks back at the user.
- the confirmation means detects and waits, Make a prompt utterance at the timing to check whether you intend to continue the dialogue. If the response is not obtained even if the reminder is uttered multiple times, or if the waiting time exceeds a predetermined value or the user has left the dialogue device, the dialogue is terminated.
- the dialogue can be terminated, so that the voice dialogue device can be prevented from talking into the air.
- you do not get a response after a single prompting utterance other prompting utterances that do not end the conversation immediately, or that you have waited for a sufficient amount of time from the user, etc. Wait for the condition to be met and end the conversation. For this reason, it does not give the user the impression that the conversation has ended suddenly.
- a natural conversation can be made as if the user had approached the voice dialogue device and made a prompt utterance in response.
- the voice interactive device guides the user, the distance from the user is increased. If you call from the speaker or the voice interaction device looks back to the user, you can prevent the user from naturally leaving the distance while keeping pace with the user's pace.
- FIG. 4 is a diagram showing an example of the interest level evaluation table in the embodiment.
- FIG. 5 is a diagram showing an example of a standby condition evaluation table in the embodiment.
- FIG. 6 Diagram showing the attitude control algorithm of the guidance robot in the embodiment
- FIG. 7 A flow chart showing an algorithm for maintaining contact with the user during guidance movement in the embodiment.
- FIGS. 1 to 7 will be described with reference to the guidance robot 2 as an example and its operation.
- 4 is a user and the person with whom the guidance robot 2 interacts
- 6 is a laser range sensor that detects the distance to the user 4 and the direction viewed from the guidance robot 2.
- CCD camera 8 Takes a face image, etc., and recognizes the direction of the face and line of sight in the image recognition unit 10 to recognize whether the user is looking at the guidance robot 2 or the other side .
- the user detection unit 12 evaluates the degree of interest of the user regarding the dialogue with the guidance robot 2 based on the recognition results of the sensor 6 and the image recognition unit 10.
- the degree of interest may be a two-level evaluation such as presence / absence, or a multi-level evaluation such as 0 to 10 or the like.
- the user's interest can be either instantaneous values at each point in time or the past interest level can be accumulated. If the past interest degree is accumulated, if the user shows interest for a long time, then the user's interest gradually increases. The degree of interest can be reduced.
- [0015] 14 is a microphone, 16 is an amplifier, and the voice recognition unit 18 recognizes the user's voice.
- 20 is a speaker, 22 is its amplifier, and the dialogue control unit 24 changes the scenario according to the degree of user interest from the user detection unit 12 and the user's voice input from the voice recognition unit 18. And controls the dialog by interrupting and ending the dialog.
- Reference numeral 26 denotes a scenario storage unit, which stores dialog scenarios and confirms the degree of interest of the user as a confirmation point in the scenario.
- the history storage unit 28 stores the history of scenario processing and prevents the same part of the scenario from being repeatedly processed.
- the history storage unit 28 changes the branch destination and jump destination in the scenario according to the history of the conversation up to now, changes the expression even when processing the same address in the scenario, and the first and second times Change the expression of the reminder utterance. For example, in the first reminder, “Can I continue the explanation?”, In the second reminder, “Let's finish the story?”
- Reference numeral 30 denotes an action control unit that causes the guidance robot 2 to walk with its feet and wheels via the walking control unit 32 and controls a hand and a neck (not shown) via the posture control unit 34.
- the control units 30 to 34 may not be provided.
- FIG. 2 shows the main processing in the embodiment.
- the scenario selection process 41 controls the branching of the scenario according to the voice input from the user or the user's answer to the reminder process 45.
- the reminder process 45 is an input that there is no voice input from the user, and that the user's interest remains low.
- the guidance process 42 guidance may be given to the user using a speaker, and another dialogue may be performed instead of the guidance.
- the confirmation process 43 uses the user detection unit 12 to confirm the degree of interest of the user. When the degree of interest of the user at the confirmation point is less than or equal to a predetermined value, the confirmation process 43 starts the standby process 44 and waits for a predetermined time.
- the prompting process 45 utters a reminder, and if a response to continue the dialogue is obtained, the guidance is continued, or the address is assumed to be of interest to the user. To jump. Further, when a reply indicating that the dialogue is to be terminated is obtained, the guidance is terminated. If an answer meaning continuation of the dialog cannot be obtained, the process further waits. If the user utters a reminder for a predetermined number of times, waits for a predetermined period of time, or the user leaves, the dialog is terminated in a termination process 46. Note that when the end of the dialogue is selected in the scenario selection process 41 or when the guidance is terminated, the dialogue is terminated in the termination process 46.
- Each process in Fig. 2 can be interpreted as an instruction. In that case, Fig. 2 can be said to show the voice dialogue program of the embodiment.
- FIG. 3 shows an algorithm for confirming the degree of interest of the user.
- guidance is provided within the university department plan, and guidance is provided in Step 11.
- the distance from the user to the guidance robot (a short distance is a factor of great interest), the force in front of the guidance robot (a factor of great interest), It is diagonally in front (a factor with a high degree of interest), it is on the side (a factor with a low level of interest), or the face direction and line of sight (a factor with a high degree of interest when looking at the robot side, the line of sight is off And the degree of interest) (step 12).
- step 13 the degree of interest is evaluated, and if the degree of interest is high, the scenario is continued in step 14, and if the degree of interest is low, the process waits (step 15).
- a reminder is made (steps 18 and 19). If the user's interest level does not increase, a reminder utterance is made every predetermined time from the start of standby (step 20). Then, if an answer is obtained that means that you want to continue the dialogue in response to the reminder (step 21), the scenario continues. The reminder utterance is "Let's continue the explanation?" If there is a positive response, or if a positive response is detected from the user's gesture, etc., the willingness of dialogue is assumed.
- step 22 if there is an answer that means the end of the dialogue, such as “No” or “Okay,” it ends (step 22). If there is no affirmative answer or no answer to the end of the dialogue, the process returns to the standby of step 15. You may change the message of the reminder utterance between the first time and the second time. For example, if you are interrupting an explanation for the faculty A, the first reminder will ask “Do you want to continue the explanation?” The second reminder will ask, “Do you want to explain other faculties?” In addition, it is okay to suggest a scenario change to the user. Detected that the user's interest has declined, such as when the willingness to talk is not displayed even after multiple reminder utterances, when the waiting time exceeds the specified time, or when the user leaves If so (step 16), terminate the conversation in step 17.
- FIG. 4 shows a table 50 for evaluating the degree of interest of the user.
- the user's face or line of sight is evaluated based on whether the user's distance or user orientation is front, diagonally front, or side force, such as whether the user is approaching, stopping, or retreating. Calculate the degree.
- the user's face is slanted and the distance is slightly far is the factor of interest, and the orientation is front and approaching is the factor of great interest.
- FIG. 6 shows an attitude control algorithm of the guidance robot in the embodiment.
- Guidance mouth Bot 2 has a face and head attached, and can change the direction around the neck to see and look back at the user. For one or more users, it detects the distance and orientation, the orientation of the user's face, and whether it is approaching or moving toward an increasing distance (step 25). From these, the degree of interest is evaluated for each user, and the robot looks at that user by turning his head or looking toward the user with the highest interest (step 26). When there is only one person, look at the user, except when the user moves away.
- FIG. 7 shows an algorithm in the case where the guidance robot guides the user while walking. While walking, the guidance robot stops at important points and interacts with the user. Guidance is given, but the details are omitted.
- the guidance robot checks the distance to the user while moving (step 31). If the distance is excessive, it stops or slows down and waits for the user to approach (step 33). If the guidance robot stops or slows down and the distance to the user is too long (step 34), rotate the guidance robot's neck and call backwards or simply call Encourage the user to come closer by looking back (step 35). While maintaining the distance to the user within the appropriate range, guidance is given with the user, and if the user is lost (step 36), the guidance is terminated (step 37).
- the will of the dialogue can be confirmed naturally, taking into account the distance to the user, the orientation with the user, whether the user is approaching, the user's face and line of sight.
Landscapes
- Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Manipulator (AREA)
Abstract
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2008522337A JP4505862B2 (ja) | 2006-06-26 | 2007-05-09 | 音声対話装置と音声対話方法及びそのプログラム |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2006175497 | 2006-06-26 | ||
JP2006-175497 | 2006-06-26 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2008001549A1 true WO2008001549A1 (fr) | 2008-01-03 |
Family
ID=38845321
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2007/059586 WO2008001549A1 (fr) | 2006-06-26 | 2007-05-09 | Dispositif audio interactif, procédé audio interactif, et programme correspondant |
Country Status (2)
Country | Link |
---|---|
JP (1) | JP4505862B2 (fr) |
WO (1) | WO2008001549A1 (fr) |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2016076007A (ja) * | 2014-10-03 | 2016-05-12 | 株式会社Nttドコモ | 対話装置および対話方法 |
JP2017049471A (ja) * | 2015-09-03 | 2017-03-09 | カシオ計算機株式会社 | 対話制御装置、対話制御方法及びプログラム |
JP2018013545A (ja) * | 2016-07-19 | 2018-01-25 | トヨタ自動車株式会社 | 音声対話装置および発話制御方法 |
WO2018034169A1 (fr) * | 2016-08-17 | 2018-02-22 | ソニー株式会社 | Dispositif et procédé de commande de dialogue |
WO2019107144A1 (fr) * | 2017-11-28 | 2019-06-06 | ソニー株式会社 | Dispositif et procédé de traitement d'informations |
JP2020038397A (ja) * | 2019-12-09 | 2020-03-12 | カシオ計算機株式会社 | 対話制御装置、対話制御方法及びプログラム |
WO2023112745A1 (fr) * | 2021-12-17 | 2023-06-22 | ソニーグループ株式会社 | Procédé de traitement d'informations, dispositif de traitement d'informations et programme de traitement d'informations |
JP7313518B1 (ja) | 2022-07-07 | 2023-07-24 | 株式会社エクサウィザーズ | 評価方法、評価装置、および、評価プログラム |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11220008B2 (en) * | 2017-07-18 | 2022-01-11 | Panasonic Intellectual Property Management Co., Ltd. | Apparatus, method, non-transitory computer-readable recording medium storing program, and robot |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0588691A (ja) * | 1991-09-30 | 1993-04-09 | Toshiba Corp | 音声対話装置 |
JPH08234789A (ja) * | 1995-02-27 | 1996-09-13 | Sharp Corp | 統合認識対話装置 |
JP2001188551A (ja) * | 1999-12-28 | 2001-07-10 | Sony Corp | 情報処理装置および方法、並びに記録媒体 |
JP2001306292A (ja) * | 2000-04-18 | 2001-11-02 | Casio Comput Co Ltd | アクション応答システムおよびそのプログラム記録媒体 |
JP2004192653A (ja) * | 1997-02-28 | 2004-07-08 | Toshiba Corp | マルチモーダルインタフェース装置およびマルチモーダルインタフェース方法 |
JP2005237561A (ja) * | 2004-02-25 | 2005-09-08 | Canon Inc | 情報処理装置及び方法 |
JP2007088803A (ja) * | 2005-09-22 | 2007-04-05 | Hitachi Ltd | 情報処理装置 |
JP2007160442A (ja) * | 2005-12-12 | 2007-06-28 | Honda Motor Co Ltd | 移動ロボット |
-
2007
- 2007-05-09 JP JP2008522337A patent/JP4505862B2/ja not_active Expired - Fee Related
- 2007-05-09 WO PCT/JP2007/059586 patent/WO2008001549A1/fr active Application Filing
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0588691A (ja) * | 1991-09-30 | 1993-04-09 | Toshiba Corp | 音声対話装置 |
JPH08234789A (ja) * | 1995-02-27 | 1996-09-13 | Sharp Corp | 統合認識対話装置 |
JP2004192653A (ja) * | 1997-02-28 | 2004-07-08 | Toshiba Corp | マルチモーダルインタフェース装置およびマルチモーダルインタフェース方法 |
JP2001188551A (ja) * | 1999-12-28 | 2001-07-10 | Sony Corp | 情報処理装置および方法、並びに記録媒体 |
JP2001306292A (ja) * | 2000-04-18 | 2001-11-02 | Casio Comput Co Ltd | アクション応答システムおよびそのプログラム記録媒体 |
JP2005237561A (ja) * | 2004-02-25 | 2005-09-08 | Canon Inc | 情報処理装置及び方法 |
JP2007088803A (ja) * | 2005-09-22 | 2007-04-05 | Hitachi Ltd | 情報処理装置 |
JP2007160442A (ja) * | 2005-12-12 | 2007-06-28 | Honda Motor Co Ltd | 移動ロボット |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2016076007A (ja) * | 2014-10-03 | 2016-05-12 | 株式会社Nttドコモ | 対話装置および対話方法 |
JP2017049471A (ja) * | 2015-09-03 | 2017-03-09 | カシオ計算機株式会社 | 対話制御装置、対話制御方法及びプログラム |
JP2018013545A (ja) * | 2016-07-19 | 2018-01-25 | トヨタ自動車株式会社 | 音声対話装置および発話制御方法 |
WO2018034169A1 (fr) * | 2016-08-17 | 2018-02-22 | ソニー株式会社 | Dispositif et procédé de commande de dialogue |
JPWO2018034169A1 (ja) * | 2016-08-17 | 2019-06-13 | ソニー株式会社 | 対話制御装置および方法 |
EP3503091A4 (fr) * | 2016-08-17 | 2019-08-07 | Sony Corporation | Dispositif et procédé de commande de dialogue |
US11183170B2 (en) | 2016-08-17 | 2021-11-23 | Sony Corporation | Interaction control apparatus and method |
JP7036015B2 (ja) | 2016-08-17 | 2022-03-15 | ソニーグループ株式会社 | 対話制御装置および方法 |
WO2019107144A1 (fr) * | 2017-11-28 | 2019-06-06 | ソニー株式会社 | Dispositif et procédé de traitement d'informations |
JP2020038397A (ja) * | 2019-12-09 | 2020-03-12 | カシオ計算機株式会社 | 対話制御装置、対話制御方法及びプログラム |
WO2023112745A1 (fr) * | 2021-12-17 | 2023-06-22 | ソニーグループ株式会社 | Procédé de traitement d'informations, dispositif de traitement d'informations et programme de traitement d'informations |
JP7313518B1 (ja) | 2022-07-07 | 2023-07-24 | 株式会社エクサウィザーズ | 評価方法、評価装置、および、評価プログラム |
Also Published As
Publication number | Publication date |
---|---|
JPWO2008001549A1 (ja) | 2009-11-26 |
JP4505862B2 (ja) | 2010-07-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2008001549A1 (fr) | Dispositif audio interactif, procédé audio interactif, et programme correspondant | |
US7069221B2 (en) | Non-target barge-in detection | |
JP6171617B2 (ja) | 応答対象音声判定装置、応答対象音声判定方法および応答対象音声判定プログラム | |
US10019992B2 (en) | Speech-controlled actions based on keywords and context thereof | |
JP5601419B2 (ja) | エレベーターの呼び登録装置 | |
WO2010013371A1 (fr) | Système de reconnaissance vocale de dialogue, procédé de reconnaissance vocale de dialogue et support d'enregistrement pour stocker un programme de reconnaissance vocale de dialogue | |
JP2005022065A5 (fr) | ||
US10800043B2 (en) | Interaction apparatus and method for determining a turn-taking behavior using multimodel information | |
JP2008509455A (ja) | ユーザとシステムとの間の通信方法及びシステム | |
WO2008069187A1 (fr) | Dispositif de support de présentation, procédé et programme | |
TW200809768A (en) | Method of driving a speech recognition system | |
JP5431282B2 (ja) | 音声対話装置、方法、プログラム | |
JP2008256802A (ja) | 音声認識装置および音声認識方法 | |
JP2006123136A (ja) | コミュニケーションロボット | |
CN115088033A (zh) | 代表对话中的人参与者生成的合成语音音频数据 | |
Selfridge et al. | Continuously predicting and processing barge-in during a live spoken dialogue task | |
JP2007155986A (ja) | 音声認識装置および音声認識装置を備えたロボット | |
JP4491438B2 (ja) | 音声対話装置、音声対話方法、およびプログラム | |
JP2004333543A (ja) | 音声対話システム及び音声対話方法 | |
JP2004234631A (ja) | ユーザと対話型実体エージェントとの間の対話を管理するシステムおよび対話型実体エージェントによるユーザとの対話を管理する方法 | |
JPH09269889A (ja) | 対話装置 | |
JP2019132997A (ja) | 音声処理装置、方法およびプログラム | |
JP3199972B2 (ja) | あいづち応答のある対話装置 | |
JP6748565B2 (ja) | 音声対話システム及び音声対話方法 | |
JP6941856B2 (ja) | 対話ロボットおよびロボット制御プログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
DPE2 | Request for preliminary examination filed before expiration of 19th month from priority date (pct application filed from 20040101) | ||
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 07743021 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2008522337 Country of ref document: JP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
NENP | Non-entry into the national phase |
Ref country code: RU |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1485/CHENP/2009 Country of ref document: IN |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 07743021 Country of ref document: EP Kind code of ref document: A1 |