WO2008001549A1 - Dispositif audio interactif, procédé audio interactif, et programme correspondant - Google Patents

Dispositif audio interactif, procédé audio interactif, et programme correspondant Download PDF

Info

Publication number
WO2008001549A1
WO2008001549A1 PCT/JP2007/059586 JP2007059586W WO2008001549A1 WO 2008001549 A1 WO2008001549 A1 WO 2008001549A1 JP 2007059586 W JP2007059586 W JP 2007059586W WO 2008001549 A1 WO2008001549 A1 WO 2008001549A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
dialogue
interest
confirmation
voice
Prior art date
Application number
PCT/JP2007/059586
Other languages
English (en)
Japanese (ja)
Inventor
Toshiki Moriguchi
Tsuyoshi Nakano
Yasutaka Shindoh
Original Assignee
Murata Kikai Kabushiki Kaisha
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Murata Kikai Kabushiki Kaisha filed Critical Murata Kikai Kabushiki Kaisha
Priority to JP2008522337A priority Critical patent/JP4505862B2/ja
Publication of WO2008001549A1 publication Critical patent/WO2008001549A1/fr

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue

Definitions

  • Voice dialogue apparatus voice dialogue method and program thereof
  • the present invention relates to a voice dialogue apparatus such as a guidance robot, a voice dialogue method, and a dialogue program, which prevents the dialogue apparatus from speaking into the air and enables natural conversation with a person. It is in.
  • Patent Document 1 Patent No. 3273620 discloses that a voice interaction device monitors human behavior, speaks when it stops in front of the interaction device for a predetermined time or more, and terminates the conversation when the user leaves. Disclosure.
  • Patent Document 1 Japanese Patent No. 3273620
  • An object of the present invention is to interrupt a dialogue when the user's interest is reduced during the dialogue, to resume the dialogue when the user's interest is recovered, and to terminate the dialogue when the user's interest does not return. is there.
  • An additional problem in the present invention is to confirm the user's intention to interact with the natural point in the conversation scenario, and to enable the user to interrupt the conversation with an uninterested user.
  • An additional problem with the present invention is to make the interaction with the user more natural.
  • An additional problem with the present invention is to eliminate the situation where the dialog device advances by ignoring the user when the voice dialog device guides the user.
  • the present invention is a voice interactive device including a speaker and a microphone
  • And ending means for ending the dialogue when the degree of waiting exceeds a predetermined value.
  • the present invention is also a voice interaction method using a speaker and a microphone
  • the user's interest is confirmed by detecting the distance and direction to the user,
  • While waiting utter the speaker power reminder to the user, confirm the willingness to continue the dialogue, continue the dialogue, and if the intention to continue the dialogue cannot be confirmed, wait further.
  • the dialogue is terminated when the degree of waiting exceeds a predetermined value.
  • the present invention is a program for a voice interaction device comprising a speaker, a microphone, and a sensor for detecting a distance and direction to a user,
  • a voice command according to the confirmation result of the confirmation command is used so that the confirmation command for confirming the user's interest by the sensor and the conversation is continued if the user's interest is high, and waits if the user's interest is low.
  • a command for switching the operation of the interactive device is used so that the confirmation command for confirming the user's interest by the sensor and the conversation is continued if the user's interest is high, and waits if the user's interest is low.
  • a scenario of a dialog with the user describing the confirmation point of the user's interest is stored, and when the user's interest is low at the stored confirmation point, the dialog is interrupted and waited.
  • the reminder means utter the reminder when the distance to the user decreases. Do.
  • a moving means is provided so that the user can move with the user, and when the user is away from the user, the user is called from the speaker or the voice interactive device looks back at the user.
  • the confirmation means detects and waits, Make a prompt utterance at the timing to check whether you intend to continue the dialogue. If the response is not obtained even if the reminder is uttered multiple times, or if the waiting time exceeds a predetermined value or the user has left the dialogue device, the dialogue is terminated.
  • the dialogue can be terminated, so that the voice dialogue device can be prevented from talking into the air.
  • you do not get a response after a single prompting utterance other prompting utterances that do not end the conversation immediately, or that you have waited for a sufficient amount of time from the user, etc. Wait for the condition to be met and end the conversation. For this reason, it does not give the user the impression that the conversation has ended suddenly.
  • a natural conversation can be made as if the user had approached the voice dialogue device and made a prompt utterance in response.
  • the voice interactive device guides the user, the distance from the user is increased. If you call from the speaker or the voice interaction device looks back to the user, you can prevent the user from naturally leaving the distance while keeping pace with the user's pace.
  • FIG. 4 is a diagram showing an example of the interest level evaluation table in the embodiment.
  • FIG. 5 is a diagram showing an example of a standby condition evaluation table in the embodiment.
  • FIG. 6 Diagram showing the attitude control algorithm of the guidance robot in the embodiment
  • FIG. 7 A flow chart showing an algorithm for maintaining contact with the user during guidance movement in the embodiment.
  • FIGS. 1 to 7 will be described with reference to the guidance robot 2 as an example and its operation.
  • 4 is a user and the person with whom the guidance robot 2 interacts
  • 6 is a laser range sensor that detects the distance to the user 4 and the direction viewed from the guidance robot 2.
  • CCD camera 8 Takes a face image, etc., and recognizes the direction of the face and line of sight in the image recognition unit 10 to recognize whether the user is looking at the guidance robot 2 or the other side .
  • the user detection unit 12 evaluates the degree of interest of the user regarding the dialogue with the guidance robot 2 based on the recognition results of the sensor 6 and the image recognition unit 10.
  • the degree of interest may be a two-level evaluation such as presence / absence, or a multi-level evaluation such as 0 to 10 or the like.
  • the user's interest can be either instantaneous values at each point in time or the past interest level can be accumulated. If the past interest degree is accumulated, if the user shows interest for a long time, then the user's interest gradually increases. The degree of interest can be reduced.
  • [0015] 14 is a microphone, 16 is an amplifier, and the voice recognition unit 18 recognizes the user's voice.
  • 20 is a speaker, 22 is its amplifier, and the dialogue control unit 24 changes the scenario according to the degree of user interest from the user detection unit 12 and the user's voice input from the voice recognition unit 18. And controls the dialog by interrupting and ending the dialog.
  • Reference numeral 26 denotes a scenario storage unit, which stores dialog scenarios and confirms the degree of interest of the user as a confirmation point in the scenario.
  • the history storage unit 28 stores the history of scenario processing and prevents the same part of the scenario from being repeatedly processed.
  • the history storage unit 28 changes the branch destination and jump destination in the scenario according to the history of the conversation up to now, changes the expression even when processing the same address in the scenario, and the first and second times Change the expression of the reminder utterance. For example, in the first reminder, “Can I continue the explanation?”, In the second reminder, “Let's finish the story?”
  • Reference numeral 30 denotes an action control unit that causes the guidance robot 2 to walk with its feet and wheels via the walking control unit 32 and controls a hand and a neck (not shown) via the posture control unit 34.
  • the control units 30 to 34 may not be provided.
  • FIG. 2 shows the main processing in the embodiment.
  • the scenario selection process 41 controls the branching of the scenario according to the voice input from the user or the user's answer to the reminder process 45.
  • the reminder process 45 is an input that there is no voice input from the user, and that the user's interest remains low.
  • the guidance process 42 guidance may be given to the user using a speaker, and another dialogue may be performed instead of the guidance.
  • the confirmation process 43 uses the user detection unit 12 to confirm the degree of interest of the user. When the degree of interest of the user at the confirmation point is less than or equal to a predetermined value, the confirmation process 43 starts the standby process 44 and waits for a predetermined time.
  • the prompting process 45 utters a reminder, and if a response to continue the dialogue is obtained, the guidance is continued, or the address is assumed to be of interest to the user. To jump. Further, when a reply indicating that the dialogue is to be terminated is obtained, the guidance is terminated. If an answer meaning continuation of the dialog cannot be obtained, the process further waits. If the user utters a reminder for a predetermined number of times, waits for a predetermined period of time, or the user leaves, the dialog is terminated in a termination process 46. Note that when the end of the dialogue is selected in the scenario selection process 41 or when the guidance is terminated, the dialogue is terminated in the termination process 46.
  • Each process in Fig. 2 can be interpreted as an instruction. In that case, Fig. 2 can be said to show the voice dialogue program of the embodiment.
  • FIG. 3 shows an algorithm for confirming the degree of interest of the user.
  • guidance is provided within the university department plan, and guidance is provided in Step 11.
  • the distance from the user to the guidance robot (a short distance is a factor of great interest), the force in front of the guidance robot (a factor of great interest), It is diagonally in front (a factor with a high degree of interest), it is on the side (a factor with a low level of interest), or the face direction and line of sight (a factor with a high degree of interest when looking at the robot side, the line of sight is off And the degree of interest) (step 12).
  • step 13 the degree of interest is evaluated, and if the degree of interest is high, the scenario is continued in step 14, and if the degree of interest is low, the process waits (step 15).
  • a reminder is made (steps 18 and 19). If the user's interest level does not increase, a reminder utterance is made every predetermined time from the start of standby (step 20). Then, if an answer is obtained that means that you want to continue the dialogue in response to the reminder (step 21), the scenario continues. The reminder utterance is "Let's continue the explanation?" If there is a positive response, or if a positive response is detected from the user's gesture, etc., the willingness of dialogue is assumed.
  • step 22 if there is an answer that means the end of the dialogue, such as “No” or “Okay,” it ends (step 22). If there is no affirmative answer or no answer to the end of the dialogue, the process returns to the standby of step 15. You may change the message of the reminder utterance between the first time and the second time. For example, if you are interrupting an explanation for the faculty A, the first reminder will ask “Do you want to continue the explanation?” The second reminder will ask, “Do you want to explain other faculties?” In addition, it is okay to suggest a scenario change to the user. Detected that the user's interest has declined, such as when the willingness to talk is not displayed even after multiple reminder utterances, when the waiting time exceeds the specified time, or when the user leaves If so (step 16), terminate the conversation in step 17.
  • FIG. 4 shows a table 50 for evaluating the degree of interest of the user.
  • the user's face or line of sight is evaluated based on whether the user's distance or user orientation is front, diagonally front, or side force, such as whether the user is approaching, stopping, or retreating. Calculate the degree.
  • the user's face is slanted and the distance is slightly far is the factor of interest, and the orientation is front and approaching is the factor of great interest.
  • FIG. 6 shows an attitude control algorithm of the guidance robot in the embodiment.
  • Guidance mouth Bot 2 has a face and head attached, and can change the direction around the neck to see and look back at the user. For one or more users, it detects the distance and orientation, the orientation of the user's face, and whether it is approaching or moving toward an increasing distance (step 25). From these, the degree of interest is evaluated for each user, and the robot looks at that user by turning his head or looking toward the user with the highest interest (step 26). When there is only one person, look at the user, except when the user moves away.
  • FIG. 7 shows an algorithm in the case where the guidance robot guides the user while walking. While walking, the guidance robot stops at important points and interacts with the user. Guidance is given, but the details are omitted.
  • the guidance robot checks the distance to the user while moving (step 31). If the distance is excessive, it stops or slows down and waits for the user to approach (step 33). If the guidance robot stops or slows down and the distance to the user is too long (step 34), rotate the guidance robot's neck and call backwards or simply call Encourage the user to come closer by looking back (step 35). While maintaining the distance to the user within the appropriate range, guidance is given with the user, and if the user is lost (step 36), the guidance is terminated (step 37).
  • the will of the dialogue can be confirmed naturally, taking into account the distance to the user, the orientation with the user, whether the user is approaching, the user's face and line of sight.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Manipulator (AREA)

Abstract

Une distance et une direction sont confirmées à un utilisateur lors d'une étape de vérification dans un scénario interactif. Si l'utilisateur est assis à proximité, à l'avant ou dans la direction antérieure oblique, l'interaction se poursuit, mais dans le cas contraire, l'interaction est interrompue et lorsque la distance jusqu'à l'utilisateur se réduit, une émission sonore lui rappelle cette interaction. L'interaction reprend si l'utilisateur répond à cette émission sonore, mais s'achève si l'obtention de plusieurs réponses est impossible. Ainsi, l'intérêt de l'utilisateur peut être confirmé pendant l'interaction, et en cas de faible intérêt, un terme peut être mis à cette interaction, après l'étape d'interruption d'interaction ou l'étape de rappel.
PCT/JP2007/059586 2006-06-26 2007-05-09 Dispositif audio interactif, procédé audio interactif, et programme correspondant WO2008001549A1 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2008522337A JP4505862B2 (ja) 2006-06-26 2007-05-09 音声対話装置と音声対話方法及びそのプログラム

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2006175497 2006-06-26
JP2006-175497 2006-06-26

Publications (1)

Publication Number Publication Date
WO2008001549A1 true WO2008001549A1 (fr) 2008-01-03

Family

ID=38845321

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2007/059586 WO2008001549A1 (fr) 2006-06-26 2007-05-09 Dispositif audio interactif, procédé audio interactif, et programme correspondant

Country Status (2)

Country Link
JP (1) JP4505862B2 (fr)
WO (1) WO2008001549A1 (fr)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2016076007A (ja) * 2014-10-03 2016-05-12 株式会社Nttドコモ 対話装置および対話方法
JP2017049471A (ja) * 2015-09-03 2017-03-09 カシオ計算機株式会社 対話制御装置、対話制御方法及びプログラム
JP2018013545A (ja) * 2016-07-19 2018-01-25 トヨタ自動車株式会社 音声対話装置および発話制御方法
WO2018034169A1 (fr) * 2016-08-17 2018-02-22 ソニー株式会社 Dispositif et procédé de commande de dialogue
WO2019107144A1 (fr) * 2017-11-28 2019-06-06 ソニー株式会社 Dispositif et procédé de traitement d'informations
JP2020038397A (ja) * 2019-12-09 2020-03-12 カシオ計算機株式会社 対話制御装置、対話制御方法及びプログラム
WO2023112745A1 (fr) * 2021-12-17 2023-06-22 ソニーグループ株式会社 Procédé de traitement d'informations, dispositif de traitement d'informations et programme de traitement d'informations
JP7313518B1 (ja) 2022-07-07 2023-07-24 株式会社エクサウィザーズ 評価方法、評価装置、および、評価プログラム

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11220008B2 (en) * 2017-07-18 2022-01-11 Panasonic Intellectual Property Management Co., Ltd. Apparatus, method, non-transitory computer-readable recording medium storing program, and robot

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0588691A (ja) * 1991-09-30 1993-04-09 Toshiba Corp 音声対話装置
JPH08234789A (ja) * 1995-02-27 1996-09-13 Sharp Corp 統合認識対話装置
JP2001188551A (ja) * 1999-12-28 2001-07-10 Sony Corp 情報処理装置および方法、並びに記録媒体
JP2001306292A (ja) * 2000-04-18 2001-11-02 Casio Comput Co Ltd アクション応答システムおよびそのプログラム記録媒体
JP2004192653A (ja) * 1997-02-28 2004-07-08 Toshiba Corp マルチモーダルインタフェース装置およびマルチモーダルインタフェース方法
JP2005237561A (ja) * 2004-02-25 2005-09-08 Canon Inc 情報処理装置及び方法
JP2007088803A (ja) * 2005-09-22 2007-04-05 Hitachi Ltd 情報処理装置
JP2007160442A (ja) * 2005-12-12 2007-06-28 Honda Motor Co Ltd 移動ロボット

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0588691A (ja) * 1991-09-30 1993-04-09 Toshiba Corp 音声対話装置
JPH08234789A (ja) * 1995-02-27 1996-09-13 Sharp Corp 統合認識対話装置
JP2004192653A (ja) * 1997-02-28 2004-07-08 Toshiba Corp マルチモーダルインタフェース装置およびマルチモーダルインタフェース方法
JP2001188551A (ja) * 1999-12-28 2001-07-10 Sony Corp 情報処理装置および方法、並びに記録媒体
JP2001306292A (ja) * 2000-04-18 2001-11-02 Casio Comput Co Ltd アクション応答システムおよびそのプログラム記録媒体
JP2005237561A (ja) * 2004-02-25 2005-09-08 Canon Inc 情報処理装置及び方法
JP2007088803A (ja) * 2005-09-22 2007-04-05 Hitachi Ltd 情報処理装置
JP2007160442A (ja) * 2005-12-12 2007-06-28 Honda Motor Co Ltd 移動ロボット

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2016076007A (ja) * 2014-10-03 2016-05-12 株式会社Nttドコモ 対話装置および対話方法
JP2017049471A (ja) * 2015-09-03 2017-03-09 カシオ計算機株式会社 対話制御装置、対話制御方法及びプログラム
JP2018013545A (ja) * 2016-07-19 2018-01-25 トヨタ自動車株式会社 音声対話装置および発話制御方法
WO2018034169A1 (fr) * 2016-08-17 2018-02-22 ソニー株式会社 Dispositif et procédé de commande de dialogue
JPWO2018034169A1 (ja) * 2016-08-17 2019-06-13 ソニー株式会社 対話制御装置および方法
EP3503091A4 (fr) * 2016-08-17 2019-08-07 Sony Corporation Dispositif et procédé de commande de dialogue
US11183170B2 (en) 2016-08-17 2021-11-23 Sony Corporation Interaction control apparatus and method
JP7036015B2 (ja) 2016-08-17 2022-03-15 ソニーグループ株式会社 対話制御装置および方法
WO2019107144A1 (fr) * 2017-11-28 2019-06-06 ソニー株式会社 Dispositif et procédé de traitement d'informations
JP2020038397A (ja) * 2019-12-09 2020-03-12 カシオ計算機株式会社 対話制御装置、対話制御方法及びプログラム
WO2023112745A1 (fr) * 2021-12-17 2023-06-22 ソニーグループ株式会社 Procédé de traitement d'informations, dispositif de traitement d'informations et programme de traitement d'informations
JP7313518B1 (ja) 2022-07-07 2023-07-24 株式会社エクサウィザーズ 評価方法、評価装置、および、評価プログラム

Also Published As

Publication number Publication date
JPWO2008001549A1 (ja) 2009-11-26
JP4505862B2 (ja) 2010-07-21

Similar Documents

Publication Publication Date Title
WO2008001549A1 (fr) Dispositif audio interactif, procédé audio interactif, et programme correspondant
US7069221B2 (en) Non-target barge-in detection
JP6171617B2 (ja) 応答対象音声判定装置、応答対象音声判定方法および応答対象音声判定プログラム
US10019992B2 (en) Speech-controlled actions based on keywords and context thereof
JP5601419B2 (ja) エレベーターの呼び登録装置
WO2010013371A1 (fr) Système de reconnaissance vocale de dialogue, procédé de reconnaissance vocale de dialogue et support d'enregistrement pour stocker un programme de reconnaissance vocale de dialogue
JP2005022065A5 (fr)
US10800043B2 (en) Interaction apparatus and method for determining a turn-taking behavior using multimodel information
JP2008509455A (ja) ユーザとシステムとの間の通信方法及びシステム
WO2008069187A1 (fr) Dispositif de support de présentation, procédé et programme
TW200809768A (en) Method of driving a speech recognition system
JP5431282B2 (ja) 音声対話装置、方法、プログラム
JP2008256802A (ja) 音声認識装置および音声認識方法
JP2006123136A (ja) コミュニケーションロボット
CN115088033A (zh) 代表对话中的人参与者生成的合成语音音频数据
Selfridge et al. Continuously predicting and processing barge-in during a live spoken dialogue task
JP2007155986A (ja) 音声認識装置および音声認識装置を備えたロボット
JP4491438B2 (ja) 音声対話装置、音声対話方法、およびプログラム
JP2004333543A (ja) 音声対話システム及び音声対話方法
JP2004234631A (ja) ユーザと対話型実体エージェントとの間の対話を管理するシステムおよび対話型実体エージェントによるユーザとの対話を管理する方法
JPH09269889A (ja) 対話装置
JP2019132997A (ja) 音声処理装置、方法およびプログラム
JP3199972B2 (ja) あいづち応答のある対話装置
JP6748565B2 (ja) 音声対話システム及び音声対話方法
JP6941856B2 (ja) 対話ロボットおよびロボット制御プログラム

Legal Events

Date Code Title Description
DPE2 Request for preliminary examination filed before expiration of 19th month from priority date (pct application filed from 20040101)
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 07743021

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 2008522337

Country of ref document: JP

NENP Non-entry into the national phase

Ref country code: DE

NENP Non-entry into the national phase

Ref country code: RU

WWE Wipo information: entry into national phase

Ref document number: 1485/CHENP/2009

Country of ref document: IN

122 Ep: pct application non-entry in european phase

Ref document number: 07743021

Country of ref document: EP

Kind code of ref document: A1