WO2016136208A1 - Voice interaction device, voice interaction system, control method of voice interaction device - Google Patents

Voice interaction device, voice interaction system, control method of voice interaction device Download PDF

Info

Publication number
WO2016136208A1
WO2016136208A1 PCT/JP2016/000855 JP2016000855W WO2016136208A1 WO 2016136208 A1 WO2016136208 A1 WO 2016136208A1 JP 2016000855 W JP2016000855 W JP 2016000855W WO 2016136208 A1 WO2016136208 A1 WO 2016136208A1
Authority
WO
WIPO (PCT)
Prior art keywords
term
voice
unit
dialogue
user
Prior art date
Application number
PCT/JP2016/000855
Other languages
French (fr)
Japanese (ja)
Inventor
釜井 孝浩
宇佐見 陽
中西 雅浩
Original Assignee
パナソニックIpマネジメント株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by パナソニックIpマネジメント株式会社 filed Critical パナソニックIpマネジメント株式会社
Publication of WO2016136208A1 publication Critical patent/WO2016136208A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/10Speech classification or search using distance or distortion measures between unknown speech and reference templates
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue

Definitions

  • the present disclosure relates to a voice dialogue apparatus, a voice dialogue system, and a control method for the voice dialogue apparatus.
  • Patent Document 1 discloses a dialog sequence recognition device that presents a vocabulary group expected to be input next so that the user can visually recognize it based on information input from the user. This prevents the inconvenience that the user is at a loss due to erroneous recognition of the dialogue.
  • This disclosure provides a speech dialogue apparatus that modifies the content of dialogue with a user by a simple method.
  • the voice dialogue apparatus is a voice dialogue apparatus that performs voice dialogue with a user, and includes an acquisition unit, a plurality of holding units, a storage unit, and a change unit.
  • the acquisition unit acquires utterance data indicating the content of the utterance by the user's voice.
  • Each of the plurality of holding units associates the term included in the utterance data with the attribute of the term.
  • the storage unit stores a history of terms held by the plurality of holding units.
  • the changing unit refers to the history stored in the storage unit, so that the plurality of holding units at the past time point become the terms held by the plurality of holding units. Change the terminology of the holding part.
  • the past time point is a past time point specified by the control term.
  • the speech dialogue apparatus is effective for correcting the content of dialogue with the user by a simple method.
  • FIG. 1 is a block diagram illustrating a configuration of a voice interaction apparatus and a voice interaction system according to an embodiment.
  • FIG. 2 is a diagram for explaining presentation by the voice interaction system according to the embodiment.
  • FIG. 3 is a diagram for explaining a dialogue sequence and history information according to the embodiment.
  • FIG. 4 is a flowchart of a main process performed by the voice interaction apparatus according to the embodiment.
  • FIG. 5 is a flowchart of restoration processing by the voice interaction apparatus according to the embodiment.
  • FIG. 6 is a flowchart of restoration point setting processing by the voice interaction apparatus according to the embodiment.
  • FIG. 7 is a diagram for explaining a dialogue sequence and history information according to the embodiment.
  • FIG. 8 is a block diagram showing a configuration of a voice interactive apparatus according to a modification of the embodiment.
  • FIG. 9 is a flowchart showing a control method of the voice interactive apparatus according to the modification of the embodiment.
  • the voice dialogue apparatus performs voice dialogue with the user, generates and corrects dialogue information indicating the content of the dialogue with the user, and outputs the dialogue information to an external processing device. . Further, the voice interaction device acquires the processing result from the external processing device and presents it to the user, and further continues the dialogue with the user. As described above, the voice interaction device sequentially presents the processing results to the user while generating and correcting the interaction information based on the interaction with the user.
  • the voice interaction device is useful when an operation such as key input by the user or touching the panel is impossible or difficult.
  • applications such as a car navigation device that searches for information while sequentially receiving instructions by the user's voice when the user is driving a car or the like. It is also useful in a voice interaction device that does not have a user interface such as a key or a panel.
  • FIG. 1 is a block diagram showing a configuration of a voice interaction device 20 and a voice interaction system 1 according to the present embodiment.
  • the voice dialogue system 1 includes a display device 10, a speaker 11, a voice synthesis unit 12, a microphone 13, a voice recognition unit 14, a voice dialogue device 20, and a task processing unit 40. Is provided.
  • the display device 10 is a display device having a display screen.
  • the display device 10 displays an image on the display screen based on the display data acquired from the voice interaction device 20.
  • the display device 10 is realized by, for example, a car navigation device, a smartphone (high-function mobile phone terminal), a mobile phone terminal, a mobile information terminal, a display, or a PC (Personal Computer).
  • the display device 10 is shown as an example of a device that displays an image based on information presented by the voice interaction device 20, a speaker that outputs information presented by the voice interaction device 20 as a voice instead of the display device 10. May be used. This speaker may be shared with the speaker 11 described later.
  • Speaker 11 is a speaker that outputs sound.
  • the speaker 11 outputs sound based on the sound signal acquired from the sound synthesizer 12. The sound output from the speaker 11 is heard by the user.
  • the speech synthesis unit 12 is a processing unit that converts a response sentence into a speech signal.
  • the voice synthesizing unit 12 acquires a response sentence, which is information transmitted from the voice dialogue apparatus 20 to the user, from the voice dialogue apparatus 20, and generates a voice signal to be output by the speaker based on the obtained response sentence.
  • the speaker 11 and the voice synthesis unit 12 may be provided inside the voice dialogue apparatus 20 as one function of the voice dialogue apparatus 20 or may be provided outside the voice dialogue apparatus 20. Further, the voice synthesizer 12 may be realized as a so-called cloud server so as to be able to communicate with the voice interaction device 20 via a network such as the Internet. In that case, the connection between the voice synthesizer 12 and the voice interaction device 20 and the connection between the voice synthesizer 12 and the speaker 11 are made through a communication path via the Internet or the like.
  • the microphone 13 is a microphone that acquires sound.
  • the microphone 13 acquires the user's voice and outputs an audio signal based on the acquired voice.
  • the voice recognition unit 14 is a processing unit that generates speech data by performing voice recognition on the user's voice.
  • the voice recognition unit 14 acquires the voice signal generated by the microphone 13 and performs voice recognition processing on the acquired voice signal, thereby generating utterance data of the user's utterance.
  • the utterance data is information transmitted from the user to the voice interaction device 20, and is expressed by characters (text) such as “I want to eat Chinese”. Note that since the speech recognition process converts a speech signal into text information, it can also be referred to as a text conversion process.
  • the microphone 13 and the voice recognition unit 14 may be provided inside the voice dialogue device 20 as one function of the voice dialogue device 20 as in the voice synthesis unit 12 or the like, or provided outside the voice dialogue device 20. May be.
  • the voice recognition unit 14 may be realized as a cloud server like the voice synthesis unit 12.
  • the task processing unit 40 is a processing unit that performs processing based on the content of the dialogue between the user and the voice interaction device 20, and outputs information indicating the processing result or related information.
  • the processing by the task processing unit 40 may be any information processing based on the content of the dialogue.
  • the task processing unit 40 may execute a search process for searching a Web page of a restaurant that matches the content of the conversation from a Web page on the Internet, and output the search result. This case will be described below.
  • the unit of execution of processing by the task processing unit 40 is also referred to as a task.
  • the task processing unit 40 corresponds to a processing unit according to the present disclosure.
  • processing for accumulating the contents of the dialogue as data may be executed, and information indicating the success or failure of the processing may be output.
  • the task processing unit 40 may identify an electric device to be controlled among a plurality of electric devices based on the content of the dialogue, and may output specific information or information on the operation of the electric device.
  • the voice dialogue device 20 is a processing device that performs voice dialogue with the user.
  • the spoken dialogue apparatus 20 generates and corrects dialogue information indicating the content of the dialogue with the user, and outputs the dialogue information to the task processing unit 40.
  • the voice interaction device 20 acquires a processing result from the task processing unit 40, presents the acquired processing result to the user, and further continues the dialog with the user.
  • the voice interaction device 20 includes a response sentence generation unit 21, an utterance data acquisition unit 22, a sequence control unit 23, a task control unit 24, an operation unit 25, an analysis unit 26, a memory 27, and a task result analysis unit. 28 and a presentation control unit 29.
  • the response sentence generation unit 21 is a processing unit that acquires a response instruction from the sequence control unit 23 and generates a response sentence based on the acquired response instruction.
  • the response sentence is information transmitted from the voice interaction device 20 to the user. Specifically, a sentence that prompts the user to speak, such as “Please specify a region”, a conflict with the user ’s speech, such as “Acknowledged”, or a voice, such as “Search” This is a sentence explaining the operation of the dialogue apparatus 20. What kind of response instruction is given at what time will be described in detail later.
  • the utterance data acquisition unit 22 is a processing unit that acquires the utterance data of the user's utterance from the voice recognition unit 14.
  • the microphone 13 and the voice recognition unit 14 generate utterance data indicating the content of the utterance
  • the utterance data acquisition unit 22 acquires the generated utterance data.
  • the utterance data acquired by the utterance data acquisition unit 22 may include control terms for changing the content of the dialogue to that at the past time. Utterance data including control terms is also referred to as control utterance data.
  • the utterance data acquisition unit 22 corresponds to one function of the acquisition unit according to the present disclosure.
  • the sequence control unit 23 is a processing unit that realizes a dialogue with the user by controlling a dialogue sequence of the dialogue between the voice dialogue apparatus 20 and the user.
  • the dialogue sequence is data in which utterances by the user in the dialogue and responses by the voice dialogue apparatus 20 are arranged in time series. Note that the sequence control unit 23 corresponds to one function of the acquisition unit according to the present disclosure.
  • the sequence control unit 23 acquires the utterance data of the user's utterance from the utterance data acquisition unit 22. Then, based on the acquired utterance data, the previous interaction sequence with the user, or the processing result acquired from the task result analysis unit 28, an instruction to create a response sentence to be presented to the user (hereinafter referred to as “response”). Is also referred to as “instruction”, and is sent to the response sentence generation unit 21. What kind of response instruction is generated in what case by the sequence control unit 23 will be specifically described later.
  • sequence control unit 23 extracts terms (also referred to as utterance terms) from the acquired utterance data. Furthermore, the sequence control unit 23 stores the extracted term in the slot 31 associated with the attribute of the term via the operation unit 25.
  • the term refers to a relatively short word such as a word. For example, one noun or one adjective corresponds to one term.
  • the task control unit 24 is a processing unit that outputs the content of the dialogue between the voice interactive device 20 and the user to the task processing unit 40 and causes the task processing unit 40 to execute a process based on the output content of the dialogue. Specifically, the task control unit 24 outputs the terms held in the plurality of slots 31 to the task processing unit 40. Further, the task control unit 24 determines whether or not a predetermined condition regarding the state of the plurality of slots 31 is satisfied, and the term held by the plurality of slots 31 is determined only when the predetermined condition is satisfied. You may make it output to the task process part 40.
  • FIG. The task control unit 24 corresponds to one function of the external processing control unit according to the present disclosure.
  • the operation unit 25 is a processing unit that adds, deletes, or changes information indicating the content of the dialogue stored in the memory 27. Specifically, the operation unit 25 changes the term held in the slot 31 when the utterance data acquired by the utterance data acquisition unit 22 includes a control term for controlling dialogue information. That is, the operation unit 25 refers to the history table 32 and changes the term held in each of the plurality of slots 31 to the term held in the slot 31 at the past time point specified by the control term. To do. Further, the operation unit 25 may set a restoration point in a predetermined record on the history table 32 in response to an instruction from the task result analysis unit 28. The operation unit 25 corresponds to one function of the acquisition unit according to the present disclosure and one function of the change unit according to the present disclosure.
  • the analysis unit 26 is a processing unit that analyzes the slot 31 or the history table 32 in the memory 27 and notifies the sequence control unit 23 according to the analysis result. Specifically, the analysis unit 26 determines whether or not each of the slots of the essential slot group of the slots 31 holds the term, and when all the slots of the essential slot group hold the term. Notifies the sequence control unit 23 to that effect.
  • the analysis unit 26 corresponds to one function of the changing unit according to the present disclosure.
  • the analysis unit 26 uses the operation unit 25 to perform a restoration process for restoring the content of the dialogue to a past time point.
  • the analysis unit 26 determines whether there are a plurality of restoration points set in the history table 32. If it is determined that there are a plurality of restoration points, A condition for selecting one of them is sent to the sequence control unit 23. Specific processing contents of the restoration processing will be described in detail later.
  • the memory 27 is a storage device that stores dialogue contents. Specifically, the memory 27 has a slot 31 and a history table 32.
  • the slot 31 is a storage area for holding dialogue information indicating the content of the dialogue, and a plurality of slots are provided in the voice dialogue device 20.
  • Each of the plurality of slots 31 is associated with a term attribute, and holds a term having an attribute associated with the slot 31.
  • the entire terms stored in each of the slots 31 indicate the dialogue information.
  • Each slot 31 holds one term. When a new term is held in the slot 31 in a state where one term is held, the one term held before that is deleted from the slot 31.
  • the term attribute is information indicating the nature, feature or category of the term.
  • the dish name, area, budget, existence of a private room, existence of a parking lot, required time on foot from the nearest station, whether or not chartering is possible, or Information such as whether or not a night view is visible can be used as an attribute.
  • holding a term in the slot 31 can also be expressed as storing or registering a term in the slot 31.
  • the area of the slot 31 in the memory 27 corresponds to a holding unit according to the present disclosure.
  • the slot 31 may be provided with two types, that is, an essential slot and an optional slot.
  • the essential slot is a slot 31 in which the task control unit 24 does not output a term to the task processing unit 40 unless the essential slot holds a term.
  • the option slot is a slot 31 in which the task control unit 24 outputs the term to the task processing unit 40 if all the essential slots hold the term even if the option slot does not hold the term. It is. For example, when a search task is executed as task processing, when the task control unit 24 outputs the terms held in all slots 31 to the task processing unit 40, all slots included in the essential slot group hold the terms. The output may be made only when it is. Whether the slot 31 is an essential slot or an optional slot is predetermined for each slot 31. If the above two types are not provided and there is only one type, all of the slots 31 may be required slots or optional slots. Which of these may be determined as appropriate based on the processing of the task processing unit 40 or the content of the dialogue.
  • the history table 32 is a table showing the history of terms held by the plurality of slots 31. Specifically, the history table 32 is a table in which the terms held in the past by the plurality of slots 31 and the terms currently held are stored in time series. By holding a new term in the slot 31, even when the term held immediately before is deleted from the slot 31, the deleted term remains in the history table 32.
  • the history table 32 may store information indicating the time at that time (for example, a time stamp) together with the terms held by the plurality of slots 31 in the past.
  • the history table 32 may store only terms held by a plurality of slots 31 in the past.
  • the area where the history table 32 is stored corresponds to a storage unit according to the present disclosure.
  • the task result analysis unit 28 is a processing unit that acquires a processing result by the task processing unit 40 and analyzes the acquired processing result.
  • the task result analysis unit 28 analyzes the acquired processing result and passes the analysis result to the sequence control unit 23. This analysis result is used when the operation unit 25 determines whether or not to set a restoration point at a time corresponding to the current time in the history table 32.
  • the task result analysis unit 28 corresponds to one function of the external processing control unit according to the present disclosure.
  • the task result analysis unit 28 acquires the title and URL (Uniform Resource Locator) of the Web page on which the searched information is posted as a result of the restaurant search process by the task processing unit 40. Further, the task result analysis unit 28 analyzes the result of the search process and calculates the number of searched information. Then, the task result analysis unit 28 may set the restoration point only when the number of retrieved information is the number suitable for browsing by the user (for example, about 1 to 30 cases). In addition, the task result analysis unit 28 prohibits setting a restoration point when the number of retrieved information items is not suitable for browsing by the user, such as 0 or 100 or more. It may be.
  • URL Uniform Resource Locator
  • the task result analysis unit 28 may set a restoration point when all of the slots of the essential slot group hold the term, or hold it while the slot 31 holds the term.
  • a restoration point may be set at a point in time when the state changes to a state that holds a term different from the existing term.
  • the presentation control unit 29 is a processing unit that generates presentation data to be presented to the user by the display device 10 and outputs the presentation data to the display device 10.
  • the presentation control unit 29 acquires the processing result from the task processing unit 40, arranges the position on the screen of the display device 10 so that the user can browse the processing result effectively, and outputs it to the display device 10
  • the presentation data is output to the display device 10 after being converted into a suitable data format.
  • part or all of the functions of the voice interaction device 20 and the task processing unit 40 may be realized as a cloud server, like the voice synthesis unit 12 and the like.
  • FIG. 2 is an explanatory diagram of presentation by the voice interaction system 1 according to the present embodiment.
  • the explanatory diagram shown in FIG. 2 is an example of an image displayed on the display screen when the display device 10 presents the processing result by the task processing unit 40 to the user.
  • Character strings 201 to 205 indicating attributes are displayed on the left side of the display screen. Character strings 201 to 205 are character strings indicating attributes of the plurality of slots 31.
  • the terms 211 to 215 are displayed on the right side of the display screen.
  • the terms 211 to 215 are terms held in the slots 31 associated with the attributes of the character strings 201 to 205, respectively.
  • a character string 206 and result information 216 are shown on the lower side of the display screen.
  • the character string 206 is a character string indicating that what is displayed below the character string 206 is a search result.
  • the result information 216 is information indicating a result of the restaurant search performed by the task processing unit 40 based on the terms 211 to 215.
  • the content of the dialogue and the result information that is the processing result by the task processing unit 40 based on the content of the dialogue are displayed on the display device 10, and the user knows the processing result in which the content of the dialogue is reflected. Can do.
  • the image displayed on the display screen is not limited to that shown in FIG. 2, and the displayed information, the presence / absence of display such as its arrangement, and the display position may be arbitrarily changed.
  • FIG. 3 is a first explanatory diagram of a dialogue sequence and history information according to the present embodiment.
  • FIG. 3 shows a dialogue sequence 310, a history table 320, and a search result 330 together with the time series of the dialogue sequence. Note that one row shown in FIG. 3 corresponds to one time point. This line is also called a record.
  • the history table 320 is an example of the history table 32.
  • the dialogue sequence 310 is data in which utterances by the user in the dialogue and responses by the voice dialogue apparatus 20 are arranged in time series.
  • the time information 311 is time information (time stamp) indicating the time when the user uttered or responded by the voice interaction apparatus 20.
  • the utterance 312 is utterance data indicating the utterance by the user at the time. Specifically, the utterance 312 is utterance data indicating the utterance by the user's voice acquired by the utterance data acquisition unit 22 via the microphone 13 and the voice recognition unit 14.
  • the response 313 is a response sentence indicating a response by the voice interaction device 20 at the time. Specifically, the response 313 is generated by the response sentence generation unit 21 in response to a response instruction from the sequence control unit 23.
  • the history table 320 is an example of the history table 32, and includes information on a mandatory slot group 321, an optional slot group 322, an action 323, and a restoration point 324. As shown in FIG. 3, the history table 320 is associated with the dialogue sequence 310 in time series.
  • the essential slot group 321 is a term held in an essential slot among the slots 31 at the time.
  • the essential slot group 321 includes, for example, terms of attributes of “dishes name”, “region”, and “budget”.
  • the option slot group 322 is a term held in the option slot of the slots 31 at the time.
  • the option slot group 322 includes, for example, attribute terms of “presence / absence of private room” and “presence / absence of parking lot”.
  • the action 323 is information indicating processing executed by the voice interaction apparatus 20 at the time point, and a plurality of information may be stored. For example, when a new term is held in a slot 31 with a certain attribute, the name of the attribute and a character string “register” are set at the time point to indicate that. In addition, when the task control unit 24 outputs a term to the task processing unit 40 to search for information, a character string “search” is set. Further, when the operation unit 25 changes the term held in the slot 31 to that at the past time point, the character string “restore” is set.
  • the restoration point 324 is information indicating whether or not a restoration point is set at the time. At the time when the restoration point is set, “1” is set.
  • the task result analysis unit 28 determines whether or not a restoration point is set at the time. When the task result analysis unit 28 determines that the restoration point is set at the time, the operation unit 25 sets the restoration point to the restoration point 324 at the time.
  • the search result 330 is the number of search processing results by the task processing unit 40 at the time.
  • the search result 330 is set by the task result analysis unit 28.
  • FIG. 3 shows an interactive sequence when the user sequentially searches for a restaurant under different search conditions while changing the search conditions.
  • FIG. 3 shows a dialogue sequence when the content of the dialogue is changed to the content of the dialogue at the past time intended by the user.
  • the terms included in the user's utterance are sequentially acquired by the utterance data acquisition unit 22 and the like, and each of the acquired terms is stored in the slot 31 corresponding to the attribute of the term. .
  • the first search processing based on the term held in the slot 31 is performed by the task processing unit 40. This is performed when the term is stored in all the slots 31 included in the essential slot group at the time corresponding to the record R7.
  • search processing based on the terms held in the slot 31 is performed.
  • the search processing is sequentially performed while changing the search word so that the search result desired by the user can be obtained.
  • the user makes a control utterance to return the content of the dialogue to a past time. This is because the search result at the time corresponding to the record R14 or R16 was 0, and the user intended to return to the search conditions of the past time before the number of searches became 0. Is.
  • the voice dialogue apparatus 20 can return the content of the dialogue to a past time point based on the speech by the user's voice, and continuously execute a new dialogue from the state.
  • the voice interaction device can correct the content of the dialogue with the user by a simple method.
  • FIG. 4 is a flowchart of main processing by the voice interaction apparatus 20 according to the present embodiment.
  • the microphone 13 acquires the voice of the user's utterance and generates a voice signal based on the acquired voice.
  • the voice of the utterance by the user may be a voice including a term for restaurant search such as “I want to eat Chinese”, or the slot 31 holds “return to the guard”. It may be a voice including a term for changing the term to be used at a past time.
  • step S102 the voice recognition unit 14 performs voice recognition processing on the voice signal generated by the microphone 13 in step S101, thereby generating utterance data of the user's utterance.
  • step S103 the utterance data acquisition unit 22 acquires the utterance data generated by the voice recognition unit 14 in step S102.
  • step S104 the sequence control unit 23 determines whether or not the utterance data acquired by the utterance data acquisition unit 22 in step S103 is empty.
  • step S104 determines that the utterance data is empty in step S104 ("Y" in step S104)
  • the process proceeds to step S105.
  • it is determined that the utterance data is not empty (“N” in step S104)
  • the process proceeds to step S121.
  • the sequence control unit 23 stores the term included in the utterance data in the slot 31 using the operation unit 25. Specifically, the sequence control unit 23 determines the attribute of the term for each of the terms included in the utterance data, and stores the term in the slot 31 having an attribute that matches the attribute of the term. For example, the sequence control unit 23 determines that the term “Chinese” included in the utterance data “Chinese wants to eat” is a term having a dish name attribute, and the term “Chinese” is a slot having a dish name attribute. 31. At this time, when the term stored in the slot 31 is an abbreviation or common name of the original name, the sequence control unit 23 converts the original name into the original name and stores it in the slot 31. Good. Specifically, the sequence control unit 23 may determine that the term “Chinese” is an abbreviation of “Chinese cuisine” and store “Chinese cuisine” in the slot 31.
  • step S106 the operation unit 25 and the presentation control unit 29 display the terms held in the slot 31 by the display device 10.
  • step S107 the operation unit 25 or the like performs a restoration process for restoring the content of the dialogue by changing the content of the dialogue to that at the past time when necessary. Details of the restoration process will be described later in detail.
  • step S108 the analysis unit 26 determines whether the term is stored in all the slots 31 of the essential slot group, that is, whether all the slots 31 of the essential slot group hold the term.
  • step S108 determines that the term is stored in all the slots 31 in step S108 ("Y" in step S108)
  • the process proceeds to step S109.
  • the analysis unit 26 determines that no term is stored in all the slots 31 (“N” in step S108), that is, if at least one slot 31 in the essential slot group is empty, the step The process proceeds to S122.
  • step S109 the sequence control unit 23 gives the task control unit 24 an execution instruction for causing the task processing unit 40 to execute the task processing.
  • the operation unit 25 records in the history table 32 that the search task has been executed. Specifically, the operation unit 25 sets “search” to the current action 323 in the history table 320.
  • step S110 the task control unit 24 outputs the term held in the slot 31 to the task processing unit 40 based on the execution instruction from the sequence control unit 23 in step S109, and performs search processing on the task processing unit 40. Let it run.
  • the task processing unit 40 acquires the term output by the task control unit 24, performs a search process using the acquired term as a search term, and outputs a search result.
  • step S111 the presentation control unit 29 acquires the search result output by the task processing unit 40 in step S110, and presents the acquired search result to the user in the display device 10 (for example, FIG. 2).
  • the display mode is output to the display device 10.
  • the display device 10 acquires the search result output by the presentation control unit 29 and displays it on the display screen.
  • step S112 the task result analysis unit 28 acquires the search result output by the task processing unit 40 in step S110, and performs restoration point setting processing based on the acquired search result. Details of the restoration point setting process will be described later in detail.
  • step S113 the sequence control unit 23 gives a response instruction to prompt the user for the next utterance to the response sentence generation unit 21.
  • step S114 the response text generation unit 21 generates a response text based on the response instruction.
  • the response sentence generation unit 21 outputs the generated response sentence to the speech synthesizer 12, and outputs the response sentence as a sound from the speaker 11 to allow the user to listen.
  • step S101 is executed again.
  • step S121 the sequence control unit 23 gives a response instruction to the response sentence generation unit 21 to prompt the user to re-utter (perform the same utterance as the previous time).
  • the fact that the utterance data is determined to be empty in step S104 means that the voice recognition unit 14 cannot acquire the utterance data from the sound although the microphone 13 has acquired some sound. . Therefore, it is expected that utterance data can be acquired by requesting the user to perform the same utterance as the previous time.
  • step S122 the sequence control unit 23 gives a response instruction for prompting the user to speak next to the response sentence generating unit 21. For example, when there is a slot 31 that does not hold a term among the slots 31 included in the essential slot group, the sequence control unit 23 causes the user to utter the term that the slot 31 that does not hold the term should hold. A response instruction is generated to generate a response sentence.
  • FIG. 5 is a flowchart of restoration processing by the voice interaction apparatus according to the present embodiment.
  • the flowchart shown in FIG. 5 shows the details of the process of step S107 in FIG. 4, and when the utterance data includes a control term, the term held in the slot 31 is changed to that at the past time point. This shows the processing to be performed.
  • the operation unit 25 determines whether or not the utterance data acquired by the utterance data acquisition unit 22 includes a first term and a second term described later. In addition, when the operation unit 25 determines that the first term and the second term are included, the operation unit 25 refers to the history table to display the term held in each of the plurality of slots 31 in the past time point. It changes to the term which 31 had.
  • the past time point is a time point when the slot 31 (corresponding to the correspondence holding unit according to the present disclosure) associated with the attribute of the second term among the plurality of slots 31 holds the second term. .
  • step S201 the sequence control unit 23 determines whether the utterance data acquired from the utterance data acquisition unit 22 includes a restoration term (also referred to as a first term).
  • the restoration term is a predetermined term indicating that the dialogue information is changed to a past time point, for example, “return to (to)” or “not (to)”. It is a thing.
  • step S201 determines that the restoration term is included in step S201 ("Y" in step S201)
  • the process proceeds to step S202.
  • N no restoration term is included
  • step S202 the analysis unit 26 acquires a term (also referred to as a second term) included in a portion excluding the restoration term in the utterance data, and extracts a restoration point from the history table 32 based on the obtained term. Specifically, the analysis unit 26 determines the attribute of the acquired term, and among the restoration points included in the history table 32, the term held in the slot 31 corresponding to the acquired term attribute is Extract restore points that match the terms you have selected. It can also be said that the utterance data including the first term and the second term is control utterance data. A plurality of setting points may be extracted.
  • step S203 the analysis unit 26 determines whether or not there is one restoration point extracted in step S202.
  • step S203 If the analysis unit 26 determines that there is one restoration point in step S203 ("Y" in step S203), the process proceeds to step S204. On the other hand, when the analysis unit 26 determines that there is not one restoration point (“N” in step S203), the process proceeds to step S211.
  • step S204 the operation unit 25 refers to the history table 32 and changes the term held in the slot 31 to the term held in the slot 31 at the point of the restoration point extracted in step S202. To do. That is, the operation unit 25 changes the terminology held in the plurality of slots 31 so as to return to the term at the time of the restoration point. In addition, the operation unit 25 sets “restore” as the action when the history table 320 changes to the term at the time of the restoration point. Note that the operation unit 25 does not have a term held in the slot 31 at the time of the restoration point, that is, if the slot 31 holds no term at the time of the restoration point, the slot 31 Terminate the term.
  • step S ⁇ b> 211 the sequence control unit 23 gives a response instruction to the response sentence generation unit 21 for a response for prompting the user to extract only one restoration point.
  • a control utterance such as “Return to Moriguchi”
  • the sequence control unit 23 responds with a response “Would it return to the place searched with the parking lot?” Give instructions.
  • step S211 when the user makes an utterance that specifies one of the two restoration points, one restoration point is extracted in step S202 executed from the next main process (FIG. 4). Step S204 is executed.
  • an attribute name that is an attribute name may be used instead of the second term. That is, the operation unit 25 determines whether or not the first term and the attribute name are included in the utterance data acquired by the utterance data acquisition unit 22. When the operation unit 25 determines that the first term and the attribute name are included, the operation unit 25 refers to the history table, and holds the term held in each of the plurality of slots 31 in the past time point. You may make it change into the vocabulary which was doing.
  • the past time point holds the term currently held in the slot 31 (corresponding to the correspondence holding unit according to the present disclosure) associated with the attribute indicated by the attribute name among the plurality of slots 31. It is the time immediately before.
  • FIG. 6 is a flowchart of restoration point setting processing by the voice interaction apparatus according to the present embodiment.
  • the flowchart shown in FIG. 6 shows details of the process in step S112 in FIG.
  • step S301 the operation unit 25 branches the process based on a condition for setting a restoration point.
  • condition C “condition C” in step S301”
  • condition D “condition D” in step S301”
  • the process proceeds to step S303.
  • the case where there are two conditions is shown as an example, but the same processing is possible even when there are three or more conditions.
  • step S302 the operation unit 25 sets a restoration point at the current time point in the history table 320.
  • step S303 the operation unit 25 acquires a search result that is an analysis result of the task result analysis unit 28, and determines whether or not the number of pieces of searched information is zero.
  • step S303 If the number of information retrieved in step S303 is 0 (“Y” in step S303), the operation unit 25 ends a series of processes without setting a restoration point at this time. That is, the operation unit 25 prohibits setting a restoration point when the information included in the information search result is 0 even when the information search result is acquired. On the other hand, when the number of retrieved information is not 0 (“N” in step S303), the process proceeds to step S302.
  • the restoration point may not be set at this point as in the case of zero.
  • FIG. 7 is a second explanatory diagram of history information according to the present embodiment.
  • FIG. 7 shows a dialogue sequence in a dialogue in which a user sequentially searches for a restaurant under different search conditions while changing the search conditions. This is an example of a dialogue sequence when the dialogue content is changed to that at a past time when the dialogue content is different from the user's intention due to misrecognition of voice or the like.
  • FIG. 7 shows a dialogue sequence 310 and the like as in FIG.
  • the terms included in the user's utterance are sequentially acquired by the utterance data acquisition unit 22 and the like, and each of the acquired terms is stored in the slot 31 corresponding to the attribute of the term. .
  • the task processing unit 40 performs an initial search process based on the term held in the slot 31. This is performed when the term is stored in all the slots 31 included in the essential slot group at the time corresponding to the record R5.
  • a search process based on the terms stored in the slot 31 is performed.
  • the search processing is sequentially performed while changing the search word so that the search result desired by the user can be obtained.
  • the term held in the slot 31 is changed to something different from the user's intention due to erroneous recognition by the voice recognition unit 14.
  • the user uttered “parking lot (Chushajomo)” with the intention of adding a parking lot as a search condition. ) ”.
  • the term “Chinese cuisine” is stored in the slot 31 of the attribute of the dish name at the time corresponding to the record R 12.
  • the user uttered “Chuka-janakute-itaria” instead of “Chuka-janakute-itaria” with the intention of correcting the search condition.
  • Iritani is stored in the slot 31 of the regional attribute at the time corresponding to the record R15.
  • the user makes an utterance for returning the content of the dialogue to the past time. This is because the term held in the slot 31 at the time corresponding to the record R12 or R14 has been changed unlike the user's intention, and the user tries to return to the search condition at the past time before the change was made. Is intended.
  • the voice interaction apparatus can return the content of the conversation to a past time point based on the utterance by the user, and continuously execute a new conversation from that state.
  • the voice interaction device can correct the content of the dialogue with the user by a simple method.
  • FIG. 8 is a block diagram showing a configuration of a voice interactive apparatus 20A according to a modification of the present embodiment.
  • the voice interaction apparatus 20 ⁇ / b> A that performs dialogue with the user by voice includes a plurality of holding units 103, a storage unit 104, an acquisition unit 101, and a change unit 102.
  • the plurality of holding units 103 hold dialogue information indicating the content of the dialogue.
  • Each of the plurality of holding units 103 is associated with a term attribute, and holds a term having the attribute.
  • the storage unit 104 stores a history of terms held by the plurality of holding units 103.
  • the acquisition unit 101 acquires utterance data indicating the contents of utterances by the user's voice.
  • the acquisition unit 101 holds the utterance term included in the acquired utterance data in the holding unit 103 associated with the attribute of the utterance term among the plurality of holding units 103.
  • the change unit 102 refers to the storage unit 104 when the acquisition unit 101 acquires control utterance data that includes a control term for controlling the conversation information. change. Specifically, the term held by each of the plurality of holding units 103 is changed to the term held by the holding unit 103 at the past time point specified by the control term.
  • the voice interaction device 20A further outputs the interaction information to a processing unit that performs processing based on the interaction information, using the terms held by each of the plurality of storage units 103 as interaction information, and processes the response as an output response.
  • An external processing control unit 105 that acquires information indicating the result of the above may be provided.
  • the processing unit may perform an information search using a term related to the acquired dialogue information as a search term, and the external processing control unit 105 may acquire a result of the information search as a response.
  • the voice interactive apparatus 20A may further include a presentation control unit 106 for presenting a result of the information search acquired by the external processing control unit 105 to the user.
  • FIG. 9 is a flowchart showing a control method of the voice interactive apparatus 20A according to a modification of the present embodiment.
  • the control method of the voice interaction apparatus 20 ⁇ / b> A that performs voice interaction with the user includes an acquisition step and a change step.
  • the acquisition step utterance data indicating the content of the utterance by the user's voice is acquired (step S 401), and the utterance term included in the acquired utterance data is associated with the attribute of the utterance term in the plurality of holding units 103.
  • Held by the holding unit 103 step S402).
  • the change step when the utterance data acquired in the acquisition step includes a control term for controlling the conversation information, the plurality of holding units 103 holds the history stored in the storage unit 104 with reference to the history stored in the storage unit 104 Change the terminology used. Specifically, the term held by each of the plurality of holding units 103 is changed to the term held by the holding unit 103 at the past time point specified by the control term (step S403).
  • the voice interaction device 20A according to this modification has the same effect as the voice interaction device 20.
  • the voice dialogue apparatus 20 is a voice dialogue apparatus 20 that performs dialogue with a user by voice, and includes a plurality of slots 31 for holding dialogue information indicating the contents of the dialogue.
  • a history table 32 an utterance data acquisition unit 22, and an operation unit 25.
  • Each of the plurality of slots 31 is associated with a term attribute, and holds a term having the associated attribute.
  • the history table 32 stores the history of terms held by the plurality of slots 31.
  • the utterance data acquisition unit 22 acquires utterance data indicating the content of the utterance by the user's voice, and the utterance terms included in the acquired utterance data are assigned to the slots associated with the utterance term attributes among the plurality of slots 31. Hold.
  • the operation unit 25 refers to the history stored in the history table 32 and refers to the plurality of slots 31. Change the terminology held by. Specifically, the term held in each of the plurality of slots 31 is changed to the term held in the slot 31 at the past time point specified by the control term.
  • the voice dialogue device 20 can change the dialogue information to that at the past time point based on the voice of the user, that is, the dialogue information can be returned to the past state.
  • the past time point is a time point determined by the user's voice. Therefore, the user can return the conversation information, which is the content of the conversation with the voice interaction apparatus 20, to the one at the past time point by uttering by voice including the control terms for specifying the past time point. .
  • the voice interaction device 20 can correct the content of the dialogue with the user by a simple method.
  • the voice dialogue apparatus 20 is characterized in that the content of the dialogue with the user is corrected by a simple method by performing control based on the voice of the user. Since it is difficult for the user to grasp the content of the dialogue in time series in the voice dialogue with the conventional voice dialogue apparatus, it is difficult to perform an operation of returning the content of the dialogue to a past time point desired by the user. Since the voice interaction device 20 according to the present embodiment performs control based on the user's voice, the content of the dialogue can be returned to a past time point desired by the user. And it is thought that the superiority of the control based on the user's voice increases as the content of the dialogue becomes more complicated, that is, the number of terms increases.
  • the above-described correction method is highly advantageous. Because, in the case of a voice dialogue apparatus having less than ten holding units as shown in the present embodiment, the term held by the holding unit instead of returning the content of the dialogue with the user to a past time point It is practically possible to reset and reset the terminology from the beginning. However, in the case where the voice interactive apparatus includes several tens or more holding units, it is complicated to reset the terms held by the holding unit from the beginning, which is a heavy burden on the user. It is hard to say that it is possible. In such a case, since the voice interaction device 20 can return the content of the dialog with the user to a past time, it is possible to restart the dialog from a past time desired by the user without resetting from the beginning. There are advantages you can do.
  • control terms may include a first term that is a predetermined term indicating that the dialogue information is changed to a past time point, and a second term that is different from the predetermined term.
  • the operation unit 25 determines whether or not both the first term and the second term are included in the utterance data acquired by the utterance data acquisition unit 22.
  • the operation unit 25 refers to the history and changes the term held in each of the plurality of slots 31. Specifically, when the slot 31 associated with the attribute of the second term among the plurality of slots 31 holds the second term, the term is changed to the term held by each slot 31.
  • the voice interaction device 20 can return the interaction information to a specific point in the past desired by the user by recognizing the control term including the first term and the second term spoken by the user. . As described above, the voice interaction device 20 can more easily identify the past time point to be referred to when returning the content of the dialogue with the user to the past state based on the user's voice.
  • the operation unit 25 may set a restoration point at the above time when the state of the plurality of slots 31 at a certain time on the history stored in the history table 32 satisfies a predetermined condition. Based on the set restoration point, the operation unit 25 changes the term held by the plurality of slots 31 to the term held by the slot 31 at the past time point.
  • the past time is the time when the slot 31 associated with the attribute of the second term among the plurality of slots 31 holds the second term among the times when the restoration point is set.
  • the voice interaction apparatus 20 determines whether or not a restoration point should be set at that time from the state of the plurality of slots 31 at each time stored in the history table 32.
  • a restoration point By appropriately setting the restoration points using predetermined conditions, it is possible to narrow down the time points at which the terms held by the holding unit are to be changed later. Thereby, when changing the term which a holding
  • the voice interaction device 20 may further include a task control unit 24.
  • the task control unit 24 outputs the term held in each of the plurality of slots 31 as dialogue information to the task processing unit 40 that performs processing based on the dialogue information.
  • the task processing unit 40 performs processing based on the output of the task control unit 24.
  • the task control unit 24 acquires information indicating the result of the processing by the task processing unit 40 as a response to the output to the task processing unit 40.
  • the voice interaction apparatus 20 presents the result of processing the terms held by the plurality of holding units by the external processing unit to the user. Therefore, the user can acquire the processing result reflecting the content of the dialogue with the voice dialogue apparatus 20.
  • the task processing unit 40 performs an information search using the acquired term as a search term
  • the task control unit 24 acquires a result of the information search as a response
  • the voice interaction device 20 further includes an external processing control unit. You may provide the presentation control part 29 for showing the acquired information search result to a user.
  • the voice interaction device 20 can acquire the result of the search processing based on the content of the dialogue as a result of the processing by the external processing unit and present it to the user.
  • the operation unit 25 may set a restoration point at the time when the task control unit 24 acquires the information search result in the history.
  • the voice interaction device 20 can return the term held by the holding unit to that at the time when the information search is performed by using the restoration point.
  • the time when the information search is performed is also the time when the result is obtained, and is the time when the user can easily specify in the dialogue.
  • the voice interaction apparatus 20 can return the term held by the holding unit to the one at the time when the user can easily specify intuitively.
  • a more appropriate point in time can be presented as a restoration point candidate.
  • the operation unit 25 determines that the information search result is zero when the information search result includes 0 information. Setting a restore point may be prohibited.
  • the voice interaction apparatus 20 can exclude the time point when the result of the information search is 0 from the time point when the restoration point is set.
  • the voice interaction device 20 can return the content of the dialog with the user to a time useful for the user.
  • the operation unit 25 is specified by the user among the two or more restoration points. You may change the terminology using a restore point.
  • the voice interaction apparatus 20 can return the content of the dialogue with the user to the one at the past time by using one restoration point specified by the user among a plurality of restoration points. Thereby, the user can select the time point that the user considers best from the time point when the voice interaction device 20 determines to be appropriate, and can return to the interaction information at the time point of selection.
  • the voice interaction device 20 may further include a response sentence generation unit 21 that generates a response sentence for accepting one restoration point used for changing a term out of two or more restoration points from the user.
  • the user can know from the response sentence that there are a plurality of candidates when the voice dialogue device 20 returns the contents of the dialogue.
  • the user specifies the time point for returning the content of the dialogue by responding to the response sentence. That is, the voice interactive apparatus 20 causes the user to specify one restoration point from among a plurality of restoration points.
  • the voice interaction device 20 can specifically accept the designation of the restoration point from the user, and return the content of the dialogue with the user to that at the past time.
  • the control term may include a first term that is a predetermined term indicating that the dialogue information is changed to a past time point and an attribute name that is an attribute name of the acquired term. Good.
  • the operation unit 25 determines whether or not the utterance data acquired by the utterance data acquisition unit 22 includes the first term and the attribute name. When the operation unit 25 determines that the first term and the attribute name are included in the utterance data, the operation unit 25 refers to the history and changes the term held in each of the plurality of slots 31. Specifically, the term held in each slot 31 immediately before holding the term currently held in the slot 31 associated with the attribute indicated by the attribute name among the plurality of slots 31 change.
  • the voice interaction device 20 specifically specifies a past time point using the name of the attribute with which the holding unit is associated. Even if the user does not specify a specific condition, the user can specify the point in time when the content of the dialogue is returned simply by specifying the attribute name. Thus, the voice interaction device 20 can correct the content of the dialogue with the user by a more specific method.
  • the voice dialogue system 1 performs voice dialogue with the user.
  • the voice dialogue system 1 includes a plurality of slots 31 for holding dialogue information indicating the contents of dialogue, a history table 32, an utterance data acquisition unit 22, an operation unit 25, a microphone 13, and a voice recognition unit 14.
  • the task processing unit 40, the speech synthesis unit 12, the speaker 11, and the display device 10 are provided.
  • Each of the plurality of slots 31 is associated with a term attribute, and holds a term having the associated attribute.
  • the history table 32 stores the history of terms held by the plurality of slots 31.
  • the utterance data acquisition unit 22 acquires utterance data indicating the content of the utterance by the user's voice, and the utterance term included in the acquired utterance data is associated with the attribute of the utterance term among the plurality of slots 31. To hold.
  • the operation unit 25 changes the terms held by the plurality of slots 31. Specifically, the operation unit 25 refers to the history stored in the history table 32 and determines the term held in each of the plurality of slots 31 at the past time point specified by the control term. It changes to the term which 31 had.
  • the microphone 13 acquires the user's voice and generates a voice signal.
  • the speech recognition unit 14 generates speech data acquired by the speech data acquisition unit 22 by performing speech recognition processing on the speech signal generated by the microphone 13.
  • the task processing unit 40 acquires dialogue information held by the plurality of slots 31, performs predetermined processing on the acquired dialogue information, and outputs information indicating the processing result.
  • the speech synthesizer 12 generates a response sentence for an utterance by the user's voice, and generates a speech signal by performing a speech synthesis process on the generated response sentence.
  • the speaker 11 outputs the voice signal generated by the voice synthesizer 12 as voice.
  • the display device 10 displays the processing result output by the task processing unit 40.
  • the method for controlling the voice interaction apparatus can be used for controlling the voice interaction apparatus 20 that performs voice conversation with the user.
  • the voice interaction device 20 includes a plurality of slots 31 for holding dialogue information indicating the content of the dialogue and a history table 32.
  • Each of the plurality of slots 31 is associated with a term attribute, and holds a term having the associated attribute.
  • the history table 32 stores the history of terms held by the plurality of slots 31.
  • the control method of the voice interaction apparatus 20 according to the present embodiment includes an acquisition step and a change step.
  • the acquisition step utterance data indicating the content of the utterance by the user's voice is acquired, and the utterance term included in the acquired utterance data is held in the slot 31 associated with the attribute of the utterance term among the plurality of slots 31.
  • the change step when the utterance data acquired in the acquisition step includes a control term for controlling the dialogue information, the term held in the plurality of slots 31 is changed. Specifically, referring to the history stored in the history table 32, the term held in each of the plurality of slots 31 is held in the slot 31 at the past time point specified by the control term. Change to a different term.
  • the present disclosure is useful as a voice dialogue apparatus that can correct the content of dialogue with the user by a simple method.
  • the present disclosure can be applied to an application of a car navigation device, a smartphone (high-function mobile phone terminal), a mobile phone terminal, a mobile information terminal, or a PC (Personal Computer).

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Acoustics & Sound (AREA)
  • Theoretical Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Multimedia (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Databases & Information Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Machine Translation (AREA)

Abstract

This voice interaction device (20A) for interacting with a user by voice, is provided with an acquisition unit (101) which acquires utterance data indicating the content of a speech utterance of a user, multiple storage units (103) which associate and store words contained in the utterance data and attributes of those words, a memory unit (104) which stores the history of words stored by the storage units (103), and a modification unit (102). If prescribed control words are included in the utterance data, the modification unit (102) refers to the history stored in the memory unit (104) and modifies the multiple storage units (103) to store the words that were stored by the storage units (103) at a point in time in the past specified by the control words.

Description

音声対話装置、音声対話システム、および、音声対話装置の制御方法Spoken dialogue apparatus, spoken dialogue system, and control method of spoken dialogue apparatus
 本開示は、音声対話装置、音声対話システム、および、音声対話装置の制御方法に関する。 The present disclosure relates to a voice dialogue apparatus, a voice dialogue system, and a control method for the voice dialogue apparatus.
 特許文献1は、利用者から入力された情報に基づいて、次に入力されると期待される語彙群を利用者が視認可能なように提示する対話列認識装置を開示する。これにより、対話の誤認識に起因して利用者が途方に暮れるという不都合が防止される。 Patent Document 1 discloses a dialog sequence recognition device that presents a vocabulary group expected to be input next so that the user can visually recognize it based on information input from the user. This prevents the inconvenience that the user is at a loss due to erroneous recognition of the dialogue.
特開2001-34292号公報JP 2001-34292 A
 本開示は、ユーザとの対話の内容を簡易な方法により修正する音声対話装置を提供する。 This disclosure provides a speech dialogue apparatus that modifies the content of dialogue with a user by a simple method.
 本開示における音声対話装置は、ユーザとの音声による対話を行う音声対話装置であって、取得部と、複数の保持部と、記憶部と、変更部とを備える。取得部は、ユーザの音声による発話の内容を示す発話データを取得する。複数の保持部はそれぞれ、発話データに含まれる用語と、同用語の属性とを対応づけて。記憶部は、複数の保持部が保持する用語の履歴を記憶する。変更部は、発話データに所定の制御用語が含まれる場合に、記憶部が記憶している履歴を参照し、過去の時点において複数の保持部が保持していた用語となるように、複数の保持部の用語を変更する。ここで、過去の時点とは、制御用語により特定される過去の時点である。 The voice dialogue apparatus according to the present disclosure is a voice dialogue apparatus that performs voice dialogue with a user, and includes an acquisition unit, a plurality of holding units, a storage unit, and a change unit. The acquisition unit acquires utterance data indicating the content of the utterance by the user's voice. Each of the plurality of holding units associates the term included in the utterance data with the attribute of the term. The storage unit stores a history of terms held by the plurality of holding units. When a predetermined control term is included in the utterance data, the changing unit refers to the history stored in the storage unit, so that the plurality of holding units at the past time point become the terms held by the plurality of holding units. Change the terminology of the holding part. Here, the past time point is a past time point specified by the control term.
 本開示における音声対話装置は、ユーザとの対話の内容を簡易な方法により修正するのに有効である。 The speech dialogue apparatus according to the present disclosure is effective for correcting the content of dialogue with the user by a simple method.
図1は、実施の形態に係る音声対話装置及び音声対話システムの構成を示すブロック図である。FIG. 1 is a block diagram illustrating a configuration of a voice interaction apparatus and a voice interaction system according to an embodiment. 図2は、実施の形態に係る音声対話システムによる提示について説明する図である。FIG. 2 is a diagram for explaining presentation by the voice interaction system according to the embodiment. 図3は、実施の形態に係る対話シーケンス及び履歴情報について説明する図である。FIG. 3 is a diagram for explaining a dialogue sequence and history information according to the embodiment. 図4は、実施の形態に係る音声対話装置によるメイン処理のフロー図である。FIG. 4 is a flowchart of a main process performed by the voice interaction apparatus according to the embodiment. 図5は、実施の形態に係る音声対話装置による復元処理のフロー図である。FIG. 5 is a flowchart of restoration processing by the voice interaction apparatus according to the embodiment. 図6は、実施の形態に係る音声対話装置による復元ポイント設定処理のフロー図である。FIG. 6 is a flowchart of restoration point setting processing by the voice interaction apparatus according to the embodiment. 図7は、実施の形態に係る対話シーケンス及び履歴情報について説明する図である。FIG. 7 is a diagram for explaining a dialogue sequence and history information according to the embodiment. 図8は、実施の形態の変形例に係る音声対話装置の構成を示すブロック図である。FIG. 8 is a block diagram showing a configuration of a voice interactive apparatus according to a modification of the embodiment. 図9は、実施の形態の変形例に係る音声対話装置の制御方法を示すフロー図である。FIG. 9 is a flowchart showing a control method of the voice interactive apparatus according to the modification of the embodiment.
 以下、適宜図面を参照しながら、実施の形態を詳細に説明する。但し、必要以上に詳細な説明は省略する場合がある。例えば、既によく知られた事項の詳細説明や実質的に同一の構成に対する重複説明を省略する場合がある。これは、以下の説明が不必要に冗長になるのを避け、当業者の理解を容易にするためである。 Hereinafter, embodiments will be described in detail with reference to the drawings as appropriate. However, more detailed description than necessary may be omitted. For example, detailed descriptions of already well-known matters and repeated descriptions for substantially the same configuration may be omitted. This is to avoid the following description from becoming unnecessarily redundant and to facilitate understanding by those skilled in the art.
 なお、発明者(ら)は、当業者が本開示を十分に理解するために添付図面および以下の説明を提供するのであって、これらによって請求の範囲に記載の主題を限定することを意図するものではない。 In addition, the inventor (s) provides the accompanying drawings and the following description in order for those skilled in the art to fully understand the present disclosure, and these are intended to limit the claimed subject matter. It is not a thing.
 (実施の形態)
 本実施の形態において、ユーザとの対話の内容を簡易な方法により修正する音声対話装置について説明する。本実施の形態に係る音声対話装置は、ユーザとの音声による対話を行うものであり、ユーザとの対話の内容を示す対話情報を生成及び修正し、その対話情報を外部の処理装置に出力する。また、音声対話装置は、外部の処理装置から処理結果を取得しユーザに提示し、さらにユーザとの対話を継続する。このように、音声対話装置は、ユーザとの対話に基づいて、対話情報を生成及び修正しながら、順次、処理結果をユーザに提示するものである。
(Embodiment)
In the present embodiment, a voice dialogue apparatus for correcting the contents of dialogue with a user by a simple method will be described. The voice dialogue apparatus according to the present embodiment performs voice dialogue with the user, generates and corrects dialogue information indicating the content of the dialogue with the user, and outputs the dialogue information to an external processing device. . Further, the voice interaction device acquires the processing result from the external processing device and presents it to the user, and further continues the dialogue with the user. As described above, the voice interaction device sequentially presents the processing results to the user while generating and correcting the interaction information based on the interaction with the user.
 なお、音声対話装置は、ユーザによるキー入力又はパネルへの接触などの操作が不可能又は困難である場合に有用である。例えば、ユーザが自動車等を運転しているときにユーザの音声による指示を順次受けながら情報検索をするカーナビゲーション装置などの用途があり得る。また、キー又はパネルのようなユーザインタフェースを有さない音声対話装置でも有用である。 Note that the voice interaction device is useful when an operation such as key input by the user or touching the panel is impossible or difficult. For example, there may be applications such as a car navigation device that searches for information while sequentially receiving instructions by the user's voice when the user is driving a car or the like. It is also useful in a voice interaction device that does not have a user interface such as a key or a panel.
 [1-1.構成]
 図1は、本実施の形態に係る音声対話装置20及び音声対話システム1の構成を示すブロック図である。
[1-1. Constitution]
FIG. 1 is a block diagram showing a configuration of a voice interaction device 20 and a voice interaction system 1 according to the present embodiment.
 図1に示されるように、音声対話システム1は、表示装置10と、スピーカ11と、音声合成部12と、マイク13と、音声認識部14と、音声対話装置20と、タスク処理部40とを備える。 As shown in FIG. 1, the voice dialogue system 1 includes a display device 10, a speaker 11, a voice synthesis unit 12, a microphone 13, a voice recognition unit 14, a voice dialogue device 20, and a task processing unit 40. Is provided.
 表示装置10は、表示画面を備える表示装置である。表示装置10は、音声対話装置20から取得する表示データに基づいて表示画面に映像を表示する。表示装置10は、例えば、カーナビゲーション装置、スマートフォン(高機能携帯電話端末)、携帯電話端末、携帯情報端末、ディスプレイ、又は、PC(Personal Computer)などにより実現される。なお、表示装置10は、音声対話装置20が提示する情報に基づく映像を表示する装置の例として示したが、表示装置10の代わりに、音声対話装置20が提示する情報を音声として出力するスピーカを用いてもよい。このスピーカは、後述のスピーカ11と共用してもよい。 The display device 10 is a display device having a display screen. The display device 10 displays an image on the display screen based on the display data acquired from the voice interaction device 20. The display device 10 is realized by, for example, a car navigation device, a smartphone (high-function mobile phone terminal), a mobile phone terminal, a mobile information terminal, a display, or a PC (Personal Computer). Although the display device 10 is shown as an example of a device that displays an image based on information presented by the voice interaction device 20, a speaker that outputs information presented by the voice interaction device 20 as a voice instead of the display device 10. May be used. This speaker may be shared with the speaker 11 described later.
 スピーカ11は、音声を出力するスピーカである。スピーカ11は、音声合成部12から取得する音声信号に基づいて音声を出力する。スピーカ11が出力した音声は、ユーザに聴取される。 Speaker 11 is a speaker that outputs sound. The speaker 11 outputs sound based on the sound signal acquired from the sound synthesizer 12. The sound output from the speaker 11 is heard by the user.
 音声合成部12は、応答文を音声信号に変換する処理部である。音声合成部12は、音声対話装置20からユーザへ伝達する情報である応答文を音声対話装置20から取得し、スピーカにより出力するための音声信号を、取得した応答文に基づいて生成する。 The speech synthesis unit 12 is a processing unit that converts a response sentence into a speech signal. The voice synthesizing unit 12 acquires a response sentence, which is information transmitted from the voice dialogue apparatus 20 to the user, from the voice dialogue apparatus 20, and generates a voice signal to be output by the speaker based on the obtained response sentence.
 なお、スピーカ11及び音声合成部12は、音声対話装置20の一機能として音声対話装置20の内部に備えられてもよいし、音声対話装置20の外部に備えられてもよい。また、音声合成部12は、音声対話装置20とインターネット等のネットワーク経由で通信可能なように、いわゆるクラウドサーバとして実現されてもよい。その場合、音声合成部12と音声対話装置20との接続、及び、音声合成部12とスピーカ11との接続は、インターネット等を介した通信路を通じてなされる。 The speaker 11 and the voice synthesis unit 12 may be provided inside the voice dialogue apparatus 20 as one function of the voice dialogue apparatus 20 or may be provided outside the voice dialogue apparatus 20. Further, the voice synthesizer 12 may be realized as a so-called cloud server so as to be able to communicate with the voice interaction device 20 via a network such as the Internet. In that case, the connection between the voice synthesizer 12 and the voice interaction device 20 and the connection between the voice synthesizer 12 and the speaker 11 are made through a communication path via the Internet or the like.
 マイク13は、音声を取得するマイクロホンである。マイク13は、ユーザの音声を取得し、取得した音声に基づく音声信号を出力する。 The microphone 13 is a microphone that acquires sound. The microphone 13 acquires the user's voice and outputs an audio signal based on the acquired voice.
 音声認識部14は、ユーザの音声を対象として音声認識を行うことで、発話データを生成する処理部である。音声認識部14は、マイク13が生成した音声信号を取得し、取得した音声信号に対して音声認識処理を施すことで、ユーザによる発話の発話データを生成する。発話データは、ユーザから音声対話装置20へ伝達する情報であり、「中華が食べたい」というように、文字(テキスト)で表現されるものである。なお、音声認識処理は、音声信号をテキスト情報に変換するものであるので、テキスト変換処理ということもできる。 The voice recognition unit 14 is a processing unit that generates speech data by performing voice recognition on the user's voice. The voice recognition unit 14 acquires the voice signal generated by the microphone 13 and performs voice recognition processing on the acquired voice signal, thereby generating utterance data of the user's utterance. The utterance data is information transmitted from the user to the voice interaction device 20, and is expressed by characters (text) such as “I want to eat Chinese”. Note that since the speech recognition process converts a speech signal into text information, it can also be referred to as a text conversion process.
 なお、マイク13及び音声認識部14は、音声合成部12等と同様、音声対話装置20の一機能として音声対話装置20の内部に備えられてもよいし、音声対話装置20の外部に備えられてもよい。また、音声認識部14は、音声合成部12同様、クラウドサーバとして実現されてもよい。 Note that the microphone 13 and the voice recognition unit 14 may be provided inside the voice dialogue device 20 as one function of the voice dialogue device 20 as in the voice synthesis unit 12 or the like, or provided outside the voice dialogue device 20. May be. In addition, the voice recognition unit 14 may be realized as a cloud server like the voice synthesis unit 12.
 タスク処理部40は、ユーザと音声対話装置20との対話の内容に基づいて処理を行い、その処理結果を示す情報又はその関連情報を出力する処理部である。タスク処理部40による処理は、対話の内容に基づく情報処理であればどのようなものであってもよい。例えば、タスク処理部40は、インターネット上のWebページから、対話の内容に適合するレストランのWebページを検索する検索処理を実行し、その検索結果を出力するものとしてもよい。この場合を以下で説明する。なお、タスク処理部40による処理の実行単位のことをタスクともいう。また、タスク処理部40は、本開示に係る処理部に相当する。 The task processing unit 40 is a processing unit that performs processing based on the content of the dialogue between the user and the voice interaction device 20, and outputs information indicating the processing result or related information. The processing by the task processing unit 40 may be any information processing based on the content of the dialogue. For example, the task processing unit 40 may execute a search process for searching a Web page of a restaurant that matches the content of the conversation from a Web page on the Internet, and output the search result. This case will be described below. Note that the unit of execution of processing by the task processing unit 40 is also referred to as a task. The task processing unit 40 corresponds to a processing unit according to the present disclosure.
 なお、タスク処理部40による処理の他の例として、対話の内容をデータとして蓄積する処理を実行し、その処理の成否を示す情報を出力するものとしてもよい。また、タスク処理部40は、対話の内容に基づいて複数の電気機器のうち制御対象の電気機器を特定し、その電気機器の固有情報又は動作に関する情報を出力するものとしてもよい。 As another example of processing by the task processing unit 40, processing for accumulating the contents of the dialogue as data may be executed, and information indicating the success or failure of the processing may be output. In addition, the task processing unit 40 may identify an electric device to be controlled among a plurality of electric devices based on the content of the dialogue, and may output specific information or information on the operation of the electric device.
 音声対話装置20は、ユーザとの音声による対話を行う処理装置である。音声対話装置20は、ユーザとの対話の内容を示す対話情報を生成及び修正し、その対話情報をタスク処理部40に出力する。また、音声対話装置20は、タスク処理部40から処理結果を取得し、取得した処理結果をユーザに提示し、さらにユーザとの対話を継続する。 The voice dialogue device 20 is a processing device that performs voice dialogue with the user. The spoken dialogue apparatus 20 generates and corrects dialogue information indicating the content of the dialogue with the user, and outputs the dialogue information to the task processing unit 40. In addition, the voice interaction device 20 acquires a processing result from the task processing unit 40, presents the acquired processing result to the user, and further continues the dialog with the user.
 音声対話装置20は、応答文生成部21と、発話データ取得部22と、シーケンス制御部23と、タスク制御部24と、操作部25と、解析部26と、メモリ27と、タスク結果解析部28と、提示制御部29とを備える。 The voice interaction device 20 includes a response sentence generation unit 21, an utterance data acquisition unit 22, a sequence control unit 23, a task control unit 24, an operation unit 25, an analysis unit 26, a memory 27, and a task result analysis unit. 28 and a presentation control unit 29.
 応答文生成部21は、シーケンス制御部23から応答指示を取得し、取得した応答指示に基づいて応答文を生成する処理部である。応答文は、音声対話装置20からユーザへ伝達する情報である。具体的には、「地域を指定下さい」というようなユーザに対して発話を促すための文章、「承知しました」というようなユーザの発話に対する相槌、又は、「検索します」というような音声対話装置20の動作を説明する文章である。どのようなときにどのような応答指示をするかについては、後で詳細に説明する。 The response sentence generation unit 21 is a processing unit that acquires a response instruction from the sequence control unit 23 and generates a response sentence based on the acquired response instruction. The response sentence is information transmitted from the voice interaction device 20 to the user. Specifically, a sentence that prompts the user to speak, such as “Please specify a region”, a conflict with the user ’s speech, such as “Acknowledged”, or a voice, such as “Search” This is a sentence explaining the operation of the dialogue apparatus 20. What kind of response instruction is given at what time will be described in detail later.
 発話データ取得部22は、ユーザによる発話の発話データを音声認識部14から取得する処理部である。ユーザの音声による発話がなされた場合、マイク13及び音声認識部14により、上記発話の内容を示す発話データが生成され、この生成された発話データを発話データ取得部22が取得する。また、発話データ取得部22が取得する発話データは、対話の内容を過去の時点におけるものに変更するための制御用語を含むこともある。制御用語を含む発話データのことを制御発話データともいう。なお、発話データ取得部22は、本開示に係る取得部の一機能に相当する。 The utterance data acquisition unit 22 is a processing unit that acquires the utterance data of the user's utterance from the voice recognition unit 14. When the user's voice is uttered, the microphone 13 and the voice recognition unit 14 generate utterance data indicating the content of the utterance, and the utterance data acquisition unit 22 acquires the generated utterance data. Further, the utterance data acquired by the utterance data acquisition unit 22 may include control terms for changing the content of the dialogue to that at the past time. Utterance data including control terms is also referred to as control utterance data. Note that the utterance data acquisition unit 22 corresponds to one function of the acquisition unit according to the present disclosure.
 シーケンス制御部23は、音声対話装置20とユーザとの対話の対話シーケンスを制御することで、ユーザとの対話を実現する処理部である。ここで、対話シーケンスとは、対話におけるユーザによる発話と音声対話装置20による応答とを時系列で並べたデータのことである。なお、シーケンス制御部23は、本開示に係る取得部の一機能に相当する。 The sequence control unit 23 is a processing unit that realizes a dialogue with the user by controlling a dialogue sequence of the dialogue between the voice dialogue apparatus 20 and the user. Here, the dialogue sequence is data in which utterances by the user in the dialogue and responses by the voice dialogue apparatus 20 are arranged in time series. Note that the sequence control unit 23 corresponds to one function of the acquisition unit according to the present disclosure.
 具体的には、シーケンス制御部23は、ユーザによる発話の発話データを発話データ取得部22から取得する。そして、取得した発話データ、これまでのユーザとの対話シーケンス、又は、タスク結果解析部28から取得する処理結果に基づいて、次にユーザに提示すべき応答文を作成する指示(以降、「応答指示」ともいう)を生成し、応答文生成部21に送る。シーケンス制御部23がどのような場合にどのような応答指示を生成するかについては、後で具体的に説明する。 Specifically, the sequence control unit 23 acquires the utterance data of the user's utterance from the utterance data acquisition unit 22. Then, based on the acquired utterance data, the previous interaction sequence with the user, or the processing result acquired from the task result analysis unit 28, an instruction to create a response sentence to be presented to the user (hereinafter referred to as “response”). Is also referred to as “instruction”, and is sent to the response sentence generation unit 21. What kind of response instruction is generated in what case by the sequence control unit 23 will be specifically described later.
 また、シーケンス制御部23は、取得した発話データから用語(発話用語ともいう)を抽出する。さらにシーケンス制御部23は、操作部25を介して、抽出した用語をその用語の属性に対応付けられたスロット31に格納する。ここで、用語とは、単語のように比較的短い語のことをいい、例えば、1つの名詞、又は、1つの形容詞などが1つの用語に相当する。 Also, the sequence control unit 23 extracts terms (also referred to as utterance terms) from the acquired utterance data. Furthermore, the sequence control unit 23 stores the extracted term in the slot 31 associated with the attribute of the term via the operation unit 25. Here, the term refers to a relatively short word such as a word. For example, one noun or one adjective corresponds to one term.
 タスク制御部24は、音声対話装置20とユーザとの対話の内容をタスク処理部40に出力し、出力した対話の内容に基づく処理をタスク処理部40に実行させる処理部である。具体的には、タスク制御部24は、複数のスロット31が保持している用語をタスク処理部40に出力する。また、タスク制御部24は、複数のスロット31の状態についての所定の条件が満たされるか否かを判定し、所定の条件が満たされる場合にのみ、複数のスロット31が保持している用語をタスク処理部40に出力するようにしてもよい。なお、タスク制御部24は、本開示に係る外部処理制御部の一機能に相当する。 The task control unit 24 is a processing unit that outputs the content of the dialogue between the voice interactive device 20 and the user to the task processing unit 40 and causes the task processing unit 40 to execute a process based on the output content of the dialogue. Specifically, the task control unit 24 outputs the terms held in the plurality of slots 31 to the task processing unit 40. Further, the task control unit 24 determines whether or not a predetermined condition regarding the state of the plurality of slots 31 is satisfied, and the term held by the plurality of slots 31 is determined only when the predetermined condition is satisfied. You may make it output to the task process part 40. FIG. The task control unit 24 corresponds to one function of the external processing control unit according to the present disclosure.
 操作部25は、メモリ27に格納されている対話の内容を示す情報を追加、削除又は変更する処理部である。具体的には、操作部25は、発話データ取得部22が取得した発話データに、対話情報を制御するための制御用語が含まれる場合に、スロット31が保持している用語を変更する。すなわち、操作部25は、履歴テーブル32を参照して、複数のスロット31のそれぞれが保持している用語を、制御用語により特定される過去の時点において当該スロット31が保持していた用語に変更する。また、操作部25は、タスク結果解析部28からの指示を受けて、履歴テーブル32上の所定のレコードに復元ポイントを設定してもよい。なお、操作部25は、本開示に係る取得部の一機能、及び、本開示に係る変更部の一機能に相当する。 The operation unit 25 is a processing unit that adds, deletes, or changes information indicating the content of the dialogue stored in the memory 27. Specifically, the operation unit 25 changes the term held in the slot 31 when the utterance data acquired by the utterance data acquisition unit 22 includes a control term for controlling dialogue information. That is, the operation unit 25 refers to the history table 32 and changes the term held in each of the plurality of slots 31 to the term held in the slot 31 at the past time point specified by the control term. To do. Further, the operation unit 25 may set a restoration point in a predetermined record on the history table 32 in response to an instruction from the task result analysis unit 28. The operation unit 25 corresponds to one function of the acquisition unit according to the present disclosure and one function of the change unit according to the present disclosure.
 解析部26は、メモリ27内のスロット31又は履歴テーブル32を解析し、解析結果に応じた通知をシーケンス制御部23に行う処理部である。具体的には、解析部26は、スロット31のうちの必須スロット群のスロットそれぞれが用語を保持しているか否かを判定し、必須スロット群のすべてのスロットが用語を保持している場合には、その旨をシーケンス制御部23に通知する。なお、解析部26は、本開示に係る変更部の一機能に相当する。 The analysis unit 26 is a processing unit that analyzes the slot 31 or the history table 32 in the memory 27 and notifies the sequence control unit 23 according to the analysis result. Specifically, the analysis unit 26 determines whether or not each of the slots of the essential slot group of the slots 31 holds the term, and when all the slots of the essential slot group hold the term. Notifies the sequence control unit 23 to that effect. The analysis unit 26 corresponds to one function of the changing unit according to the present disclosure.
 また、解析部26は、操作部25を利用して、対話の内容を過去の時点に復元するための復元処理を行う。解析部26は、復元処理を行う際に、履歴テーブル32内に設定された復元ポイントが複数あるか否かを判定し、複数の復元ポイントがあると判定した場合には、複数の復元ポイントの中から1つを選択するための条件をシーケンス制御部23に送る。復元処理の具体的な処理内容については後で詳しく説明する。 In addition, the analysis unit 26 uses the operation unit 25 to perform a restoration process for restoring the content of the dialogue to a past time point. When performing the restoration process, the analysis unit 26 determines whether there are a plurality of restoration points set in the history table 32. If it is determined that there are a plurality of restoration points, A condition for selecting one of them is sent to the sequence control unit 23. Specific processing contents of the restoration processing will be described in detail later.
 メモリ27は、対話の内容を記憶している記憶装置である。具体的には、メモリ27は、スロット31及び履歴テーブル32を有する。 The memory 27 is a storage device that stores dialogue contents. Specifically, the memory 27 has a slot 31 and a history table 32.
 スロット31は、対話の内容を示す対話情報を保持するための記憶領域であり、音声対話装置20に複数備えられる。複数のスロット31は、それぞれが用語の属性に対応付けられており、それぞれが当該スロット31に対応付けられた属性を有する用語を保持する。そして、スロット31のそれぞれに格納された用語全体が、上記対話情報を示している。スロット31は、それぞれが1つの用語を保持する。そして、スロット31は、1つの用語を保持している状態において新たな用語を保持した場合には、その前に保持していた1つの用語はスロット31上からは消去される。 The slot 31 is a storage area for holding dialogue information indicating the content of the dialogue, and a plurality of slots are provided in the voice dialogue device 20. Each of the plurality of slots 31 is associated with a term attribute, and holds a term having an attribute associated with the slot 31. The entire terms stored in each of the slots 31 indicate the dialogue information. Each slot 31 holds one term. When a new term is held in the slot 31 in a state where one term is held, the one term held before that is deleted from the slot 31.
 ここで、用語の属性とは、当該用語の性質、特徴又はカテゴリを示す情報のことである。例えば、タスク処理部40の処理がレストラン検索の場合、料理名、地域、予算、個室の有無、駐車場の有無、最寄駅からの徒歩での所要時間、貸切が可能か否か、又は、夜景が見えるか否かというような情報を属性として用いることができる。なお、スロット31が用語を保持することを、スロット31に用語が格納される、又は、登録される、と表現することもできる。なお、メモリ27のうちのスロット31の領域は、本開示に係る保持部に相当する。 Here, the term attribute is information indicating the nature, feature or category of the term. For example, when the processing of the task processing unit 40 is a restaurant search, the dish name, area, budget, existence of a private room, existence of a parking lot, required time on foot from the nearest station, whether or not chartering is possible, or Information such as whether or not a night view is visible can be used as an attribute. Note that holding a term in the slot 31 can also be expressed as storing or registering a term in the slot 31. Note that the area of the slot 31 in the memory 27 corresponds to a holding unit according to the present disclosure.
 また、スロット31には、必須スロット及びオプションスロットという2つの種別が設けられていてもよい。必須スロットとは、当該必須スロットが用語を保持していないとタスク制御部24がタスク処理部40に用語を出力しないスロット31のことである。また、オプションスロットとは、当該オプションスロットが用語を保持していなくても、すべての必須スロットが用語を保持していればタスク制御部24がタスク処理部40に用語を出力するスロット31のことである。例えば、タスク処理として検索タスクを実行させる場合、すべてのスロット31が保持している用語をタスク制御部24がタスク処理部40に出力する際、必須スロット群に含まれるすべてのスロットが用語を保持している場合に限り出力するようにしてもよい。スロット31が、必須スロット及びオプションスロットのうちのどちらであるかは、スロット31ごとに予め定められている。なお、上記2つの種別が設けられず、種別が1つだけである場合には、スロット31の全てを必須スロットとしてもよいし、オプションスロットとしてもよい。これらのどちらにするかは、タスク処理部40の処理、又は、対話の内容に基づいて適宜定められてよい。 In addition, the slot 31 may be provided with two types, that is, an essential slot and an optional slot. The essential slot is a slot 31 in which the task control unit 24 does not output a term to the task processing unit 40 unless the essential slot holds a term. The option slot is a slot 31 in which the task control unit 24 outputs the term to the task processing unit 40 if all the essential slots hold the term even if the option slot does not hold the term. It is. For example, when a search task is executed as task processing, when the task control unit 24 outputs the terms held in all slots 31 to the task processing unit 40, all slots included in the essential slot group hold the terms. The output may be made only when it is. Whether the slot 31 is an essential slot or an optional slot is predetermined for each slot 31. If the above two types are not provided and there is only one type, all of the slots 31 may be required slots or optional slots. Which of these may be determined as appropriate based on the processing of the task processing unit 40 or the content of the dialogue.
 履歴テーブル32は、複数のスロット31が保持する用語の履歴を示すテーブルである。具体的には、履歴テーブル32は、複数のスロット31が過去に保持していた用語、及び、現在保持している用語が時系列で収められたテーブルである。スロット31が新たな用語を保持することで、その直前に保持していた用語をスロット31上から消去した場合でも、その消去された用語は、履歴テーブル32には残されている。 The history table 32 is a table showing the history of terms held by the plurality of slots 31. Specifically, the history table 32 is a table in which the terms held in the past by the plurality of slots 31 and the terms currently held are stored in time series. By holding a new term in the slot 31, even when the term held immediately before is deleted from the slot 31, the deleted term remains in the history table 32.
 なお、履歴テーブル32には、過去に複数のスロット31が保持した用語と共に、その時点での時刻を示す情報(例えば、タイムスタンプ)が格納されてもよい。また、時間の進みと共にレコードを追加的に格納するという前提があれば、履歴テーブル32には、過去に複数のスロット31が保持した用語だけが格納されてもよい。なお、メモリ27のうち、履歴テーブル32が記憶された領域は、本開示に係る記憶部に相当する。 Note that the history table 32 may store information indicating the time at that time (for example, a time stamp) together with the terms held by the plurality of slots 31 in the past. In addition, if there is a premise that records are additionally stored as time progresses, the history table 32 may store only terms held by a plurality of slots 31 in the past. In the memory 27, the area where the history table 32 is stored corresponds to a storage unit according to the present disclosure.
 タスク結果解析部28は、タスク処理部40による処理結果を取得し、取得した処理結果を解析する処理部である。タスク結果解析部28は、タスク処理部40から処理結果を取得した場合には、取得した処理結果を解析し、解析結果をシーケンス制御部23に渡す。なお、この解析結果は、履歴テーブル32のうちの現在時刻に対応する時点に復元ポイントを設定するか否かを操作部25が判定する際に用いられる。なお、タスク結果解析部28は、本開示に係る外部処理制御部の一機能に相当する。 The task result analysis unit 28 is a processing unit that acquires a processing result by the task processing unit 40 and analyzes the acquired processing result. When the task result analysis unit 28 acquires the processing result from the task processing unit 40, the task result analysis unit 28 analyzes the acquired processing result and passes the analysis result to the sequence control unit 23. This analysis result is used when the operation unit 25 determines whether or not to set a restoration point at a time corresponding to the current time in the history table 32. The task result analysis unit 28 corresponds to one function of the external processing control unit according to the present disclosure.
 例えば、タスク結果解析部28は、タスク処理部40によるレストラン検索処理の結果として、検索された情報が掲載されたWebページのタイトル及びURL(Uniform Resource Locator)を取得する。また、タスク結果解析部28は、検索処理の結果を解析し、検索された情報の件数を算出する。そして、タスク結果解析部28は、検索された情報の件数が、ユーザによる閲覧に適した件数(例えば1件~30件程度)である場合にのみ復元ポイントを設定するようにしてもよい。また、タスク結果解析部28は、検索された情報の件数が、0件、又は、100件以上というように、ユーザによる閲覧に適さない件数である場合に復元ポイントを設定することを禁止するようにしてもよい。 For example, the task result analysis unit 28 acquires the title and URL (Uniform Resource Locator) of the Web page on which the searched information is posted as a result of the restaurant search process by the task processing unit 40. Further, the task result analysis unit 28 analyzes the result of the search process and calculates the number of searched information. Then, the task result analysis unit 28 may set the restoration point only when the number of retrieved information is the number suitable for browsing by the user (for example, about 1 to 30 cases). In addition, the task result analysis unit 28 prohibits setting a restoration point when the number of retrieved information items is not suitable for browsing by the user, such as 0 or 100 or more. It may be.
 また、タスク結果解析部28は、必須スロット群のスロットの全てが用語を保持した時点に復元ポイントを設定するようにしてもよいし、スロット31が用語を保持している状態において、保持している用語と異なる用語を保持する状態に変わる時点に復元ポイントを設定してもよい。 In addition, the task result analysis unit 28 may set a restoration point when all of the slots of the essential slot group hold the term, or hold it while the slot 31 holds the term. A restoration point may be set at a point in time when the state changes to a state that holds a term different from the existing term.
 提示制御部29は、表示装置10によりユーザに提示するための提示データを生成し、表示装置10に出力する処理部である。提示制御部29は、タスク処理部40から処理結果を取得し、ユーザに効果的に処理結果を閲覧させるために表示装置10の画面上の位置を整え、また、表示装置10に出力するのに適したデータ形式に変換した上で、提示データを表示装置10に出力する。 The presentation control unit 29 is a processing unit that generates presentation data to be presented to the user by the display device 10 and outputs the presentation data to the display device 10. The presentation control unit 29 acquires the processing result from the task processing unit 40, arranges the position on the screen of the display device 10 so that the user can browse the processing result effectively, and outputs it to the display device 10 The presentation data is output to the display device 10 after being converted into a suitable data format.
 なお、音声対話装置20の一部又は全部の機能、及び、タスク処理部40は、音声合成部12等同様、クラウドサーバとして実現されてもよい。 Note that part or all of the functions of the voice interaction device 20 and the task processing unit 40 may be realized as a cloud server, like the voice synthesis unit 12 and the like.
 図2は、本実施の形態に係る音声対話システム1による提示の説明図である。図2に示される説明図は、タスク処理部40による処理結果を表示装置10がユーザに提示するときの表示画面に表示される画像の一例である。 FIG. 2 is an explanatory diagram of presentation by the voice interaction system 1 according to the present embodiment. The explanatory diagram shown in FIG. 2 is an example of an image displayed on the display screen when the display device 10 presents the processing result by the task processing unit 40 to the user.
 表示画面内の左側には、属性を示す文字列201~205が表示されている。文字列201~205は、複数のスロット31それぞれの属性を示す文字列である。 * Character strings 201 to 205 indicating attributes are displayed on the left side of the display screen. Character strings 201 to 205 are character strings indicating attributes of the plurality of slots 31.
 表示画面内の右側には、用語211~215が表示されている。用語211~215は、それぞれ、文字列201~205の属性に対応付けられたスロット31が保持している用語である。 The terms 211 to 215 are displayed on the right side of the display screen. The terms 211 to 215 are terms held in the slots 31 associated with the attributes of the character strings 201 to 205, respectively.
 表示画面内の下側には、文字列206及び結果情報216が示されている。文字列206は、文字列206の下方に表示されるものが検索結果であることを示す文字列である。結果情報216は、用語211~215に基づいてタスク処理部40がレストラン検索を行った結果を示す情報である。 A character string 206 and result information 216 are shown on the lower side of the display screen. The character string 206 is a character string indicating that what is displayed below the character string 206 is a search result. The result information 216 is information indicating a result of the restaurant search performed by the task processing unit 40 based on the terms 211 to 215.
 このように、対話の内容と、その対話の内容に基づくタスク処理部40による処理結果である結果情報とが表示装置10に表示され、ユーザは、対話の内容が反映された処理結果を知ることができる。 Thus, the content of the dialogue and the result information that is the processing result by the task processing unit 40 based on the content of the dialogue are displayed on the display device 10, and the user knows the processing result in which the content of the dialogue is reflected. Can do.
 なお、表示画面に表示される画像は、図2に示されるものに限定されるわけではなく、表示される情報、その配置などの表示の有無、表示位置は、任意に変更されてよい。 It should be noted that the image displayed on the display screen is not limited to that shown in FIG. 2, and the displayed information, the presence / absence of display such as its arrangement, and the display position may be arbitrarily changed.
 図3は、本実施の形態に係る対話シーケンス及び履歴情報の第一の説明図である。 FIG. 3 is a first explanatory diagram of a dialogue sequence and history information according to the present embodiment.
 図3には、対話シーケンス310、履歴テーブル320、及び、検索結果330が、対話シーケンスの時系列に併せて示されている。なお、図3に示される一列は、1つの時点に対応している。この一列のことをレコードともいう。履歴テーブル320は、履歴テーブル32の一例である。 FIG. 3 shows a dialogue sequence 310, a history table 320, and a search result 330 together with the time series of the dialogue sequence. Note that one row shown in FIG. 3 corresponds to one time point. This line is also called a record. The history table 320 is an example of the history table 32.
 対話シーケンス310は、対話におけるユーザによる発話と音声対話装置20による応答とを時系列で並べたデータである。 The dialogue sequence 310 is data in which utterances by the user in the dialogue and responses by the voice dialogue apparatus 20 are arranged in time series.
 時刻情報311は、ユーザによる発話又は音声対話装置20による応答があった時刻を示す時刻情報(タイムスタンプ)である。 The time information 311 is time information (time stamp) indicating the time when the user uttered or responded by the voice interaction apparatus 20.
 発話312は、当該時刻におけるユーザによる発話を示す発話データである。具体的には、発話312は、発話データ取得部22が、マイク13及び音声認識部14を介して取得したユーザの音声による発話を示す発話データである。 The utterance 312 is utterance data indicating the utterance by the user at the time. Specifically, the utterance 312 is utterance data indicating the utterance by the user's voice acquired by the utterance data acquisition unit 22 via the microphone 13 and the voice recognition unit 14.
 応答313は、当該時刻における音声対話装置20による応答を示す応答文である。具体的には、応答313は、応答文生成部21が、シーケンス制御部23からの応答指示を受けて生成するものである。 The response 313 is a response sentence indicating a response by the voice interaction device 20 at the time. Specifically, the response 313 is generated by the response sentence generation unit 21 in response to a response instruction from the sequence control unit 23.
 履歴テーブル320は、履歴テーブル32の一例であって、必須スロット群321と、オプションスロット群322と、アクション323と、復元ポイント324との各情報を有する。図3に示されるように、履歴テーブル320は、時系列において対話シーケンス310と対応付けられている。 The history table 320 is an example of the history table 32, and includes information on a mandatory slot group 321, an optional slot group 322, an action 323, and a restoration point 324. As shown in FIG. 3, the history table 320 is associated with the dialogue sequence 310 in time series.
 必須スロット群321は、スロット31のうちの必須スロットに、当該時点において保持されていた用語である。必須スロット群321には、例えば、「料理名」、「地域」及び「予算」の属性の用語が含まれる。 The essential slot group 321 is a term held in an essential slot among the slots 31 at the time. The essential slot group 321 includes, for example, terms of attributes of “dishes name”, “region”, and “budget”.
 オプションスロット群322は、スロット31のうちのオプションスロットに、当該時点において保持されていた用語である。オプションスロット群322には、例えば、「個室の有無」及び「駐車場の有無」の属性の用語が含まれる。 The option slot group 322 is a term held in the option slot of the slots 31 at the time. The option slot group 322 includes, for example, attribute terms of “presence / absence of private room” and “presence / absence of parking lot”.
 アクション323は、当該時点において音声対話装置20が実行した処理を示す情報であり、複数の情報が格納されることもある。例えば、ある属性のスロット31に新たな用語を保持させた場合には、そのことを示すために、その属性の名称と、「登録」の文字列とが当該時点に設定される。また、タスク制御部24がタスク処理部40に用語を出力して情報検索をさせた時点には、「検索」の文字列が設定される。また、操作部25が、スロット31が保持している用語を過去の時点におけるものに変更した時点には、「復元」の文字列が設定される。 The action 323 is information indicating processing executed by the voice interaction apparatus 20 at the time point, and a plurality of information may be stored. For example, when a new term is held in a slot 31 with a certain attribute, the name of the attribute and a character string “register” are set at the time point to indicate that. In addition, when the task control unit 24 outputs a term to the task processing unit 40 to search for information, a character string “search” is set. Further, when the operation unit 25 changes the term held in the slot 31 to that at the past time point, the character string “restore” is set.
 復元ポイント324は、当該時点に復元ポイントが設定されているか否かを示す情報である。復元ポイントが設定されている時点には、「1」が設定されている。当該時点に復元ポイントが設定されるか否かは、タスク結果解析部28により判定される。タスク結果解析部28が当該時点に復元ポイントを設定すると判定した場合、操作部25は、当該時点の復元ポイント324に復元ポイントを設定する。 The restoration point 324 is information indicating whether or not a restoration point is set at the time. At the time when the restoration point is set, “1” is set. The task result analysis unit 28 determines whether or not a restoration point is set at the time. When the task result analysis unit 28 determines that the restoration point is set at the time, the operation unit 25 sets the restoration point to the restoration point 324 at the time.
 検索結果330は、当該時点におけるタスク処理部40による検索処理の結果の件数である。検索結果330は、タスク結果解析部28により設定されるものである。 The search result 330 is the number of search processing results by the task processing unit 40 at the time. The search result 330 is set by the task result analysis unit 28.
 図3は、ユーザが、検索条件を変えながら、順次、異なる検索条件でレストラン検索を行う場合の対話シーケンスを示したものである。図3においては、対話の内容を、ユーザが意図する過去の時点における対話の内容に変更する場合の対話シーケンスが示されている。 FIG. 3 shows an interactive sequence when the user sequentially searches for a restaurant under different search conditions while changing the search conditions. FIG. 3 shows a dialogue sequence when the content of the dialogue is changed to the content of the dialogue at the past time intended by the user.
 レコードR1~R7に対応する時点において、順次、ユーザによる発話に含まれる用語が発話データ取得部22等により取得され、取得された用語のそれぞれが当該用語の属性に対応したスロット31に格納される。 At the time corresponding to the records R1 to R7, the terms included in the user's utterance are sequentially acquired by the utterance data acquisition unit 22 and the like, and each of the acquired terms is stored in the slot 31 corresponding to the attribute of the term. .
 レコードR8に対応する時点において、スロット31が保持している用語に基づいた最初の検索処理がタスク処理部40により行われる。これは、レコードR7に対応する時点で必須スロット群に含まれるスロット31の全てに用語が格納されたことを契機として行われたものである。 At the time corresponding to the record R8, the first search processing based on the term held in the slot 31 is performed by the task processing unit 40. This is performed when the term is stored in all the slots 31 included in the essential slot group at the time corresponding to the record R7.
 レコードR9~R16に対応する時点において、スロット31が保持している用語に基づいた検索処理が行われる。これは、ユーザが所望する検索結果が得られるように検索語を変えながら、順次、検索処理がなされたものである。 At the time corresponding to the records R9 to R16, search processing based on the terms held in the slot 31 is performed. In this case, the search processing is sequentially performed while changing the search word so that the search result desired by the user can be obtained.
 レコードR17に対応する時点において、対話の内容を過去の時点に戻すための制御発話がユーザによりなされる。これは、レコードR14又はR16に対応する時点での検索結果が0件であったので、検索件数が0件になる前の過去の時点の検索条件に戻そうと、ユーザが意図して行ったものである。 At the time corresponding to the record R17, the user makes a control utterance to return the content of the dialogue to a past time. This is because the search result at the time corresponding to the record R14 or R16 was 0, and the user intended to return to the search conditions of the past time before the number of searches became 0. Is.
 レコードR18~R20において、スロット31のそれぞれが保持する用語が、レコードR10におけるものに復元される。 In the records R18 to R20, the terms held in the slots 31 are restored to those in the record R10.
 このようにすることで、音声対話装置20は、対話の内容を、ユーザの音声による発話に基づいた過去の時点に戻し、その状態から新たな対話を継続的に実行することができる。このように、音声対話装置は、ユーザとの対話の内容を簡易な方法により修正することができる。 By doing in this way, the voice dialogue apparatus 20 can return the content of the dialogue to a past time point based on the speech by the user's voice, and continuously execute a new dialogue from the state. In this way, the voice interaction device can correct the content of the dialogue with the user by a simple method.
 [1-2.動作]
 以上のように構成された音声対話装置20及び音声対話システム1について、その動作を以下に説明する。
[1-2. Operation]
The operations of the voice interaction device 20 and the voice interaction system 1 configured as described above will be described below.
 図4は、本実施の形態に係る音声対話装置20によるメイン処理のフロー図である。 FIG. 4 is a flowchart of main processing by the voice interaction apparatus 20 according to the present embodiment.
 ステップS101において、マイク13は、ユーザによる発話の音声を取得し、取得した音声に基づいて音声信号を生成する。ここで、ユーザによる発話の音声とは、例えば「中華が食べたい」というようにレストラン検索のための用語を含む音声であってもよいし、「守口に戻して」というようにスロット31が保持する用語を過去の時点におけるものに変更するための用語を含む音声であってもよい。 In step S101, the microphone 13 acquires the voice of the user's utterance and generates a voice signal based on the acquired voice. Here, the voice of the utterance by the user may be a voice including a term for restaurant search such as “I want to eat Chinese”, or the slot 31 holds “return to the guard”. It may be a voice including a term for changing the term to be used at a past time.
 ステップS102において、音声認識部14は、ステップS101でマイク13が生成した音声信号に対して音声認識処理を行うことで、ユーザによる発話の発話データを生成する。 In step S102, the voice recognition unit 14 performs voice recognition processing on the voice signal generated by the microphone 13 in step S101, thereby generating utterance data of the user's utterance.
 ステップS103において、発話データ取得部22は、ステップS102で音声認識部14が生成した発話データを取得する。 In step S103, the utterance data acquisition unit 22 acquires the utterance data generated by the voice recognition unit 14 in step S102.
 ステップS104において、シーケンス制御部23は、ステップS103で発話データ取得部22が取得した発話データが空(から)であるか否かを判定する。 In step S104, the sequence control unit 23 determines whether or not the utterance data acquired by the utterance data acquisition unit 22 in step S103 is empty.
 ステップS104で発話データが空であるとシーケンス制御部23が判定した場合(ステップS104で「Y」)、ステップS105に進む。一方、発話データが空でないと判定した場合(ステップS104で「N」)、ステップS121に進む。 If the sequence control unit 23 determines that the utterance data is empty in step S104 ("Y" in step S104), the process proceeds to step S105. On the other hand, if it is determined that the utterance data is not empty (“N” in step S104), the process proceeds to step S121.
 ステップS105において、シーケンス制御部23は、操作部25を利用して発話データに含まれる用語をスロット31に格納する。具体的には、シーケンス制御部23は、発話データに含まれる用語のそれぞれについて当該用語の属性を判定し、当該用語の属性に一致する属性を有するスロット31に当該用語を格納する。例えば、シーケンス制御部23は、発話データ「中華が食べたい」に含まれる用語「中華」が、料理名の属性を有する用語であると判定し、用語「中華」を料理名の属性を有するスロット31に格納する。なお、このとき、シーケンス制御部23は、スロット31に格納される用語が本来の名称の略称又は俗称等であるような場合には、本来の名称に変換した上でスロット31に格納してもよい。具体的には、シーケンス制御部23は、用語「中華」が「中華料理」を短縮した名称(略称)であると判定し、スロット31に「中華料理」を格納するようにしてもよい。 In step S105, the sequence control unit 23 stores the term included in the utterance data in the slot 31 using the operation unit 25. Specifically, the sequence control unit 23 determines the attribute of the term for each of the terms included in the utterance data, and stores the term in the slot 31 having an attribute that matches the attribute of the term. For example, the sequence control unit 23 determines that the term “Chinese” included in the utterance data “Chinese wants to eat” is a term having a dish name attribute, and the term “Chinese” is a slot having a dish name attribute. 31. At this time, when the term stored in the slot 31 is an abbreviation or common name of the original name, the sequence control unit 23 converts the original name into the original name and stores it in the slot 31. Good. Specifically, the sequence control unit 23 may determine that the term “Chinese” is an abbreviation of “Chinese cuisine” and store “Chinese cuisine” in the slot 31.
 ステップS106において、操作部25及び提示制御部29は、スロット31が保持している用語を表示装置10により表示する。 In step S106, the operation unit 25 and the presentation control unit 29 display the terms held in the slot 31 by the display device 10.
 ステップS107において、操作部25等は、必要な場合に、対話の内容を過去の時点におけるものに変更することで、対話の内容を復元するための復元処理を行う。復元処理の詳細については、後で詳細に説明する。 In step S107, the operation unit 25 or the like performs a restoration process for restoring the content of the dialogue by changing the content of the dialogue to that at the past time when necessary. Details of the restoration process will be described later in detail.
 ステップS108において、解析部26は、必須スロット群の全てのスロット31に用語が格納されているか否か、つまり、必須スロット群の全てのスロット31が用語を保持しているか否かを判定する。 In step S108, the analysis unit 26 determines whether the term is stored in all the slots 31 of the essential slot group, that is, whether all the slots 31 of the essential slot group hold the term.
 ステップS108において全てのスロット31に用語が格納されたと解析部26が判定した場合(ステップS108で「Y」)、ステップS109に進む。一方、全てのスロット31に用語が格納されていないと解析部26が判定した場合(ステップS108で「N」)、つまり、必須スロット群のうちの少なくとも1つのスロット31が空である場合、ステップS122に進む。 If the analysis unit 26 determines that the term is stored in all the slots 31 in step S108 ("Y" in step S108), the process proceeds to step S109. On the other hand, if the analysis unit 26 determines that no term is stored in all the slots 31 (“N” in step S108), that is, if at least one slot 31 in the essential slot group is empty, the step The process proceeds to S122.
 ステップS109において、シーケンス制御部23は、タスク処理をタスク処理部40に実行させるための実行指示をタスク制御部24に行う。このとき、操作部25は、履歴テーブル32に検索タスクを実行したことを記録する。具体的には、操作部25は、履歴テーブル320における現時点のアクション323に「検索」を設定する。 In step S109, the sequence control unit 23 gives the task control unit 24 an execution instruction for causing the task processing unit 40 to execute the task processing. At this time, the operation unit 25 records in the history table 32 that the search task has been executed. Specifically, the operation unit 25 sets “search” to the current action 323 in the history table 320.
 ステップS110において、タスク制御部24は、ステップS109でのシーケンス制御部23による実行指示に基づいて、スロット31が保持している用語をタスク処理部40に出力し、タスク処理部40に検索処理を実行させる。タスク処理部40は、タスク制御部24が出力した用語を取得し、取得した用語を検索語として用いて検索処理を行い、検索結果を出力する。 In step S110, the task control unit 24 outputs the term held in the slot 31 to the task processing unit 40 based on the execution instruction from the sequence control unit 23 in step S109, and performs search processing on the task processing unit 40. Let it run. The task processing unit 40 acquires the term output by the task control unit 24, performs a search process using the acquired term as a search term, and outputs a search result.
 ステップS111において、提示制御部29は、ステップS110でタスク処理部40が出力した検索結果を取得し、取得した検索結果を、表示装置10によりユーザに提示するのに適切な形式(例えば、図2のような表示態様)にして表示装置10に出力する。表示装置10は、提示制御部29が出力した検索結果を取得し、表示画面に表示する。 In step S111, the presentation control unit 29 acquires the search result output by the task processing unit 40 in step S110, and presents the acquired search result to the user in the display device 10 (for example, FIG. 2). The display mode is output to the display device 10. The display device 10 acquires the search result output by the presentation control unit 29 and displays it on the display screen.
 ステップS112において、タスク結果解析部28は、ステップS110でタスク処理部40が出力した検索結果を取得し、取得した検索結果に基づいて復元ポイント設定処理を行う。復元ポイント設定処理の詳細については、後で詳細に説明する。 In step S112, the task result analysis unit 28 acquires the search result output by the task processing unit 40 in step S110, and performs restoration point setting processing based on the acquired search result. Details of the restoration point setting process will be described later in detail.
 ステップS113において、シーケンス制御部23は、ユーザに対して次の発話を促すための応答指示を、応答文生成部21に対して行う。 In step S113, the sequence control unit 23 gives a response instruction to prompt the user for the next utterance to the response sentence generation unit 21.
 ステップS114において、応答文生成部21は、応答指示に基づいて応答文を生成する。また、応答文生成部21は、生成した応答文を音声合成部12に出力し、当該応答文を音声としてスピーカ11より出力し、ユーザに聴取させる。 In step S114, the response text generation unit 21 generates a response text based on the response instruction. In addition, the response sentence generation unit 21 outputs the generated response sentence to the speech synthesizer 12, and outputs the response sentence as a sound from the speaker 11 to allow the user to listen.
 ステップS114の処理が終了したら、再びステップS101の処理を実行する。 When the process of step S114 is completed, the process of step S101 is executed again.
 ステップS121において、シーケンス制御部23は、ユーザに対して再発話(前回と同じ発話を行うこと)を促すための応答指示を、応答文生成部21に対して行う。ステップS104で発話データが空と判定されたことは、マイク13が何らかの音を取得したにもかかわらずその音から音声認識部14が発話データを取得することができなかったことを意味している。よって、ユーザに対して前回と同じ発話を行うことを要請することで、発話データを取得することができると期待される。 In step S121, the sequence control unit 23 gives a response instruction to the response sentence generation unit 21 to prompt the user to re-utter (perform the same utterance as the previous time). The fact that the utterance data is determined to be empty in step S104 means that the voice recognition unit 14 cannot acquire the utterance data from the sound although the microphone 13 has acquired some sound. . Therefore, it is expected that utterance data can be acquired by requesting the user to perform the same utterance as the previous time.
 ステップS122において、シーケンス制御部23は、ユーザに対して次の発話を促すための応答指示を、応答文生成部21に対して行う。シーケンス制御部23は、例えば、必須スロット群に含まれるスロット31のうち、用語を保持していないものがある場合に、用語を保持していないスロット31が保持すべき用語をユーザに発話させるための応答文を生成する応答指示を行う。 In step S122, the sequence control unit 23 gives a response instruction for prompting the user to speak next to the response sentence generating unit 21. For example, when there is a slot 31 that does not hold a term among the slots 31 included in the essential slot group, the sequence control unit 23 causes the user to utter the term that the slot 31 that does not hold the term should hold. A response instruction is generated to generate a response sentence.
 図5は、本実施の形態に係る音声対話装置による復元処理のフロー図である。図5に示されるフロー図は、図4におけるステップS107の処理を詳細に示すものであり、発話データに制御用語が含まれている場合にスロット31が保持する用語を過去の時点におけるものに変更する処理を示すものである。 FIG. 5 is a flowchart of restoration processing by the voice interaction apparatus according to the present embodiment. The flowchart shown in FIG. 5 shows the details of the process of step S107 in FIG. 4, and when the utterance data includes a control term, the term held in the slot 31 is changed to that at the past time point. This shows the processing to be performed.
 より具体的には、操作部25は、発話データ取得部22が取得した発話データに後述する第一用語及び第二用語が含まれるか否かを判定する。さらに、操作部25は、第一用語及び第二用語が含まれると判定した場合に、履歴テーブルを参照して、複数のスロット31のそれぞれが保持している用語を、過去の時点において当該スロット31が保持していた用語に変更する。ここで、過去の時点とは、複数のスロット31のうち第二用語の属性に対応付けられたスロット31(本開示に係る対応保持部に相当)が第二用語を保持していた時点である。 More specifically, the operation unit 25 determines whether or not the utterance data acquired by the utterance data acquisition unit 22 includes a first term and a second term described later. In addition, when the operation unit 25 determines that the first term and the second term are included, the operation unit 25 refers to the history table to display the term held in each of the plurality of slots 31 in the past time point. It changes to the term which 31 had. Here, the past time point is a time point when the slot 31 (corresponding to the correspondence holding unit according to the present disclosure) associated with the attribute of the second term among the plurality of slots 31 holds the second term. .
 ステップS201において、シーケンス制御部23は、発話データ取得部22から取得した発話データに、復元用語(第一用語ともいう)が含まれているか否かを判定する。ここで、復元用語とは、対話情報を過去の時点に変更することを示す予め定められた用語であり、例えば、「(~に)戻して」、又は、「(~)じゃない」というようなものである。 In step S201, the sequence control unit 23 determines whether the utterance data acquired from the utterance data acquisition unit 22 includes a restoration term (also referred to as a first term). Here, the restoration term is a predetermined term indicating that the dialogue information is changed to a past time point, for example, “return to (to)” or “not (to)”. It is a thing.
 ステップS201で復元用語が含まれているとシーケンス制御部23が判定した場合(ステップS201で「Y」)には、ステップS202に進む。一方、復元用語が含まれていないと判定した場合(ステップS201で「N」)、図5に示される一連の処理を終了する。 If the sequence control unit 23 determines that the restoration term is included in step S201 ("Y" in step S201), the process proceeds to step S202. On the other hand, if it is determined that no restoration term is included (“N” in step S201), the series of processes shown in FIG. 5 is terminated.
 ステップS202において、解析部26は、発話データのうち復元用語を除く部分に含まれる用語(第二用語ともいう)を取得し、取得した用語に基づいて履歴テーブル32から復元ポイントを抽出する。具体的には、解析部26は、取得した用語の属性を判定し、履歴テーブル32に含まれる復元ポイントのうち、取得した用語の属性に対応するスロット31が保持していた用語が、当該取得した用語に一致する復元ポイントを抽出する。なお、上記第一用語及び上記第二用語が含まれている発話データが、制御発話データであるということもできる。なお、複数の設定ポイントが抽出されてもよい。 In step S202, the analysis unit 26 acquires a term (also referred to as a second term) included in a portion excluding the restoration term in the utterance data, and extracts a restoration point from the history table 32 based on the obtained term. Specifically, the analysis unit 26 determines the attribute of the acquired term, and among the restoration points included in the history table 32, the term held in the slot 31 corresponding to the acquired term attribute is Extract restore points that match the terms you have selected. It can also be said that the utterance data including the first term and the second term is control utterance data. A plurality of setting points may be extracted.
 ステップS203において、解析部26は、ステップS202で抽出した復元ポイントが1個であるか否かを判定する。 In step S203, the analysis unit 26 determines whether or not there is one restoration point extracted in step S202.
 ステップS203で復元ポイントが1個であると解析部26が判定した場合(ステップS203で「Y」)、ステップS204に進む。一方、復元ポイントが1個でないと解析部26が判定した場合(ステップS203で「N」)、ステップS211に進む。 If the analysis unit 26 determines that there is one restoration point in step S203 ("Y" in step S203), the process proceeds to step S204. On the other hand, when the analysis unit 26 determines that there is not one restoration point (“N” in step S203), the process proceeds to step S211.
 ステップS204において、操作部25は、履歴テーブル32を参照して、スロット31が保持している用語を、ステップS202で1個抽出された復元ポイントの時点においてスロット31に保持していた用語に変更する。つまり、操作部25は、複数のスロット31が保持している用語を、復元ポイントの時点におけるものに戻すように変更する。また、操作部25は、履歴テーブル320において、復元ポイントの時点における用語に変更した時点に、アクションとして「復元」を設定する。なお、操作部25は、復元ポイントの時点でスロット31が保持していた用語がなかった、つまり、復元ポイントの時点でスロット31が何も用語を保持していなかった場合には、スロット31が用語を保持しない状態にする。 In step S204, the operation unit 25 refers to the history table 32 and changes the term held in the slot 31 to the term held in the slot 31 at the point of the restoration point extracted in step S202. To do. That is, the operation unit 25 changes the terminology held in the plurality of slots 31 so as to return to the term at the time of the restoration point. In addition, the operation unit 25 sets “restore” as the action when the history table 320 changes to the term at the time of the restoration point. Note that the operation unit 25 does not have a term held in the slot 31 at the time of the restoration point, that is, if the slot 31 holds no term at the time of the restoration point, the slot 31 Terminate the term.
 ステップS211において、シーケンス制御部23は、ユーザに対して復元ポイントが1個だけ抽出されるようにするための発話を促すための応答についての応答指示を、応答文生成部21に対して行う。例えば、履歴テーブル320において、ユーザから「守口に戻して」というような制御発話が取得されたとき、この制御発話から特定される復元ポイントの候補が2個ある。ユーザがこの2個の復元ポイントのどちらを意図しているかを指示する発話を促すために、シーケンス制御部23は、「駐車場付で検索したところまで戻せばいいですか」という応答についての応答指示を行う。 In step S <b> 211, the sequence control unit 23 gives a response instruction to the response sentence generation unit 21 for a response for prompting the user to extract only one restoration point. For example, in the history table 320, when a control utterance such as “Return to Moriguchi” is acquired from the user, there are two restoration point candidates specified from the control utterance. In order to urge the user to indicate which of the two restoration points is intended, the sequence control unit 23 responds with a response “Would it return to the place searched with the parking lot?” Give instructions.
 ステップS211の後、ユーザにより2個の復元ポイントのいずれかを特定する発話がなされた場合には、次回のメイン処理(図4)から実行されるステップS202では、復元ポイントが1個抽出され、ステップS204が実行されることになる。 After step S211, when the user makes an utterance that specifies one of the two restoration points, one restoration point is extracted in step S202 executed from the next main process (FIG. 4). Step S204 is executed.
 なお、上記において、第二用語に代えて、属性の名称である属性名を用いてもよい。つまり、操作部25は、発話データ取得部22が取得した発話データに第一用語及び属性名が含まれるか否かを判定する。操作部25は、第一用語及び属性名が含まれると判定した場合に、履歴テーブルを参照して、複数のスロット31のそれぞれが保持している用語を、当該スロット31が過去の時点において保持していた用語に変更するようにしてもよい。ここで、過去の時点とは、複数のスロット31のうち上記属性名により示される属性に対応付けられたスロット31(本開示に係る対応保持部に相当)が現在保持している用語を保持する直前の時点である。 In the above, instead of the second term, an attribute name that is an attribute name may be used. That is, the operation unit 25 determines whether or not the first term and the attribute name are included in the utterance data acquired by the utterance data acquisition unit 22. When the operation unit 25 determines that the first term and the attribute name are included, the operation unit 25 refers to the history table, and holds the term held in each of the plurality of slots 31 in the past time point. You may make it change into the vocabulary which was doing. Here, the past time point holds the term currently held in the slot 31 (corresponding to the correspondence holding unit according to the present disclosure) associated with the attribute indicated by the attribute name among the plurality of slots 31. It is the time immediately before.
 図6は、本実施の形態に係る音声対話装置による復元ポイント設定処理のフロー図である。図6に示されるフロー図は、図4におけるステップS112の処理を詳細に示すものである。 FIG. 6 is a flowchart of restoration point setting processing by the voice interaction apparatus according to the present embodiment. The flowchart shown in FIG. 6 shows details of the process in step S112 in FIG.
 ステップS301において、操作部25は、復元ポイントを設定するための条件に基づいて処理を分岐する。上記条件が、「検索を実行した時点」(条件C)である場合(ステップS301で「条件C」)、ステップS302に進む。一方、上記条件が、『「検索を実行した時点」かつ「検索結果が有効」』(条件D)である場合(ステップS301で「条件D」)、ステップS303に進む。なお、ここでは条件が2つの場合を例として示したが、3つ以上の条件がある場合でも同様の処理が可能である。 In step S301, the operation unit 25 branches the process based on a condition for setting a restoration point. When the above condition is “when search is executed” (condition C) (“condition C” in step S301), the process proceeds to step S302. On the other hand, when the above condition is ““ time when search is performed ”and“ search result is valid ”” (condition D) (“condition D” in step S301), the process proceeds to step S303. Here, the case where there are two conditions is shown as an example, but the same processing is possible even when there are three or more conditions.
 ステップS302において、操作部25は、履歴テーブル320における現在の時点に復元ポイントを設定する。 In step S302, the operation unit 25 sets a restoration point at the current time point in the history table 320.
 ステップS303において、操作部25は、タスク結果解析部28の解析結果である検索結果を取得し、検索された情報の件数が0件であるか否かを判定する。 In step S303, the operation unit 25 acquires a search result that is an analysis result of the task result analysis unit 28, and determines whether or not the number of pieces of searched information is zero.
 ステップS303で検索された情報の件数が0件である場合(ステップS303で「Y」)、操作部25は、この時点に復元ポイントを設定せずに一連の処理を終了する。すなわち、操作部25は、情報検索の結果を取得した時点であっても、情報検索の結果に含まれる情報が0件であった時点には、復元ポイントを設定することを禁止する。一方、検索された情報の件数が0件でない場合(ステップS303で「N」)、ステップS302に進む。 If the number of information retrieved in step S303 is 0 (“Y” in step S303), the operation unit 25 ends a series of processes without setting a restoration point at this time. That is, the operation unit 25 prohibits setting a restoration point when the information included in the information search result is 0 even when the information search result is acquired. On the other hand, when the number of retrieved information is not 0 (“N” in step S303), the process proceeds to step S302.
 なお、検索された情報の件数がユーザによる閲覧に適さない数(例えば100件以上)である場合にも、0件である場合と同様、この時点に復元ポイントを設定しないことにしてもよい。 Note that even when the number of retrieved information is not suitable for browsing by the user (for example, 100 or more), the restoration point may not be set at this point as in the case of zero.
 図7は、本実施の形態に係る履歴情報の第二の説明図である。図7は、ユーザが、検索条件を変えながら、順次、異なる検索条件でレストラン検索を行うための対話における対話シーケンスである。これは、音声の誤認識などに起因して対話の内容がユーザの意図と異なるものとなった場合に、対話の内容を過去の時点におけるものに変更する場合の対話シーケンスの一例である。 FIG. 7 is a second explanatory diagram of history information according to the present embodiment. FIG. 7 shows a dialogue sequence in a dialogue in which a user sequentially searches for a restaurant under different search conditions while changing the search conditions. This is an example of a dialogue sequence when the dialogue content is changed to that at a past time when the dialogue content is different from the user's intention due to misrecognition of voice or the like.
 図7には、図3と同様、対話シーケンス310等が示されている。 FIG. 7 shows a dialogue sequence 310 and the like as in FIG.
 レコードR1~R5に対応する時点において、順次、ユーザによる発話に含まれる用語が発話データ取得部22等により取得され、取得された用語のそれぞれが当該用語の属性に対応したスロット31に格納される。 At the time corresponding to the records R1 to R5, the terms included in the user's utterance are sequentially acquired by the utterance data acquisition unit 22 and the like, and each of the acquired terms is stored in the slot 31 corresponding to the attribute of the term. .
 レコードR6に対応する時点において、スロット31が保持している用語に基づいた最初の検索処理がタスク処理部40により行われる。これは、レコードR5に対応する時点で必須スロット群に含まれるスロット31の全てに用語が格納されたことを契機として行われたものである。 At the time corresponding to the record R6, the task processing unit 40 performs an initial search process based on the term held in the slot 31. This is performed when the term is stored in all the slots 31 included in the essential slot group at the time corresponding to the record R5.
 レコードR7~R14に対応する時点において、スロット31に格納された用語に基づいた検索処理が行われる。これは、ユーザが所望する検索結果が得られるように検索語を変えながら、順次、検索処理がなされたものである。 At the time corresponding to the records R7 to R14, a search process based on the terms stored in the slot 31 is performed. In this case, the search processing is sequentially performed while changing the search word so that the search result desired by the user can be obtained.
 この対話の中で、音声認識部14による誤認識によりスロット31が保持している用語がユーザの意図と異なるものに変更されている。具体的には、レコードR11に対応する時点において、ユーザが検索条件として駐車場を追加する意図で「駐車場も(Chushajomo)」と発話したものの、これを音声認識部14が「中華料理(Chukaryori)」と誤認識している。音声認識部14が誤認識したことにより、レコードR12に対応する時点において、用語「中華料理」が料理名の属性のスロット31に格納される。また、レコードR13に対応する時点において、ユーザが検索条件を修正する意図で「中華じゃなくてイタリア(Chuka-janakute-itaria)」と発話したものの、これを音声認識部14が「入谷(Iriya)」すなわち地名である入谷と誤認識している。音声認識部14が誤認識したことにより、レコードR15に対応する時点において、用語「入谷」が地域の属性のスロット31に格納される。 In this dialogue, the term held in the slot 31 is changed to something different from the user's intention due to erroneous recognition by the voice recognition unit 14. Specifically, at the time corresponding to the record R11, the user uttered “parking lot (Chushajomo)” with the intention of adding a parking lot as a search condition. ) ”. Due to the erroneous recognition by the voice recognition unit 14, the term “Chinese cuisine” is stored in the slot 31 of the attribute of the dish name at the time corresponding to the record R 12. At the time corresponding to the record R13, the user uttered “Chuka-janakute-itaria” instead of “Chuka-janakute-itaria” with the intention of correcting the search condition. "In other words, I misunderstood the place name Iriya. Due to the erroneous recognition by the voice recognition unit 14, the term “Iritani” is stored in the slot 31 of the regional attribute at the time corresponding to the record R15.
 レコードR15に対応する時点において、対話の内容を過去の時点に戻すための発話がユーザによりなされる。これは、レコードR12又はR14に対応する時点でスロット31が保持している用語がユーザの意図と異なり変更されたので、この変更が行われる前の過去の時点の検索条件に戻そうと、ユーザが意図して行ったものである。 At the time corresponding to the record R15, the user makes an utterance for returning the content of the dialogue to the past time. This is because the term held in the slot 31 at the time corresponding to the record R12 or R14 has been changed unlike the user's intention, and the user tries to return to the search condition at the past time before the change was made. Is intended.
 レコードR15~R16に対応する時点において、スロット31のそれぞれが保持している用語が、レコードR10におけるものに復元される。 At the time corresponding to the records R15 to R16, the terms held in the slots 31 are restored to those in the record R10.
 このようにすることで、音声対話装置は、対話の内容を、ユーザによる発話に基づいた過去の時点に戻し、その状態から新たな対話を継続的に実行することができる。このように、音声対話装置は、ユーザとの対話の内容を簡易な方法により修正することができる。 In this way, the voice interaction apparatus can return the content of the conversation to a past time point based on the utterance by the user, and continuously execute a new conversation from that state. In this way, the voice interaction device can correct the content of the dialogue with the user by a simple method.
 [1-3.変形例]
 図8は、本実施の形態の変形例に係る音声対話装置20Aの構成を示すブロック図である。
[1-3. Modified example]
FIG. 8 is a block diagram showing a configuration of a voice interactive apparatus 20A according to a modification of the present embodiment.
 図8に示されるように、ユーザとの音声による対話を行う音声対話装置20Aは、複数の保持部103と、記憶部104と、取得部101と、変更部102とを備える。複数の保持部103は、対話の内容を示す対話情報を保持する。また、複数の保持部103は、それぞれが用語の属性に対応付けられており、その属性を有する用語を保持する。記憶部104は、複数の保持部103が保持する用語の履歴を記憶する。取得部101は、ユーザの音声による発話の内容を示す発話データを取得する。さらに、取得部101は、取得した発話データに含まれる発話用語を、複数の保持部103のうちその発話用語の属性に対応付けられた保持部103に保持させる。変更部102は、取得部101が対話情報を制御するための制御用語が含まれる制御発話データを取得した場合に、記憶部104を参照して、複数の保持部103が保持している用語を変更する。具体的には、複数の保持部103のそれぞれが保持している用語を、制御用語により特定される過去の時点において当該保持部103が保持していた用語に変更する。 As shown in FIG. 8, the voice interaction apparatus 20 </ b> A that performs dialogue with the user by voice includes a plurality of holding units 103, a storage unit 104, an acquisition unit 101, and a change unit 102. The plurality of holding units 103 hold dialogue information indicating the content of the dialogue. Each of the plurality of holding units 103 is associated with a term attribute, and holds a term having the attribute. The storage unit 104 stores a history of terms held by the plurality of holding units 103. The acquisition unit 101 acquires utterance data indicating the contents of utterances by the user's voice. Furthermore, the acquisition unit 101 holds the utterance term included in the acquired utterance data in the holding unit 103 associated with the attribute of the utterance term among the plurality of holding units 103. The change unit 102 refers to the storage unit 104 when the acquisition unit 101 acquires control utterance data that includes a control term for controlling the conversation information. change. Specifically, the term held by each of the plurality of holding units 103 is changed to the term held by the holding unit 103 at the past time point specified by the control term.
 なお、音声対話装置20Aは、さらに、複数の保持部103のそれぞれが保持している用語を対話情報として、対話情報に基づいて処理を行う処理部に対話情報を出力し、出力の応答として処理の結果を示す情報を取得する外部処理制御部105を備えてもよい。 The voice interaction device 20A further outputs the interaction information to a processing unit that performs processing based on the interaction information, using the terms held by each of the plurality of storage units 103 as interaction information, and processes the response as an output response. An external processing control unit 105 that acquires information indicating the result of the above may be provided.
 なお、処理部は、取得した対話情報に係る用語を検索語として情報検索を実行し、外部処理制御部105は、情報検索の結果を応答として取得してもよい。音声対話装置20Aは、さらに、外部処理制御部105が取得した情報検索の結果をユーザに提示するための提示制御部106を備えてもよい。 Note that the processing unit may perform an information search using a term related to the acquired dialogue information as a search term, and the external processing control unit 105 may acquire a result of the information search as a response. The voice interactive apparatus 20A may further include a presentation control unit 106 for presenting a result of the information search acquired by the external processing control unit 105 to the user.
 図9は、本実施の形態の変形例に係る音声対話装置20Aの制御方法を示すフロー図である。 FIG. 9 is a flowchart showing a control method of the voice interactive apparatus 20A according to a modification of the present embodiment.
 図9に示されるように、ユーザとの音声による対話を行う音声対話装置20Aの制御方法は、取得ステップと、変更ステップとを含む。取得ステップにおいては、ユーザの音声による発話の内容を示す発話データを取得し(ステップS401)、取得した発話データに含まれる発話用語を、複数の保持部103のうち発話用語の属性に対応付けられた保持部103に保持させる(ステップS402)。変更ステップにおいては、取得ステップで取得した発話データに、対話情報を制御するための制御用語が含まれる場合に、記憶部104が記憶している履歴を参照して、複数の保持部103が保持している用語を変更する。具体的には、複数の保持部103のそれぞれが保持している用語を、制御用語により特定される過去の時点において当該保持部103が保持していた用語に変更する(ステップS403)。 As shown in FIG. 9, the control method of the voice interaction apparatus 20 </ b> A that performs voice interaction with the user includes an acquisition step and a change step. In the acquisition step, utterance data indicating the content of the utterance by the user's voice is acquired (step S 401), and the utterance term included in the acquired utterance data is associated with the attribute of the utterance term in the plurality of holding units 103. Held by the holding unit 103 (step S402). In the change step, when the utterance data acquired in the acquisition step includes a control term for controlling the conversation information, the plurality of holding units 103 holds the history stored in the storage unit 104 with reference to the history stored in the storage unit 104 Change the terminology used. Specifically, the term held by each of the plurality of holding units 103 is changed to the term held by the holding unit 103 at the past time point specified by the control term (step S403).
 本変形例に係る音声対話装置20Aは、音声対話装置20と同様の効果を奏する。 The voice interaction device 20A according to this modification has the same effect as the voice interaction device 20.
 [1-4.効果等]
 以上のように、本実施の形態に係る音声対話装置20は、ユーザとの音声による対話を行う音声対話装置20であって、対話の内容を示す対話情報を保持するための複数のスロット31と、履歴テーブル32と、発話データ取得部22と、操作部25とを備える。複数のスロット31は、それぞれが用語の属性に対応付けられており、対応付けられた属性を有する用語を保持する。履歴テーブル32は、複数のスロット31が保持する用語の履歴を記憶する。発話データ取得部22は、ユーザの音声による発話の内容を示す発話データを取得し、取得した発話データに含まれる発話用語を、複数のスロット31のうち発話用語の属性に対応付けられたスロットに保持させる。操作部25は、発話データ取得部22が取得した発話データに、対話情報を制御するための制御用語が含まれる場合に、履歴テーブル32が記憶している履歴を参照して、複数のスロット31が保持している用語を変更する。具体的には、複数のスロット31のそれぞれが保持している用語を、制御用語により特定される過去の時点において当該スロット31が保持していた用語に変更する。
[1-4. Effect]
As described above, the voice dialogue apparatus 20 according to the present embodiment is a voice dialogue apparatus 20 that performs dialogue with a user by voice, and includes a plurality of slots 31 for holding dialogue information indicating the contents of the dialogue. A history table 32, an utterance data acquisition unit 22, and an operation unit 25. Each of the plurality of slots 31 is associated with a term attribute, and holds a term having the associated attribute. The history table 32 stores the history of terms held by the plurality of slots 31. The utterance data acquisition unit 22 acquires utterance data indicating the content of the utterance by the user's voice, and the utterance terms included in the acquired utterance data are assigned to the slots associated with the utterance term attributes among the plurality of slots 31. Hold. When the utterance data acquired by the utterance data acquisition unit 22 includes a control term for controlling dialogue information, the operation unit 25 refers to the history stored in the history table 32 and refers to the plurality of slots 31. Change the terminology held by. Specifically, the term held in each of the plurality of slots 31 is changed to the term held in the slot 31 at the past time point specified by the control term.
 これによれば、音声対話装置20は、ユーザの音声に基づいて、対話情報を過去の時点におけるものに変更する、つまり、対話情報を過去の状態に戻すことができる。ここで、過去の時点とは、ユーザの音声により定められる時点である。よって、ユーザは、過去の時点を特定するための制御用語を含む音声による発話をすることで、音声対話装置20との対話の内容である対話情報を、過去の時点におけるものに戻すことができる。このように、音声対話装置20は、ユーザとの対話の内容を簡易な方法により修正することができる。 According to this, the voice dialogue device 20 can change the dialogue information to that at the past time point based on the voice of the user, that is, the dialogue information can be returned to the past state. Here, the past time point is a time point determined by the user's voice. Therefore, the user can return the conversation information, which is the content of the conversation with the voice interaction apparatus 20, to the one at the past time point by uttering by voice including the control terms for specifying the past time point. . Thus, the voice interaction device 20 can correct the content of the dialogue with the user by a simple method.
 特に、音声対話装置20は、ユーザの音声に基づいた制御を行うことにより、ユーザとの対話の内容を簡易な方法により修正する点に特徴を有する。ユーザは、従来の音声対話装置との音声対話では対話の内容を時系列で把握することが難しいので、対話の内容をユーザが希望する過去の時点に戻すという操作を行うことが難しい。本実施の形態に係る音声対話装置20は、ユーザの音声に基づいた制御を行うので、対話の内容をユーザが希望する過去の時点に戻すことができる。そして、対話の内容が複雑になる、つまり、用語の数が増えるほど、ユーザの音声に基づく制御の優位性が高まると考えられる。 In particular, the voice dialogue apparatus 20 is characterized in that the content of the dialogue with the user is corrected by a simple method by performing control based on the voice of the user. Since it is difficult for the user to grasp the content of the dialogue in time series in the voice dialogue with the conventional voice dialogue apparatus, it is difficult to perform an operation of returning the content of the dialogue to a past time point desired by the user. Since the voice interaction device 20 according to the present embodiment performs control based on the user's voice, the content of the dialogue can be returned to a past time point desired by the user. And it is thought that the superiority of the control based on the user's voice increases as the content of the dialogue becomes more complicated, that is, the number of terms increases.
 また、技術進化に伴い対話情報がより複雑になる場合、例えば、保持部が数十個又はそれ以上あるような場合には、上記のような修正の方法の優位性が高い。なぜなら、本実施の形態に示すような十個に満たない保持部を備える音声対話装置であれば、ユーザとの対話の内容を過去の時点に戻すことを行う代わりに、保持部が保持する用語をリセットし、最初から用語を設定し直すことも現実的に可能である。しかし、音声対話装置が数十個又はそれ以上の保持部を備える場合には、保持部が保持する用語を最初から設定し直すことは煩雑であり、ユーザにとって大きな負担となるので、現実的に可能とは言い難い。このような場合に、音声対話装置20は、ユーザとの対話の内容を過去の時点に戻すことができるので、最初から設定し直すことなく、ユーザが希望する過去の時点から対話をやり直すことができる利点がある。 Also, when the dialogue information becomes more complicated as the technology evolves, for example, when there are several dozen or more holding units, the above-described correction method is highly advantageous. Because, in the case of a voice dialogue apparatus having less than ten holding units as shown in the present embodiment, the term held by the holding unit instead of returning the content of the dialogue with the user to a past time point It is practically possible to reset and reset the terminology from the beginning. However, in the case where the voice interactive apparatus includes several tens or more holding units, it is complicated to reset the terms held by the holding unit from the beginning, which is a heavy burden on the user. It is hard to say that it is possible. In such a case, since the voice interaction device 20 can return the content of the dialog with the user to a past time, it is possible to restart the dialog from a past time desired by the user without resetting from the beginning. There are advantages you can do.
 また、制御用語は、対話情報を過去の時点に変更することを示す予め定められた用語である第一用語と、上記予め定められた用語とは異なる第二用語とを含むものとしてもよい。このとき、操作部25は、発話データ取得部22が取得した発話データに第一用語及び第二用語が両方とも含まれるか否かを判定する。操作部25は、第一用語及び第二用語が両方とも発話データに含まれると判定した場合に、履歴を参照して、複数のスロット31のそれぞれが保持している用語を変更する。具体的には、複数のスロット31のうち第二用語の属性に対応付けられたスロット31が第二用語を保持していた時点において、それぞれのスロット31が保持していた用語に変更する。 Further, the control terms may include a first term that is a predetermined term indicating that the dialogue information is changed to a past time point, and a second term that is different from the predetermined term. At this time, the operation unit 25 determines whether or not both the first term and the second term are included in the utterance data acquired by the utterance data acquisition unit 22. When the operation unit 25 determines that both the first term and the second term are included in the utterance data, the operation unit 25 refers to the history and changes the term held in each of the plurality of slots 31. Specifically, when the slot 31 associated with the attribute of the second term among the plurality of slots 31 holds the second term, the term is changed to the term held by each slot 31.
 これによれば、音声対話装置20は、ユーザが発話する第一用語と第二用語とを含む制御用語を認識することによって、ユーザが希望する過去の特定の時点に対話情報を戻すことができる。このように、音声対話装置20は、ユーザの音声に基づいてユーザとの対話の内容を過去の状態に戻す際に、参照する過去の時点をより容易に特定できる。 According to this, the voice interaction device 20 can return the interaction information to a specific point in the past desired by the user by recognizing the control term including the first term and the second term spoken by the user. . As described above, the voice interaction device 20 can more easily identify the past time point to be referred to when returning the content of the dialogue with the user to the past state based on the user's voice.
 また、操作部25は、履歴テーブル32が記憶している履歴上のある時点における複数のスロット31の状態が所定の条件を満たす場合に、上記時点に復元ポイントを設定してもよい。操作部25は、設定された復元ポイントに基づいて、複数のスロット31が保持している用語を当該スロット31が過去の時点において保持していた用語に変更する。ここで過去の時点とは、復元ポイントが設定された時点のうち、複数のスロット31のうち第二用語の属性に対応付けられたスロット31が第二用語を保持していた時点である。 In addition, the operation unit 25 may set a restoration point at the above time when the state of the plurality of slots 31 at a certain time on the history stored in the history table 32 satisfies a predetermined condition. Based on the set restoration point, the operation unit 25 changes the term held by the plurality of slots 31 to the term held by the slot 31 at the past time point. Here, the past time is the time when the slot 31 associated with the attribute of the second term among the plurality of slots 31 holds the second term among the times when the restoration point is set.
 これによれば、音声対話装置20は、履歴テーブル32が記憶している各時点における複数のスロット31の状態から、その時点に復元ポイントを設定すべきかを判定する。復元ポイントを所定の条件を用いて適切に設定することにより、後に保持部が保持する用語を変更する対象となる時点を絞り込んでおくことができる。これにより、音声対話装置20は、保持部が保持する用語を変更する際に、所定の条件により絞り込まれた、より適切な過去の時点に対話の状態を戻すことができる。 According to this, the voice interaction apparatus 20 determines whether or not a restoration point should be set at that time from the state of the plurality of slots 31 at each time stored in the history table 32. By appropriately setting the restoration points using predetermined conditions, it is possible to narrow down the time points at which the terms held by the holding unit are to be changed later. Thereby, when changing the term which a holding | maintenance part hold | maintains, the voice interactive apparatus 20 can return the state of a dialog to the more suitable past time point narrowed down by predetermined conditions.
 また、音声対話装置20は、さらに、タスク制御部24を備えてもよい。タスク制御部24は、複数のスロット31のそれぞれが保持している用語を対話情報として、対話情報に基づいて処理を行うタスク処理部40に出力する。タスク処理部40は、タスク制御部24の出力に基づいて処理を行う。タスク制御部24は、タスク処理部40への出力に対する応答として、タスク処理部40による処理の結果を示す情報を取得する。 Further, the voice interaction device 20 may further include a task control unit 24. The task control unit 24 outputs the term held in each of the plurality of slots 31 as dialogue information to the task processing unit 40 that performs processing based on the dialogue information. The task processing unit 40 performs processing based on the output of the task control unit 24. The task control unit 24 acquires information indicating the result of the processing by the task processing unit 40 as a response to the output to the task processing unit 40.
 これによれば、音声対話装置20は、複数の保持部が保持する用語を外部の処理部により処理した結果をユーザに提示する。よって、ユーザは、音声対話装置20との対話の内容を反映した処理結果を取得することができる。 According to this, the voice interaction apparatus 20 presents the result of processing the terms held by the plurality of holding units by the external processing unit to the user. Therefore, the user can acquire the processing result reflecting the content of the dialogue with the voice dialogue apparatus 20.
 また、タスク処理部40は、取得した用語を検索語として情報検索を実行し、タスク制御部24は、情報検索の結果を応答として取得し、音声対話装置20は、さらに、外部処理制御部が取得した情報検索の結果をユーザに提示するための提示制御部29を備えてもよい。 In addition, the task processing unit 40 performs an information search using the acquired term as a search term, the task control unit 24 acquires a result of the information search as a response, and the voice interaction device 20 further includes an external processing control unit. You may provide the presentation control part 29 for showing the acquired information search result to a user.
 これによれば、音声対話装置20は、外部の処理部による処理の結果として、対話の内容に基づいた検索処理の結果を取得し、ユーザに提示することができる。 According to this, the voice interaction device 20 can acquire the result of the search processing based on the content of the dialogue as a result of the processing by the external processing unit and present it to the user.
 また、操作部25は、履歴において、タスク制御部24が情報検索の結果を取得した時点に、復元ポイントを設定してもよい。 Further, the operation unit 25 may set a restoration point at the time when the task control unit 24 acquires the information search result in the history.
 これによれば、音声対話装置20は、復元ポイントを用いて、保持部が保持する用語を、情報検索を行った時点におけるものに戻すことができる。情報検索を行った時点は、その結果が得られる時点でもあり、対話の中でユーザが特定しやすい時点である。このように復元ポイントを設定することで、音声対話装置20は、保持部が保持する用語を、ユーザが直感的に特定しやすい時点におけるものに戻すことができる。また、ユーザに対して、対話情報を戻す時点を特定する発話を促す情報を提示するとき、より適切な時点を復元ポイントの候補として提示できる。 According to this, the voice interaction device 20 can return the term held by the holding unit to that at the time when the information search is performed by using the restoration point. The time when the information search is performed is also the time when the result is obtained, and is the time when the user can easily specify in the dialogue. By setting the restoration point in this way, the voice interaction apparatus 20 can return the term held by the holding unit to the one at the time when the user can easily specify intuitively. In addition, when presenting information prompting the user to specify the point in time when the dialogue information is to be returned, a more appropriate point in time can be presented as a restoration point candidate.
 また、操作部25は、履歴において、タスク制御部24が情報検索の結果を取得した時点であっても、その情報検索の結果に含まれる情報が0件であった場合には、当該時点に復元ポイントを設定することを禁止してもよい。 In addition, even when the task control unit 24 obtains the information search result in the history, the operation unit 25 determines that the information search result is zero when the information search result includes 0 information. Setting a restore point may be prohibited.
 これによれば、音声対話装置20は、情報検索の結果が0件であった時点を、復元ポイントを設定する時点から除外することができる。ユーザが対話の状態を戻そうとする場合、情報検索の結果が1件以上あった時点にするのが有用と考えられる。よって、音声対話装置20は、ユーザとの対話の内容を、ユーザにとって有用な時点におけるものに戻すことができる。 According to this, the voice interaction apparatus 20 can exclude the time point when the result of the information search is 0 from the time point when the restoration point is set. When the user wants to return the dialog state, it is considered useful to set the time when there are one or more information search results. Therefore, the voice interaction device 20 can return the content of the dialog with the user to a time useful for the user.
 また、操作部25は、複数のスロット31のそれぞれが保持している用語を変更する際に、履歴上に2以上の復元ポイントがある場合には、2以上の復元ポイントのうちユーザにより特定される復元ポイントを用いて、用語を変更してもよい。 In addition, when there are two or more restoration points on the history when the terms held by each of the plurality of slots 31 are changed, the operation unit 25 is specified by the user among the two or more restoration points. You may change the terminology using a restore point.
 これによれば、音声対話装置20は、複数ある復元ポイントのうちユーザが特定する1つの復元ポイントを用いて、ユーザとの対話の内容を過去の時点におけるものに戻すことができる。これにより、ユーザは、音声対話装置20が適切と判定した時点から、ユーザが最良と考える時点を選択し、選択した時点の対話情報に戻すことができる。 According to this, the voice interaction apparatus 20 can return the content of the dialogue with the user to the one at the past time by using one restoration point specified by the user among a plurality of restoration points. Thereby, the user can select the time point that the user considers best from the time point when the voice interaction device 20 determines to be appropriate, and can return to the interaction information at the time point of selection.
 また、音声対話装置20は、さらに、2以上の復元ポイントのうち用語を変更するのに用いる1つの復元ポイントをユーザから受け付けるための応答文を生成する応答文生成部21を備えてもよい。 Further, the voice interaction device 20 may further include a response sentence generation unit 21 that generates a response sentence for accepting one restoration point used for changing a term out of two or more restoration points from the user.
 これによれば、ユーザは、音声対話装置20が対話の内容を戻す時点の候補が複数あることを、応答文により知ることができる。ユーザは、応答文に対して応答することにより、対話の内容を戻す時点を特定する。すなわち、音声対話装置20は、複数ある復元ポイントのうちからユーザにより1つの復元ポイントを特定させる。これにより、音声対話装置20は、具体的にユーザから復元ポイントの指定を受け付け、ユーザとの対話の内容を過去の時点におけるものに戻すことができる。 According to this, the user can know from the response sentence that there are a plurality of candidates when the voice dialogue device 20 returns the contents of the dialogue. The user specifies the time point for returning the content of the dialogue by responding to the response sentence. That is, the voice interactive apparatus 20 causes the user to specify one restoration point from among a plurality of restoration points. As a result, the voice interaction device 20 can specifically accept the designation of the restoration point from the user, and return the content of the dialogue with the user to that at the past time.
 また、制御用語は、対話情報を過去の時点に変更することを示す予め定められた用語である第一用語と、取得された用語の属性の名称である属性名とを含むものであってもよい。操作部25は、発話データ取得部22が取得した発話データに第一用語及び属性名が含まれるか否かを判定する。操作部25は、発話データに第一用語及び属性名が含まれると判定した場合に、履歴を参照して、複数のスロット31のそれぞれが保持している用語を変更する。具体的には、複数のスロット31のうち属性名により示される属性に対応付けられたスロット31が現在保持している用語を保持する直前の時点において、それぞれのスロット31が保持していた用語に変更する。 The control term may include a first term that is a predetermined term indicating that the dialogue information is changed to a past time point and an attribute name that is an attribute name of the acquired term. Good. The operation unit 25 determines whether or not the utterance data acquired by the utterance data acquisition unit 22 includes the first term and the attribute name. When the operation unit 25 determines that the first term and the attribute name are included in the utterance data, the operation unit 25 refers to the history and changes the term held in each of the plurality of slots 31. Specifically, the term held in each slot 31 immediately before holding the term currently held in the slot 31 associated with the attribute indicated by the attribute name among the plurality of slots 31 change.
 これによれば、音声対話装置20は、取得部が取得した制御用語に基づいて、保持部が対応付けられている属性の名称を用いて具体的に過去の時点を特定する。ユーザは、具体的な条件を指定しなくても、属性名を指定するだけで対話の内容を戻す時点を特定することができる。このように、音声対話装置20は、ユーザとの対話の内容を、より具体的な方法により修正することができる。 According to this, based on the control terms acquired by the acquisition unit, the voice interaction device 20 specifically specifies a past time point using the name of the attribute with which the holding unit is associated. Even if the user does not specify a specific condition, the user can specify the point in time when the content of the dialogue is returned simply by specifying the attribute name. Thus, the voice interaction device 20 can correct the content of the dialogue with the user by a more specific method.
 また、本実施の形態に係る音声対話システム1は、ユーザとの音声による対話を行う。音声対話システム1は、対話の内容を示す対話情報を保持するための複数のスロット31と、履歴テーブル32と、発話データ取得部22と、操作部25と、マイク13と、音声認識部14と、タスク処理部40と、音声合成部12と、スピーカ11と、表示装置10とを備える。複数のスロット31のそれぞれは、用語の属性に対応付けられており、対応付けられた属性を有する用語を保持する。履歴テーブル32は、複数のスロット31が保持する用語の履歴を記憶する。発話データ取得部22は、ユーザの音声による発話の内容を示す発話データを取得し、取得した発話データに含まれる発話用語を、複数のスロット31のうち発話用語の属性に対応付けられたスロット31に保持させる。操作部25は、発話データ取得部22が取得した発話データに、対話情報を制御するための制御用語が含まれる場合に、複数のスロット31が保持している用語を変更する。具体的には、操作部25は、履歴テーブル32が記憶している履歴を参照して、複数のスロット31のそれぞれが保持している用語を、制御用語により特定される過去の時点において当該スロット31が保持していた用語に変更する。マイク13は、ユーザの音声を取得して音声信号を生成する。音声認識部14は、マイク13が生成した音声信号に対して音声認識処理を施すことで、発話データ取得部22により取得される発話データを生成する。タスク処理部40は、複数のスロット31が保持している対話情報を取得し、取得した対話情報に対して所定の処理を施し、処理の結果を示す情報を出力する。音声合成部12は、ユーザの音声による発話に対する応答文を生成し、生成した応答文に対して音声合成処理を施すことで音声信号を生成する。スピーカ11は、音声合成部12が生成した音声信号を音声として出力する。表示装置10は、タスク処理部40が出力した処理の結果を表示する。 Also, the voice dialogue system 1 according to the present embodiment performs voice dialogue with the user. The voice dialogue system 1 includes a plurality of slots 31 for holding dialogue information indicating the contents of dialogue, a history table 32, an utterance data acquisition unit 22, an operation unit 25, a microphone 13, and a voice recognition unit 14. The task processing unit 40, the speech synthesis unit 12, the speaker 11, and the display device 10 are provided. Each of the plurality of slots 31 is associated with a term attribute, and holds a term having the associated attribute. The history table 32 stores the history of terms held by the plurality of slots 31. The utterance data acquisition unit 22 acquires utterance data indicating the content of the utterance by the user's voice, and the utterance term included in the acquired utterance data is associated with the attribute of the utterance term among the plurality of slots 31. To hold. When the utterance data acquired by the utterance data acquisition unit 22 includes control terms for controlling the conversation information, the operation unit 25 changes the terms held by the plurality of slots 31. Specifically, the operation unit 25 refers to the history stored in the history table 32 and determines the term held in each of the plurality of slots 31 at the past time point specified by the control term. It changes to the term which 31 had. The microphone 13 acquires the user's voice and generates a voice signal. The speech recognition unit 14 generates speech data acquired by the speech data acquisition unit 22 by performing speech recognition processing on the speech signal generated by the microphone 13. The task processing unit 40 acquires dialogue information held by the plurality of slots 31, performs predetermined processing on the acquired dialogue information, and outputs information indicating the processing result. The speech synthesizer 12 generates a response sentence for an utterance by the user's voice, and generates a speech signal by performing a speech synthesis process on the generated response sentence. The speaker 11 outputs the voice signal generated by the voice synthesizer 12 as voice. The display device 10 displays the processing result output by the task processing unit 40.
 これにより、上記音声対話装置20と同様の効果を奏する。 As a result, the same effect as that of the voice dialogue apparatus 20 is obtained.
 また、本実施の形態に係る音声対話装置の制御方法は、ユーザとの音声による対話を行う音声対話装置20の制御に用いることができる。音声対話装置20は、対話の内容を示す対話情報を保持するための複数のスロット31と、履歴テーブル32とを備える。複数のスロット31のそれぞれは、用語の属性に対応付けられており、対応付けられた属性を有する用語を保持する。履歴テーブル32は、複数のスロット31が保持する用語の履歴を記憶する。本実施の形態に係る音声対話装置20の制御方法は、取得ステップと、変更ステップとを含む。取得ステップにおいては、ユーザの音声による発話の内容を示す発話データを取得し、取得した発話データに含まれる発話用語を、複数のスロット31のうち発話用語の属性に対応付けられたスロット31に保持させる。変更ステップにおいては、取得ステップで取得した発話データに、対話情報を制御するための制御用語が含まれる場合に、複数のスロット31が保持している用語を変更する。具体的には、履歴テーブル32が記憶している履歴を参照して、複数のスロット31のそれぞれが保持している用語を、制御用語により特定される過去の時点において当該スロット31が保持していた用語に変更する。 Also, the method for controlling the voice interaction apparatus according to the present embodiment can be used for controlling the voice interaction apparatus 20 that performs voice conversation with the user. The voice interaction device 20 includes a plurality of slots 31 for holding dialogue information indicating the content of the dialogue and a history table 32. Each of the plurality of slots 31 is associated with a term attribute, and holds a term having the associated attribute. The history table 32 stores the history of terms held by the plurality of slots 31. The control method of the voice interaction apparatus 20 according to the present embodiment includes an acquisition step and a change step. In the acquisition step, utterance data indicating the content of the utterance by the user's voice is acquired, and the utterance term included in the acquired utterance data is held in the slot 31 associated with the attribute of the utterance term among the plurality of slots 31. Let In the change step, when the utterance data acquired in the acquisition step includes a control term for controlling the dialogue information, the term held in the plurality of slots 31 is changed. Specifically, referring to the history stored in the history table 32, the term held in each of the plurality of slots 31 is held in the slot 31 at the past time point specified by the control term. Change to a different term.
 これにより、上記音声対話装置20と同様の効果を奏する。 As a result, the same effect as that of the voice dialogue apparatus 20 is obtained.
 以上のように、本開示における技術の例示として、実施の形態を説明した。そのために、添付図面および詳細な説明を提供した。 As described above, the embodiments have been described as examples of the technology in the present disclosure. For this purpose, the accompanying drawings and detailed description are provided.
 したがって、添付図面および詳細な説明に記載された構成要素の中には、課題解決のために必須な構成要素だけでなく、上記実装を例示するために、課題解決のためには必須でない構成要素も含まれ得る。そのため、それらの必須ではない構成要素が添付図面や詳細な説明に記載されていることをもって、直ちに、それらの必須ではない構成要素が必須であるとの認定をするべきではない。 Accordingly, among the components described in the accompanying drawings and the detailed description, not only the components essential for solving the problem, but also the components not essential for solving the problem in order to illustrate the above implementation. May also be included. Therefore, it should not be immediately recognized that these non-essential components are essential as those non-essential components are described in the accompanying drawings and detailed description.
 また、上述の実施の形態は、本開示における技術を例示するためのものであるから、請求の範囲またはその均等の範囲において種々の変更、置き換え、付加、省略などを行うことができる。 In addition, since the above-described embodiment is for illustrating the technique in the present disclosure, various modifications, replacements, additions, omissions, and the like can be performed within the scope of the claims or an equivalent scope thereof.
 本開示は、簡易な方法により、ユーザとの対話の内容を修正することができる音声対話装置として有用である。例えば、本開示は、カーナビゲーション装置、スマートフォン(高機能携帯電話端末)、携帯電話端末、携帯情報端末、又は、PC(Personal Computer)のアプリケーションに適用することができる。 The present disclosure is useful as a voice dialogue apparatus that can correct the content of dialogue with the user by a simple method. For example, the present disclosure can be applied to an application of a car navigation device, a smartphone (high-function mobile phone terminal), a mobile phone terminal, a mobile information terminal, or a PC (Personal Computer).
 1  音声対話システム
 10  表示装置
 11  スピーカ
 12  音声合成部
 13  マイク
 14  音声認識部
 20,20A  音声対話装置
 21  応答文生成部
 22  発話データ取得部
 23  シーケンス制御部
 24  タスク制御部
 25  操作部
 26  解析部
 27  メモリ
 28  タスク結果解析部
 29,106  提示制御部
 31  スロット
 32,320  履歴テーブル
 40  タスク処理部
 101  取得部
 102  変更部
 103  保持部
 104  記憶部
 105  外部処理制御部
 310  対話シーケンス
 311  時刻情報
 312  発話
 313  応答
 321  必須スロット群
 322  オプションスロット群
 323  アクション
 324  復元ポイント
 330  検索結果
DESCRIPTION OF SYMBOLS 1 Voice dialogue system 10 Display apparatus 11 Speaker 12 Voice synthesizer 13 Microphone 14 Voice recognition part 20, 20A Voice dialogue apparatus 21 Response sentence production | generation part 22 Utterance data acquisition part 23 Sequence control part 24 Task control part 25 Operation part 26 Analysis part 27 Memory 28 Task result analysis unit 29, 106 Presentation control unit 31 Slot 32, 320 History table 40 Task processing unit 101 Acquisition unit 102 Change unit 103 Holding unit 104 Storage unit 105 External processing control unit 310 Dialog sequence 311 Time information 312 Utterance 313 Response 321 Required slot group 322 Optional slot group 323 Action 324 Restore point 330 Search result

Claims (13)

  1. ユーザとの音声による対話を行う音声対話装置であって、
    前記ユーザの音声による発話の内容を示す発話データを取得する取得部と、
    それぞれが、前記発話データに含まれる用語と、前記用語の属性とを対応づけて保持する複数の保持部と、
    前記複数の保持部が保持する用語の履歴を記憶する記憶部と、
    前記発話データに所定の制御用語が含まれる場合に、前記記憶部が記憶している前記履歴を参照して、前記制御用語により特定される過去の時点において前記複数の保持部が保持していた用語となるように、前記複数の保持部の用語を変更する変更部とを備える
    音声対話装置。
    A voice dialogue device that performs voice dialogue with a user,
    An acquisition unit for acquiring utterance data indicating the content of the utterance of the user's voice;
    A plurality of holding units each holding a term included in the utterance data and the attribute of the term in association with each other;
    A storage unit for storing a history of terms held by the plurality of holding units;
    When a predetermined control term is included in the utterance data, referring to the history stored in the storage unit, the plurality of holding units held in the past time point specified by the control term A spoken dialogue apparatus comprising: a changing unit that changes the terms of the plurality of holding units so as to be terms.
  2. 前記制御用語は、
    前記ユーザとの対話情報を過去の時点に変更することを示す予め定められた用語である第一用語と、前記予め定められた用語とは異なる第二用語とを含み、
    前記変更部は、
    前記発話データに前記第一用語及び前記第二用語が含まれるか否かを判定し、前記第一用語及び前記第二用語が含まれると判定した場合に、前記履歴を参照して、前記複数の保持部のうち前記第二用語の属性に対応する属性を保持する対応保持部を特定し、前記対応保持部が前記第二用語を保持していた時点において前記複数の保持部が保持していた用語となるように、前記複数の保持部の用語を変更する
    請求項1に記載の音声対話装置。
    The control term is:
    A first term that is a predetermined term indicating that the dialogue information with the user is changed to a past time point, and a second term that is different from the predetermined term,
    The changing unit is
    It is determined whether or not the first term and the second term are included in the utterance data, and when it is determined that the first term and the second term are included, the plurality of the plurality of the utterance data are referred to by referring to the history. The corresponding holding unit that holds the attribute corresponding to the attribute of the second term is specified, and the plurality of holding units hold the second holding term when the corresponding holding unit holds the second term. The spoken dialogue apparatus according to claim 1, wherein the terms of the plurality of holding units are changed so as to be different terms.
  3. 前記変更部は、
    前記記憶部が記憶している前記履歴上のある時点における複数の保持部の状態が所定の条件を満たす場合に、前記時点に復元ポイントを設定し、
    前記復元ポイントが設定された時点のうち前記対応保持部が前記第二用語を保持していた時点において前記複数の保持部が保持していた用語となるように、前記複数の保持部の用語を変更する
    請求項2に記載の音声対話装置。
    The changing unit is
    When the state of a plurality of holding units at a certain point on the history stored in the storage unit satisfies a predetermined condition, a restoration point is set at the point in time,
    The terms of the plurality of holding units are set such that the corresponding holding unit is the term held by the plurality of holding units when the corresponding holding unit holds the second term among the time points when the restoration points are set. The spoken dialogue apparatus according to claim 2 to be changed.
  4. さらに、
    前記複数の保持部が保持している用語を含む前記対話情報を処理する処理部から処理の結果を取得し、前記処理の結果を示す情報を取得する外部処理制御部を備える
    請求項3に記載の音声対話装置。
    further,
    The external processing control part which acquires the result of a process from the process part which processes the said dialogue information containing the term which these holding | maintenance parts hold | maintain, and acquires the information which shows the result of the said process is provided. Voice interaction device.
  5. 前記処理部は、前記対話情報に係る用語を検索語として情報検索を実行し、
    前記外部処理制御部は、前記情報検索の結果を前記処理の結果として取得し、
    さらに、
    前記外部処理制御部が取得した前記情報検索の結果を前記ユーザに提示するための提示制御部を備える
    請求項4に記載の音声対話装置。
    The processing unit performs an information search using a term related to the dialogue information as a search word,
    The external processing control unit acquires a result of the information search as a result of the processing,
    further,
    The voice interaction apparatus according to claim 4, further comprising a presentation control unit for presenting the information search result acquired by the external processing control unit to the user.
  6. 前記変更部は、前記履歴において、前記外部処理制御部が前記情報検索の結果を取得した時点に、前記復元ポイントを設定する
    請求項5に記載の音声対話装置。
    The voice interaction device according to claim 5, wherein the changing unit sets the restoration point at the time when the external processing control unit acquires the information search result in the history.
  7. 前記変更部は、前記履歴において、前記外部処理制御部が前記情報検索の結果を取得した時点であっても、前記情報検索の結果に含まれる情報が0件であった時点には、前記復元ポイントを設定することを禁止する
    請求項6に記載の音声対話装置。
    In the history, the changing unit is configured to restore the restoration when the information included in the information search result is 0 even when the external processing control unit acquires the information search result. The voice interactive apparatus according to claim 6, wherein setting of points is prohibited.
  8. 前記変更部は、前記履歴上に2以上の前記復元ポイントがある場合には、前記2以上の復元ポイントのうち前記ユーザにより特定される復元ポイントにおいて前記複数の保持部が保持していた用語となるように、前記複数の保持部の用語を変更する
    請求項3に記載の音声対話装置。
    When there are two or more restoration points on the history, the changing unit is a term held by the plurality of holding units at a restoration point specified by the user among the two or more restoration points. The spoken dialogue apparatus according to claim 3, wherein the terms of the plurality of holding units are changed.
  9. さらに、
    前記2以上の復元ポイントのうち前記複数の保持部の用語を変更するのに用いる1つの復元ポイントの選択を前記ユーザから受け付けるための応答文を生成する応答文生成部を備える
    請求項8に記載の音声対話装置。
    further,
    The response sentence generation part which generates the response sentence for accepting selection of one restoration point used for changing the term of a plurality of above-mentioned restoration parts among the two or more restoration points from the user. Voice interaction device.
  10. 前記制御用語は、
    前記ユーザとの対話情報を過去の時点に変更することを示す予め定められた用語である第一用語と、前記属性の名称である属性名とを含み、
    前記変更部は、
    前記発話データに前記第一用語及び前記属性名が含まれるか否かを判定し、前記第一用語及び前記属性名が含まれると判定した場合に、前記履歴を参照して、前記複数の保持部のうち前記属性名により示される属性を保持する対応保持部を特定し、前記対応保持部が現在保持している用語を保持する直前の時点において前記複数の保持部が保持していた用語となるように、前記複数の保持部の用語を変更する
    請求項1に記載の音声対話装置。
    The control term is:
    A first term that is a predetermined term indicating that the dialogue information with the user is changed to a past time point, and an attribute name that is a name of the attribute,
    The changing unit is
    It is determined whether or not the first term and the attribute name are included in the utterance data, and when it is determined that the first term and the attribute name are included, the plurality of holdings are referred to the history. Identifying the correspondence holding unit that holds the attribute indicated by the attribute name, and the term held by the plurality of holding units immediately before holding the term currently held by the correspondence holding unit; The spoken dialogue apparatus according to claim 1, wherein the terms of the plurality of holding units are changed.
  11. ユーザとの音声による対話を行う音声対話装置であって、
    前記ユーザとの対話の内容を示す対話情報を保持するための複数の保持部と、
    前記ユーザとの対話を制御する制御回路と、を備え、
    前記複数の保持部は、
    それぞれが用語の属性に対応付けられ、
    前記制御回路は、
    前記ユーザの音声による発話の内容を示す発話データを取得し、
    取得した前記発話データに含まれる発話用語を、前記複数の保持部のうち前記発話用語の属性に対応付けられた保持部に保持させ、
    前記発話データに所定の制御用語が含まれる場合に、前記制御用語により特定される過去の時点において前記複数の保持部が保持していた用語となるように、前記複数の保持部の用語を変更する
    音声対話装置。
    A voice dialogue device that performs voice dialogue with a user,
    A plurality of holding units for holding dialogue information indicating the content of dialogue with the user;
    A control circuit for controlling a dialogue with the user,
    The plurality of holding portions are
    Each is associated with a term attribute,
    The control circuit includes:
    Obtaining utterance data indicating the content of the utterance of the user's voice;
    The utterance term included in the acquired utterance data is held in a holding unit associated with the attribute of the utterance term among the plurality of holding units,
    When a predetermined control term is included in the utterance data, the terminology of the plurality of holding units is changed so that the term is held by the plurality of holding units at a past time specified by the control term. A voice interaction device.
  12. 請求項1から11のいずれか1項に記載の音声対話装置と、
    前記ユーザの音声を取得して音声信号を生成するマイクと、
    前記マイクが生成した前記音声信号に対して音声認識処理を施すことで、前記取得部により取得される前記発話データを生成する音声認識部と、
    前記複数の保持部が保持している用語を含む対話情報を取得し、取得した前記対話情報に対して所定の処理を施し、前記処理の結果を出力する処理部と、
    前記ユーザの音声による発話に対する応答文を生成し、生成した前記応答文に対して音声合成処理を施すことで音声信号を生成する音声合成部と、
    前記音声合成部が生成した前記音声信号を音声として出力するスピーカと、
    前記処理部が出力した前記処理の結果を表示する表示装置とを備える
    音声対話システム。
    The voice interaction device according to any one of claims 1 to 11,
    A microphone for acquiring the user's voice and generating a voice signal;
    A voice recognition unit that generates the utterance data acquired by the acquisition unit by performing a voice recognition process on the voice signal generated by the microphone;
    A processing unit that acquires dialogue information including terms held by the plurality of holding units, performs predetermined processing on the acquired dialogue information, and outputs a result of the processing;
    A speech synthesizer that generates a response signal to an utterance by the user's voice and generates a speech signal by performing speech synthesis processing on the generated response statement;
    A speaker that outputs the voice signal generated by the voice synthesizer as voice;
    A speech dialogue system comprising: a display device that displays a result of the processing output by the processing unit.
  13. ユーザとの音声による対話を行う音声対話装置の制御方法であって、
    前記音声対話装置は、
    それぞれが、前記ユーザの音声による発話の内容を示す発話データに含まれる用語と、前記用語の属性とを対応づけて保持する複数の保持部と、
    前記複数の保持部が保持する用語の履歴を記憶する記憶部とを備え、
    前記制御方法は、
    前記発話データを取得する取得ステップと、
    前記取得ステップで取得した前記発話データに、所定の制御用語が含まれる場合に、前記記憶部が記憶している前記履歴を参照して、前記制御用語により特定される過去の時点において前記複数の保持部が保持していた用語となるように、前記複数の保持部の用語を変更する変更ステップとを含む
    制御方法。
    A method for controlling a voice dialogue apparatus that performs voice dialogue with a user,
    The voice interaction device
    A plurality of holding units that hold the terms included in the utterance data indicating the content of the utterances by the user's voice and the attributes of the terms in association with each other,
    A storage unit that stores a history of terms held by the plurality of holding units,
    The control method is:
    An acquisition step of acquiring the utterance data;
    When a predetermined control term is included in the utterance data obtained in the obtaining step, the history stored in the storage unit is referred to, and the plurality of the plurality of utterance data at the past time point specified by the control term And a changing step of changing the terms of the plurality of holding units so that the terms held by the holding unit are obtained.
PCT/JP2016/000855 2015-02-27 2016-02-18 Voice interaction device, voice interaction system, control method of voice interaction device WO2016136208A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2015-039573 2015-02-27
JP2015039573A JP2018063272A (en) 2015-02-27 2015-02-27 Voice dialogue apparatus, voice dialogue system, and control method of voice dialogue apparatus

Publications (1)

Publication Number Publication Date
WO2016136208A1 true WO2016136208A1 (en) 2016-09-01

Family

ID=56788243

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2016/000855 WO2016136208A1 (en) 2015-02-27 2016-02-18 Voice interaction device, voice interaction system, control method of voice interaction device

Country Status (2)

Country Link
JP (1) JP2018063272A (en)
WO (1) WO2016136208A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019220791A1 (en) * 2018-05-14 2019-11-21 株式会社Nttドコモ Dialogue device

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH086955A (en) * 1994-06-16 1996-01-12 Canon Inc Device and method for retrieving information
JP2001022779A (en) * 1999-07-09 2001-01-26 Nissan Motor Co Ltd Interactive information retrieval device, method for interactive information retrieval using computer, and computer-readable medium where program performing interactive information retrieval is recorded
JP2003223187A (en) * 2001-11-20 2003-08-08 Koninkl Philips Electronics Nv Method of operating speech dialogue system
JP2008506156A (en) * 2004-07-06 2008-02-28 ボクシィファイ, インコーポレイテッド Multi-slot interaction system and method
US20120081371A1 (en) * 2009-05-01 2012-04-05 Inci Ozkaragoz Dialog design tool and method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH086955A (en) * 1994-06-16 1996-01-12 Canon Inc Device and method for retrieving information
JP2001022779A (en) * 1999-07-09 2001-01-26 Nissan Motor Co Ltd Interactive information retrieval device, method for interactive information retrieval using computer, and computer-readable medium where program performing interactive information retrieval is recorded
JP2003223187A (en) * 2001-11-20 2003-08-08 Koninkl Philips Electronics Nv Method of operating speech dialogue system
JP2008506156A (en) * 2004-07-06 2008-02-28 ボクシィファイ, インコーポレイテッド Multi-slot interaction system and method
US20120081371A1 (en) * 2009-05-01 2012-04-05 Inci Ozkaragoz Dialog design tool and method

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019220791A1 (en) * 2018-05-14 2019-11-21 株式会社Nttドコモ Dialogue device
JPWO2019220791A1 (en) * 2018-05-14 2021-02-12 株式会社Nttドコモ Dialogue device
JP7033195B2 (en) 2018-05-14 2022-03-09 株式会社Nttドコモ Dialogue device

Also Published As

Publication number Publication date
JP2018063272A (en) 2018-04-19

Similar Documents

Publication Publication Date Title
US10733983B2 (en) Parameter collection and automatic dialog generation in dialog systems
US11004444B2 (en) Systems and methods for enhancing user experience by communicating transient errors
JP6588637B2 (en) Learning personalized entity pronunciation
KR101418163B1 (en) Speech recognition repair using contextual information
US9002708B2 (en) Speech recognition system and method based on word-level candidate generation
CN105592343B (en) Display device and method for question and answer
US10860289B2 (en) Flexible voice-based information retrieval system for virtual assistant
US9484034B2 (en) Voice conversation support apparatus, voice conversation support method, and computer readable medium
JP6726354B2 (en) Acoustic model training using corrected terms
US10838954B1 (en) Identifying user content
WO2016136207A1 (en) Voice interaction device, voice interaction system, control method of voice interaction device, and program
JP2015176099A (en) Dialog system construction assist system, method, and program
CN109326284B (en) Voice search method, apparatus and storage medium
JP5616390B2 (en) Response generation apparatus, response generation method, and response generation program
JP6622165B2 (en) Dialog log analysis apparatus, dialog log analysis method and program
JPWO2016143131A1 (en) Dialog support apparatus, method and program, and terminal
WO2019035373A1 (en) Information processing device, information processing method, and program
US20210065708A1 (en) Information processing apparatus, information processing system, information processing method, and program
WO2016136208A1 (en) Voice interaction device, voice interaction system, control method of voice interaction device
JP2015143866A (en) Voice recognition apparatus, voice recognition system, voice recognition method, and voice recognition program
JPWO2005076259A1 (en) Voice input system, voice input method, and voice input program
WO2022271555A1 (en) Early invocation for contextual data processing
KR20230075386A (en) Method and apparatus for speech signal processing
JP2015096923A (en) Information processing device and method

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16754959

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

NENP Non-entry into the national phase

Ref country code: JP

122 Ep: pct application non-entry in european phase

Ref document number: 16754959

Country of ref document: EP

Kind code of ref document: A1