WO2014181524A1 - Conversation processing system and program - Google Patents

Conversation processing system and program Download PDF

Info

Publication number
WO2014181524A1
WO2014181524A1 PCT/JP2014/002348 JP2014002348W WO2014181524A1 WO 2014181524 A1 WO2014181524 A1 WO 2014181524A1 JP 2014002348 W JP2014002348 W JP 2014002348W WO 2014181524 A1 WO2014181524 A1 WO 2014181524A1
Authority
WO
WIPO (PCT)
Prior art keywords
conversation
unit
user
algorithm
category
Prior art date
Application number
PCT/JP2014/002348
Other languages
French (fr)
Japanese (ja)
Inventor
孫 正義
筒井 多圭志
康介 朝長
賢之 鎌谷
輝 稲葉
Original Assignee
ソフトバンクモバイル株式会社
ソフネック株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ソフトバンクモバイル株式会社, ソフネック株式会社 filed Critical ソフトバンクモバイル株式会社
Publication of WO2014181524A1 publication Critical patent/WO2014181524A1/en

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/04Real-time or near real-time messaging, e.g. instant messaging [IM]
    • H04L51/046Interoperability with other network applications or services
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/226Procedures used during a speech recognition process, e.g. man-machine dialogue using non-speech characteristics
    • G10L2015/227Procedures used during a speech recognition process, e.g. man-machine dialogue using non-speech characteristics of the speaker; Human-factor methodology
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/63Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for estimating an emotional state

Definitions

  • the present invention relates to a conversation processing system and a program.
  • Patent Document 1 Japanese Patent Application Laid-Open No. 2011-253389
  • a conversation system that can execute a flexible conversation according to the user's condition is desired.
  • a voice acquisition unit that acquires a user's voice
  • an emotion recognition unit that recognizes a user's emotion based on the voice acquired by the voice acquisition unit
  • each of a plurality of types of emotions
  • a first conversation algorithm storage unit that stores a first conversation algorithm (for example, emotion conversation algorithm) in association with the first conversation stored in the first conversation algorithm storage unit in association with the emotion recognized by the emotion recognition unit.
  • a conversation processing system includes a conversation algorithm selection unit that selects an algorithm and a conversation execution unit that executes a conversation with a user according to a first conversation algorithm selected by the conversation algorithm selection unit.
  • the conversation processing system includes a second conversation algorithm storage unit that stores an execution condition in association with each of a plurality of second conversation algorithms (for example, a conditional conversation algorithm), and a plurality of second conversation algorithm storage units that are stored.
  • An execution condition determining unit that determines that any one of the execution conditions is satisfied, and the conversation algorithm selecting unit is associated with the execution condition determined to be satisfied by the execution condition determining unit.
  • the second conversation algorithm stored in the second conversation algorithm storage unit may be selected, and the conversation execution unit may execute the conversation with the user according to the second conversation algorithm selected by the conversation algorithm selection unit.
  • the voice acquisition unit may acquire the user's voice while the conversation execution unit is executing a conversation with the user according to the second conversation algorithm.
  • the conversation according to the second conversation algorithm selected by may be interrupted, and the conversation according to the first conversation algorithm selected by the conversation algorithm selection unit may be started.
  • the second conversation algorithm storage unit may further store a priority in association with each of the plurality of second conversation algorithms, and the emotion recognition unit may select the second conversation selected by the conversation algorithm selection unit.
  • the conversation execution unit is executing a conversation with the user according to the algorithm
  • the user's emotion may be recognized based on the voice acquired by the voice acquisition unit, and the conversation processing system may recognize the user's emotion recognized by the emotion recognition unit.
  • You may further provide the 1st priority change part which changes the priority memorize
  • the conversation processing system selects one category from a plurality of categories based on an utterance content storage unit that stores a plurality of utterance contents in association with each of a plurality of categories and a voice acquired by the voice acquisition unit And a conversation execution unit that executes a conversation with the user based on a plurality of utterance contents stored in association with one category selected by the category selection unit.
  • the voice acquisition unit may acquire the voice of the user while the conversation execution unit executes a conversation with the user based on a plurality of utterance contents stored in association with one category. Interrupts the conversation with the user based on a plurality of utterance contents stored in association with one category, and starts a conversation according to the first conversation algorithm selected by the conversation algorithm selection unit.
  • the utterance content storage unit may further store a priority in association with each of the plurality of categories, and the emotion recognition unit may use the plurality of utterance contents stored in association with one category.
  • the conversation execution unit is executing a conversation with the user
  • the user's emotion may be recognized based on the voice acquired by the voice acquisition unit, and the conversation processing system recognizes the user's emotion recognized by the emotion recognition unit.
  • a second priority changing unit that changes the priority stored in association with one category may be further provided.
  • the category selection unit determines the priority when switching the conversation with the user executed by the conversation execution unit according to a plurality of utterance contents stored in association with the selected category. On the basis of this, a category other than one category may be selected, and the conversation execution unit may associate a plurality of categories stored in the utterance content storage unit in association with a category other than the one category selected by the category selection unit. Depending on the utterance content, a conversation with the user may be started.
  • the plurality of categories may have a hierarchical structure, and the category selection unit is executed by the conversation execution unit according to the plurality of utterance contents stored in association with the selected one category.
  • a category adjacent to one category may be selected, and the conversation execution unit associates the category with the category adjacent to the one category selected by the category selection unit.
  • a conversation with the user may be started based on a plurality of utterance contents stored in the storage unit.
  • the conversation processing system includes: a voice output control unit that outputs a first utterance content included in a plurality of utterance contents stored in association with one category selected by the category selection unit; and a first utterance content And a conversation data generation unit that generates conversation data including the first response content of the user 10 with respect to the first utterance content.
  • the conversation execution unit may include the first utterance content and the first utterance content.
  • the voice acquisition unit acquires the voice that matches the first utterance content
  • the first response content may be output by voice
  • the voice acquisition unit The second response content corresponding to the first response content output by the voice may be acquired, and the conversation data generation unit generates the conversation data associated in the order of the first utterance content, the first response content, and the second response content. It's okay.
  • the second conversation algorithm stored in the second conversation algorithm storage unit of the first information terminal possessed by the first user is a standard in which the similarity with the profile of the first user is determined in advance.
  • An algorithm sharing processing unit for copying to a second conversation algorithm storage unit of a second information terminal possessed by a second user having a profile exceeding the threshold may be further provided.
  • An example of the communication environment of the information terminal 100 is shown schematically.
  • An example of the conversation process by the information terminal 100 is shown schematically.
  • the other example of the conversation process by the information terminal 100 is shown schematically.
  • the function structure of the information terminal 100 is shown schematically.
  • movement flow by the information terminal 100 is shown roughly.
  • movement flow by the information terminal 100 is shown roughly.
  • An example of the conversation algorithm corresponding to pleasure is shown roughly.
  • An example of the conversation algorithm corresponding to anger is shown roughly.
  • An example of the conversation algorithm corresponding to sadness is shown roughly.
  • An example of the hierarchical structure of the category 44 is shown schematically.
  • FIG. 1 schematically shows an example of a communication environment of the information terminal 100.
  • the information terminal 100 has a voice input function and a voice output function, and executes a conversation with the user 10.
  • the information terminal 100 is a mobile phone such as a smartphone, for example.
  • the information terminal 100 may execute a conversation with the user 10 alone. Further, the information terminal 100 may execute a conversation with the user 10 in cooperation with the server 200 that can communicate via the communication network 20.
  • the information terminal 100 may be an example of a conversation processing system. Further, the conversation processing system may be configured by the information terminal 100 and the server 200.
  • the information terminal 100 may be any device that has a voice input function and a voice output function.
  • the information terminal 100 may be a tablet terminal, a PC, a home appliance, a car, a car navigation system, a robot, a stuffed toy, or the like. It may be.
  • FIG. 2 schematically shows an example of conversation processing by the information terminal 100.
  • the information terminal 100 may execute a conversation with the user 10 in accordance with the conversation algorithm stored in the conversation algorithm DB 30.
  • the information terminal 100 may hold the conversation algorithm DB 30.
  • the server 200 may hold
  • the conversation algorithm DB 30 may be held by devices other than the information terminal 100 and the server 200.
  • the information terminal 100 may execute the conversation algorithm corresponding to the execution condition when the execution condition registered in the execution condition table 32 is satisfied. For example, the information terminal 100 executes the “program notification application” when there is a television program of a genre that the user 10 likes in the near future. For example, the information terminal 100 receives a TV program data providing service and acquires the profile data of the user 10 to determine that there is a TV program of a genre that the user 10 likes soon.
  • the “program notification application” is a conversation for notifying the user 10 that a TV program of a genre that the user 10 likes is broadcast based on the TV program data, or answering an inquiry from the user 10 regarding the program content. It may be an algorithm. For example, the “program notification application” may give a voice notification that there is a music program from 20:00 today, and may respond with “silver bomb will come out” in response to an inquiry from the user 10 about who will appear.
  • the “commuting assistance application” may be a conversation algorithm for notifying the user 10 of events related to commuting and answering inquiries from the user 10 regarding commuting.
  • the information terminal 100 executes the “commuting assistance application” when the operation information data received from the server that provides the train operation information indicates the delay of the commuter train.
  • the “event notification application” may be a conversation algorithm that notifies the user 10 that the event registered in the calendar is about to be held, or answers an inquiry from the user 10 regarding the contents of the event.
  • the information terminal 100 executes the “event notification application” on the day before the event registered in the calendar, for example.
  • the “umbrella alert app” may be a conversation algorithm that makes a voice notification that prompts the user 10 to hold an umbrella or answers an inquiry from the user 10 regarding the weather according to the probability of precipitation.
  • the information terminal 100 executes the “umbrella alert app” when the weather forecast data received from a server that provides weather forecast information occupies a precipitation probability of 30% or more.
  • the “blood group fortune-telling application” may be a conversation algorithm for voice notification of the content of fortune-telling for the blood type designated by the user 10.
  • the information terminal 100 executes the “blood type fortune-telling app” when receiving an instruction to start the “blood type fortune-telling app”.
  • these conversation algorithms are examples, and the conversation algorithm DB 30 may store other conversation algorithms.
  • the execution condition table 32 may store priorities in association with each conversation algorithm.
  • the information terminal 100 may select a conversation algorithm based on the priority. For example, when there are a plurality of conversation algorithms that satisfy the execution condition, the information terminal 100 may execute a conversation algorithm with a high priority or may execute a conversation algorithm in descending order of priority. Further, the information terminal 100 may select a conversation algorithm having a higher priority than the conversation algorithm currently being executed when the topic is switched halfway.
  • FIG. 3 schematically shows another example of conversation processing by the information terminal 100.
  • the information terminal 100 may execute a conversation with the user 10 based on the plurality of utterance contents 46 stored in the conversation DB 40. Further, the information terminal 100 may execute a conversation with the user 10 based on the plurality of utterance contents 48 stored in the Q & ADB 42.
  • the conversation terminal 40 and the Q & ADB 42 may be held by the information terminal 100.
  • the conversation DB 40 and the Q & ADB 42 may be held by the server 200.
  • the conversation DB 40 and the Q & ADB 42 may be held by devices other than the information terminal 100 and the server 200.
  • the conversation DB 40 stores a plurality of utterance contents 46 in association with each of the plurality of categories 44.
  • the information terminal 100 may select one category from a plurality of categories based on the voice of the user 10. For example, when the information terminal 100 determines that the voice of the user 10 matches any of the plurality of utterance contents 46, the information terminal 100 puts the category 44 including the utterance contents 46 into a selected state. For example, if the information terminal 100 determines that the keyword included in the voice of the user 10 matches the name of the category 44, the information terminal 100 may select the category 44.
  • the information terminal 100 Select “Entertainment: Musician: Silver Bomber”. Further, the information terminal 100 determines that the voice of the user 10 does not match any of the plurality of utterance contents 46 associated with “Entertainment: Musician: Silver Bomber”. If the keyword is included, “Entertainment: Musician: Silver Bomber” may be selected.
  • the information terminal 100 may output one of a plurality of utterance contents 46 associated with the selected category 44 as a response to the inquiry from the user 10. For example, the information terminal 100 may respond with “It is an air band” in response to the user 10's question “It ’s fun! Thereby, a highly relevant response can be made to the voice of the user 10.
  • the conversation DB 40 may store the priority 45 in association with each of the plurality of categories 44.
  • the information terminal 100 may select another category 44 based on the priority 45 when switching the selected category 44 to another category 44.
  • the information terminal 100 selects the category 44 without being based on the voice of the user 10, such as when the information terminal 100 speaks to the user 10, the information terminal 100 selects the category 44 based on the priority 45. It's okay.
  • the Q & ADB 42 stores a plurality of utterance contents 48 in association with each of the plurality of categories 44. Each of the plurality of utterance contents 48 may be associated with a question.
  • the information terminal 100 may output the utterance content 48 corresponding to the question as a voice. .
  • FIG. 3 if “Entertainment: Musician: Silver Bomber” is selected and the voice of the user 10 is “Who is the vocalist?”, The information terminal 100 displays “Kirishu”. "It's right”.
  • the information terminal 100 executes a conversation with the user 10 using the conversation algorithm DB 30, the conversation DB 40, the Q & ADB 42, and the like.
  • the information terminal 100 uses an emotion recognition result for the voice of the user 10 in order to realize a more flexible conversation.
  • FIG. 4 schematically shows a functional configuration of the information terminal 100.
  • conversation algorithm DB 30, conversation DB 40, Q & ADB 42, execution condition determination unit 112, condition data acquisition unit 116, conversation algorithm selection unit 118, conversation execution unit 120, voice output control unit 122, voice acquisition unit 124, category selection unit 126 The information terminal 100 including the emotion recognition unit 128, the priority change unit 130, the utterance content registration unit 132, the conversation data generation unit 134, and the algorithm sharing processing unit 136 will be described.
  • the conversation algorithm DB 30 stores a plurality of conversation algorithms.
  • the conversation algorithm DB 30 stores an emotion conversation algorithm 36 associated with each of a plurality of emotion types.
  • the emotion conversation algorithm 36 may be an example of a first conversation algorithm.
  • the conversation algorithm DB 30 may be an example of a first conversation algorithm storage unit.
  • the conversation algorithm DB 30 stores a plurality of conditional conversation algorithms 34 each associated with an execution condition and an execution condition table 32.
  • the conditional conversation algorithm 34 may be an example of a second conversation algorithm.
  • the conversation algorithm DB 30 may be an example of a second conversation algorithm storage unit.
  • the conversation DB 40 stores a plurality of utterance contents 46 in association with each of the plurality of categories 44.
  • the Q & ADB 42 stores a plurality of combinations of the question and the utterance content 48 in association with each of the plurality of categories 44.
  • the conversation DB 40 and the Q & ADB 42 may be an example of an utterance content storage unit.
  • the execution condition determination unit 112 determines that any of a plurality of execution conditions registered in the execution condition table 32 is satisfied.
  • the execution condition determination unit 112 refers to the condition data acquired by the condition data acquisition unit 116 that acquires the condition data related to the execution condition and the execution condition table 32, so that one of the plurality of execution conditions is satisfied. You may judge that.
  • condition data acquisition unit 116 acquires the genre of the television program that the user 10 likes from the profile data of the user 10.
  • the condition data acquisition unit 116 acquires television program data via the communication network 20.
  • the execution condition determination unit 112 may determine that the execution condition of the “program notification application” is satisfied from the TV program genre that the user 10 likes, the TV program data, and the execution condition table 32.
  • the conversation algorithm selection unit 118 selects the conditional conversation algorithm 34 stored in the conversation algorithm DB 30 in association with the execution condition determined to be satisfied by the execution condition determination unit 112.
  • the conversation execution unit 120 executes a conversation with the user 10 in accordance with the conditional conversation algorithm 34 selected by the conversation algorithm selection unit 118.
  • the conversation execution unit 120 executes a conversation with the user 10 based on the utterance contents 46 and the utterance contents 48 stored in the conversation DB 40 and the Q & ADB 42.
  • the conversation execution unit 120 applies a speech synthesis technique to the utterance contents included in the conditional conversation algorithm 34, the utterance contents 46 and the utterance contents 48 stored in the conversation DB 40 and the Q & ADB 42, thereby providing a voice output control unit.
  • the voice is output to 122.
  • the conversation execution unit 120 may cause the voice output control unit 122 to output the voice data.
  • the conversation execution unit 120 may recognize the voice of the user 10 by applying voice recognition technology to the voice of the user 10 acquired by the voice acquisition unit 124. As described above, the conversation execution unit 120 may execute the conversation with the user 10 by controlling the voice output control unit 122 and the voice acquisition unit 124.
  • the voice output control unit 122 outputs voice according to the control of the conversation execution unit 120.
  • the voice output control unit 122 may output the voice data and voice synthesis data specified by the conversation execution unit 120 as voices using a speaker or the like.
  • the voice acquisition unit 124 acquires the voice of the user 10 using a microphone or the like.
  • the voice acquisition unit 124 may transmit the acquired voice to the conversation execution unit 120.
  • the voice acquisition unit 124 may transmit the acquired voice to the category selection unit 126.
  • the category selection unit 126 sets one category 44 among the plurality of categories 44 based on the voice acquired by the voice acquisition unit 124. For example, when the category selection unit 126 determines that the voice acquired by the voice acquisition unit 124 matches any of the plurality of utterance contents 46 stored in the conversation DB 40, the category selection unit 126 selects the category 44 including the utterance contents 46 as a selected state. To do. For example, the category selection unit 126 determines that they match when the degree of coincidence between the voice recognition result of the voice acquired by the voice acquisition unit 124 and the utterance content 46 is higher than a predetermined threshold.
  • the category selection unit 126 determines that the keyword included in the voice acquired by the voice acquisition unit 124 matches the name of the category 44, the category selection unit 126 sets the category 44 in a selected state. For example, the category selection unit 126 determines that a match is found when the degree of matching between the keyword included in the voice acquired by the voice acquisition unit 124 and the name of the category 44 is higher than a predetermined threshold.
  • the conversation execution unit 120 may execute a conversation with the user 10 using the plurality of utterance contents 46 and utterance contents 48 stored in the conversation DB 40 and the Q & ADB 42 in association with the category 44 selected by the category selection unit 126.
  • the voice acquisition unit 124 may transmit the acquired voice to the emotion recognition unit 128.
  • the emotion recognition unit 128 recognizes the emotion of the user 10 based on the voice of the user 10 acquired by the voice acquisition unit 124.
  • the emotion recognition unit 128 may recognize the emotion of the user 10 using existing voice emotion recognition technology.
  • the emotion recognition unit 128 may recognize the emotion of the user 10 based on the prosody, sound quality, and phoneme of the user 10 voice.
  • the emotion recognition unit 128 may recognize the emotion of the user 10 based on the voice recognition result of the user 10 voice.
  • the emotion recognition unit 128 may recognize the emotion of the user 10 by combining these. For example, emotions such as joy, anger, and sadness may be recognized by the emotion recognition unit 128.
  • the emotion recognition unit 128 may output result data indicating that fact.
  • the emotion recognition unit 128 may transmit the emotion recognition result to the conversation algorithm selection unit 118.
  • the conversation algorithm selection unit 118 selects the emotion conversation algorithm 36 stored in the conversation algorithm DB 30 in association with the emotion recognized by the emotion recognition unit 128. For example, when the emotion recognition unit 128 recognizes pleasure, the conversation algorithm selection unit 118 selects the emotion conversation algorithm 36 stored in association with pleasure.
  • the conversation algorithm selection unit 118 may transmit the selected emotion conversation algorithm 36 to the conversation execution unit 120.
  • the conversation execution unit 120 may execute a conversation with the user 10 according to the emotion conversation algorithm 36 selected by the conversation algorithm selection unit 118. For example, when the conversation algorithm selection unit 118 selects the emotion conversation algorithm 36 corresponding to pleasure, the conversation execution unit 120 executes a conversation with the user 10 according to the emotion conversation algorithm 36 corresponding to pleasure. Thereby, the conversation execution part 120 can perform the conversation suitable for the emotion of the user 10 recognized during the conversation with the user 10.
  • the emotion recognition unit 128 may further transmit the emotion recognition result to the priority changing unit 130.
  • the priority changing unit 130 may change the priority associated with the conditional conversation algorithm 34 based on the emotion recognition result. For example, when the emotion recognition unit 128 recognizes a positive emotion while executing one conditional conversation algorithm 34, the priority changing unit 130 improves the priority of the one conditional conversation algorithm 34. Further, the priority changing unit 130 may lower the priority of the one conditional conversation algorithm 34 when the emotion recognition unit 128 recognizes a negative emotion while executing the one conditional conversation algorithm 34.
  • the priority changing unit 130 improves the priority of the “program notification app” when the emotion recognition unit 128 recognizes a positive emotion while executing the “program notification app”. For example, when the emotion recognition unit 128 recognizes a negative emotion while executing the “umbrella alert app”, the priority change unit 130 decreases the priority of the “umbrella alert app”. Thereby, according to a user's emotion, the priority of the conditional conversation algorithm 34 can be changed appropriately.
  • the priority changing unit 130 may change the priority 45 associated with the category 44 based on the emotion recognition result. For example, when the emotion recognition unit 128 recognizes a positive emotion in a state where one category 44 is selected, the priority changing unit 130 improves the priority of the one category 44. Moreover, the priority change part 130 reduces the priority of the said one category 44, when the emotion recognition part 128 recognizes a negative emotion in the state from which the one category 44 was selected. Thereby, the priority of the category 44 in which the user 10 has a positive emotion can be improved, and the priority of the category 44 in which the user 10 has a negative emotion can be reduced.
  • the utterance content registration unit 132 additionally registers the voice acquired by the voice acquisition unit 124 as the utterance content 46 corresponding to the one category 44 after the category selection unit 126 selects one category. For example, when “Entertainment: Musician: Silver Explosion” is selected and the user's 10 voice “It was formed in 2004” is acquired, the utterance content registration unit 132 displays “Entertainment: Musician: Silver”. As an utterance content 46 corresponding to “explosion”, “you were formed in 2004” is additionally registered. Thereby, conversation DB40 can be enriched.
  • the conversation data generation unit 134 generates conversation data indicating a conversation flow.
  • the conversation data indicating the conversation flow includes continuous utterance contents and response contents.
  • the conversation data may be a conversation flow such as “Member B seems to be handsome”, “Hey, is that so”, and “Muscle is also amazing”.
  • the conversation data generation unit 134 first performs voice output control of the first utterance content 46 (eg, “Member B seems to be handsome”) from the plurality of utterance contents 46 stored in association with the selected category 44.
  • the unit 122 outputs a sound.
  • the conversation execution unit 120 acquires the first response content (eg, “Hey, that's right”) by the user 10 with respect to the first utterance content 46.
  • the conversation data generation unit 134 generates conversation data including the first utterance content 46 and the first response content.
  • the conversation execution unit 120 After generating the conversation data, the conversation execution unit 120 outputs the first response content as a voice when the voice matching the first utterance content 46 is acquired from the user 10. As a result, a human-like natural response can be made. Furthermore, the conversation execution unit 120 acquires the second response content of the user 10 with respect to the first response content output by voice (for example, “muscles are also amazing”). Then, the conversation data generation unit 134 generates conversation data by ordering the first utterance content 46, the first response content, and the second response content. Thereby, the conversation data which can continue receiving and answering like a human being can be generated.
  • conversation data may be generated by the two information terminals 100.
  • the conversation data generation unit 134 generates conversation data including the first utterance content 46 and the first response content, and then transmits the conversation data to another information terminal 100. Then, in another information terminal 100, when the voice that matches the first utterance content 46 is acquired, the first response content is output as voice, the second response content is acquired, and the first utterance content 46, first The response contents and the second response contents are ordered to generate conversation data.
  • the algorithm sharing processing unit 136 copies the conditional conversation algorithm 34 stored in the conversation algorithm DB 30 to the conversation algorithm DB 30 included in another information terminal 100. For example, the algorithm sharing processing unit 136 adds the conditional conversation algorithm 34 to the conversation algorithm DB 30 of the information terminal 100 possessed by another user who has a profile whose similarity with the profile of the user 10 exceeds a predetermined standard. make a copy. For example, when the address, age, and gender match, the algorithm sharing processing unit 136 may determine that the profile similarity exceeds a standard.
  • the algorithm sharing processing unit 136 may copy, for example, a trial version of the conditional conversation algorithm 34 to the conversation algorithm DB 30 of the other information terminal 100. Thereby, the user 10 possessing the other information terminal 100 can try the conditional conversation algorithm 34 and can be used as a reference for purchasing the conditional conversation algorithm 34.
  • the server 200 may have the same functional configuration as that shown in FIG.
  • the audio output control unit 122 may control the information terminal 100 so that the information terminal 100 outputs the audio.
  • the voice acquisition unit 124 may receive the voice of the user 10 from the information terminal 100.
  • the server 200 when the server 200 includes the conversation data generation unit 134, the server 200 first selects one information terminal 100 at random from the plurality of information terminals 100, and outputs the first utterance content 46 by voice. Next, the information terminal 100 is made to acquire the first response content, and conversation data including the first utterance content 46 and the first response content is generated. Next, the conversation of a plurality of information terminals 100 is monitored, and the information terminal 100 that has acquired the same voice as the first utterance content 46 is specified. Then, the identified information terminal 100 is caused to output the first response content by voice and acquire the second response content. As a result, the server 200 can generate conversation data that can continue to be accepted and answered like humans.
  • the server 200 when the server 200 includes the algorithm sharing processing unit 136, the server 200 first arbitrarily selects one information terminal 100. Then, the server 200 identifies the information terminal 100 possessed by another user having a profile whose similarity with the profile of the user 10 of the selected information terminal 100 exceeds a predetermined standard, and performs copy processing Execute.
  • FIG. 5 schematically shows an example of an operation flow by the information terminal 100.
  • the operation flow illustrated in FIG. 5 may be started when an execution instruction for conversation processing according to the present embodiment is received.
  • step 502 step may be abbreviated as S
  • the execution condition determination unit 112 determines whether any of a plurality of execution conditions registered in the execution condition table 32 is satisfied. to decide. If it is determined in S502 that the condition is satisfied, the process proceeds to S504.
  • the conversation algorithm selection unit 118 selects the conditional conversation algorithm 34 stored in the conversation algorithm DB 30 in association with the execution condition determined to be satisfied by the execution condition determination unit 112.
  • the conversation execution unit 120 acquires the utterance content.
  • the conversation execution unit 120 may acquire the utterance content registered in association with the selected conditional conversation algorithm 34 from the conversation algorithm DB 30.
  • the conversation algorithm DB 30 may hold an utterance content that is first voice-output when the conditional conversation algorithm 34 is selected.
  • the conversation algorithm DB 30 may hold utterance contents indicating a response to the voice of the user 10.
  • the response to the voice of the user 10 is registered in advance by, for example, the administrator of the conversation algorithm DB 30.
  • the administrator of the conversation algorithm DB 30 may register the voice of the user 10 assumed in advance and the utterance content indicating the response in the conversation algorithm DB 30.
  • the conversation execution unit 120 causes the voice output control unit 122 to output the utterance content acquired in S506.
  • the conversation execution unit 120 determines whether or not a conversation end instruction from the user 10 has been received. If an end instruction has not been received, the process proceeds to S512.
  • the voice acquisition unit 124 acquires the voice uttered by the user 10 with respect to the voice output in S508.
  • the emotion recognition unit 128 recognizes the emotion of the user 10 based on the voice of the user 10 acquired by the voice acquisition unit 124.
  • conversation execution unit 120 determines whether or not the emotion recognized in S514 matches a predetermined emotion.
  • the predetermined emotion is, for example, an emotion associated with the emotion conversation algorithm 36.
  • the predetermined emotion may be an emotion designated in advance among a plurality of emotions.
  • the process proceeds to S518. If it is determined that the emotion does not match, the process returns to S508. In S518, the priority changing unit 130 changes the priority of the conditional conversation algorithm 34 selected in S504 based on the emotion recognized in 514.
  • the conversation execution unit 120 interrupts the conversation according to the conditional conversation algorithm 34, and the conversation with the user 10 is performed in accordance with the emotion conversation algorithm 36 stored in the conversation algorithm DB 30 in association with the emotion recognized in S514. Execute.
  • the process returns to S508. If it is not determined in S522 to continue the original conversation, the process ends.
  • the conversation execution unit 120 determines whether the utterance content can be acquired.
  • the conversation execution unit 120 may determine whether or not the utterance content corresponding to the voice acquired in S512 can be acquired from the conversation algorithm DB 30.
  • the conversation algorithm selected in S504 is “blood group fortune-telling app”, and in S508, “What is your blood type?” Is output as a voice, whereas a pre-registered user voice, for example “ When the voice “A type” is acquired in S512, the conversation execution unit 120 determines that the utterance content has been acquired.
  • the user voice that is not registered in advance for example, the voice of “What is your blood type?” Is acquired in S512, the conversation execution unit 120 cannot acquire the utterance content.
  • FIG. 6 schematically shows an example of an operation flow by the information terminal 100.
  • the operation flow illustrated in FIG. 6 may be started when it is determined in S524 of FIG. 5 that the utterance content has not been acquired. Further, the operation flow illustrated in FIG. 6 may be started at an arbitrary timing at which the voice from the user 10 is acquired, for example.
  • the category selection unit 126 selects one category 44 among the plurality of categories 44 based on the voice of the user 10.
  • the conversation execution unit 120 acquires one utterance content 46 from the plurality of utterance contents 46 associated with the selected category 44 stored in the conversation DB 40.
  • the conversation execution unit 120 causes the audio output control unit 122 to output the utterance content 46 acquired in S604.
  • step S ⁇ b> 608 the conversation execution unit 120 determines whether a conversation end instruction from the user 10 has been received. If an end instruction has not been received, the process proceeds to S610.
  • the voice acquisition unit 124 acquires the user's voice with respect to the voice output in S606.
  • the emotion recognition unit 128 recognizes the emotion of the user 10 based on the voice acquired in S610.
  • conversation execution unit 120 determines whether or not the emotion recognized in S612 matches a predetermined emotion.
  • the predetermined emotion is, for example, an emotion associated with the emotion conversation algorithm 36.
  • the predetermined emotion may be an emotion designated in advance among a plurality of emotions.
  • the process proceeds to S616. If it is determined that the emotion does not match, the process proceeds to S622. In S616, the priority changing unit 130 changes the priority of the category 44 selected in S602 based on the emotion recognized in S612.
  • the conversation execution unit 120 interrupts the conversation based on the plurality of utterance contents 46 associated with the selected category 44, and the emotion conversation stored in the conversation algorithm DB 30 in association with the emotion recognized in S612.
  • a conversation with the user 10 is executed according to the algorithm 36.
  • the process proceeds to S622. If it is not determined in S620 to continue the original conversation, the process ends.
  • the conversation execution unit 120 determines whether the utterance content 46 can be acquired.
  • the conversation execution unit 120 may determine whether or not the utterance content 46 can be acquired from the conversation DB 40 based on the selected category 44. For example, the conversation execution unit 120 determines that the utterance content 46 has been acquired when one utterance content 46 to be output by voice is obtained from the plurality of utterance contents 46 associated with the category 44 in the selected state. If it cannot be acquired, it is determined that the utterance content 46 cannot be acquired. For example, in one conversation process, the conversation execution unit 120 outputs all of the plurality of utterance contents 46 associated with the selected category 44, and when there is no unoutput utterance contents 46, the utterance contents 46 are displayed. It is determined that it could not be acquired.
  • the process returns to S606, and the conversation execution unit 120 causes the voice output control unit 122 to output the utterance content 46 acquired in S622. If it is determined in S622 that the utterance content 46 has not been acquired, the process proceeds to S624.
  • the execution condition determination unit 112 determines whether any of a plurality of execution conditions registered in the execution condition table 32 is satisfied. If it is determined that any one of the execution conditions is satisfied, the process proceeds to S504 in FIG.
  • the information terminal 100 may advance the conversation with the user 10 by appropriately switching between the conversation process using the conversation algorithm DB 30 and the conversation process using the conversation DB 40 and the Q & ADB 42.
  • the process proceeds to S626.
  • the conversation execution unit 120 executes error processing. For example, the conversation execution unit 120 notifies the user 10 that the conversation process is to be terminated, and then terminates the conversation process. Note that the conversation execution unit 120 may continue the conversation process by notifying the user 10 to urge other utterances and acquiring a new voice from the user 10. In this case, for example, after notifying the user 10 of another utterance in S626, the process may return to S602 or S610.
  • FIG. 7 schematically shows an example of the emotion conversation algorithm 36 corresponding to pleasure.
  • the conversation execution unit 120 causes the audio output control unit 122 to output thanks for appreciation that the user 10 feels joy and to inquire whether or not to continue the conversation.
  • the audio output control unit 122 outputs “Thank you. Do you want to continue?”
  • the voice acquisition unit 124 acquires the voice uttered by the user 10 with respect to the voice output in S702. Then, the conversation execution unit 120 determines whether or not the user 10 wishes to continue the conversation based on the voice of the user 10 acquired by the voice acquisition unit 124. When the conversation execution unit 120 determines that it is desired, the process proceeds to S706, and when it is determined that it is not desired, the process proceeds to S710.
  • the conversation execution unit 120 causes the voice output control unit 122 to output a voice expressing joy that the user 10 has selected to continue the conversation. For example, the audio output control unit 122 outputs “I am happy”.
  • the conversation execution unit 120 determines to continue the original conversation. Then return. That is, if it is the operation
  • the conversation execution unit 120 causes the audio output control unit 122 to output the content that the user 10 does not wish to continue the conversation and the content that indicates that the conversation is to be ended.
  • the voice output control unit 122 outputs a voice saying “Yes, today.
  • the conversation in which the user 10 feels pleasure can be continued. Further, when the user 10 is satisfied, the conversation can be appropriately terminated without continuing the conversation persistently.
  • FIG. 8 schematically shows an example of a conversation algorithm corresponding to anger.
  • the conversation execution unit 120 causes the audio output control unit 122 to output a sound corresponding to the user 10 feeling angry.
  • the audio output control unit 122 outputs an audio message “Acha, have you done it again?”.
  • the voice acquisition unit 124 acquires the voice uttered by the user 10 with respect to the voice output in S802. Then, the conversation execution unit 120 determines whether or not the user 10 is angry based on the voice of the user 10 acquired by the voice acquisition unit 124. When the conversation execution unit 120 determines that it feels anger, the process proceeds to S806, and when it determines that it does not feel anger, the process proceeds to S810.
  • the conversation execution unit 120 causes the voice output control unit 122 to output a voice indicating that voice output is not performed for a while. For example, the voice output control unit 122 outputs a voice saying “Please keep quiet for a while”. In S808, the conversation execution unit 120 shifts to a standby state for waiting for conversation.
  • the conversation execution unit 120 causes the voice output control unit 122 to output a voice in response to an erroneous emotion recognition result.
  • the voice output control unit 122 outputs a voice “That?”.
  • FIG. 9 schematically shows an example of a conversation algorithm corresponding to sadness.
  • the conversation execution unit 120 causes the audio output control unit 122 to output a sound for confirming whether the user 10 feels sadness.
  • the voice output control unit 122 outputs a voice saying “That's bad glue”.
  • the voice acquisition unit 124 acquires the voice uttered by the user 10 with respect to the voice output in S902. Then, the conversation execution unit 120 determines whether the user 10 feels sadness based on the voice of the user 10 acquired by the voice acquisition unit 124. When the conversation execution unit 120 determines that it is felt, the process proceeds to S906, and when it is determined that it is not felt, the process proceeds to S910.
  • the conversation execution unit 120 causes the audio output control unit 122 to output a voice that notifies the user 10 that the topic is to be switched.
  • the audio output control unit 122 outputs, for example, “let's talk about another topic”.
  • the conversation execution unit 120 switches conversations. For example, when one category 44 is in a selected state, the conversation execution unit 120 puts another category 44 into a selected state. When a plurality of categories 44 have a hierarchical structure, the conversation execution unit 120 may select a category 44 adjacent to one category 44 that has been selected.
  • the conversation execution unit 120 may select a category 44 having a higher priority than the one category 44 that has been selected. Further, the conversation execution unit 120 may select a category 44 selected at random from the plurality of categories 44.
  • the conversation execution unit 120 determines to continue the original conversation. Then return. That is, if it is the operation
  • the emotion conversation algorithm 36 corresponding to sadness the conversation can be switched in order to make the user 10 have a more enjoyable emotion.
  • FIG. 10 schematically shows an example of the hierarchical structure of the category 44.
  • a priority 54 is assigned to each of the plurality of category names 52.
  • the category selection unit 126 may select the adjacent “entertainment: musician: group B”.
  • the conversation execution unit 120 may cause the audio output control unit 122 to output a sound including the upper category name 52.
  • the voice output control unit 122 outputs a voice saying “Speaking of musicians, it is group B”. In this way, by switching to the adjacent category 44, it is possible to execute a conversation related to a category that is highly related to the currently selected category.
  • the category selection unit 126 may select a category 44 other than the currently selected “entertainment: musician: group A” based on the priority. For example, the category selection unit 126 may select “TV: mail order program” having a higher priority than the currently selected category 44. Thereby, it is possible to switch to a conversation with higher priority for the user 10.
  • the category selection unit 126 may select a category 44 other than the currently selected “entertainment: musician: group A” based on the profile of the user 10. For example, when “marathon” is registered as a hobby in the profile of the user 10, the category selection unit 126 may select “sports general: athletics: marathon”. Thereby, it can switch to the conversation suitable for the user's 10 profile.
  • each unit of the information terminal 100 may be realized by hardware or may be realized by software. Further, it may be realized by a combination of hardware and software.
  • a computer may function as a part of the information terminal 100 by executing a program on the information terminal 100.
  • the program may be stored in a computer-readable medium, or may be stored in a storage device connected to a network.
  • an information processing apparatus having a general configuration including a data processing device having a CPU, ROM, RAM, communication interface, etc., an input device, an output device, and a storage device, the operation of each part of the information terminal 100 is defined.
  • the information terminal 100 may be realized by starting software or a program.
  • a program that is installed in a computer and causes the computer to function as a part of the information terminal 100 according to the present embodiment includes a module that defines the operation of each unit of the information terminal 100. These programs or modules work on the CPU or the like to cause the computer to function as each unit of the information terminal 100. Information processing described in these programs functions as a specific means in which software and the various hardware resources described above cooperate with each other by being read by a computer. A specific measurement device according to the purpose of use can be constructed by realizing calculation or processing of information according to the purpose of use of the computer in the present embodiment by these specific means.
  • the server 200 includes a data unit having a CPU, a ROM, a RAM, a communication interface, an input unit such as a keyboard, a touch panel, and a microphone, an output unit such as a display and a speaker, and a storage unit such as a memory and an HDD.
  • the information processing apparatus having a general configuration provided may be realized by activating software or a program that defines the operation of each unit of the server 200.
  • the server 200 may be a virtual server or a cloud system.

Abstract

A conversation system that can perform flexible conversations in accordance with a user state is desired, and to this end, provided is a conversation processing system that is equipped with: a voice acquisition unit that acquires the voice of a user; an emotion recognition unit that recognizes the emotions of the user on the basis of the voice acquired by the voice acquisition unit; a first conversation algorithm memory unit that stores first conversation algorithms correlating with each type of a plurality of emotions; a conversation algorithm selection unit that makes a selection, in correlation with the emotions recognized by the emotion recognition unit, from among the first conversation algorithms being stored in the first conversation algorithm memory unit; and a conversation performance unit that performs conversations with the user, in accordance with the first conversation algorithms selected by the conversation algorithm selection unit.

Description

会話処理システム及びプログラムConversation processing system and program
 本発明は、会話処理システム及びプログラムに関する。 The present invention relates to a conversation processing system and a program.
 従来、ユーザと音声会話をする会話システムが知られていた。(例えば、特許文献1参照)。
 [先行技術文献]
 [特許文献]
 [特許文献1]特開2011-253389号公報
Conventionally, a conversation system for voice conversation with a user has been known. (For example, refer to Patent Document 1).
[Prior art documents]
[Patent Literature]
[Patent Document 1] Japanese Patent Application Laid-Open No. 2011-253389
 ユーザの状態に応じて柔軟な会話を実行できる会話システムが望まれている。 A conversation system that can execute a flexible conversation according to the user's condition is desired.
 本発明の第1の態様においては、ユーザの音声を取得する音声取得部と、音声取得部が取得した音声に基づいてユーザの感情を認識する感情認識部と、複数の感情の種類のそれぞれに対応付けて第1会話アルゴリズム(例えば、感情会話アルゴリズム)を記憶する第1会話アルゴリズム記憶部と、感情認識部が認識した感情に対応付けて第1会話アルゴリズム記憶部に記憶されている第1会話アルゴリズムを選択する会話アルゴリズム選択部と、会話アルゴリズム選択部が選択した第1会話アルゴリズムに従って、ユーザとの会話を実行する会話実行部とを備える会話処理システムが提供される。 In the first aspect of the present invention, a voice acquisition unit that acquires a user's voice, an emotion recognition unit that recognizes a user's emotion based on the voice acquired by the voice acquisition unit, and each of a plurality of types of emotions A first conversation algorithm storage unit that stores a first conversation algorithm (for example, emotion conversation algorithm) in association with the first conversation stored in the first conversation algorithm storage unit in association with the emotion recognized by the emotion recognition unit. A conversation processing system is provided that includes a conversation algorithm selection unit that selects an algorithm and a conversation execution unit that executes a conversation with a user according to a first conversation algorithm selected by the conversation algorithm selection unit.
 上記会話処理システムは、複数の第2会話アルゴリズム(例えば、条件会話アルゴリズム)のそれぞれに対応付けて実行条件を記憶する第2会話アルゴリズム記憶部と、第2会話アルゴリズム記憶部が記憶している複数の実行条件のいずれかが満たされたことを判断する実行条件判断部とをさらに備えてよく、会話アルゴリズム選択部は、実行条件判断部によって満たされたことが判断された実行条件に対応付けて第2会話アルゴリズム記憶部に記憶されている第2会話アルゴリズムを選択してよく、会話実行部は、会話アルゴリズム選択部が選択した第2会話アルゴリズムに従って、ユーザとの会話を実行してよい。 The conversation processing system includes a second conversation algorithm storage unit that stores an execution condition in association with each of a plurality of second conversation algorithms (for example, a conditional conversation algorithm), and a plurality of second conversation algorithm storage units that are stored. An execution condition determining unit that determines that any one of the execution conditions is satisfied, and the conversation algorithm selecting unit is associated with the execution condition determined to be satisfied by the execution condition determining unit. The second conversation algorithm stored in the second conversation algorithm storage unit may be selected, and the conversation execution unit may execute the conversation with the user according to the second conversation algorithm selected by the conversation algorithm selection unit.
 上記会話処理システムにおいて、音声取得部は、会話実行部が第2会話アルゴリズムに従ってユーザとの会話を実行している間に、ユーザの音声を取得してよく、会話実行部は、会話アルゴリズム選択部が選択した第2会話アルゴリズムに従った会話を中断し、会話アルゴリズム選択部が選択した第1会話アルゴリズムに従った会話を開始してよい。上記会話処理システムにおいて、第2会話アルゴリズム記憶部は、複数の第2会話アルゴリズムのそれぞれに対応付けて優先度をさらに記憶してよく、感情認識部は、会話アルゴリズム選択部が選択した第2会話アルゴリズムに従って会話実行部がユーザとの会話を実行している間に音声取得部が取得した音声に基づいてユーザの感情を認識してよく、上記会話処理システムは、感情認識部が認識したユーザの感情に基づいて、会話アルゴリズム選択部が選択した第2会話アルゴリズムに対応付けて記憶されている優先度を変更する第1優先度変更部をさらに備えてよい。 In the conversation processing system, the voice acquisition unit may acquire the user's voice while the conversation execution unit is executing a conversation with the user according to the second conversation algorithm. The conversation according to the second conversation algorithm selected by may be interrupted, and the conversation according to the first conversation algorithm selected by the conversation algorithm selection unit may be started. In the conversation processing system, the second conversation algorithm storage unit may further store a priority in association with each of the plurality of second conversation algorithms, and the emotion recognition unit may select the second conversation selected by the conversation algorithm selection unit. While the conversation execution unit is executing a conversation with the user according to the algorithm, the user's emotion may be recognized based on the voice acquired by the voice acquisition unit, and the conversation processing system may recognize the user's emotion recognized by the emotion recognition unit. You may further provide the 1st priority change part which changes the priority memorize | stored in association with the 2nd conversation algorithm which the conversation algorithm selection part selected based on emotion.
 上記会話処理システムは、複数のカテゴリのそれぞれに対応付けて複数の発話内容を記憶する発話内容記憶部と、音声取得部が取得した音声に基づいて、複数のカテゴリのうちの一のカテゴリを選択状態にするカテゴリ選択部とをさらに備えてよく、会話実行部は、カテゴリ選択部が選択状態にした一のカテゴリに対応付けて記憶された複数の発話内容によって、ユーザとの会話を実行してよく、音声取得部は、会話実行部が一のカテゴリに対応付けて記憶された複数の発話内容によりユーザとの会話を実行している間に、ユーザの音声を取得してよく、会話実行部は、一のカテゴリに対応付けて記憶された複数の発話内容によるユーザとの会話を中断し、会話アルゴリズム選択部が選択した第1会話アルゴリズムに従った会話を開始してよい。上記会話処理システムにおいて、発話内容記憶部は、複数のカテゴリのそれぞれに対応付けて優先度をさらに記憶してよく、感情認識部は、一のカテゴリに対応付けて記憶された複数の発話内容によって会話実行部がユーザとの会話を実行している間に音声取得部が取得した音声に基づいてユーザの感情を認識してよく、上記会話処理システムは、感情認識部が認識したユーザの感情に基づいて、一のカテゴリに対応付けて記憶されている優先度を変更する第2優先度変更部を更に備えてよい。 The conversation processing system selects one category from a plurality of categories based on an utterance content storage unit that stores a plurality of utterance contents in association with each of a plurality of categories and a voice acquired by the voice acquisition unit And a conversation execution unit that executes a conversation with the user based on a plurality of utterance contents stored in association with one category selected by the category selection unit. The voice acquisition unit may acquire the voice of the user while the conversation execution unit executes a conversation with the user based on a plurality of utterance contents stored in association with one category. Interrupts the conversation with the user based on a plurality of utterance contents stored in association with one category, and starts a conversation according to the first conversation algorithm selected by the conversation algorithm selection unit. There. In the above conversation processing system, the utterance content storage unit may further store a priority in association with each of the plurality of categories, and the emotion recognition unit may use the plurality of utterance contents stored in association with one category. While the conversation execution unit is executing a conversation with the user, the user's emotion may be recognized based on the voice acquired by the voice acquisition unit, and the conversation processing system recognizes the user's emotion recognized by the emotion recognition unit. A second priority changing unit that changes the priority stored in association with one category may be further provided.
 上記会話処理システムにおいて、カテゴリ選択部は、選択状態にした一のカテゴリに対応付けて記憶された複数の発話内容によって会話実行部が実行しているユーザとの会話を切り替える場合に、優先度に基づいて、一のカテゴリ以外のカテゴリを選択状態にしてよく、会話実行部は、カテゴリ選択部により選択状態にされた一のカテゴリ以外のカテゴリに対応付けて発話内容記憶部に記憶された複数の発話内容によって、ユーザとの会話を開始してよい。上記会話処理システムにおいて、複数のカテゴリは階層構造を有してよく、カテゴリ選択部は、選択状態にした一のカテゴリに対応付けて記憶された複数の発話内容によって会話実行部が実行しているユーザとの会話を切り替える場合に、一のカテゴリに隣接するカテゴリを選択状態にしてよく、会話実行部は、カテゴリ選択部により選択状態にされた一のカテゴリに隣接するカテゴリに対応付けて発話内容記憶部に記憶された複数の発話内容によって、ユーザとの会話を開始してよい。 In the above conversation processing system, the category selection unit determines the priority when switching the conversation with the user executed by the conversation execution unit according to a plurality of utterance contents stored in association with the selected category. On the basis of this, a category other than one category may be selected, and the conversation execution unit may associate a plurality of categories stored in the utterance content storage unit in association with a category other than the one category selected by the category selection unit. Depending on the utterance content, a conversation with the user may be started. In the conversation processing system, the plurality of categories may have a hierarchical structure, and the category selection unit is executed by the conversation execution unit according to the plurality of utterance contents stored in association with the selected one category. When switching a conversation with a user, a category adjacent to one category may be selected, and the conversation execution unit associates the category with the category adjacent to the one category selected by the category selection unit. A conversation with the user may be started based on a plurality of utterance contents stored in the storage unit.
 上記会話処理システムは、カテゴリ選択部が一のカテゴリを選択状態にした後に、音声取得部が取得した音声を、選択状態の一のカテゴリに発話内容として対応づけて発話内容記憶部に登録する発話内容登録部をさらに備えてよい。上記会話処理システムは、カテゴリ選択部によって選択状態にされた一のカテゴリに対応付けて記憶された複数の発話内容に含まれる第1発話内容を音声出力させる音声出力制御部と、第1発話内容と、第1発話内容に対するユーザ10の第1応答内容とを含む会話データを生成する会話データ生成部とをさらに備えてよく、会話実行部は、会話データ生成部が第1発話内容と第1応答内容とを含む会話データを生成した後に、音声取得部が第1発話内容に一致する音声を取得した場合に、第1応答内容を前記音声出力してよく、音声取得部は、会話実行部が音声出力した第1応答内容に対する第2応答内容を取得してよく、会話データ生成部は、第1発話内容、第1応答内容、第2応答内容の順で対応付けた会話データを生成してよい。 In the conversation processing system, after the category selection unit selects one category, the speech acquired by the voice acquisition unit is associated with the one category in the selected state as utterance content and registered in the utterance content storage unit. A content registration unit may be further provided. The conversation processing system includes: a voice output control unit that outputs a first utterance content included in a plurality of utterance contents stored in association with one category selected by the category selection unit; and a first utterance content And a conversation data generation unit that generates conversation data including the first response content of the user 10 with respect to the first utterance content. The conversation execution unit may include the first utterance content and the first utterance content. After the conversation data including the response content is generated, when the voice acquisition unit acquires the voice that matches the first utterance content, the first response content may be output by voice, and the voice acquisition unit The second response content corresponding to the first response content output by the voice may be acquired, and the conversation data generation unit generates the conversation data associated in the order of the first utterance content, the first response content, and the second response content. It's okay.
 上記会話処理システムは、第1のユーザが所持する第1情報端末の第2会話アルゴリズム記憶部が記憶する第2会話アルゴリズムを、第1のユーザが有するプロファイルとの類似度が予め定められた基準を超えているプロファイルを有する第2のユーザが所持する第2情報端末の第2会話アルゴリズム記憶部にコピーするアルゴリズム共有処理部をさらに備えてよい。 In the above conversation processing system, the second conversation algorithm stored in the second conversation algorithm storage unit of the first information terminal possessed by the first user is a standard in which the similarity with the profile of the first user is determined in advance. An algorithm sharing processing unit for copying to a second conversation algorithm storage unit of a second information terminal possessed by a second user having a profile exceeding the threshold may be further provided.
 本発明の第2の態様においては、コンピュータを、上記会話処理システムとして機能させるためのプログラムが提供される。 In the second aspect of the present invention, there is provided a program for causing a computer to function as the conversation processing system.
 なお、上記の発明の概要は、本発明の必要な特徴の全てを列挙したものではない。また、これらの特徴群のサブコンビネーションもまた、発明となりうる。 Note that the above summary of the invention does not enumerate all the necessary features of the present invention. In addition, a sub-combination of these feature groups can also be an invention.
情報端末100の通信環境の一例を概略的に示す。An example of the communication environment of the information terminal 100 is shown schematically. 情報端末100による会話処理の一例を概略的に示す。An example of the conversation process by the information terminal 100 is shown schematically. 情報端末100による会話処理の他の例を概略的に示す。The other example of the conversation process by the information terminal 100 is shown schematically. 情報端末100の機能構成を概略的に示す。The function structure of the information terminal 100 is shown schematically. 情報端末100による動作フローの一例を概略的に示す。An example of the operation | movement flow by the information terminal 100 is shown roughly. 情報端末100による動作フローの一例を概略的に示す。An example of the operation | movement flow by the information terminal 100 is shown roughly. 喜びに対応する会話アルゴリズムの一例を概略的に示す。An example of the conversation algorithm corresponding to pleasure is shown roughly. 怒りに対応する会話アルゴリズムの一例を概略的に示す。An example of the conversation algorithm corresponding to anger is shown roughly. 悲しみに対応する会話アルゴリズムの一例を概略的に示す。An example of the conversation algorithm corresponding to sadness is shown roughly. カテゴリ44の階層構造の一例を概略的に示す。An example of the hierarchical structure of the category 44 is shown schematically.
 以下、発明の実施の形態を通じて本発明を説明するが、以下の実施形態は請求の範囲にかかる発明を限定するものではない。また、実施形態の中で説明されている特徴の組み合わせの全てが発明の解決手段に必須であるとは限らない。 Hereinafter, the present invention will be described through embodiments of the invention. However, the following embodiments do not limit the invention according to the claims. In addition, not all the combinations of features described in the embodiments are essential for the solving means of the invention.
 図1は、情報端末100の通信環境の一例を概略的に示す。情報端末100は、音声入力機能及び音声出力機能を備え、ユーザ10との会話を実行する。情報端末100は、例えば、スマートフォンなどの携帯電話である。情報端末100は、単体でユーザ10との会話を実行してよい。また、情報端末100は、通信網20を介して通信可能なサーバ200などと連携して、ユーザ10との会話を実行してもよい。 FIG. 1 schematically shows an example of a communication environment of the information terminal 100. The information terminal 100 has a voice input function and a voice output function, and executes a conversation with the user 10. The information terminal 100 is a mobile phone such as a smartphone, for example. The information terminal 100 may execute a conversation with the user 10 alone. Further, the information terminal 100 may execute a conversation with the user 10 in cooperation with the server 200 that can communicate via the communication network 20.
 情報端末100は、会話処理システムの一例であってよい。また、情報端末100及びサーバ200によって会話処理システムが構成されてもよい。なお、情報端末100は、音声入力機能及び音声出力機能を有する装置であればどのような装置であってもよく、例えば、タブレット端末、PC、家電、自動車、カーナビゲーション、ロボット、及びぬいぐるみなどであってよい。 The information terminal 100 may be an example of a conversation processing system. Further, the conversation processing system may be configured by the information terminal 100 and the server 200. The information terminal 100 may be any device that has a voice input function and a voice output function. For example, the information terminal 100 may be a tablet terminal, a PC, a home appliance, a car, a car navigation system, a robot, a stuffed toy, or the like. It may be.
 図2は、情報端末100による会話処理の一例を概略的に示す。情報端末100は、会話アルゴリズムDB30に記憶された会話アルゴリズムに従って、ユーザ10との会話を実行してよい。会話アルゴリズムDB30は、情報端末100が保持してよい。また、会話アルゴリズムDB30は、サーバ200が保持してもよい。また、会話アルゴリズムDB30は、情報端末100及びサーバ200以外の装置が保持してもよい。 FIG. 2 schematically shows an example of conversation processing by the information terminal 100. The information terminal 100 may execute a conversation with the user 10 in accordance with the conversation algorithm stored in the conversation algorithm DB 30. The information terminal 100 may hold the conversation algorithm DB 30. Moreover, the server 200 may hold | maintain conversation algorithm DB30. The conversation algorithm DB 30 may be held by devices other than the information terminal 100 and the server 200.
 情報端末100は、実行条件テーブル32に登録された実行条件が満たされた場合に、当該実行条件に対応する会話アルゴリズムを実行してよい。例えば、情報端末100は、ユーザ10の好きなジャンルのテレビ番組が近いうちにある場合に「番組お知らせアプリ」を実行する。情報端末100は、例えば、テレビ番組データの提供サービスを受け、ユーザ10のプロファイルデータを取得することにより、ユーザ10が好きなジャンルのテレビ番組が近いうちにあることを判断する。 The information terminal 100 may execute the conversation algorithm corresponding to the execution condition when the execution condition registered in the execution condition table 32 is satisfied. For example, the information terminal 100 executes the “program notification application” when there is a television program of a genre that the user 10 likes in the near future. For example, the information terminal 100 receives a TV program data providing service and acquires the profile data of the user 10 to determine that there is a TV program of a genre that the user 10 likes soon.
 「番組お知らせアプリ」は、テレビ番組データに基づいて、ユーザ10が好きなジャンルのテレビ番組が放送されることをユーザ10に音声通知したり、番組内容に対するユーザ10の問い合わせに回答したりする会話アルゴリズムであってよい。例えば、「番組お知らせアプリ」は、今日20時から音楽番組があることを音声通知してよく、ユーザ10による誰が出るのかの問い合わせに対して「銀爆が出るよ」と応答してよい。 The “program notification application” is a conversation for notifying the user 10 that a TV program of a genre that the user 10 likes is broadcast based on the TV program data, or answering an inquiry from the user 10 regarding the program content. It may be an algorithm. For example, the “program notification application” may give a voice notification that there is a music program from 20:00 today, and may respond with “silver bomb will come out” in response to an inquiry from the user 10 about who will appear.
 「通勤補助アプリ」は、通勤関連のイベントをユーザ10に音声通知したり、通勤に関するユーザ10からの問い合わせに回答したりする会話アルゴリズムであってよい。情報端末100は、例えば、電車の運行情報を提供するサーバから受信した運行情報データが、通勤電車の遅延を示す場合に「通勤補助アプリ」を実行する。「イベント通知アプリ」は、カレンダーに登録されたイベントの開催が近づいていることをユーザ10に音声通知したり、イベントの内容に関するユーザ10からの問い合わせに回答したりする会話アルゴリズムであってよい。情報端末100は、例えば、カレンダーに登録されたイベントの開催前日に「イベント通知アプリ」を実行する。 The “commuting assistance application” may be a conversation algorithm for notifying the user 10 of events related to commuting and answering inquiries from the user 10 regarding commuting. For example, the information terminal 100 executes the “commuting assistance application” when the operation information data received from the server that provides the train operation information indicates the delay of the commuter train. The “event notification application” may be a conversation algorithm that notifies the user 10 that the event registered in the calendar is about to be held, or answers an inquiry from the user 10 regarding the contents of the event. The information terminal 100 executes the “event notification application” on the day before the event registered in the calendar, for example.
 「傘アラートアプリ」は、降水確率に応じて、ユーザ10に傘を持つことを促す音声通知をしたり、天気に関するユーザ10からの問い合わせに回答したりする会話アルゴリズムであってよい。情報端末100は、例えば、天気予報の情報を提供するサーバから受信した天気予報データが、降水確率30%以上を占めす場合に「傘アラートアプリ」を実行する。「血液型占いアプリ」は、ユーザ10に指定された血液型に対する占いの内容を音声通知する会話アルゴリズムであってよい。情報端末100は、例えば、「血液型占いアプリ」の開始指示を受け付けた場合に、「血液型占いアプリ」を実行する。なお、これらの会話アルゴリズムは例示であり、会話アルゴリズムDB30は他の会話アルゴリズムを記憶していてもよい。 The “umbrella alert app” may be a conversation algorithm that makes a voice notification that prompts the user 10 to hold an umbrella or answers an inquiry from the user 10 regarding the weather according to the probability of precipitation. For example, the information terminal 100 executes the “umbrella alert app” when the weather forecast data received from a server that provides weather forecast information occupies a precipitation probability of 30% or more. The “blood group fortune-telling application” may be a conversation algorithm for voice notification of the content of fortune-telling for the blood type designated by the user 10. For example, the information terminal 100 executes the “blood type fortune-telling app” when receiving an instruction to start the “blood type fortune-telling app”. Note that these conversation algorithms are examples, and the conversation algorithm DB 30 may store other conversation algorithms.
 実行条件テーブル32は、各会話アルゴリズムに対応付けて、優先度を記憶していてよい。情報端末100は、優先度に基づいて会話アルゴリズムを選択してよい。例えば、情報端末100は、実行条件を満たす会話アルゴリズムが複数ある場合に、優先度が高い会話アルゴリズムを実行したり、優先度の高い順に会話アルゴリズムを実行したりしてよい。また、情報端末100は、話題を途中で切り替える場合に、現在実行中の会話アルゴリズムよりも優先度が高い会話アルゴリズムを選択してよい。 The execution condition table 32 may store priorities in association with each conversation algorithm. The information terminal 100 may select a conversation algorithm based on the priority. For example, when there are a plurality of conversation algorithms that satisfy the execution condition, the information terminal 100 may execute a conversation algorithm with a high priority or may execute a conversation algorithm in descending order of priority. Further, the information terminal 100 may select a conversation algorithm having a higher priority than the conversation algorithm currently being executed when the topic is switched halfway.
 図3は、情報端末100による会話処理の他の例を概略的に示す。情報端末100は、会話DB40に記憶された複数の発話内容46により、ユーザ10との会話を実行してよい。また、情報端末100は、Q&ADB42に記憶された複数の発話内容48により、ユーザ10との会話を実行してよい。 FIG. 3 schematically shows another example of conversation processing by the information terminal 100. The information terminal 100 may execute a conversation with the user 10 based on the plurality of utterance contents 46 stored in the conversation DB 40. Further, the information terminal 100 may execute a conversation with the user 10 based on the plurality of utterance contents 48 stored in the Q & ADB 42.
 会話DB40及びQ&ADB42は、情報端末100が保持してよい。また、会話DB40及びQ&ADB42は、サーバ200が保持してもよい。また、会話DB40及びQ&ADB42は、情報端末100及びサーバ200以外の装置が保持してもよい。 The conversation terminal 40 and the Q & ADB 42 may be held by the information terminal 100. The conversation DB 40 and the Q & ADB 42 may be held by the server 200. The conversation DB 40 and the Q & ADB 42 may be held by devices other than the information terminal 100 and the server 200.
 会話DB40は、複数のカテゴリ44のそれぞれに対応付けて複数の発話内容46を記憶する。情報端末100は、ユーザ10の音声に基づいて、複数のカテゴリから一のカテゴリを選択状態にしてよい。例えば、情報端末100は、ユーザ10の音声が複数の発話内容46のいずれかに一致すると判断した場合、当該発話内容46を含むカテゴリ44を選択状態にする。また、例えば、情報端末100は、ユーザ10の音声に含まれるキーワードと、カテゴリ44の名称とが一致すると判断した場合、当該カテゴリ44を選択状態にしてよい。 The conversation DB 40 stores a plurality of utterance contents 46 in association with each of the plurality of categories 44. The information terminal 100 may select one category from a plurality of categories based on the voice of the user 10. For example, when the information terminal 100 determines that the voice of the user 10 matches any of the plurality of utterance contents 46, the information terminal 100 puts the category 44 including the utterance contents 46 into a selected state. For example, if the information terminal 100 determines that the keyword included in the voice of the user 10 matches the name of the category 44, the information terminal 100 may select the category 44.
 図3に示す例では、ユーザ10の音声「銀爆おもしろいよね」が、「芸能:ミュージシャン:銀爆(シルバーボンバー)」に対応付けられた発話内容46に一致することから、情報端末100は、「芸能:ミュージシャン:銀爆(シルバーボンバー)」を選択状態にする。また、情報端末100は、ユーザ10の音声が「芸能:ミュージシャン:銀爆(シルバーボンバー)」に対応付けられた複数の発話内容46のいずれにも一致しない場合であっても、「銀爆」というキーワードを含んでいた場合には、「芸能:ミュージシャン:銀爆(シルバーボンバー)」を選択状態にしてよい。 In the example shown in FIG. 3, since the voice of the user 10 “Silver explosion is interesting” matches the utterance content 46 associated with “Entertainment: Musician: Silver bomb (silver bomber)”, the information terminal 100 Select “Entertainment: Musician: Silver Bomber”. Further, the information terminal 100 determines that the voice of the user 10 does not match any of the plurality of utterance contents 46 associated with “Entertainment: Musician: Silver Bomber”. If the keyword is included, “Entertainment: Musician: Silver Bomber” may be selected.
 情報端末100は、ユーザ10の問いかけに対する応答として、選択状態のカテゴリ44に対応付けられた複数の発話内容46のいずれかを音声出力してよい。例えば、情報端末100は「銀爆おもしろいよね」というユーザ10の問いかけに対して、「エアーバンドなんだよね」と応答してよい。これにより、ユーザ10の音声に対して関連性の高い応答をすることができる。 The information terminal 100 may output one of a plurality of utterance contents 46 associated with the selected category 44 as a response to the inquiry from the user 10. For example, the information terminal 100 may respond with “It is an air band” in response to the user 10's question “It ’s fun! Thereby, a highly relevant response can be made to the voice of the user 10.
 会話DB40は、複数のカテゴリ44のそれぞれに対応付けて優先度45を記憶してよい。情報端末100は、選択中のカテゴリ44を他のカテゴリ44に切り替える場合に、優先度45に基づいて他のカテゴリ44を選択してよい。また、例えば、情報端末100が主体的にユーザ10に話しかける場合など、ユーザ10の音声に基づかずにカテゴリ44を選択する場合に、情報端末100は、優先度45に基づいてカテゴリ44を選択してよい。 The conversation DB 40 may store the priority 45 in association with each of the plurality of categories 44. The information terminal 100 may select another category 44 based on the priority 45 when switching the selected category 44 to another category 44. In addition, for example, when the information terminal 100 selects the category 44 without being based on the voice of the user 10, such as when the information terminal 100 speaks to the user 10, the information terminal 100 selects the category 44 based on the priority 45. It's okay.
 Q&ADB42は、複数のカテゴリ44のそれぞれに対応付けて、複数の発話内容48を記憶する。複数の発話内容48のそれぞれは、質問に対応付けられていてよい。情報端末100は、取得したユーザ10の音声が、選択状態のカテゴリ44に対応付けられた複数の質問のいずれかに一致すると判断した場合、当該質問に対応する発話内容48を音声出力してよい。図3に示す例では、仮に「芸能:ミュージシャン:銀爆(シルバーボンバー)」が選択状態であり、ユーザ10の音声が「ボーカル誰だっけ?」であった場合、情報端末100は、「キリシューだよね」を音声出力する。 The Q & ADB 42 stores a plurality of utterance contents 48 in association with each of the plurality of categories 44. Each of the plurality of utterance contents 48 may be associated with a question. When the information terminal 100 determines that the acquired voice of the user 10 matches one of the plurality of questions associated with the selected category 44, the information terminal 100 may output the utterance content 48 corresponding to the question as a voice. . In the example shown in FIG. 3, if “Entertainment: Musician: Silver Bomber” is selected and the voice of the user 10 is “Who is the vocalist?”, The information terminal 100 displays “Kirishu”. "It's right".
 上述したように、情報端末100は、会話アルゴリズムDB30、会話DB40、及びQ&ADB42などを用いて、ユーザ10との会話を実行する。本実施形態に係る情報端末100は、さらに柔軟な会話を実現するべく、ユーザ10の音声に対する感情認識結果を用いる。 As described above, the information terminal 100 executes a conversation with the user 10 using the conversation algorithm DB 30, the conversation DB 40, the Q & ADB 42, and the like. The information terminal 100 according to the present embodiment uses an emotion recognition result for the voice of the user 10 in order to realize a more flexible conversation.
 図4は、情報端末100の機能構成を概略的に示す。ここでは、会話アルゴリズムDB30、会話DB40、Q&ADB42、実行条件判断部112、条件データ取得部116、会話アルゴリズム選択部118、会話実行部120、音声出力制御部122、音声取得部124、カテゴリ選択部126、感情認識部128、優先度変更部130、発話内容登録部132、会話データ生成部134、及びアルゴリズム共有処理部136を備える情報端末100について説明する。 FIG. 4 schematically shows a functional configuration of the information terminal 100. Here, conversation algorithm DB 30, conversation DB 40, Q & ADB 42, execution condition determination unit 112, condition data acquisition unit 116, conversation algorithm selection unit 118, conversation execution unit 120, voice output control unit 122, voice acquisition unit 124, category selection unit 126 The information terminal 100 including the emotion recognition unit 128, the priority change unit 130, the utterance content registration unit 132, the conversation data generation unit 134, and the algorithm sharing processing unit 136 will be described.
 会話アルゴリズムDB30は、複数の会話アルゴリズムを記憶する。会話アルゴリズムDB30は、複数の感情の種類のそれぞれに対応付けられた感情会話アルゴリズム36を記憶する。感情会話アルゴリズム36は、第1会話アルゴリズムの一例であってよい。会話アルゴリズムDB30は、第1会話アルゴリズム記憶部の一例であってよい。 The conversation algorithm DB 30 stores a plurality of conversation algorithms. The conversation algorithm DB 30 stores an emotion conversation algorithm 36 associated with each of a plurality of emotion types. The emotion conversation algorithm 36 may be an example of a first conversation algorithm. The conversation algorithm DB 30 may be an example of a first conversation algorithm storage unit.
 また、会話アルゴリズムDB30は、図2で説明したように、それぞれに実行条件が対応付けられた複数の条件会話アルゴリズム34と、実行条件テーブル32とを記憶する。条件会話アルゴリズム34は、第2会話アルゴリズムの一例であってよい。会話アルゴリズムDB30は第2会話アルゴリズム記憶部の一例であってよい。 Further, as described in FIG. 2, the conversation algorithm DB 30 stores a plurality of conditional conversation algorithms 34 each associated with an execution condition and an execution condition table 32. The conditional conversation algorithm 34 may be an example of a second conversation algorithm. The conversation algorithm DB 30 may be an example of a second conversation algorithm storage unit.
 会話DB40は、複数のカテゴリ44のそれぞれに対応付けて複数の発話内容46を記憶する。Q&ADB42は、複数のカテゴリ44のそれぞれに対応付けて、質問と発話内容48との組み合わせを複数記憶する。会話DB40及びQ&ADB42は、発話内容記憶部の一例であってよい。 The conversation DB 40 stores a plurality of utterance contents 46 in association with each of the plurality of categories 44. The Q & ADB 42 stores a plurality of combinations of the question and the utterance content 48 in association with each of the plurality of categories 44. The conversation DB 40 and the Q & ADB 42 may be an example of an utterance content storage unit.
 実行条件判断部112は、実行条件テーブル32に登録された複数の実行条件のいずれかが満たされたことを判断する。実行条件判断部112は、実行条件に関連する条件データを取得する条件データ取得部116が取得した条件データと、実行条件テーブル32とを参照することによって、複数の実行条件のいずれかが満たされたことを判断してよい。 The execution condition determination unit 112 determines that any of a plurality of execution conditions registered in the execution condition table 32 is satisfied. The execution condition determination unit 112 refers to the condition data acquired by the condition data acquisition unit 116 that acquires the condition data related to the execution condition and the execution condition table 32, so that one of the plurality of execution conditions is satisfied. You may judge that.
 例えば、条件データ取得部116は、ユーザ10のプロファイルデータから、ユーザ10が好きなテレビ番組のジャンルを取得する。また、条件データ取得部116は、通信網20を介して、テレビ番組データを取得する。実行条件判断部112は、ユーザ10が好きなテレビ番組のジャンルと、テレビ番組データと、実行条件テーブル32とから、「番組お知らせアプリ」の実行条件が満たされたと判断してよい。 For example, the condition data acquisition unit 116 acquires the genre of the television program that the user 10 likes from the profile data of the user 10. The condition data acquisition unit 116 acquires television program data via the communication network 20. The execution condition determination unit 112 may determine that the execution condition of the “program notification application” is satisfied from the TV program genre that the user 10 likes, the TV program data, and the execution condition table 32.
 会話アルゴリズム選択部118は、実行条件判断部112によって満たされたことが判断された実行条件に対応付けて会話アルゴリズムDB30に記憶されている条件会話アルゴリズム34を選択する。 The conversation algorithm selection unit 118 selects the conditional conversation algorithm 34 stored in the conversation algorithm DB 30 in association with the execution condition determined to be satisfied by the execution condition determination unit 112.
 会話実行部120は、会話アルゴリズム選択部118が選択した条件会話アルゴリズム34に従って、ユーザ10との会話を実行する。また、会話実行部120は、会話DB40及びQ&ADB42に記憶された発話内容46及び発話内容48により、ユーザ10との会話を実行する。 The conversation execution unit 120 executes a conversation with the user 10 in accordance with the conditional conversation algorithm 34 selected by the conversation algorithm selection unit 118. The conversation execution unit 120 executes a conversation with the user 10 based on the utterance contents 46 and the utterance contents 48 stored in the conversation DB 40 and the Q & ADB 42.
 会話実行部120は、例えば、条件会話アルゴリズム34に含まれる発話内容、会話DB40及びQ&ADB42に記憶された発話内容46及び発話内容48に対して、音声合成技術を適用することによって、音声出力制御部122に音声出力させる。また、会話実行部120は、会話アルゴリズムDB30、会話DB40、及びQ&ADB42に音声データが記憶されている場合には、当該音声データを音声出力制御部122に音声出力させてもよい。会話実行部120は、音声取得部124が取得したユーザ10の音声に音声認識技術を適用することによって、ユーザ10の音声を認識してよい。このように会話実行部120は、音声出力制御部122及び音声取得部124を制御することによって、ユーザ10との会話を実行してよい。 For example, the conversation execution unit 120 applies a speech synthesis technique to the utterance contents included in the conditional conversation algorithm 34, the utterance contents 46 and the utterance contents 48 stored in the conversation DB 40 and the Q & ADB 42, thereby providing a voice output control unit. The voice is output to 122. In addition, when voice data is stored in the conversation algorithm DB 30, the conversation DB 40, and the Q & ADB 42, the conversation execution unit 120 may cause the voice output control unit 122 to output the voice data. The conversation execution unit 120 may recognize the voice of the user 10 by applying voice recognition technology to the voice of the user 10 acquired by the voice acquisition unit 124. As described above, the conversation execution unit 120 may execute the conversation with the user 10 by controlling the voice output control unit 122 and the voice acquisition unit 124.
 音声出力制御部122は、会話実行部120の制御に従って、音声を出力する。音声出力制御部122は、会話実行部120により指定された音声データ及び音声合成データを、スピーカなどによって音声出力してよい。 The voice output control unit 122 outputs voice according to the control of the conversation execution unit 120. The voice output control unit 122 may output the voice data and voice synthesis data specified by the conversation execution unit 120 as voices using a speaker or the like.
 音声取得部124は、マイクなどによってユーザ10の音声を取得する。音声取得部124は、取得した音声を会話実行部120に送信してよい。また、音声取得部124は、取得した音声をカテゴリ選択部126に送信してよい。 The voice acquisition unit 124 acquires the voice of the user 10 using a microphone or the like. The voice acquisition unit 124 may transmit the acquired voice to the conversation execution unit 120. The voice acquisition unit 124 may transmit the acquired voice to the category selection unit 126.
 カテゴリ選択部126は、音声取得部124が取得した音声に基づいて、複数のカテゴリ44のうちの一のカテゴリ44を選択状態にする。カテゴリ選択部126は、例えば、音声取得部124が取得した音声が、会話DB40に記憶された複数の発話内容46のいずれかと一致すると判断した場合、当該発話内容46を含むカテゴリ44を選択状態にする。カテゴリ選択部126は、例えば、音声取得部124が取得した音声の音声認識結果と、発話内容46との一致度が予め定められた閾値より高い場合に、一致すると判断する。 The category selection unit 126 sets one category 44 among the plurality of categories 44 based on the voice acquired by the voice acquisition unit 124. For example, when the category selection unit 126 determines that the voice acquired by the voice acquisition unit 124 matches any of the plurality of utterance contents 46 stored in the conversation DB 40, the category selection unit 126 selects the category 44 including the utterance contents 46 as a selected state. To do. For example, the category selection unit 126 determines that they match when the degree of coincidence between the voice recognition result of the voice acquired by the voice acquisition unit 124 and the utterance content 46 is higher than a predetermined threshold.
 また、カテゴリ選択部126は、例えば、音声取得部124が取得した音声に含まれるキーワードと、カテゴリ44の名称とが一致すると判断した場合、当該カテゴリ44を選択状態にする。カテゴリ選択部126は、例えば、音声取得部124が取得した音声に含まれるキーワードと、カテゴリ44の名称との一致度が予め定められた閾値より高い場合に、一致すると判断する。 In addition, for example, when the category selection unit 126 determines that the keyword included in the voice acquired by the voice acquisition unit 124 matches the name of the category 44, the category selection unit 126 sets the category 44 in a selected state. For example, the category selection unit 126 determines that a match is found when the degree of matching between the keyword included in the voice acquired by the voice acquisition unit 124 and the name of the category 44 is higher than a predetermined threshold.
 会話実行部120は、カテゴリ選択部126が選択状態にしたカテゴリ44に対応付けて会話DB40及びQ&ADB42が記憶する複数の発話内容46及び発話内容48により、ユーザ10との会話を実行してよい。 The conversation execution unit 120 may execute a conversation with the user 10 using the plurality of utterance contents 46 and utterance contents 48 stored in the conversation DB 40 and the Q & ADB 42 in association with the category 44 selected by the category selection unit 126.
 音声取得部124は、取得した音声を感情認識部128に送信してよい。感情認識部128は、音声取得部124が取得したユーザ10の音声に基づいて、ユーザ10の感情を認識する。 The voice acquisition unit 124 may transmit the acquired voice to the emotion recognition unit 128. The emotion recognition unit 128 recognizes the emotion of the user 10 based on the voice of the user 10 acquired by the voice acquisition unit 124.
 感情認識部128は、既存の音声感情認識技術を用いてユーザ10の感情を認識してよい。感情認識部128は、ユーザ10の音声の韻律、音質、及び音韻などに基づいてユーザ10の感情を認識してよい。感情認識部128は、ユーザ10の音声の音声認識結果に基づいてユーザ10の感情を認識してもよい。感情認識部128は、これらを組み合わせてユーザ10の感情を認識してもよい。感情認識部128によって、例えば、喜び、怒り、及び悲しみなどの感情が認識されてよい。感情認識部128は、ユーザ10の音声からいずれの感情も認識できなかった場合には、その旨を示す結果データを出力してよい。 The emotion recognition unit 128 may recognize the emotion of the user 10 using existing voice emotion recognition technology. The emotion recognition unit 128 may recognize the emotion of the user 10 based on the prosody, sound quality, and phoneme of the user 10 voice. The emotion recognition unit 128 may recognize the emotion of the user 10 based on the voice recognition result of the user 10 voice. The emotion recognition unit 128 may recognize the emotion of the user 10 by combining these. For example, emotions such as joy, anger, and sadness may be recognized by the emotion recognition unit 128. When the emotion recognition unit 128 cannot recognize any emotion from the voice of the user 10, the emotion recognition unit 128 may output result data indicating that fact.
 感情認識部128は、感情認識結果を会話アルゴリズム選択部118に送信してよい。会話アルゴリズム選択部118は、感情認識部128が認識した感情に対応付けて会話アルゴリズムDB30に記憶されている感情会話アルゴリズム36を選択する。会話アルゴリズム選択部118は、例えば、感情認識部128が喜びを認識した場合、喜びに対応付けて記憶された感情会話アルゴリズム36を選択する。会話アルゴリズム選択部118は、選択した感情会話アルゴリズム36を会話実行部120に送信してよい。 The emotion recognition unit 128 may transmit the emotion recognition result to the conversation algorithm selection unit 118. The conversation algorithm selection unit 118 selects the emotion conversation algorithm 36 stored in the conversation algorithm DB 30 in association with the emotion recognized by the emotion recognition unit 128. For example, when the emotion recognition unit 128 recognizes pleasure, the conversation algorithm selection unit 118 selects the emotion conversation algorithm 36 stored in association with pleasure. The conversation algorithm selection unit 118 may transmit the selected emotion conversation algorithm 36 to the conversation execution unit 120.
 会話実行部120は、会話アルゴリズム選択部118が選択した感情会話アルゴリズム36に従って、ユーザ10との会話を実行してよい。例えば、会話実行部120は、会話アルゴリズム選択部118が喜びに対応した感情会話アルゴリズム36を選択した場合、喜びに対応した感情会話アルゴリズム36に従って、ユーザ10との会話を実行する。これにより、会話実行部120は、ユーザ10との会話中に認識したユーザ10の感情に適した会話を実行することができる。 The conversation execution unit 120 may execute a conversation with the user 10 according to the emotion conversation algorithm 36 selected by the conversation algorithm selection unit 118. For example, when the conversation algorithm selection unit 118 selects the emotion conversation algorithm 36 corresponding to pleasure, the conversation execution unit 120 executes a conversation with the user 10 according to the emotion conversation algorithm 36 corresponding to pleasure. Thereby, the conversation execution part 120 can perform the conversation suitable for the emotion of the user 10 recognized during the conversation with the user 10.
 感情認識部128は、さらに、感情認識結果を優先度変更部130に送信してよい。優先度変更部130は、感情認識結果に基づいて、条件会話アルゴリズム34に対応付けられた優先度を変更してよい。例えば、優先度変更部130は、一の条件会話アルゴリズム34を実行中に、感情認識部128が正の感情を認識した場合、当該一の条件会話アルゴリズム34の優先度を向上させる。また、優先度変更部130は、一の条件会話アルゴリズム34を実行中に、感情認識部128が負の感情を認識した場合、当該一の条件会話アルゴリズム34の優先度を低下させてよい。 The emotion recognition unit 128 may further transmit the emotion recognition result to the priority changing unit 130. The priority changing unit 130 may change the priority associated with the conditional conversation algorithm 34 based on the emotion recognition result. For example, when the emotion recognition unit 128 recognizes a positive emotion while executing one conditional conversation algorithm 34, the priority changing unit 130 improves the priority of the one conditional conversation algorithm 34. Further, the priority changing unit 130 may lower the priority of the one conditional conversation algorithm 34 when the emotion recognition unit 128 recognizes a negative emotion while executing the one conditional conversation algorithm 34.
 例えば、優先度変更部130は、「番組お知らせアプリ」を実行中に、感情認識部128が正の感情を認識した場合、「番組お知らせアプリ」の優先度を向上する。また、優先度変更部130は、例えば、「傘アラートアプリ」を実行中に、感情認識部128が負の感情を認識した場合、「傘アラートアプリ」の優先度を低下する。これにより、ユーザの感情に応じて、条件会話アルゴリズム34の優先度を適切に変更することができる。 For example, the priority changing unit 130 improves the priority of the “program notification app” when the emotion recognition unit 128 recognizes a positive emotion while executing the “program notification app”. For example, when the emotion recognition unit 128 recognizes a negative emotion while executing the “umbrella alert app”, the priority change unit 130 decreases the priority of the “umbrella alert app”. Thereby, according to a user's emotion, the priority of the conditional conversation algorithm 34 can be changed appropriately.
 また、優先度変更部130は、感情認識結果に基づいてカテゴリ44に対応付けられた優先度45を変更してよい。例えば、優先度変更部130は、一のカテゴリ44が選択された状態で、感情認識部128が正の感情を認識した場合、当該一のカテゴリ44の優先度を向上させる。また、優先度変更部130は、一のカテゴリ44が選択された状態で、感情認識部128が負の感情を認識した場合、当該一のカテゴリ44の優先度を低下させる。これにより、ユーザ10が正の感情を抱くカテゴリ44の優先度を向上させ、ユーザ10が負の感情を抱くカテゴリ44の優先度を低下させることができる。 Also, the priority changing unit 130 may change the priority 45 associated with the category 44 based on the emotion recognition result. For example, when the emotion recognition unit 128 recognizes a positive emotion in a state where one category 44 is selected, the priority changing unit 130 improves the priority of the one category 44. Moreover, the priority change part 130 reduces the priority of the said one category 44, when the emotion recognition part 128 recognizes a negative emotion in the state from which the one category 44 was selected. Thereby, the priority of the category 44 in which the user 10 has a positive emotion can be improved, and the priority of the category 44 in which the user 10 has a negative emotion can be reduced.
 発話内容登録部132は、カテゴリ選択部126が一のカテゴリを選択状態にした後に、音声取得部124が取得した音声を、当該一のカテゴリ44に対応する発話内容46として追加登録する。例えば「芸能:ミュージシャン:銀爆」が選択されている状態で、ユーザ10の「2004年に結成されたんだってね」という音声を取得した場合、発話内容登録部132は、「芸能:ミュージシャン:銀爆」に対応付ける発話内容46として、「2004年に結成されたんだってね」を追加登録する。これにより、会話DB40を充実させることができる。 The utterance content registration unit 132 additionally registers the voice acquired by the voice acquisition unit 124 as the utterance content 46 corresponding to the one category 44 after the category selection unit 126 selects one category. For example, when “Entertainment: Musician: Silver Explosion” is selected and the user's 10 voice “It was formed in 2004” is acquired, the utterance content registration unit 132 displays “Entertainment: Musician: Silver”. As an utterance content 46 corresponding to “explosion”, “you were formed in 2004” is additionally registered. Thereby, conversation DB40 can be enriched.
 会話データ生成部134は、会話の流れを示す会話データを生成する。会話の流れを示す会話データは、連続する発話内容と応答内容を含む。例えば、会話データは「メンバーBってイケメンらしいよ」、「へぇ、そうなんだ」、「筋肉もすごいらしいよ」という会話の流れであってよい。 The conversation data generation unit 134 generates conversation data indicating a conversation flow. The conversation data indicating the conversation flow includes continuous utterance contents and response contents. For example, the conversation data may be a conversation flow such as “Member B seems to be handsome”, “Hey, is that so”, and “Muscle is also amazing”.
 会話データ生成部134は、まず、選択状態のカテゴリ44に対応付けて記憶された複数の発話内容46から、第1発話内容46(例.「メンバーBってイケメンらしいよ」)を音声出力制御部122に音声出力させる。次に、会話実行部120が、第1発話内容46に対するユーザ10による第1応答内容(例.「へぇ、そうなんだ」)を取得する。そして、会話データ生成部134は、第1発話内容46と、第1応答内容とを含む会話データを生成する。 The conversation data generation unit 134 first performs voice output control of the first utterance content 46 (eg, “Member B seems to be handsome”) from the plurality of utterance contents 46 stored in association with the selected category 44. The unit 122 outputs a sound. Next, the conversation execution unit 120 acquires the first response content (eg, “Hey, that's right”) by the user 10 with respect to the first utterance content 46. Then, the conversation data generation unit 134 generates conversation data including the first utterance content 46 and the first response content.
 当該会話データを生成した後、会話実行部120は、第1発話内容46に一致する音声をユーザ10から取得した場合に、第1応答内容を音声出力する。これにより、人間らしい自然な応答をすることができる。またさらに、会話実行部120は、音声出力した第1応答内容に対する、ユーザ10の第2応答内容(例.「筋肉もすごいらしいよ」)を取得する。そして、会話データ生成部134は、第1発話内容46、第1応答内容、第2応答内容を順序付けて会話データを生成する。これにより、人間らしい受け答えを続けることができる会話データを生成できる。 After generating the conversation data, the conversation execution unit 120 outputs the first response content as a voice when the voice matching the first utterance content 46 is acquired from the user 10. As a result, a human-like natural response can be made. Furthermore, the conversation execution unit 120 acquires the second response content of the user 10 with respect to the first response content output by voice (for example, “muscles are also amazing”). Then, the conversation data generation unit 134 generates conversation data by ordering the first utterance content 46, the first response content, and the second response content. Thereby, the conversation data which can continue receiving and answering like a human being can be generated.
 なお、ここでは、一のユーザ10に対して会話データを生成する例を挙げて説明したが、二人のユーザ10に対して会話データを生成してもよい。また、2台の情報端末100によって、会話データを生成してもよい。例えば、会話データ生成部134は、第1発話内容46と第1応答内容を含む会話データを生成した後、当該会話データを他の情報端末100に送信する。そして、他の情報端末100において、第1発話内容46に一致する音声を取得した場合に、第1応答内容を音声出力し、第2応答内容を取得して、第1発話内容46、第1応答内容、第2応答内容を順序付けて会話データを生成する。 In addition, although the example which produces | generates conversation data with respect to the one user 10 was given and demonstrated here, you may produce | generate conversation data with respect to the two users 10. FIG. Further, conversation data may be generated by the two information terminals 100. For example, the conversation data generation unit 134 generates conversation data including the first utterance content 46 and the first response content, and then transmits the conversation data to another information terminal 100. Then, in another information terminal 100, when the voice that matches the first utterance content 46 is acquired, the first response content is output as voice, the second response content is acquired, and the first utterance content 46, first The response contents and the second response contents are ordered to generate conversation data.
 アルゴリズム共有処理部136は、会話アルゴリズムDB30が記憶する条件会話アルゴリズム34を、他の情報端末100が有する会話アルゴリズムDB30にコピーする。例えば、アルゴリズム共有処理部136は、ユーザ10のプロファイルとの類似度が予め定められた基準を超えているプロファイルを有する他のユーザが所持する情報端末100の会話アルゴリズムDB30に、条件会話アルゴリズム34をコピーする。アルゴリズム共有処理部136は、例えば、住所、年齢、性別が一致する場合に、プロファイルの類似度が基準を超えていると判断してよい。 The algorithm sharing processing unit 136 copies the conditional conversation algorithm 34 stored in the conversation algorithm DB 30 to the conversation algorithm DB 30 included in another information terminal 100. For example, the algorithm sharing processing unit 136 adds the conditional conversation algorithm 34 to the conversation algorithm DB 30 of the information terminal 100 possessed by another user who has a profile whose similarity with the profile of the user 10 exceeds a predetermined standard. make a copy. For example, when the address, age, and gender match, the algorithm sharing processing unit 136 may determine that the profile similarity exceeds a standard.
 これにより、同じようなプロファイルを持つユーザに対して、有用な条件会話アルゴリズム34を提供することができる。なお、アルゴリズム共有処理部136は、例えば、条件会話アルゴリズム34の体験版を、当該他の情報端末100の会話アルゴリズムDB30にコピーしてよい。これにより、他の情報端末100を所持するユーザ10に、条件会話アルゴリズム34を試用させることができ、条件会話アルゴリズム34を購入する参考とさせることができる。 This makes it possible to provide a useful conditional conversation algorithm 34 for users having similar profiles. The algorithm sharing processing unit 136 may copy, for example, a trial version of the conditional conversation algorithm 34 to the conversation algorithm DB 30 of the other information terminal 100. Thereby, the user 10 possessing the other information terminal 100 can try the conditional conversation algorithm 34 and can be used as a reference for purchasing the conditional conversation algorithm 34.
 なお、サーバ200が、図4に示す機能構成と同一の機能構成を有してもよい。この場合、音声出力制御部122は、情報端末100に音声出力させるべく、情報端末100を制御してよい。また、音声取得部124は、情報端末100から、ユーザ10の音声を受信してよい。 The server 200 may have the same functional configuration as that shown in FIG. In this case, the audio output control unit 122 may control the information terminal 100 so that the information terminal 100 outputs the audio. Further, the voice acquisition unit 124 may receive the voice of the user 10 from the information terminal 100.
 また、サーバ200が会話データ生成部134を備える場合、サーバ200は、まず、複数の情報端末100からランダムに1台の情報端末100を選択して、第1発話内容46を音声出力させる。次に、当該情報端末100に、第1応答内容を取得させ、第1発話内容46と第1応答内容とを含む会話データを生成する。次に、複数の情報端末100の会話を監視して、第1発話内容46と同一の音声を取得した情報端末100を特定する。そして、特定した情報端末100に、第1応答内容を音声出力させ、第2応答内容を取得させる。これにより、サーバ200は、人間らしい受け答えをつづけることができる会話データを生成できる。 Further, when the server 200 includes the conversation data generation unit 134, the server 200 first selects one information terminal 100 at random from the plurality of information terminals 100, and outputs the first utterance content 46 by voice. Next, the information terminal 100 is made to acquire the first response content, and conversation data including the first utterance content 46 and the first response content is generated. Next, the conversation of a plurality of information terminals 100 is monitored, and the information terminal 100 that has acquired the same voice as the first utterance content 46 is specified. Then, the identified information terminal 100 is caused to output the first response content by voice and acquire the second response content. As a result, the server 200 can generate conversation data that can continue to be accepted and answered like humans.
 また、サーバ200がアルゴリズム共有処理部136を備える場合、サーバ200は、まず、任意に一の情報端末100を選択する。そして、サーバ200は、選択した情報端末100のユーザ10のプロファイルとの類似度が予め定められた基準を超えているプロファイルを有する他のユーザが所持する情報端末100とを特定して、コピー処理を実行する。 Further, when the server 200 includes the algorithm sharing processing unit 136, the server 200 first arbitrarily selects one information terminal 100. Then, the server 200 identifies the information terminal 100 possessed by another user having a profile whose similarity with the profile of the user 10 of the selected information terminal 100 exceeds a predetermined standard, and performs copy processing Execute.
 図5は、情報端末100による動作フローの一例を概略的に示す。図5に示す動作フローは、本実施形態に係る会話処理の実行指示を受け付けた場合に開始してよい。 FIG. 5 schematically shows an example of an operation flow by the information terminal 100. The operation flow illustrated in FIG. 5 may be started when an execution instruction for conversation processing according to the present embodiment is received.
 ステップ502(ステップをSと省略して表記する場合がある。)では、実行条件判断部112が、実行条件テーブル32に登録された複数の実行条件のうちのいずれかが満たされたか否かを判断する。S502で、満たされたと判断された場合、S504に進む。 In step 502 (step may be abbreviated as S), the execution condition determination unit 112 determines whether any of a plurality of execution conditions registered in the execution condition table 32 is satisfied. to decide. If it is determined in S502 that the condition is satisfied, the process proceeds to S504.
 S504では、会話アルゴリズム選択部118が、実行条件判断部112によって満たされたことが判断された実行条件に対応付けて会話アルゴリズムDB30に記憶されている条件会話アルゴリズム34を選択する。S506では、選択された条件会話アルゴリズム34に応じて、会話実行部120が発話内容を取得する。会話実行部120は、会話アルゴリズムDB30から、選択された条件会話アルゴリズム34に対応付けて登録された発話内容を取得してよい。会話アルゴリズムDB30は、条件会話アルゴリズム34が選択された場合に最初に音声出力する発話内容を保持していてよい。また、会話アルゴリズムDB30は、ユーザ10の音声に対する応答を示す発話内容を保持していてよい。ユーザ10の音声に対する応答は、例えば、会話アルゴリズムDB30の管理者によって、予め登録される。会話アルゴリズムDB30の管理者は、予め想定されるユーザ10の音声とその応答を示す発話内容とを会話アルゴリズムDB30に登録してよい。 In S504, the conversation algorithm selection unit 118 selects the conditional conversation algorithm 34 stored in the conversation algorithm DB 30 in association with the execution condition determined to be satisfied by the execution condition determination unit 112. In S506, according to the selected conditional conversation algorithm 34, the conversation execution unit 120 acquires the utterance content. The conversation execution unit 120 may acquire the utterance content registered in association with the selected conditional conversation algorithm 34 from the conversation algorithm DB 30. The conversation algorithm DB 30 may hold an utterance content that is first voice-output when the conditional conversation algorithm 34 is selected. The conversation algorithm DB 30 may hold utterance contents indicating a response to the voice of the user 10. The response to the voice of the user 10 is registered in advance by, for example, the administrator of the conversation algorithm DB 30. The administrator of the conversation algorithm DB 30 may register the voice of the user 10 assumed in advance and the utterance content indicating the response in the conversation algorithm DB 30.
 S508では、会話実行部120が、S506で取得した発話内容を音声出力制御部122に音声出力させる。S510では、会話実行部120が、ユーザ10による会話終了指示を受領したか否かを判定する。終了指示を受領していなかった場合、S512に進む。S512では、S508で出力した音声に対してユーザ10が発した音声を音声取得部124が取得する。 In S508, the conversation execution unit 120 causes the voice output control unit 122 to output the utterance content acquired in S506. In S510, the conversation execution unit 120 determines whether or not a conversation end instruction from the user 10 has been received. If an end instruction has not been received, the process proceeds to S512. In S512, the voice acquisition unit 124 acquires the voice uttered by the user 10 with respect to the voice output in S508.
 S514では、感情認識部128が、音声取得部124で取得したユーザ10の音声に基づいて、ユーザ10の感情を認識する。S516では、会話実行部120が、S514で認識された感情が予め定められた感情に一致するか否かを判断する。予め定められた感情は、例えば、感情会話アルゴリズム36が対応付けられている感情である。予め定められた感情は、複数の感情のうち事前に指定された感情であってよい。 In S514, the emotion recognition unit 128 recognizes the emotion of the user 10 based on the voice of the user 10 acquired by the voice acquisition unit 124. In S516, conversation execution unit 120 determines whether or not the emotion recognized in S514 matches a predetermined emotion. The predetermined emotion is, for example, an emotion associated with the emotion conversation algorithm 36. The predetermined emotion may be an emotion designated in advance among a plurality of emotions.
 S516で、予め定められた感情に一致すると判定された場合、S518に進み、一致しないと判定された場合、S508に戻る。S518では、優先度変更部130が、514で認識された感情に基づいて、S504で選択された条件会話アルゴリズム34の優先度を変更する。 If it is determined in S516 that the emotion matches the predetermined emotion, the process proceeds to S518. If it is determined that the emotion does not match, the process returns to S508. In S518, the priority changing unit 130 changes the priority of the conditional conversation algorithm 34 selected in S504 based on the emotion recognized in 514.
 S520では、会話実行部120が、条件会話アルゴリズム34に従った会話を中断し、S514で認識された感情に対応付けて会話アルゴリズムDB30に記憶されている感情会話アルゴリズム36に従って、ユーザ10との会話を実行する。S522では、S520で会話実行部120が感情会話アルゴリズム36を実行した結果、元の会話を継続すると判定されていた場合、S508に戻る。S522で元の会話を継続すると判定されていなかった場合、処理を終了する。 In S520, the conversation execution unit 120 interrupts the conversation according to the conditional conversation algorithm 34, and the conversation with the user 10 is performed in accordance with the emotion conversation algorithm 36 stored in the conversation algorithm DB 30 in association with the emotion recognized in S514. Execute. In S522, when it is determined that the original conversation is continued as a result of the conversation execution unit 120 executing the emotion conversation algorithm 36 in S520, the process returns to S508. If it is not determined in S522 to continue the original conversation, the process ends.
 S524では、会話実行部120が、発話内容を取得できる否かを判定する。会話実行部120は、S512で取得した音声に対応する発話内容を会話アルゴリズムDB30から取得できるか否かを判定してよい。例えば、S504で選択された会話アルゴリズムが「血液型占いアプリ」であり、S508で「あなたの血液型は何ですか?」と音声出力したのに対して、予め登録されたユーザ音声、例えば「A型です」という音声をS512で取得した場合、会話実行部120は、発話内容を取得できたと判定する。これに対して、予め登録されていないユーザ音声、例えば「そういえば、君の血液型は何だっけ?」という音声をS512で取得した場合、会話実行部120は、発話内容を取得できなかったと判定する。S524で、発話内容を取得できたと判定された場合、S508に進み、会話実行部120が、取得した発話内容を音声出力制御部122に音声出力させる。S524で、発話内容を取得できなかったと判定された場合、図6のフローに進む。 In S524, the conversation execution unit 120 determines whether the utterance content can be acquired. The conversation execution unit 120 may determine whether or not the utterance content corresponding to the voice acquired in S512 can be acquired from the conversation algorithm DB 30. For example, the conversation algorithm selected in S504 is “blood group fortune-telling app”, and in S508, “What is your blood type?” Is output as a voice, whereas a pre-registered user voice, for example “ When the voice “A type” is acquired in S512, the conversation execution unit 120 determines that the utterance content has been acquired. On the other hand, when the user voice that is not registered in advance, for example, the voice of “What is your blood type?” Is acquired in S512, the conversation execution unit 120 cannot acquire the utterance content. judge. If it is determined in S524 that the utterance content has been acquired, the process proceeds to S508, and the conversation execution unit 120 causes the audio output control unit 122 to output the acquired utterance content as audio. If it is determined in S524 that the utterance content could not be acquired, the process proceeds to the flow of FIG.
 図6は、情報端末100による動作フローの一例を概略的に示す。図6に示す動作フローは、図5のS524で、発話内容を取得できなかったと判定された場合に開始してよい。また、図6に示す動作フローは、例えば、ユーザ10からの音声を取得した任意のタイミングで開始してよい。 FIG. 6 schematically shows an example of an operation flow by the information terminal 100. The operation flow illustrated in FIG. 6 may be started when it is determined in S524 of FIG. 5 that the utterance content has not been acquired. Further, the operation flow illustrated in FIG. 6 may be started at an arbitrary timing at which the voice from the user 10 is acquired, for example.
 S602では、カテゴリ選択部126が、ユーザ10の音声に基づいて、複数のカテゴリ44のうちの一のカテゴリ44を選択状態にする。S604では、会話実行部120が、会話DB40に記憶された、選択状態のカテゴリ44に対応付けられた複数の発話内容46から一の発話内容46を取得する。 In S602, the category selection unit 126 selects one category 44 among the plurality of categories 44 based on the voice of the user 10. In S <b> 604, the conversation execution unit 120 acquires one utterance content 46 from the plurality of utterance contents 46 associated with the selected category 44 stored in the conversation DB 40.
 S606では、会話実行部120が、S604で取得した発話内容46を、音声出力制御部122に音声出力させる。S608では、会話実行部120が、ユーザ10による会話終了指示を受領したか否かを判定する。終了指示を受領していなかった場合、S610に進む。 In S606, the conversation execution unit 120 causes the audio output control unit 122 to output the utterance content 46 acquired in S604. In step S <b> 608, the conversation execution unit 120 determines whether a conversation end instruction from the user 10 has been received. If an end instruction has not been received, the process proceeds to S610.
 S610では、音声取得部124が、S606で出力した音声に対するユーザの音声を取得する。S612では、感情認識部128が、S610で取得した音声に基づいて、ユーザ10の感情を認識する。S614では、会話実行部120が、S612で認識された感情が予め定められた感情に一致するか否かを判断する。予め定められた感情は、例えば、感情会話アルゴリズム36が対応付けられている感情である。予め定められた感情は、複数の感情のうち事前に指定された感情であってよい。 In S610, the voice acquisition unit 124 acquires the user's voice with respect to the voice output in S606. In S612, the emotion recognition unit 128 recognizes the emotion of the user 10 based on the voice acquired in S610. In S614, conversation execution unit 120 determines whether or not the emotion recognized in S612 matches a predetermined emotion. The predetermined emotion is, for example, an emotion associated with the emotion conversation algorithm 36. The predetermined emotion may be an emotion designated in advance among a plurality of emotions.
 S614で、予め定められた感情に一致すると判定された場合、S616に進み、一致しないと判定された場合、S622に進む。S616では、優先度変更部130が、S612で認識された感情に基づいて、S602で選択されたカテゴリ44の優先度を変更する。 If it is determined in S614 that the emotion matches the predetermined emotion, the process proceeds to S616. If it is determined that the emotion does not match, the process proceeds to S622. In S616, the priority changing unit 130 changes the priority of the category 44 selected in S602 based on the emotion recognized in S612.
 S618では、会話実行部120が、選択状態のカテゴリ44に対応付けられた複数の発話内容46による会話を中断し、S612で認識された感情に対応付けて会話アルゴリズムDB30に記憶されている感情会話アルゴリズム36に従って、ユーザ10との会話を実行する。S620では、S618で会話実行部120が感情会話アルゴリズム36を実行した結果、元の会話を継続すると判定されていた場合、S622に進む。S620で元の会話を継続すると判定しなかった場合、処理を終了する。 In S618, the conversation execution unit 120 interrupts the conversation based on the plurality of utterance contents 46 associated with the selected category 44, and the emotion conversation stored in the conversation algorithm DB 30 in association with the emotion recognized in S612. A conversation with the user 10 is executed according to the algorithm 36. In S620, if it is determined that the original conversation is continued as a result of the conversation execution unit 120 executing the emotion conversation algorithm 36 in S618, the process proceeds to S622. If it is not determined in S620 to continue the original conversation, the process ends.
 S622では、会話実行部120が、発話内容46を取得できるか否かを判定する。会話実行部120は、選択状態のカテゴリ44に基づいて会話DB40から発話内容46を取得できるか否かを判定してよい。例えば、会話実行部120は、選択状態のカテゴリ44に対応付けられた複数の発話内容46から、音声出力する一の発話内容46を取得できた場合に、発話内容46を取得できたと判定し、取得できなかった場合に、発話内容46を取得できなかったと判定する。会話実行部120は、例えば、一の会話処理において、選択状態のカテゴリ44に対応付けられた複数の発話内容46をすべて出力し、未出力の発話内容46がなくなった場合に、発話内容46を取得できなかったと判定する。 In S622, the conversation execution unit 120 determines whether the utterance content 46 can be acquired. The conversation execution unit 120 may determine whether or not the utterance content 46 can be acquired from the conversation DB 40 based on the selected category 44. For example, the conversation execution unit 120 determines that the utterance content 46 has been acquired when one utterance content 46 to be output by voice is obtained from the plurality of utterance contents 46 associated with the category 44 in the selected state. If it cannot be acquired, it is determined that the utterance content 46 cannot be acquired. For example, in one conversation process, the conversation execution unit 120 outputs all of the plurality of utterance contents 46 associated with the selected category 44, and when there is no unoutput utterance contents 46, the utterance contents 46 are displayed. It is determined that it could not be acquired.
 S622で、発話内容46を取得できたと判定された場合、S606に戻り、会話実行部120が、S622で取得した発話内容46を、音声出力制御部122に音声出力させる。S622で、発話内容46を取得できなかったと判定された場合、S624に進む。 If it is determined in S622 that the utterance content 46 has been acquired, the process returns to S606, and the conversation execution unit 120 causes the voice output control unit 122 to output the utterance content 46 acquired in S622. If it is determined in S622 that the utterance content 46 has not been acquired, the process proceeds to S624.
 S624では、実行条件判断部112が、実行条件テーブル32に登録された複数の実行条件のうちのいずれかが満たされたか否かを判断する。いずれかの実行条件が満たされたと判定された場合、図5のS504に進む。このように、本実施形態に係る情報端末100は、会話アルゴリズムDB30を用いた会話処理と、会話DB40及びQ&ADB42を用いた会話処理とを適宜切り替えることによってユーザ10との会話を進めてよい。 In S624, the execution condition determination unit 112 determines whether any of a plurality of execution conditions registered in the execution condition table 32 is satisfied. If it is determined that any one of the execution conditions is satisfied, the process proceeds to S504 in FIG. As described above, the information terminal 100 according to the present embodiment may advance the conversation with the user 10 by appropriately switching between the conversation process using the conversation algorithm DB 30 and the conversation process using the conversation DB 40 and the Q & ADB 42.
 S624で、複数の実行条件のうちのいずれもが満たされていないと判定された場合、S626に進む。S626では、会話実行部120が、エラー処理を実行する。例えば、会話実行部120は、会話処理を終了する旨をユーザ10に通知した後、会話処理を終了する。なお、会話実行部120は、ユーザ10に他の発声を促す通知をして、ユーザ10から新たな音声を取得することにより、会話処理を継続してもよい。この場合、例えば、S626でユーザ10に他の発声を促す通知をした後、S602又はS610に戻ってよい。 If it is determined in S624 that none of the plurality of execution conditions is satisfied, the process proceeds to S626. In S626, the conversation execution unit 120 executes error processing. For example, the conversation execution unit 120 notifies the user 10 that the conversation process is to be terminated, and then terminates the conversation process. Note that the conversation execution unit 120 may continue the conversation process by notifying the user 10 to urge other utterances and acquiring a new voice from the user 10. In this case, for example, after notifying the user 10 of another utterance in S626, the process may return to S602 or S610.
 図7は、喜びに対応する感情会話アルゴリズム36の一例を概略的に示す。S702では、会話実行部120が、音声出力制御部122に、ユーザ10が喜びを感じてくれたことへの感謝と、会話を続けるか否かを問い合わせる内容の音声を出力させる。例えば、音声出力制御部122は「ありがとうございます。続けますか?」と音声出力する。 FIG. 7 schematically shows an example of the emotion conversation algorithm 36 corresponding to pleasure. In step S <b> 702, the conversation execution unit 120 causes the audio output control unit 122 to output thanks for appreciation that the user 10 feels joy and to inquire whether or not to continue the conversation. For example, the audio output control unit 122 outputs “Thank you. Do you want to continue?”
 S704では、S702で出力した音声に対してユーザ10が発した音声を、音声取得部124が取得する。そして、音声取得部124が取得したユーザ10の音声により、ユーザ10が会話を続けることを希望しているか否かを会話実行部120が判断する。会話実行部120が、希望していると判断した場合、S706に進み、希望していないと判断した場合、S710に進む。 In S704, the voice acquisition unit 124 acquires the voice uttered by the user 10 with respect to the voice output in S702. Then, the conversation execution unit 120 determines whether or not the user 10 wishes to continue the conversation based on the voice of the user 10 acquired by the voice acquisition unit 124. When the conversation execution unit 120 determines that it is desired, the process proceeds to S706, and when it is determined that it is not desired, the process proceeds to S710.
 S706では、会話実行部120が、音声出力制御部122に、ユーザ10が会話を続けることを選択してくれたことへの喜びを表現する音声を出力させる。例えば、音声出力制御部122は「嬉しいです」と音声出力する。 In S706, the conversation execution unit 120 causes the voice output control unit 122 to output a voice expressing joy that the user 10 has selected to continue the conversation. For example, the audio output control unit 122 outputs “I am happy”.
 S708では、会話実行部120が、元の会話を継続すると判定する。そして、リターンする。すなわち、図5の動作フローであればS522に進み、図6の動作フローであれば、S620に進む。 In S708, the conversation execution unit 120 determines to continue the original conversation. Then return. That is, if it is the operation | movement flow of FIG. 5, it will progress to S522, and if it is the operation | movement flow of FIG. 6, it will progress to S620.
 S710では、会話実行部120が、音声出力制御部122に、ユーザ10が会話を続けることを希望していないことを了承する内容と、会話を終了させることを示す内容とを音声出力させる。例えば、音声出力制御部122は「そうですね。今日はこの辺りですね」と音声出力する。 In S710, the conversation execution unit 120 causes the audio output control unit 122 to output the content that the user 10 does not wish to continue the conversation and the content that indicates that the conversation is to be ended. For example, the voice output control unit 122 outputs a voice saying “Yes, today.
 このように喜びに対応する感情会話アルゴリズム36によれば、ユーザ10が喜びを感じている会話を継続させることができる。また、ユーザ10が満足している場合に、会話をしつこく継続させることなく、適切に会話を終了させることができる。 Thus, according to the emotion conversation algorithm 36 corresponding to pleasure, the conversation in which the user 10 feels pleasure can be continued. Further, when the user 10 is satisfied, the conversation can be appropriately terminated without continuing the conversation persistently.
 図8は、怒りに対応する会話アルゴリズムの一例を概略的に示す。S802では、会話実行部120が、音声出力制御部122に、ユーザ10に怒りを感じさせてしまったことに対する音声を出力させる。例えば、音声出力制御部122は「あちゃー、またやってしまいましたか」と音声出力する。 FIG. 8 schematically shows an example of a conversation algorithm corresponding to anger. In step S <b> 802, the conversation execution unit 120 causes the audio output control unit 122 to output a sound corresponding to the user 10 feeling angry. For example, the audio output control unit 122 outputs an audio message “Acha, have you done it again?”.
 S804では、S802で出力した音声に対してユーザ10が発した音声を、音声取得部124が取得する。そして、音声取得部124が取得したユーザ10の音声により、ユーザ10が怒りを感じているか否かを会話実行部120が判断する。会話実行部120が、怒りを感じていると判断した場合、S806に進み、怒りを感じていないと判断した場合、S810に進む。 In S804, the voice acquisition unit 124 acquires the voice uttered by the user 10 with respect to the voice output in S802. Then, the conversation execution unit 120 determines whether or not the user 10 is angry based on the voice of the user 10 acquired by the voice acquisition unit 124. When the conversation execution unit 120 determines that it feels anger, the process proceeds to S806, and when it determines that it does not feel anger, the process proceeds to S810.
 S806では、会話実行部120が、音声出力制御部122に、しばらく音声出力を行わない旨を示す音声を出力させる。例えば、音声出力制御部122は、「しばらく静かにしておきますね」と音声出力する。S808では、会話実行部120が、会話を待機する待機状態に移行する。 In S806, the conversation execution unit 120 causes the voice output control unit 122 to output a voice indicating that voice output is not performed for a while. For example, the voice output control unit 122 outputs a voice saying “Please keep quiet for a while”. In S808, the conversation execution unit 120 shifts to a standby state for waiting for conversation.
 S810では、会話実行部120が音声出力制御部122に、感情認識結果が誤っていたことに対する音声を出力させる。例えば、音声出力制御部122は「あれあれ?」と音声出力する。このように、怒りに対応する感情会話アルゴリズム36によれば、ユーザ10が怒りを感じている場合に、会話を止めて待機状態に移行することができる。 In S810, the conversation execution unit 120 causes the voice output control unit 122 to output a voice in response to an erroneous emotion recognition result. For example, the voice output control unit 122 outputs a voice “That?”. Thus, according to the emotion conversation algorithm 36 corresponding to anger, when the user 10 feels anger, a conversation can be stopped and it can transfer to a standby state.
 図9は、悲しみに対応する会話アルゴリズムの一例を概略的に示す。S902では、会話実行部120が、音声出力制御部122に、ユーザ10が悲しみを感じているか否かを確認するための音声を出力させる。例えば、音声出力制御部122は「あれ、のりが悪いですね」と音声出力する。 FIG. 9 schematically shows an example of a conversation algorithm corresponding to sadness. In step S <b> 902, the conversation execution unit 120 causes the audio output control unit 122 to output a sound for confirming whether the user 10 feels sadness. For example, the voice output control unit 122 outputs a voice saying “That's bad glue”.
 S904では、S902で出力した音声に対してユーザ10が発した音声を、音声取得部124が取得する。そして、音声取得部124が取得したユーザ10の音声により、ユーザ10が悲しみを感じているか否かを、会話実行部120が判断する。会話実行部120が、感じていると判断した場合、S906に進み、感じていないと判断した場合、S910に進む。 In S904, the voice acquisition unit 124 acquires the voice uttered by the user 10 with respect to the voice output in S902. Then, the conversation execution unit 120 determines whether the user 10 feels sadness based on the voice of the user 10 acquired by the voice acquisition unit 124. When the conversation execution unit 120 determines that it is felt, the process proceeds to S906, and when it is determined that it is not felt, the process proceeds to S910.
 S906では、会話実行部120が、音声出力制御部122に、話題を切り替えることをユーザ10に通知する音声を出力させる。音声出力制御部122は、例えば「別の話題にしましょう」と音声出力する。 In S906, the conversation execution unit 120 causes the audio output control unit 122 to output a voice that notifies the user 10 that the topic is to be switched. The audio output control unit 122 outputs, for example, “let's talk about another topic”.
 S908では、会話実行部120が、会話を切り替える。例えば、会話実行部120は、一のカテゴリ44が選択状態になっていた場合、他のカテゴリ44を選択状態にする。会話実行部120は、複数のカテゴリ44が階層構造を有している場合には、選択状態になっていた一のカテゴリ44に隣接するカテゴリ44を選択状態にしてよい。 In S908, the conversation execution unit 120 switches conversations. For example, when one category 44 is in a selected state, the conversation execution unit 120 puts another category 44 into a selected state. When a plurality of categories 44 have a hierarchical structure, the conversation execution unit 120 may select a category 44 adjacent to one category 44 that has been selected.
 また、会話実行部120は、選択状態になっていた一のカテゴリ44よりも高い優先度を有するカテゴリ44を選択状態にしてもよい。また、会話実行部120は、複数のカテゴリ44のうち、ランダムに選択したカテゴリ44を選択状態にしてもよい。 Further, the conversation execution unit 120 may select a category 44 having a higher priority than the one category 44 that has been selected. Further, the conversation execution unit 120 may select a category 44 selected at random from the plurality of categories 44.
 S910では、会話実行部120が、元の会話を継続すると判定する。そして、リターンする。すなわち、図5の動作フローであればS522に進み、図6の動作フローであればS620に進む。このように、悲しみに対応する感情会話アルゴリズム36によれば、ユーザ10に、より楽しい感情を抱かせるべく、会話を切り替えることができる。 In S910, the conversation execution unit 120 determines to continue the original conversation. Then return. That is, if it is the operation | movement flow of FIG. 5, it will progress to S522, and if it is the operation | movement flow of FIG. 6, it will progress to S620. Thus, according to the emotion conversation algorithm 36 corresponding to sadness, the conversation can be switched in order to make the user 10 have a more enjoyable emotion.
 図10は、カテゴリ44の階層構造の一例を概略的に示す。複数のカテゴリ名52のそれぞれには、優先度54が割り当てられている。カテゴリ選択部126は、例えば、「芸能:ミュージシャン:グループA」を選択した状態で会話を切り替える場合に、隣接する「芸能:ミュージシャン:グループB」を選択状態にしてよい。この場合、会話実行部120は、上位のカテゴリ名52を含む音声を音声出力制御部122に出力させてよい。例えば、音声出力制御部122は、「ミュージシャンといえば、グループBですよね」と音声出力する。このように、隣接するカテゴリ44に切り替えることによって、現在選択されているカテゴリとの関連性が高いカテゴリに関する会話を実行できる。 FIG. 10 schematically shows an example of the hierarchical structure of the category 44. A priority 54 is assigned to each of the plurality of category names 52. For example, when the conversation is switched while “entertainment: musician: group A” is selected, the category selection unit 126 may select the adjacent “entertainment: musician: group B”. In this case, the conversation execution unit 120 may cause the audio output control unit 122 to output a sound including the upper category name 52. For example, the voice output control unit 122 outputs a voice saying “Speaking of musicians, it is group B”. In this way, by switching to the adjacent category 44, it is possible to execute a conversation related to a category that is highly related to the currently selected category.
 また、カテゴリ選択部126は、優先度に基づいて、現在選択されている「芸能:ミュージシャン:グループA」以外のカテゴリ44を選択状態にしてよい。例えば、カテゴリ選択部126は、現在選択されているカテゴリ44よりも優先度が高い「テレビ:通販番組」を選択状態にしてよい。これにより、ユーザ10にとってより優先度の高い会話に切り替えることができる。 Further, the category selection unit 126 may select a category 44 other than the currently selected “entertainment: musician: group A” based on the priority. For example, the category selection unit 126 may select “TV: mail order program” having a higher priority than the currently selected category 44. Thereby, it is possible to switch to a conversation with higher priority for the user 10.
 また、カテゴリ選択部126は、ユーザ10のプロファイルに基づいて、現在選択されている「芸能:ミュージシャン:グループA」以外のカテゴリ44を選択状態にしてよい。例えば、ユーザ10のプロファイルに趣味として「マラソン」が登録されている場合に、カテゴリ選択部126は、「スポーツ一般:陸上競技:マラソン」を選択状態にしてよい。これにより、ユーザ10のプロファイルに適した会話に切り替えることができる。 Further, the category selection unit 126 may select a category 44 other than the currently selected “entertainment: musician: group A” based on the profile of the user 10. For example, when “marathon” is registered as a hobby in the profile of the user 10, the category selection unit 126 may select “sports general: athletics: marathon”. Thereby, it can switch to the conversation suitable for the user's 10 profile.
 以上の説明において、情報端末100の各部は、ハードウエアにより実現されてもよく、ソフトウエアにより実現されてもよい。また、ハードウエアとソフトウエアとの組み合わせにより実現されてもよい。例えば、情報端末100上でプログラムが実行されることにより、コンピュータが、情報端末100の一部として機能してもよい。プログラムは、コンピュータ読み取り可能な媒体に記憶されていてもよく、ネットワークに接続された記憶装置に記憶されていてもよい。CPU、ROM、RAM、通信インターフェース等を有するデータ処理装置と、入力装置と、出力装置と、記憶装置とを備えた一般的な構成の情報処理装置において、情報端末100の各部の動作を規定したソフトウエア又はプログラムを起動することにより、情報端末100が実現されてよい。 In the above description, each unit of the information terminal 100 may be realized by hardware or may be realized by software. Further, it may be realized by a combination of hardware and software. For example, a computer may function as a part of the information terminal 100 by executing a program on the information terminal 100. The program may be stored in a computer-readable medium, or may be stored in a storage device connected to a network. In an information processing apparatus having a general configuration including a data processing device having a CPU, ROM, RAM, communication interface, etc., an input device, an output device, and a storage device, the operation of each part of the information terminal 100 is defined. The information terminal 100 may be realized by starting software or a program.
 コンピュータにインストールされ、コンピュータを本実施形態に係る情報端末100の一部として機能させるプログラムは、情報端末100の各部の動作を規定したモジュールを備える。これらのプログラム又はモジュールは、CPU等に働きかけて、コンピュータを、情報端末100の各部としてそれぞれ機能させる。これらのプログラムに記述された情報処理は、コンピュータに読込まれることにより、ソフトウエアと上述した各種のハードウエア資源とが協働した具体的手段として機能する。そして、これらの具体的手段によって、本実施形態におけるコンピュータの使用目的に応じた情報の演算又は加工を実現することにより、使用目的に応じた特有の測定装置を構築することができる。 A program that is installed in a computer and causes the computer to function as a part of the information terminal 100 according to the present embodiment includes a module that defines the operation of each unit of the information terminal 100. These programs or modules work on the CPU or the like to cause the computer to function as each unit of the information terminal 100. Information processing described in these programs functions as a specific means in which software and the various hardware resources described above cooperate with each other by being read by a computer. A specific measurement device according to the purpose of use can be constructed by realizing calculation or processing of information according to the purpose of use of the computer in the present embodiment by these specific means.
 また、サーバ200は、CPU、ROM、RAM、通信インターフェースなどを有するデータユニットと、キーボード、タッチパネル、マイクなどの入力ユニットと、ディスプレイ、スピーカなどの出力ユニットと、メモリ、HDDなどの記憶ユニットとを備えた一般的な構成の情報処理装置において、サーバ200の各部の動作を規定したソフトウェア又はプログラムを起動することにより実現されてよい。サーバ200は、仮想サーバまたはクラウドシステムであってもよい。 The server 200 includes a data unit having a CPU, a ROM, a RAM, a communication interface, an input unit such as a keyboard, a touch panel, and a microphone, an output unit such as a display and a speaker, and a storage unit such as a memory and an HDD. The information processing apparatus having a general configuration provided may be realized by activating software or a program that defines the operation of each unit of the server 200. The server 200 may be a virtual server or a cloud system.
 以上、本発明を実施の形態を用いて説明したが、本発明の技術的範囲は上記実施の形態に記載の範囲には限定されない。上記実施の形態に、多様な変更または改良を加えることが可能であることが当業者に明らかである。その様な変更または改良を加えた形態も本発明の技術的範囲に含まれ得ることが、請求の範囲の記載から明らかである。 As mentioned above, although this invention was demonstrated using embodiment, the technical scope of this invention is not limited to the range as described in the said embodiment. It will be apparent to those skilled in the art that various modifications or improvements can be added to the above-described embodiment. It is apparent from the scope of the claims that the embodiments added with such changes or improvements can be included in the technical scope of the present invention.
 請求の範囲、明細書、および図面中において示した装置、システム、プログラム、および方法における動作、手順、ステップ、および段階などの各処理の実行順序は、特段「より前に」、「先立って」などと明示しておらず、また、前の処理の出力を後の処理で用いるのでない限り、任意の順序で実現しうることに留意すべきである。請求の範囲、明細書、および図面中の動作フローに関して、便宜上「まず、」、「次に、」などを用いて説明したとしても、この順で実施することが必須であることを意味するものではない。 The execution order of each process such as operations, procedures, steps, and stages in the apparatus, system, program, and method shown in the claims, the description, and the drawings is particularly “before” or “prior”. It should be noted that it can be realized in any order unless the output of the previous process is used in the subsequent process. Regarding the operation flow in the claims, the description, and the drawings, even if it is described using “first”, “next”, etc. for the sake of convenience, it means that it is essential to carry out in this order. is not.
10 ユーザ、20 通信網、30 会話アルゴリズムDB、32 実行条件テーブル、34 条件会話アルゴリズム、36 感情会話アルゴリズム、40 会話DB、42 Q&ADB、44 カテゴリ、45 優先度、46 発話内容、48 発話内容、52 カテゴリ名、54 優先度、100 情報端末、112 実行条件判断部、116 条件データ取得部、118 会話アルゴリズム選択部、120 会話実行部、122 音声出力制御部、124 音声取得部、126 カテゴリ選択部、128 感情認識部、130 優先度変更部、132 発話内容登録部、134 会話データ生成部、136 アルゴリズム共有処理部、200 サーバ 10 users, 20 communication networks, 30 conversation algorithm DB, 32 execution condition table, 34 conditional conversation algorithm, 36 emotion conversation algorithm, 40 conversation DB, 42 Q & ADB, 44 category, 45 priority, 46 utterance content, 48 utterance content, 52 Category name, 54 priority, 100 information terminal, 112 execution condition determination unit, 116 condition data acquisition unit, 118 conversation algorithm selection unit, 120 conversation execution unit, 122 voice output control unit, 124 voice acquisition unit, 126 category selection unit, 128 emotion recognition unit, 130 priority change unit, 132 utterance content registration unit, 134 conversation data generation unit, 136 algorithm sharing processing unit, 200 server

Claims (12)

  1.  ユーザの音声を取得する音声取得部と、
     前記音声取得部が取得した音声に基づいて前記ユーザの感情を認識する感情認識部と、
     複数の感情の種類のそれぞれに対応付けて第1会話アルゴリズムを記憶する第1会話アルゴリズム記憶部と、
     前記感情認識部が認識した感情に対応付けて前記第1会話アルゴリズム記憶部に記憶されている第1会話アルゴリズムを選択する会話アルゴリズム選択部と、
     前記会話アルゴリズム選択部が選択した第1会話アルゴリズムに従って、前記ユーザとの会話を実行する会話実行部と
    を備える会話処理システム。
    An audio acquisition unit for acquiring the user's audio;
    An emotion recognition unit that recognizes the user's emotion based on the voice acquired by the voice acquisition unit;
    A first conversation algorithm storage unit that stores a first conversation algorithm in association with each of a plurality of emotion types;
    A conversation algorithm selection unit that selects a first conversation algorithm stored in the first conversation algorithm storage unit in association with the emotion recognized by the emotion recognition unit;
    A conversation processing system comprising: a conversation execution unit that executes a conversation with the user according to a first conversation algorithm selected by the conversation algorithm selection unit.
  2.  複数の第2会話アルゴリズムのそれぞれに対応付けて実行条件を記憶する第2会話アルゴリズム記憶部と、
     前記第2会話アルゴリズム記憶部が記憶している複数の実行条件のいずれかが満たされたことを判断する実行条件判断部と
    をさらに備え、
     前記会話アルゴリズム選択部は、前記実行条件判断部によって満たされたことが判断された実行条件に対応付けて前記第2会話アルゴリズム記憶部に記憶されている第2会話アルゴリズムを選択し、
     前記会話実行部は、前記会話アルゴリズム選択部が選択した第2会話アルゴリズムに従って、ユーザとの会話を実行する請求項1に記載の会話処理システム。
    A second conversation algorithm storage unit that stores execution conditions in association with each of the plurality of second conversation algorithms;
    An execution condition determination unit that determines that any one of the plurality of execution conditions stored in the second conversation algorithm storage unit is satisfied;
    The conversation algorithm selection unit selects a second conversation algorithm stored in the second conversation algorithm storage unit in association with the execution condition determined to be satisfied by the execution condition determination unit;
    The conversation processing system according to claim 1, wherein the conversation execution unit executes a conversation with a user according to a second conversation algorithm selected by the conversation algorithm selection unit.
  3.  前記音声取得部は、前記会話実行部が第2会話アルゴリズムに従って前記ユーザとの会話を実行している間に、前記ユーザの音声を取得し、
     前記会話実行部は、前記会話アルゴリズム選択部が選択した第2会話アルゴリズムに従った会話を中断し、前記会話アルゴリズム選択部が選択した第1会話アルゴリズムに従った会話を開始する、請求項2に記載の会話処理システム。
    The voice acquisition unit acquires the voice of the user while the conversation execution unit is executing a conversation with the user according to a second conversation algorithm,
    The conversation execution unit interrupts the conversation according to the second conversation algorithm selected by the conversation algorithm selection unit, and starts the conversation according to the first conversation algorithm selected by the conversation algorithm selection unit. The conversation processing system described.
  4.  前記第2会話アルゴリズム記憶部は、前記複数の第2会話アルゴリズムのそれぞれに対応付けて優先度をさらに記憶し、
     前記感情認識部は、前記会話アルゴリズム選択部が選択した第2会話アルゴリズムに従って前記会話実行部が前記ユーザとの会話を実行している間に前記音声取得部が取得した音声に基づいて前記ユーザの感情を認識し、
     前記会話処理システムは、
     前記感情認識部が認識した前記ユーザの感情に基づいて、前記会話アルゴリズム選択部が選択した第2会話アルゴリズムに対応付けて記憶されている前記優先度を変更する第1優先度変更部
    をさらに備える、請求項2又は3に記載の会話処理システム。
    The second conversation algorithm storage unit further stores a priority in association with each of the plurality of second conversation algorithms;
    The emotion recognizing unit is based on the voice acquired by the voice acquisition unit while the conversation execution unit is executing a conversation with the user according to the second conversation algorithm selected by the conversation algorithm selection unit. Recognize emotions,
    The conversation processing system includes:
    A first priority changing unit that changes the priority stored in association with the second conversation algorithm selected by the conversation algorithm selection unit based on the emotion of the user recognized by the emotion recognition unit. The conversation processing system according to claim 2 or 3.
  5.  複数のカテゴリのそれぞれに対応付けて複数の発話内容を記憶する発話内容記憶部と、
     前記音声取得部が取得した前記音声に基づいて、前記複数のカテゴリのうちの一のカテゴリを選択状態にするカテゴリ選択部と
    をさらに備え、
     前記会話実行部は、前記カテゴリ選択部が選択状態にした前記一のカテゴリに対応付けて記憶された複数の発話内容によって、前記ユーザとの会話を実行し、
     前記音声取得部は、前記会話実行部が前記一のカテゴリに対応付けて記憶された複数の発話内容により前記ユーザとの会話を実行している間に、前記ユーザの音声を取得し、
     前記会話実行部は、前記一のカテゴリに対応付けて記憶された複数の発話内容による前記ユーザとの会話を中断し、前記会話アルゴリズム選択部が選択した前記第1会話アルゴリズムに従った会話を開始する、請求項1から4のいずれか1項に記載の会話処理システム。
    An utterance content storage unit that stores a plurality of utterance contents in association with each of a plurality of categories
    A category selection unit that selects one of the plurality of categories based on the voice acquired by the voice acquisition unit;
    The conversation execution unit executes a conversation with the user according to a plurality of utterance contents stored in association with the one category selected by the category selection unit,
    The voice acquisition unit acquires the voice of the user while the conversation execution unit is executing a conversation with the user based on a plurality of utterance contents stored in association with the one category,
    The conversation execution unit interrupts a conversation with the user based on a plurality of utterance contents stored in association with the one category, and starts a conversation according to the first conversation algorithm selected by the conversation algorithm selection unit. The conversation processing system according to any one of claims 1 to 4.
  6.  前記発話内容記憶部は、前記複数のカテゴリのそれぞれに対応付けて優先度をさらに記憶し、
     前記感情認識部は、前記一のカテゴリに対応付けて記憶された複数の発話内容によって前記会話実行部が前記ユーザとの会話を実行している間に前記音声取得部が取得した音声に基づいて前記ユーザの感情を認識し、
     前記会話処理システムは、
     前記感情認識部が認識した前記ユーザの感情に基づいて、前記一のカテゴリに対応付けて記憶されている前記優先度を変更する第2優先度変更部
    を更に備える、請求項5に記載の会話処理システム。
    The utterance content storage unit further stores a priority in association with each of the plurality of categories,
    The emotion recognition unit is based on the voice acquired by the voice acquisition unit while the conversation execution unit is executing a conversation with the user based on a plurality of utterance contents stored in association with the one category. Recognizing the user's emotions,
    The conversation processing system includes:
    The conversation according to claim 5, further comprising a second priority changing unit that changes the priority stored in association with the one category based on the emotion of the user recognized by the emotion recognition unit. Processing system.
  7.  前記カテゴリ選択部は、選択状態にした前記一のカテゴリに対応付けて記憶された複数の発話内容によって前記会話実行部が実行している前記ユーザとの会話を切り替える場合に、前記優先度に基づいて、前記一のカテゴリ以外のカテゴリを選択状態にし、
     前記会話実行部は、前記カテゴリ選択部により選択状態にされた前記一のカテゴリ以外のカテゴリに対応付けて前記発話内容記憶部に記憶された複数の発話内容によって、前記ユーザとの会話を開始する、請求項6に記載の会話処理システム。
    The category selection unit is based on the priority when switching the conversation with the user being executed by the conversation execution unit according to a plurality of utterance contents stored in association with the one category selected. To select a category other than the one category,
    The conversation execution unit starts a conversation with the user based on a plurality of utterance contents stored in the utterance content storage unit in association with a category other than the one category selected by the category selection unit. The conversation processing system according to claim 6.
  8.  前記複数のカテゴリは階層構造を有し、
     前記カテゴリ選択部は、選択状態にした前記一のカテゴリに対応付けて記憶された複数の発話内容によって前記会話実行部が実行している前記ユーザとの会話を切り替える場合に、前記一のカテゴリに隣接するカテゴリを選択状態にし、
     前記会話実行部は、前記カテゴリ選択部により選択状態にされた前記一のカテゴリに隣接するカテゴリに対応付けて前記発話内容記憶部に記憶された複数の発話内容によって、前記ユーザとの会話を開始する、請求項6に記載の会話処理システム。
    The plurality of categories have a hierarchical structure;
    The category selection unit switches to the one category when switching conversations with the user being executed by the conversation execution unit according to a plurality of utterance contents stored in association with the one category selected. Select the adjacent category,
    The conversation execution unit starts a conversation with the user based on a plurality of utterance contents stored in the utterance content storage unit in association with a category adjacent to the one category selected by the category selection unit. The conversation processing system according to claim 6.
  9.  前記カテゴリ選択部が前記一のカテゴリを選択状態にした後に、前記音声取得部が取得した前記音声を、選択状態の前記一のカテゴリに前記発話内容として対応づけて前記発話内容記憶部に登録する発話内容登録部
    をさらに備える、請求項5から8のいずれか1項に記載の会話処理システム。
    After the category selection unit selects the one category, the voice acquired by the voice acquisition unit is associated with the one category in the selection state as the utterance content and is registered in the utterance content storage unit. The conversation processing system according to any one of claims 5 to 8, further comprising an utterance content registration unit.
  10.  前記カテゴリ選択部によって選択状態にされた前記一のカテゴリに対応付けて記憶された複数の発話内容に含まれる第1発話内容を音声出力させる音声出力制御部と、
     前記第1発話内容と、前記第1発話内容に対するユーザ10の第1応答内容とを含む会話データを生成する会話データ生成部と
     をさらに備え、
     前記会話実行部は、前記会話データ生成部が前記第1発話内容と前記第1応答内容とを含む前記会話データを生成した後に、前記音声取得部が前記第1発話内容に一致する音声を取得した場合に、前記第1応答内容を前記音声出力し、
     前記音声取得部は、前記会話実行部が音声出力した前記第1応答内容に対する第2応答内容を取得し、
     前記会話データ生成部は、前記第1発話内容、前記第1応答内容、前記第2応答内容の順で対応付けた会話データを生成する、請求項5から9のいずれか1項に記載の会話処理システム。
    A voice output control unit that outputs a first utterance content included in a plurality of utterance contents stored in association with the one category selected by the category selection unit;
    A conversation data generating unit that generates conversation data including the first utterance content and the first response content of the user 10 with respect to the first utterance content;
    The conversation execution unit acquires the voice that matches the first utterance content after the conversation data generation unit generates the conversation data including the first utterance content and the first response content. If so, the first response content is output as voice,
    The voice acquisition unit acquires a second response content with respect to the first response content output by the conversation execution unit as a voice,
    The conversation according to any one of claims 5 to 9, wherein the conversation data generation unit generates conversation data associated in the order of the first utterance content, the first response content, and the second response content. Processing system.
  11.  第1のユーザが所持する第1情報端末の前記第2会話アルゴリズム記憶部が記憶する第2会話アルゴリズムを、前記第1のユーザが有するプロファイルとの類似度が予め定められた基準を超えているプロファイルを有する第2のユーザが所持する第2情報端末の前記第2会話アルゴリズム記憶部にコピーするアルゴリズム共有処理部
    をさらに備える、請求項2に記載の会話処理システム。
    The similarity with the profile of the first user that exceeds the second conversation algorithm stored in the second conversation algorithm storage unit of the first information terminal possessed by the first user exceeds a predetermined standard. The conversation processing system according to claim 2, further comprising an algorithm sharing processing unit that copies to the second conversation algorithm storage unit of a second information terminal possessed by a second user having a profile.
  12.  コンピュータを、請求項1から11のいずれか1項に記載の会話処理システムとして機能させるためのプログラム。 A program for causing a computer to function as the conversation processing system according to any one of claims 1 to 11.
PCT/JP2014/002348 2013-05-09 2014-04-25 Conversation processing system and program WO2014181524A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2013099328A JP2014219594A (en) 2013-05-09 2013-05-09 Conversation processing system and program
JP2013-099328 2013-05-09

Publications (1)

Publication Number Publication Date
WO2014181524A1 true WO2014181524A1 (en) 2014-11-13

Family

ID=51867015

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2014/002348 WO2014181524A1 (en) 2013-05-09 2014-04-25 Conversation processing system and program

Country Status (2)

Country Link
JP (1) JP2014219594A (en)
WO (1) WO2014181524A1 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2016110151A (en) * 2014-12-04 2016-06-20 悠之介 北 Voice management server device, conversation voice provision method, and conversation voice provision system
CN107316654A (en) * 2017-07-24 2017-11-03 湖南大学 Emotion identification method based on DIS NV features
CN109147824A (en) * 2017-06-23 2019-01-04 卡西欧计算机株式会社 Electronic equipment, emotional information obtain system and adquisitiones and storage medium
CN110737761A (en) * 2019-09-26 2020-01-31 联想(北京)有限公司 information processing method, electronic equipment and storage medium
US10748644B2 (en) 2018-06-19 2020-08-18 Ellipsis Health, Inc. Systems and methods for mental health assessment
US11120895B2 (en) 2018-06-19 2021-09-14 Ellipsis Health, Inc. Systems and methods for mental health assessment

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6122816B2 (en) 2014-08-07 2017-04-26 シャープ株式会社 Audio output device, network system, audio output method, and audio output program
JP6572969B2 (en) * 2015-03-30 2019-09-11 富士通クライアントコンピューティング株式会社 Speech recognition apparatus, speech recognition system, and program
JP6601069B2 (en) 2015-09-01 2019-11-06 カシオ計算機株式会社 Dialog control apparatus, dialog control method, and program
JP2018021987A (en) * 2016-08-02 2018-02-08 ユニロボット株式会社 Conversation processing device and program
JP2018041230A (en) * 2016-09-06 2018-03-15 富士通株式会社 Reception support program, reception support method, reception support system and information processor
DE112017006878T5 (en) 2017-01-20 2019-10-24 Honda Motor Co., Ltd. DIALOG PROCESSING SERVER, CONTROL PROCEDURE FOR A DIALOG PROCESSING SERVER AND END DEVICE
KR102463581B1 (en) * 2017-12-05 2022-11-07 현대자동차주식회사 Dialogue processing apparatus, vehicle having the same

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002169804A (en) * 2000-12-01 2002-06-14 Namco Ltd System and method for simulated conversation, and information storage medium
JP2003329477A (en) * 2002-05-15 2003-11-19 Pioneer Electronic Corp Navigation device and interactive information providing program
JP2004109323A (en) * 2002-09-17 2004-04-08 Denso Corp Voice interaction apparatus and program
JP2004513445A (en) * 2000-10-30 2004-04-30 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ User interface / entertainment device that simulates personal interaction and responds to the user's emotional state and / or personality
JP2006171709A (en) * 2004-11-17 2006-06-29 Denso Corp Voice interactive apparatus and speech interactive method
JP2006178063A (en) * 2004-12-21 2006-07-06 Toyota Central Res & Dev Lab Inc Interactive processing device
JP2006201870A (en) * 2005-01-18 2006-08-03 Toyota Central Res & Dev Lab Inc Interactive processor
JP2009064186A (en) * 2007-09-05 2009-03-26 Mazda Motor Corp Interactive system for vehicle
JP2011108055A (en) * 2009-11-19 2011-06-02 Nippon Telegr & Teleph Corp <Ntt> Interactive system, interactive method, and interactive program

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3002136B2 (en) * 1996-08-06 2000-01-24 株式会社エイ・ティ・アール知能映像通信研究所 Emotion conversion device
JP2001215993A (en) * 2000-01-31 2001-08-10 Sony Corp Device and method for interactive processing and recording medium
JP2002123289A (en) * 2000-10-13 2002-04-26 Matsushita Electric Ind Co Ltd Voice interactive device
JP2004021121A (en) * 2002-06-19 2004-01-22 Nec Corp Voice interaction controller unit
JP2006039120A (en) * 2004-07-26 2006-02-09 Sony Corp Interactive device and interactive method, program and recording medium
JP4456537B2 (en) * 2004-09-14 2010-04-28 本田技研工業株式会社 Information transmission device
JP2009134008A (en) * 2007-11-29 2009-06-18 Toyota Central R&D Labs Inc Emotional answer generation device and emotional answer generation program

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004513445A (en) * 2000-10-30 2004-04-30 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ User interface / entertainment device that simulates personal interaction and responds to the user's emotional state and / or personality
JP2002169804A (en) * 2000-12-01 2002-06-14 Namco Ltd System and method for simulated conversation, and information storage medium
JP2003329477A (en) * 2002-05-15 2003-11-19 Pioneer Electronic Corp Navigation device and interactive information providing program
JP2004109323A (en) * 2002-09-17 2004-04-08 Denso Corp Voice interaction apparatus and program
JP2006171709A (en) * 2004-11-17 2006-06-29 Denso Corp Voice interactive apparatus and speech interactive method
JP2006178063A (en) * 2004-12-21 2006-07-06 Toyota Central Res & Dev Lab Inc Interactive processing device
JP2006201870A (en) * 2005-01-18 2006-08-03 Toyota Central Res & Dev Lab Inc Interactive processor
JP2009064186A (en) * 2007-09-05 2009-03-26 Mazda Motor Corp Interactive system for vehicle
JP2011108055A (en) * 2009-11-19 2011-06-02 Nippon Telegr & Teleph Corp <Ntt> Interactive system, interactive method, and interactive program

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2016110151A (en) * 2014-12-04 2016-06-20 悠之介 北 Voice management server device, conversation voice provision method, and conversation voice provision system
CN109147824A (en) * 2017-06-23 2019-01-04 卡西欧计算机株式会社 Electronic equipment, emotional information obtain system and adquisitiones and storage medium
CN107316654A (en) * 2017-07-24 2017-11-03 湖南大学 Emotion identification method based on DIS NV features
US10748644B2 (en) 2018-06-19 2020-08-18 Ellipsis Health, Inc. Systems and methods for mental health assessment
US11120895B2 (en) 2018-06-19 2021-09-14 Ellipsis Health, Inc. Systems and methods for mental health assessment
US11942194B2 (en) 2018-06-19 2024-03-26 Ellipsis Health, Inc. Systems and methods for mental health assessment
CN110737761A (en) * 2019-09-26 2020-01-31 联想(北京)有限公司 information processing method, electronic equipment and storage medium
CN110737761B (en) * 2019-09-26 2023-09-19 联想(北京)有限公司 Information processing method, electronic equipment and storage medium

Also Published As

Publication number Publication date
JP2014219594A (en) 2014-11-20

Similar Documents

Publication Publication Date Title
WO2014181524A1 (en) Conversation processing system and program
CN107340991B (en) Voice role switching method, device, equipment and storage medium
CN108962217B (en) Speech synthesis method and related equipment
US10540970B2 (en) Architectures and topologies for vehicle-based, voice-controlled devices
US20150348538A1 (en) Speech summary and action item generation
WO2017200072A1 (en) Dialog method, dialog system, dialog device, and program
CN109346076A (en) Interactive voice, method of speech processing, device and system
US11267121B2 (en) Conversation output system, conversation output method, and non-transitory recording medium
JP2014191030A (en) Voice recognition terminal and voice recognition method using computer terminal
WO2017200076A1 (en) Dialog method, dialog system, dialog device, and program
CN107393529A (en) Audio recognition method, device, terminal and computer-readable recording medium
CN109377979B (en) Method and system for updating welcome language
US10629199B1 (en) Architectures and topologies for vehicle-based, voice-controlled devices
JP2014167517A (en) Conversation providing system, game providing system, conversation providing method, game providing method, and program
JP7026004B2 (en) Conversation aids, conversation aid methods and programs
CN113314104A (en) Interactive object driving and phoneme processing method, device, equipment and storage medium
CN110524547B (en) Session device, robot, session device control method, and storage medium
KR102063389B1 (en) Character display device based the artificial intelligent and the display method thereof
CN114391165A (en) Voice information processing method, device, equipment and storage medium
US20220161131A1 (en) Systems and devices for controlling network applications
CN115527542A (en) Design method and device of vehicle-mounted voice assistant, terminal equipment and storage medium
WO2021196647A1 (en) Method and apparatus for driving interactive object, device, and storage medium
CN111095397A (en) Natural language data generation system and method
US20200175988A1 (en) Information providing method and information providing apparatus
KR20210015977A (en) Apparatus for realizing coversation with died person

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14794795

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 14794795

Country of ref document: EP

Kind code of ref document: A1