WO2016175354A1 - Artificial intelligence conversation device and method - Google Patents

Artificial intelligence conversation device and method Download PDF

Info

Publication number
WO2016175354A1
WO2016175354A1 PCT/KR2015/004347 KR2015004347W WO2016175354A1 WO 2016175354 A1 WO2016175354 A1 WO 2016175354A1 KR 2015004347 W KR2015004347 W KR 2015004347W WO 2016175354 A1 WO2016175354 A1 WO 2016175354A1
Authority
WO
WIPO (PCT)
Prior art keywords
response
voice
user
conversation
question
Prior art date
Application number
PCT/KR2015/004347
Other languages
French (fr)
Korean (ko)
Inventor
이영근
김승곤
임완섭
임성환
김우현
이영호
김두호
Original Assignee
주식회사 아카인텔리전스
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 주식회사 아카인텔리전스 filed Critical 주식회사 아카인텔리전스
Priority to PCT/KR2015/004347 priority Critical patent/WO2016175354A1/en
Publication of WO2016175354A1 publication Critical patent/WO2016175354A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems

Definitions

  • the present invention relates to an artificial intelligence dialogue apparatus and method for supporting a dialogue between a person and a robot.
  • Chat is a computer or a portable terminal to support the conversation with the other party over the network, it is widely used in the form of chat window messenger.
  • chat robot technologies As the necessity increases as a means of communication using natural language between humans and computers (robots) in intelligent agents, various chat robot technologies have been proposed.
  • the conventional technology simply answers based on the user's input.
  • the passive dialog engine providing only provides a user with not only a lot of heterogeneity, but also requires the user to induce a conversation, the flow of the conversation and the user's interest in the conversation are sharply dropped.
  • An object of the present invention is to provide an artificial intelligence device and method for supporting a natural conversation with a user without departing from the present invention.
  • an artificial intelligence dialog device may include an input response analysis unit analyzing an input user response, and selecting at least one response scenario among preset scenarios according to an analysis result to respond to a response and a question about a user response. And an output unit for outputting a response control unit for outputting an output command and a silent or conversation start voice, and outputting a response voice and a question voice according to the output command of the reaction control unit.
  • the artificial intelligence dialogue apparatus and method according to the present invention actively proceeds a conversation in the order of question transmission, user response reception, and response response to a user response based on a preset scenario, thereby providing only a predetermined answer according to a user input. Rather, it leads to active conversations, thereby minimizing the heterogeneity of the user by conducting conversations with the conversation engine, and enhancing the interest of the conversation.
  • FIG. 1 is a block diagram illustrating an artificial intelligence conversation apparatus according to an embodiment of the present invention.
  • FIG. 2 is a flowchart illustrating an artificial intelligence conversation method according to an embodiment of the present invention.
  • FIG. 1 is a block diagram illustrating an artificial intelligence conversation apparatus according to an embodiment of the present invention.
  • Artificial intelligence communication apparatus includes an input unit 100 for receiving a voice from the user utterance, Speech To Text (STT) unit 200 for converting the voice received by the input unit 100 into text, Input response analysis unit 300 for receiving the STT conversion result and analyzing the user response, and at least one response scenario selected from the preset scenarios according to the analysis result, and outputs the response command for the user response and the question And a output unit 500 for outputting a response control unit 400 for transmitting a silent or conversation start voice and outputting a response voice and a question voice according to an output command of the reaction control unit 400.
  • STT Speech To Text
  • Input unit 100 receives the user's voice through the microphone of the artificial intelligence chat device.
  • the artificial intelligence dialogue apparatus performs the output of the question to the user, the input of the answer from the user, the output of the response to the response to the user and the output of the next question according to the response output to the user in order.
  • the question according to an embodiment of the present invention is provided at the beginning of a conversation with a user and is expressed by a conversation start voice.
  • the conversation start voice is the first question provided through the output unit, or when the silent sound is output, the conversation proceeds in the order of the response voice and the question voice output from the user's first voice input.
  • one conversation Supports natural dialogue between the user and the AI dialog within the subject.
  • the output unit 500 outputs a conversation start voice, which is a question voice for starting a conversation, based on the application execution environment information before starting the conversation.
  • the first embodiment outputs a conversation start voice, which is a question voice, and the conversation proceeds in the order of the user's answer, response voice, and question voice output.
  • a conversation start voice which is a question voice
  • the conversation proceeds in the order of the user's answer, response voice, and question voice output.
  • silence is output
  • the voice input of is the starting point of the conversation and the conversation proceeds in the order of response voice and question voice output.
  • the application execution environment information may be at least one of a built-in scenario database, a user's personal information, a user's behavior pattern, a record of a previous conversation, and surrounding environment information. If the record of previous conversations is about a company's project, as the application runs, it outputs the question, "How did the project go today?"
  • the output unit 500 is a question to start a conversation, not a question related to the company, "good weekend. Is the weather good? ”
  • the output unit 500 does not only provide only a predetermined answer based on a user's input, but also provides a user with a question of an appropriate topic as the application is executed, thereby providing a conversation. It's a natural way to start and provide a customized conversation.
  • reaction control unit 400 not only commands to select and output a response and a question for a user response from a pre-stored list, but also generates a new response and a question for the user response. It is also possible to print.
  • the input unit 100 receives a user's voice input in response to a conversation start voice output or after a silent output, and the STT unit 200 inputs a result of converting the voice of the user into a string.
  • the answer analysis unit 300 is provided.
  • the input response analysis unit 300 analyzes whether or not the user answer converted into a character string corresponds to a predetermined answer type.
  • the structured scenario database stores and manages the selected answers, the general answers, the answers that you want to repeat, and the unrelated answers by their types.
  • the selective answer is a type in which the classification according to the answer selection is clearly defined in the user's answer to the question, and examples thereof include positive / negative, spring / summer / fall / winter, and the like.
  • the general answer is a type of ambiguity and multiple choices for a question, such as the answer to the question “What kind of exercise do you like?”.
  • Repetition is the type that the desired answer corresponds to the answer you want the output to ask again.
  • the output unit 500 re-outputs the question immediately output.
  • the artificial intelligence dialog device is to modify the scenario based on the user's answer to extract and provide the answers and questions sequentially or to query the user again the question of the category corresponding to the first question It is also possible.
  • the input response analysis unit 300 determines which category of a predetermined category the input user's answer belongs to, and displays the result. For example, when it corresponds to the optional answer type, it is determined whether the user response input through sentence analysis is affirmative or negative for the question.
  • the reaction controller 400 selects at least one reaction scenario from among scenarios stored in the instrumented scenario database according to the analysis result, and responds based on a reaction scenario according to a subject to which a user answer is applicable. And generate an output command for the question.
  • the user can perform a natural conversation with the AI conversation device corresponding to his answer without any dissatisfaction.
  • the response control unit 400 transmits a pause command signal to the output unit 500 when new voice data is received from the user during output of the response voice and the question voice of the output unit 500, and then input response analysis unit 300. Resample the responses and questions according to the analysis results according to the new voice data of).
  • reaction scenario selected by the reaction controller 400 is determined and modified in real time according to the user's response or the user's comment, and the reaction scenario is appropriately based on the instrumented scenario database. It is corrected.
  • the reaction control unit 400 changes the scenario (eg, “Would your friend get married? Where is the marriage ceremony?” Asking the first question on the topic and entering a specific event called wedding attendance). To continue the conversation).
  • the output unit 500 outputs text corresponding to the response voice and the question voice through the screen. Accordingly, even in a noisy environment in which the user cannot properly receive a voice from the output unit 500, the user may recognize the response and the question from the text output through the screen, and continue the conversation by uttering the answer. .
  • FIG. 2 is a flowchart illustrating an artificial intelligence conversation method according to an embodiment of the present invention.
  • an AI conversation method includes outputting a conversation start voice (including silence), receiving a user response (S200), analyzing a user response, and analyzing the result. Selecting at least one reaction scenario from among predetermined scenarios, extracting a response and a question based on the response scenario, and outputting a response voice and a question voice according to the extracted response and question (S400) do.
  • step S100 is a step of outputting a question voice or a silence corresponding to a conversation start question to start a conversation, and includes a structured scenario database, a user's personal information, a user's behavior pattern, and a previous conversation.
  • the conversation start question is extracted according to the application execution environment information which is at least one of the recording information, and the output information.
  • the dialogue is performed in response to the user's answer input by voice in the step S200 and in the order of questions.
  • Step S200 converts a user's answer input by voice into text, and provides a sentence for analyzing a user's answer.
  • Step S300 performs analysis by determining which type of response type the user answer corresponds to.
  • the dialogue is performed in the order of the first question, the user's answer, the response to the user's answer, and the question according to the response (when silence is output, the user's voice input, the response to the user's voice, the question according to the response)
  • preset answer types e.g., optional, general, answer you want to repeat, unrelated answer
  • the step S400 outputs the text corresponding to the response voice and the question voice through the screen, thereby providing the user with the response text and the question text visually as well as hearing, thereby supporting more accurate recognition of the user. .
  • step S600 is a step of determining whether the conversation ends with a criterion.
  • the user is confirmed to say goodbye for a predetermined time, if the user does not answer for a predetermined time or longer, the user for a predetermined time or more. If there is no answer, and if the end criterion, such as when there is no user's reply to the voice of the output unit calling the user, the conversation is terminated. From step S200 to step S500 are repeated.

Abstract

The present invention relates to an artificial intelligence conversation device and method which supports a conversation between a human and a robot. An artificial intelligence conversation device according to an aspect of the present invention comprises: an input answer analysis unit which analyzes an answer input from a user; a response control unit which selects at least one response scenario among predetermined scenarios according to the result of the analysis and transmits an output command for a response and a question to the user's answer; and an output unit which outputs a silence or a conversation start voice and outputs a response voice and a question voice according to the output command from the response control unit.

Description

인공지능 대화 장치 및 방법Artificial Intelligence Apparatus and Method
본 발명은 사람 대 로봇 간의 대화를 지원하는 인공지능 대화 장치 및 방법에 관한 것이다. The present invention relates to an artificial intelligence dialogue apparatus and method for supporting a dialogue between a person and a robot.
채팅은 컴퓨터 또는 휴대용 단말기 등을 이용하여 네트워크를 통해 상대방과 대화를 할 수 있도록 지원하는 것으로, 대화창 메신저의 형태로 널리 이용되고 있다. Chat is a computer or a portable terminal to support the conversation with the other party over the network, it is widely used in the form of chat window messenger.
그러나, 사람과 사람 사이의 채팅에서는 상대방이 없는 경우에는 채팅을 행할 수 없어, 채팅 로봇이 탄생하게 되었다. However, in a chat between a person and a person, when there is no other party, a chat cannot be performed, resulting in a chat robot.
이는 지능형 에이전트에서 사람 대 컴퓨터(로봇)간의 자연어를 이용한 의사소통 수단으로서 그 필요성이 증대됨에 따라, 다양한 채팅 로봇 기술이 제안되었다. As the necessity increases as a means of communication using natural language between humans and computers (robots) in intelligent agents, various chat robot technologies have been proposed.
종래 기술에 따른 대화 엔진의 경우, 사용자가 입력한 텍스트에 따라 기설정된 해당 답변이 제공되는 바, 사용자의 입력에 따라 대화의 주제가 급변한다. In the case of the conversation engine according to the related art, a corresponding answer preset according to the text input by the user is provided, and thus the subject of the conversation changes rapidly according to the user's input.
사람 대 로봇 간의 자연어를 이용한 대화 엔진은 사용자에게 로봇과 대화하는 이질감을 최소화시키고, 자연스러운 대화가 이루어지도록 지원하는 것이 가장 핵심적 요소라 할 수 있음에도 불구하고, 종래 기술에 따르면 사용자의 입력에 따라 단순히 답변만을 제공하는 수동형 대화 엔진이 제공됨에 따라 사용자는 이질감을 많이 느낄 뿐 아니라, 사용자가 대화를 유도하여야 하여 대화의 흐름 및 사용자의 대화에 대한 흥미도가 급격히 떨어지는 문제점이 있다. Despite the fact that the dialogue engine using natural language between humans and robots can be said to be the most important factor to minimize the heterogeneity of dialogue with the robot and to enable natural dialogue, the conventional technology simply answers based on the user's input. As the passive dialog engine providing only provides a user with not only a lot of heterogeneity, but also requires the user to induce a conversation, the flow of the conversation and the user's interest in the conversation are sharply dropped.
본 발명은 전술한 문제점을 해결하기 위하여 제안된 것으로, 질문 송신, 답변 수신, 답변에 대한 반응 및 다음 질문 송신의 순서로 대화를 진행하여, 사용자의 답변에 맞춰 다음 대화로 이어지도록 유도함으로써, 주제에서 벗어나지 않고 사용자와의 자연스러운 대화가 이루어질 수 있도록 지원하는 인공지능 대화 장치 및 방법을 제공하는 데 목적이 있다. The present invention has been proposed to solve the above-described problems, by proceeding the dialogue in the order of sending a question, receiving an answer, responding to the answer and sending the next question, by inducing the user to lead to the next conversation in response to the subject, An object of the present invention is to provide an artificial intelligence device and method for supporting a natural conversation with a user without departing from the present invention.
본 발명의 일면에 따른 인공지능 대화 장치는 입력된 사용자 답변을 분석하는 입력답변 분석부와, 분석 결과에 따라 기설정된 시나리오 중 적어도 어느 하나의 반응시나리오를 선택하여, 사용자 답변에 대한 반응 및 질문에 대한 출력 명령을 전송하는 반응 제어부 및 무음 또는 대화 시작 음성을 출력하고, 반응 제어부의 출력 명령에 따라 반응 음성 및 질문 음성을 출력하는 출력부를 포함하는 것을 특징으로 한다. According to an aspect of an exemplary embodiment, an artificial intelligence dialog device may include an input response analysis unit analyzing an input user response, and selecting at least one response scenario among preset scenarios according to an analysis result to respond to a response and a question about a user response. And an output unit for outputting a response control unit for outputting an output command and a silent or conversation start voice, and outputting a response voice and a question voice according to the output command of the reaction control unit.
본 발명에 따른 인공지능 대화 장치 및 방법은 기설정된 시나리오 기반으로 질문 송신, 사용자 답변 수신, 사용자 답변에 대한 반응 송신의 순서로 대화를 능동적으로 진행시킴으로써, 사용자의 입력에 따른 기설정된 답변만을 제공하는 것이 아니라 능동형으로 대화를 이끌어 감으로써 대화 엔진과 대화를 수행함에 따른 사용자의 이질감을 최소화시키고, 대화의 흥미도를 높이는 효과가 있다. The artificial intelligence dialogue apparatus and method according to the present invention actively proceeds a conversation in the order of question transmission, user response reception, and response response to a user response based on a preset scenario, thereby providing only a predetermined answer according to a user input. Rather, it leads to active conversations, thereby minimizing the heterogeneity of the user by conducting conversations with the conversation engine, and enhancing the interest of the conversation.
사용자로부터 수신한 답변을 유형별로 분류하고, 답변에 속하는 컴포넌트를 집합으로 구성하여 사용자 답변에 대한 분석의 신뢰성을 높이고, 그에 따라 사용자 답변에 따른 반응을 적절히 제공하여 대화의 다음 순서로 유연히 진행시키는 것이 가능한 효과가 있다. It is necessary to classify the answers received from the users by type, to organize the components belonging to the answers to improve the reliability of the analysis of the user answers, and to provide the response according to the user's answers accordingly, so as to flexibly proceed to the next order of the conversation. There is a possible effect.
본 발명의 효과는 이상에서 언급한 것들에 한정되지 않으며, 언급되지 아니한 다른 효과들은 아래의 기재로부터 당업자에게 명확하게 이해될 수 있을 것이다.The effects of the present invention are not limited to those mentioned above, and other effects that are not mentioned will be clearly understood by those skilled in the art from the following description.
도 1은 본 발명의 실시예에 따른 인공지능 대화 장치를 나타내는 블록도이다. 1 is a block diagram illustrating an artificial intelligence conversation apparatus according to an embodiment of the present invention.
도 2는 본 발명의 실시예에 따른 인공지능 대화 방법을 나타내는 순서도이다.2 is a flowchart illustrating an artificial intelligence conversation method according to an embodiment of the present invention.
본 발명의 전술한 목적 및 그 이외의 목적과 이점 및 특징, 그리고 그것들을 달성하는 방법은 첨부되는 도면과 함께 상세하게 후술되어 있는 실시예들을 참조하면 명확해질 것이다. BRIEF DESCRIPTION OF THE DRAWINGS The above and other objects, advantages and features of the present invention, and methods of achieving them will be apparent with reference to the embodiments described below in detail with the accompanying drawings.
그러나 본 발명은 이하에서 개시되는 실시예들에 한정되는 것이 아니라 서로 다른 다양한 형태로 구현될 수 있으며, 단지 이하의 실시예들은 본 발명이 속하는 기술분야에서 통상의 지식을 가진 자에게 발명의 목적, 구성 및 효과를 용이하게 알려주기 위해 제공되는 것일 뿐으로서, 본 발명의 권리범위는 청구항의 기재에 의해 정의된다. However, the present invention is not limited to the embodiments disclosed below, but may be implemented in various forms, and only the following embodiments are provided to those skilled in the art to which the present invention pertains. It is merely provided to easily inform the configuration and effects, the scope of the present invention is defined by the description of the claims.
한편, 본 명세서에서 사용된 용어는 실시예들을 설명하기 위한 것이며 본 발명을 제한하고자 하는 것은 아니다. 본 명세서에서, 단수형은 문구에서 특별히 언급하지 않는 한 복수형도 포함한다. 명세서에서 사용되는 "포함한다(comprises)" 및/또는 "포함하는(comprising)"은 언급된 구성소자, 단계, 동작 및/또는 소자가 하나 이상의 다른 구성소자, 단계, 동작 및/또는 소자의 존재 또는 추가됨을 배제하지 않는다.Meanwhile, the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. In this specification, the singular also includes the plural unless specifically stated otherwise in the phrase. As used herein, “comprises” and / or “comprising” refers to the presence of one or more other components, steps, operations and / or devices in which the mentioned components, steps, operations and / or devices are described. Or does not exclude addition.
도 1은 본 발명의 실시예에 따른 인공지능 대화 장치를 나타내는 블록도이다. 1 is a block diagram illustrating an artificial intelligence conversation apparatus according to an embodiment of the present invention.
본 발명의 실시예에 따른 인공지능 대화 장치는 사용자 발화로부터 음성을 수신하는 입력부(100)와, 입력부(100)가 수신한 음성을 텍스트로 변환하는 STT(Speech To Text)부(200)와, STT 변환 결과를 수신하여 사용자 답변을 분석하는 입력답변 분석부(300)와, 분석 결과에 따라 기설정된 시나리오 중 적어도 어느 하나의 반응시나리오를 선택하여, 사용자 답변에 대한 반응 및 질문에 대한 출력 명령을 전송하는 반응 제어부(400) 및 무음 또는 대화 시작 음성을 출력하고, 반응 제어부(400)의 출력 명령에 따라 반응 음성 및 질문 음성을 출력하는 출력부(500)를 포함한다. Artificial intelligence communication apparatus according to an embodiment of the present invention includes an input unit 100 for receiving a voice from the user utterance, Speech To Text (STT) unit 200 for converting the voice received by the input unit 100 into text, Input response analysis unit 300 for receiving the STT conversion result and analyzing the user response, and at least one response scenario selected from the preset scenarios according to the analysis result, and outputs the response command for the user response and the question And a output unit 500 for outputting a response control unit 400 for transmitting a silent or conversation start voice and outputting a response voice and a question voice according to an output command of the reaction control unit 400.
본 발명의 실시예에 따른 입력부(100)는 인공지능 대화 장치의 마이크를 통해 사용자의 음성을 수신한다. Input unit 100 according to an embodiment of the present invention receives the user's voice through the microphone of the artificial intelligence chat device.
본 발명의 실시예에 따른 인공지능 대화 장치는 사용자에게 질문 출력, 사용자로부터 답변 입력, 사용자에게 답변에 대한 반응 출력 및 사용자에게 반응 출력에 따른 다음 질문 출력을 순서대로 수행한다. The artificial intelligence dialogue apparatus according to the embodiment of the present invention performs the output of the question to the user, the input of the answer from the user, the output of the response to the response to the user and the output of the next question according to the response output to the user in order.
본 발명의 실시예에 따른 질문은 사용자와 대화를 시작하는 단계에서 제공되며, 대화 시작 음성으로 표현된다. The question according to an embodiment of the present invention is provided at the beginning of a conversation with a user and is expressed by a conversation start voice.
이 때, 대화 시작 음성은 출력부를 통하여 제공되는 최초 질문이거나, 무음이 출력되는 경우에는 사용자의 최초 음성 입력으로부터 이에 대한 반응 음성 및 질문 음성 출력의 순서로 대화가 진행되도록 지원한다. At this time, the conversation start voice is the first question provided through the output unit, or when the silent sound is output, the conversation proceeds in the order of the response voice and the question voice output from the user's first voice input.
즉, 사용자로부터 입력된 질의어에 대하여 기설정된 매칭 답변만을 제공하는 것이 아니라, 기설정된 시나리오 기반으로 사용자에게 질문을 하고, 이에 대한 사용자 답변을 분석하여 그에 대한 반응 및 다음 질문을 출력함으로써, 하나의 대화 주제 안에서 사용자와 인공지능 대화 장치 간의 자연스러운 대화가 이루어질 수 있도록 지원한다.That is, by not only providing a preset matching answer to a query input from the user, but also asking a user a question based on a preset scenario, analyzing the user's response to this, and outputting a response and the next question, one conversation Supports natural dialogue between the user and the AI dialog within the subject.
본 발명의 실시예에 따른 출력부(500)는 대화를 시작하기에 앞서 어플리케이션 실행 환경 정보에 기초하여 대화를 시작하는 질문음성인 대화 시작 음성을 출력한다. The output unit 500 according to an embodiment of the present invention outputs a conversation start voice, which is a question voice for starting a conversation, based on the application execution environment information before starting the conversation.
이 때, 제1 실시예는 질문음성인 대화 시작 음성을 출력하는 것이고, 이에 대한 사용자의 답변, 반응 음성 및 질문 음성 출력의 순서로 대화가 진행되고, 제2 실시예는 무음이 출력되어, 사용자의 음성 입력이 대화의 시작 포인트가 되어 이에 대한 반응 음성 및 질문 음성 출력의 순서로 대화가 진행된다.At this time, the first embodiment outputs a conversation start voice, which is a question voice, and the conversation proceeds in the order of the user's answer, response voice, and question voice output. In the second embodiment, silence is output, The voice input of is the starting point of the conversation and the conversation proceeds in the order of response voice and question voice output.
이 때, 어플리케이션 실행 환경 정보는 기구축된 시나리오 데이터베이스, 사용자의 개인 정보, 사용자의 행동 패턴, 이전의 대화에 대한 기록 정보, 주변 환경 정보 중 적어도 어느 하나로써, 일례로 사용자가 어플리케이션 가입에 있어 회사원임을 입력하고, 이전의 대화에 대한 기록 정보가 회사의 프로젝트에 관한 것이라면, 어플리케이션을 실행함에 따라 “오늘 그 프로젝트는 잘 진행되었나요?”라는 식으로 질문 음성을 출력하게 된다. In this case, the application execution environment information may be at least one of a built-in scenario database, a user's personal information, a user's behavior pattern, a record of a previous conversation, and surrounding environment information. If the record of previous conversations is about a company's project, as the application runs, it outputs the question, "How did the project go today?"
또한, 어플리케이션 실행 환경 정보가 “주말”이고 날씨 정보가 “맑음”이라면, 출력부(500)는 대화를 시작하는 질문으로서 회사와 관련된 질문이 아닌, “좋은 주말이에요. 날씨가 좋죠?”라는 식으로 질문 음성을 출력하게 된다. Also, if the application execution environment information is "weekend" and the weather information is "sunny", the output unit 500 is a question to start a conversation, not a question related to the company, "good weekend. Is the weather good? ”
즉, 본 발명의 실시예에 따른 출력부(500)는 사용자의 입력에 따라 기설정된 답변만을 단편적으로 제공하는 것에 그치지 아니하고, 어플리케이션이 실행됨에 따라 사용자에게 적절한 주제의 질문을 먼저 제공함으로써, 대화를 자연스럽게 시작하고 사용자 맞춤형 대화를 제공하는 효과가 있다. That is, the output unit 500 according to an embodiment of the present invention does not only provide only a predetermined answer based on a user's input, but also provides a user with a question of an appropriate topic as the application is executed, thereby providing a conversation. It's a natural way to start and provide a customized conversation.
또한, 본 발명의 실시예에 따른 반응 제어부(400)는 사용자 답변에 대한 반응 및 질문을 기저장된 목록 중 하나를 선택하여 출력하도록 명령하는 것 뿐 아니라, 사용자 답변에 대하여 새로운 반응 및 질문을 생성하여 출력하는 것 역시 가능하다.In addition, the reaction control unit 400 according to an embodiment of the present invention not only commands to select and output a response and a question for a user response from a pre-stored list, but also generates a new response and a question for the user response. It is also possible to print.
본 발명의 실시예에 따른 입력부(100)는 대화 시작 음성 출력에 반응하여 또는 무음 출력 후에 입력된 사용자의 음성을 입력받고, STT부(200)는 이러한 사용자의 음성을 문자열로 전환한 결과를 입력답변 분석부(300)로 제공한다. The input unit 100 according to an embodiment of the present invention receives a user's voice input in response to a conversation start voice output or after a silent output, and the STT unit 200 inputs a result of converting the voice of the user into a string. The answer analysis unit 300 is provided.
입력답변 분석부(300)는 문자열로 전환된 사용자 답변이 기설정된 답변 유형 중 어느 유형에 해당되는지 여부를 판별하여 분석을 수행한다. The input response analysis unit 300 analyzes whether or not the user answer converted into a character string corresponds to a predetermined answer type.
본 발명의 실시예에 따른 기구축된 시나리오 데이터베이스는 답변의 유형으로 선택형 답변, 일반 답변, 반복을 원하는 답변, 관련없는 답변을 그 유형별로 분류하여 저장 관리한다. The structured scenario database according to the embodiment of the present invention stores and manages the selected answers, the general answers, the answers that you want to repeat, and the unrelated answers by their types.
선택형 답변은 질문에 대한 사용자의 답변에 있어서, 답변 선택에 따른 분류가 명확히 정해진 유형으로, 긍/부정, 봄/여름/가을/겨울 등을 그 예시로 들 수 있다. The selective answer is a type in which the classification according to the answer selection is clearly defined in the user's answer to the question, and examples thereof include positive / negative, spring / summer / fall / winter, and the like.
일반 답변은 선택형 답변과 달리, 질문에 대해 선택할 수 있는 분류가 불명확하고 다수인 경우의 유형으로, “너는 무슨 운동을 좋아해?”라는 질문에 대한 답변을 그 예시로 들 수 있다. The general answer, unlike the optional answer, is a type of ambiguity and multiple choices for a question, such as the answer to the question “What kind of exercise do you like?”.
반복은 원하는 답변은 출력부가 출력한 직전에 나온 질문을 다시 해주길 바라는 답변에 해당하는 유형이다. 이 때, 출력부(500)는 직전에 출력한 질문을 재출력한다. Repetition is the type that the desired answer corresponds to the answer you want the output to ask again. At this time, the output unit 500 re-outputs the question immediately output.
관련없는 답변은 질문에 대해 연관성이 없는 답변으로, “너는 무슨 운동을 좋아해?”라는 질문에 대하여 “오늘은 날씨가 좋아”라는 답변이 수신되는 유형이다. 이 때, 본 발명의 실시예에 따른 인공 지능 대화 장치는 사용자의 답변에 기반하여 시나리오를 수정하여 답변 및 질문을 추출하여 순차적으로 제공하거나, 최초 질문이 해당하는 카테고리의 질문을 다시 사용자에게 질의하는 것 역시 가능하다. An irrelevant answer is an unrelated answer to a question, in which the answer is "What weather do you like?" At this time, the artificial intelligence dialog device according to an embodiment of the present invention is to modify the scenario based on the user's answer to extract and provide the answers and questions sequentially or to query the user again the question of the category corresponding to the first question It is also possible.
입력답변 분석부(300)는 입력된 사용자의 답변이 미리 정해둔 분류 중 어느 분류에 속하는지 판단하고, 그 결과를 표시한다. 일례로, 선택형 답변 유형에 해당하는 경우, 문장 분석을 통해 입력된 사용자 답변이 질문에 대한 긍정문인지 부정문인지 판단한다. The input response analysis unit 300 determines which category of a predetermined category the input user's answer belongs to, and displays the result. For example, when it corresponds to the optional answer type, it is determined whether the user response input through sentence analysis is affirmative or negative for the question.
본 발명의 실시예에 따른 반응 제어부(400)는 분석 결과에 따라 기구축된 시나리오 데이터베이스에 저장된 시나리오 중 적어도 어느 하나의 반응시나리오를 선택하고, 사용자 답변이 해당되는 주제에 따른 반응시나리오에 기초하여 반응 및 질문에 대한 출력 명령을 생성한다. The reaction controller 400 according to an embodiment of the present invention selects at least one reaction scenario from among scenarios stored in the instrumented scenario database according to the analysis result, and responds based on a reaction scenario according to a subject to which a user answer is applicable. And generate an output command for the question.
사용자의 답변에 맞춰 다음 대화로 이어지도록, 시나리오 기반으로 반응 및 질문을 추출하여 제공함으로써, 사용자는 이질감 없이 자신의 답변에 대응하는 인공지능 대화 장치와 자연스러운 대화를 수행하는 것이 가능하다. By extracting and providing responses and questions based on scenarios so as to lead to the next conversation according to the user's answer, the user can perform a natural conversation with the AI conversation device corresponding to his answer without any dissatisfaction.
반응 제어부(400)는 출력부(500)의 반응 음성 및 질문 음성 출력 중, 사용자로부터 새로운 음성 데이터가 수신된 경우, 출력부(500)로 일시 정지 명령 신호를 전송하고, 입력답변 분석부(300)의 새로운 음성 데이터에 따른 분석 결과에 따라 반응 및 질문을 재추출한다. The response control unit 400 transmits a pause command signal to the output unit 500 when new voice data is received from the user during output of the response voice and the question voice of the output unit 500, and then input response analysis unit 300. Resample the responses and questions according to the analysis results according to the new voice data of).
즉, 본 발명의 실시예에 따른 반응 제어부(400)에서 선택한 반응시나리오는 사용자의 답변 또는 사용자의 코멘트에 따라서 실시간으로 그 신뢰성이 판단되어 변형되고, 기구축된 시나리오 데이터베이스에 기초하여 반응시나리오가 적절히 수정된다. That is, the reaction scenario selected by the reaction controller 400 according to an embodiment of the present invention is determined and modified in real time according to the user's response or the user's comment, and the reaction scenario is appropriately based on the instrumented scenario database. It is corrected.
예를 들어, 최초의 질문이 “회사 잘 다녀왔어요?”이고, 회사에서 있었던 일들에 대해서 사용자와 대화가 이루어지던 도중, 사용자의 답변이 “근데 이번 주말에 결혼식장에 가야 해”인 경우, 즉 주제를 변환하는 것에 해당되는 것으로 판단되면, 반응 제어부(400)는 시나리오를 변경(예: “친구분이 결혼하세요? 결혼 식장은 어디에요?”라는 해당 주제에 관한 최초 질문을 제공하며 결혼식 참석이라는 특정 이벤트에 대한 대화를 속행함)하여 대화를 진행한다. For example, if the first question is “Have you been in the company?” And while you were talking to you about what happened at the company, your answer was “But I have to go to the wedding this weekend.” If it is determined that the topic is to be transformed, the reaction control unit 400 changes the scenario (eg, “Would your friend get married? Where is the marriage ceremony?” Asking the first question on the topic and entering a specific event called wedding attendance). To continue the conversation).
본 발명의 실시예에 따른 출력부(500)는 반응 음성 및 질문 음성에 대응되는 텍스트를 화면을 통하여 출력한다. 이에 따라, 사용자는 출력부(500)로부터 음성을 제대로 수신하지 못하는 노이즈가 심한 환경에서도, 화면을 통하여 출력되는 텍스트로부터 반응 및 질문을 인지하고, 이에 대한 답변을 발화함으로써 대화를 지속하는 것이 가능하다. The output unit 500 according to an embodiment of the present invention outputs text corresponding to the response voice and the question voice through the screen. Accordingly, even in a noisy environment in which the user cannot properly receive a voice from the output unit 500, the user may recognize the response and the question from the text output through the screen, and continue the conversation by uttering the answer. .
도 2는 본 발명의 실시예에 따른 인공지능 대화 방법을 나타내는 순서도이다.2 is a flowchart illustrating an artificial intelligence conversation method according to an embodiment of the present invention.
본 발명의 실시예에 따른 인공지능 대화 방법은 대화 시작 음성(무음을 포함함)을 출력하는 단계(S100)와, 사용자 답변을 입력 받는 단계(S200)와, 사용자 답변을 분석하고, 그 결과에 따라 기설정된 시나리오 중 적어도 어느 하나의 반응시나리오를 선택하고, 이에 기초하여 반응 및 질문을 추출하는 단계(S300) 및 추출된 반응 및 질문에 따라 반응 음성 및 질문 음성을 출력하는 단계(S400)를 포함한다. According to an embodiment of the present invention, an AI conversation method includes outputting a conversation start voice (including silence), receiving a user response (S200), analyzing a user response, and analyzing the result. Selecting at least one reaction scenario from among predetermined scenarios, extracting a response and a question based on the response scenario, and outputting a response voice and a question voice according to the extracted response and question (S400) do.
본 발명의 실시예에 따른 S100 단계는 대화를 시작하는 대화 시작 질문에 해당하는 질문 음성 또는 무음을 출력하는 단계로서, 기구축된 시나리오 데이터베이스, 사용자의 개인정보, 사용자의 행동 패턴, 이전의 대화에 대한 기록 정보 중 적어도 어느 하나인 어플리케이션 실행 환경 정보에 따라 대화 시작 질문을 추출하고, 이를 출력한다. According to an embodiment of the present invention, step S100 is a step of outputting a question voice or a silence corresponding to a conversation start question to start a conversation, and includes a structured scenario database, a user's personal information, a user's behavior pattern, and a previous conversation. The conversation start question is extracted according to the application execution environment information which is at least one of the recording information, and the output information.
즉, 사용자의 인적 사항, 날짜, 시간 등 환경 정보, 이전의 대화에 대한 기록 정보에 기초하여 사용자가 관심을 가질 만한 대화 주제에 해당하는 최초 질문을 추출하고, 이를 사용자에게 출력함으로써 대화의 시작을 능동적으로 수행한다. That is, based on the user's personal information, date, time, environment information such as the history of the previous conversation, and extracts the first question that corresponds to the subject of the conversation that the user may be interested in, and outputs it to the user to start the conversation. Actively perform
또는, S100 단계에서 무음이 출력되는 경우에는, S200 단계에서 음성으로 입력된 사용자 답변에 대한 반응 및 질문 순서로 대화가 진행되도록 지원된다. Alternatively, when the silent sound is output in step S100, the dialogue is performed in response to the user's answer input by voice in the step S200 and in the order of questions.
본 발명의 실시예에 따른 S200 단계는 음성으로 입력된 사용자 답변을 텍스트로 전환하여, 사용자 답변 분석을 위한 문장을 제공한다. Step S200 according to an embodiment of the present invention converts a user's answer input by voice into text, and provides a sentence for analyzing a user's answer.
본 발명의 실시예에 따른 S300 단계는 사용자 답변이 기설정된 답변 유형 중 어느 유형에 해당되는지 판별하여 분석을 수행한다. 본 발명에 따르면 최초 질문, 사용자의 답변, 사용자의 답변에 대한 반응, 반응에 따른 질문 순(무음이 출력되는 경우에는 사용자의 음성 입력, 사용자 음성에 대한 반응, 반응에 따른 질문 순)으로 대화가 진행되므로, 사용자 답변이 기설정된 답변 유형(예: 선택형, 일반형, 반복을 원하는 답변, 관련 없는 답변)으로 구분된 유형 중 어느 유형에 해당하는지를 판별하여, 사용자 답변에 대한 반응 및 질문을 선택하기 위한 근거 자료로 활용한다. Step S300 according to an embodiment of the present invention performs analysis by determining which type of response type the user answer corresponds to. According to the present invention, the dialogue is performed in the order of the first question, the user's answer, the response to the user's answer, and the question according to the response (when silence is output, the user's voice input, the response to the user's voice, the question according to the response) To determine which type of response the user's answer falls into, which is divided into preset answer types (e.g., optional, general, answer you want to repeat, unrelated answer), Use it as evidence.
본 발명의 실시예에 따른 S400 단계는 반응 음성 및 질문 음성에 대응되는 텍스트를 화면을 통해 출력함으로써, 청각 뿐 아니라 시각적으로 반응 텍스트 및 질문 텍스트를 사용자에게 제공함으로써, 사용자의 보다 정확한 인지를 지원한다. According to an embodiment of the present invention, the step S400 outputs the text corresponding to the response voice and the question voice through the screen, thereby providing the user with the response text and the question text visually as well as hearing, thereby supporting more accurate recognition of the user. .
본 발명의 실시예에 따른 S600 단계는 대화 종료 기준에 해당하는지 여부를 판별하는 단계로서, 사용자가 기설정된 작별 인사를 하는 것으로 확인된 경우, 일정 시간 이상 사용자의 답변이 없는 경우, 일정 시간 이상 사용자의 답변이 없고, 사용자를 호출하는 출력부의 음성에 대하여 사용자의 회신이 없는 경우 등 종료 기준에 해당하는 경우 대화는 종결되고, 대화 종료 기준에 해당하지 않는 경우에는 S200 단계로 돌아가 사용자 답변을 입력 받는 단계로부터 S200 내지 S500 단계가 반복 수행된다. According to an embodiment of the present invention, step S600 is a step of determining whether the conversation ends with a criterion. When the user is confirmed to say goodbye for a predetermined time, if the user does not answer for a predetermined time or longer, the user for a predetermined time or more. If there is no answer, and if the end criterion, such as when there is no user's reply to the voice of the output unit calling the user, the conversation is terminated. From step S200 to step S500 are repeated.
이제까지 본 발명의 실시예들을 중심으로 살펴보았다. 본 발명이 속하는 기술 분야에서 통상의 지식을 가진 자는 본 발명이 본 발명의 본질적인 특성에서 벗어나지 않는 범위에서 변형된 형태로 구현될 수 있음을 이해할 수 있을 것이다. 그러므로 개시된 실시예들은 한정적인 관점이 아니라 설명적인 관점에서 고려되어야 한다. 본 발명의 범위는 전술한 설명이 아니라 특허청구범위에 나타나 있으며, 그와 동등한 범위 내에 있는 모든 차이점은 본 발명에 포함된 것으로 해석되어야 할 것이다.So far I looked at the center of the embodiments of the present invention. Those skilled in the art will appreciate that the present invention can be implemented in a modified form without departing from the essential features of the present invention. Therefore, the disclosed embodiments should be considered in descriptive sense only and not for purposes of limitation. The scope of the present invention is shown in the claims rather than the foregoing description, and all differences within the scope will be construed as being included in the present invention.

Claims (12)

  1. 입력된 사용자 답변을 분석하는 입력답변 분석부; An input response analyzer analyzing the input user response;
    분석 결과에 따라 기설정된 시나리오 중 적어도 어느 하나의 반응시나리오를 선택하여, 상기 사용자 답변에 대한 반응 및 질문에 대한 출력 명령을 전송하는 반응 제어부; 및A response control unit for selecting at least one response scenario among preset scenarios according to an analysis result and transmitting an output command for a response to the user response and a question; And
    무음 또는 대화 시작 음성을 출력하고, 상기 반응 제어부의 출력 명령에 따라 반응 음성 및 질문 음성을 출력하는 출력부 An output unit for outputting a silent or conversation start voice and outputting a response voice and a question voice according to an output command of the response control unit
    를 포함하는 인공지능 대화 장치. Artificial intelligence communication device comprising a.
  2. 제1항에 있어서, The method of claim 1,
    상기 출력부는 어플리케이션 실행 환경정보에 기초하여 추출되며, 대화를 시작하는 질문음성인 대화 시작 음성을 출력하는 것 The output unit is extracted based on the application execution environment information, and outputs a conversation start voice which is a question voice to start a conversation.
    인 인공지능 대화 장치. AI talk device.
  3. 제2항에 있어서, The method of claim 2,
    상기 출력부는 기구축된 시나리오 데이터베이스, 사용자의 개인 정보, 사용자의 행동 패턴, 이전의 대화에 대한 기록 정보 중 적어도 어느 하나인 상기 어플리케이션 실행 환경정보에 따라 상기 대화 시작 음성을 추출하는 것The output unit extracts the conversation start voice according to the application execution environment information, which is at least one of a built-in scenario database, a user's personal information, a user's behavior pattern, and information on a previous conversation.
    인 인공지능 대화 장치. AI talk device.
  4. 제1항에 있어서, The method of claim 1,
    상기 입력답변 분석부는 상기 대화 시작 음성 출력에 반응하여 입력된 상기 사용자의 음성 또는 상기 무음 출력 후 입력된 상기 사용자의 음성을 문자열로 전환한 결과를 입력 받고, 상기 사용자 답변이 기설정된 답변 유형 중 어느 유형에 해당되는지 여부를 판별하여 분석을 수행하는 것The input response analysis unit receives a result of converting the voice of the user input in response to the conversation start voice output or the voice of the user input after the silent output into a character string, and the user response is one of preset response types. To perform an analysis by determining whether a type is applicable
    인 인공지능 대화 장치. AI talk device.
  5. 제4항에 있어서, The method of claim 4, wherein
    상기 반응 제어부는 상기 분석 결과에 따라 기구축된 시나리오 데이터베이스에 저장된 시나리오 중 적어도 어느 하나의 반응시나리오를 선택하고, 상기 사용자 답변이 해당되는 주제에 따른 반응시나리오에 기초하여 반응 및 질문에 대한 출력 명령을 전송하는 것The reaction controller selects at least one reaction scenario from among scenarios stored in the scenario database constructed according to the analysis result, and outputs a response command for a response and a question based on the reaction scenario according to the subject to which the user answer corresponds. To transmit
    인 인공지능 대화 장치. AI talk device.
  6. 제1항에 있어서, The method of claim 1,
    상기 반응 제어부는 상기 출력부의 반응 음성 및 질문 음성 출력 중, 상기 사용자로부터 새로운 음성 데이터가 수신된 경우, 상기 출력부로 일시 정지 명령 신호를 전송하고, 상기 입력답변 분석부의 상기 새로운 음성 데이터에 따른 분석 결과에 따라 반응 및 질문을 재추출하는 것The response control unit transmits a pause command signal to the output unit when new voice data is received from the user among the response voices and the question voice outputs of the output unit, and analyzes the result according to the new voice data of the input response analyzer. Reextracting responses and questions according to
    인 인공지능 대화 장치. AI talk device.
  7. 제1항에 있어서, The method of claim 1,
    상기 출력부는 반응 음성 및 질문 음성에 대응되는 텍스트를 화면을 통하여 출력하는 것 The output unit to output a text corresponding to the response voice and the question voice through the screen
    인 인공지능 대화 장치. AI talk device.
  8. (a) 무음 또는 대화 시작 음성을 출력하는 단계; (a) outputting a silent or conversation start voice;
    (b) 상기 무음 또는 대화 시작 음성 출력에 따른 사용자 답변을 입력받는 단계; (b) receiving a user response according to the silent or conversation start voice output;
    (c) 상기 사용자 답변을 분석하고, 그 결과에 따라 기설정된 시나리오 중 적어도 어느 하나의 반응시나리오를 선택하고, 이에 기초하여 반응 및 질문을 추출하는 단계; 및 (c) analyzing the user response and selecting at least one response scenario among preset scenarios based on the result, and extracting a response and a question based on the response scenario; And
    (d) 상기 추출된 반응 및 질문에 따라 반응 음성 및 질문 음성을 출력하는 단계(d) outputting a response voice and a question voice according to the extracted response and question
    를 포함하는 인공지능 대화 방법. Artificial intelligence conversation method comprising a.
  9. 제8항에 있어서, The method of claim 8,
    상기 (a) 단계는 기구축된 시나리오 데이터베이스, 사용자의 개인정보, 사용자의 행동 패턴, 이전의 대화에 대한 기록 정보 중 적어도 어느 하나인 어플리케이션 실행 환경 정보에 따라 대화 시작 음성을 출력하는 것The step (a) is to output a conversation start voice according to the application execution environment information which is at least one of a built-in scenario database, a user's personal information, a user's behavior pattern, and a record of a previous conversation.
    인 인공지능 대화 방법. Artificial intelligence conversation method.
  10. 제8항에 있어서, The method of claim 8,
    상기 (b) 단계는 음성으로 입력된 상기 사용자 답변을 텍스트로 전환하는 것 The step (b) is to convert the user response inputted by voice into text.
    인 인공지능 대화 방법. Artificial intelligence conversation method.
  11. 제8항에 있어서, The method of claim 8,
    상기 (c) 단계는 상기 사용자 답변이 기설정된 답변 유형 중 어느 유형에 해당되는지 판별하여 분석을 수행하는 것The step (c) is to determine which type of response type the user answer corresponds to and performs an analysis.
    인 인공지능 대화 방법. Artificial intelligence conversation method.
  12. 제8항에 있어서, The method of claim 8,
    상기 (d) 단계는 상기 반응 음성 및 질문 음성에 대응되는 텍스트를 화면을 통해 출력하는 것In step (d), the text corresponding to the response voice and the question voice is output through the screen.
    인 인공지능 대화 방법. Artificial intelligence conversation method.
PCT/KR2015/004347 2015-04-29 2015-04-29 Artificial intelligence conversation device and method WO2016175354A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/KR2015/004347 WO2016175354A1 (en) 2015-04-29 2015-04-29 Artificial intelligence conversation device and method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/KR2015/004347 WO2016175354A1 (en) 2015-04-29 2015-04-29 Artificial intelligence conversation device and method

Publications (1)

Publication Number Publication Date
WO2016175354A1 true WO2016175354A1 (en) 2016-11-03

Family

ID=57199748

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2015/004347 WO2016175354A1 (en) 2015-04-29 2015-04-29 Artificial intelligence conversation device and method

Country Status (1)

Country Link
WO (1) WO2016175354A1 (en)

Cited By (92)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019000326A1 (en) * 2017-06-29 2019-01-03 Microsoft Technology Licensing, Llc Generating responses in automated chatting
CN109582763A (en) * 2017-09-27 2019-04-05 韩国电子通信研究院 Answering system and method in Motion Picture Experts Group's media environment of internet of things
US10311144B2 (en) 2017-05-16 2019-06-04 Apple Inc. Emoji word sense disambiguation
US10390213B2 (en) 2014-09-30 2019-08-20 Apple Inc. Social reminders
US10395654B2 (en) 2017-05-11 2019-08-27 Apple Inc. Text normalization based on a data-driven learning network
US10417344B2 (en) 2014-05-30 2019-09-17 Apple Inc. Exemplar-based natural language processing
US10438595B2 (en) 2014-09-30 2019-10-08 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US10474753B2 (en) 2016-09-07 2019-11-12 Apple Inc. Language identification using recurrent neural networks
US10496705B1 (en) 2018-06-03 2019-12-03 Apple Inc. Accelerated task performance
US10529332B2 (en) 2015-03-08 2020-01-07 Apple Inc. Virtual assistant activation
US10580409B2 (en) 2016-06-11 2020-03-03 Apple Inc. Application integration with a digital assistant
US10592604B2 (en) 2018-03-12 2020-03-17 Apple Inc. Inverse text normalization for automatic speech recognition
US10657966B2 (en) 2014-05-30 2020-05-19 Apple Inc. Better resolution when referencing to concepts
US10681212B2 (en) 2015-06-05 2020-06-09 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US10684703B2 (en) 2018-06-01 2020-06-16 Apple Inc. Attention aware virtual assistant dismissal
US10692504B2 (en) 2010-02-25 2020-06-23 Apple Inc. User profiling for voice input processing
US10699717B2 (en) 2014-05-30 2020-06-30 Apple Inc. Intelligent assistant for home automation
US10714117B2 (en) 2013-02-07 2020-07-14 Apple Inc. Voice trigger for a digital assistant
US10720160B2 (en) 2018-06-01 2020-07-21 Apple Inc. Voice interaction at a primary device to access call functionality of a companion device
US10741185B2 (en) 2010-01-18 2020-08-11 Apple Inc. Intelligent automated assistant
US10741181B2 (en) 2017-05-09 2020-08-11 Apple Inc. User interface for correcting recognition errors
US10748546B2 (en) 2017-05-16 2020-08-18 Apple Inc. Digital assistant services based on device capabilities
US10769385B2 (en) 2013-06-09 2020-09-08 Apple Inc. System and method for inferring user intent from speech inputs
US10839159B2 (en) 2018-09-28 2020-11-17 Apple Inc. Named entity normalization in a spoken dialog system
US10878809B2 (en) 2014-05-30 2020-12-29 Apple Inc. Multi-command single utterance input method
US10892996B2 (en) 2018-06-01 2021-01-12 Apple Inc. Variable latency device coordination
US10909171B2 (en) 2017-05-16 2021-02-02 Apple Inc. Intelligent automated assistant for media exploration
US10930282B2 (en) 2015-03-08 2021-02-23 Apple Inc. Competing devices responding to voice triggers
US10942703B2 (en) 2015-12-23 2021-03-09 Apple Inc. Proactive assistance based on dialog communication between devices
US10956666B2 (en) 2015-11-09 2021-03-23 Apple Inc. Unconventional virtual assistant interactions
US11010561B2 (en) 2018-09-27 2021-05-18 Apple Inc. Sentiment prediction from textual data
US11010127B2 (en) 2015-06-29 2021-05-18 Apple Inc. Virtual assistant for media playback
US11037565B2 (en) 2016-06-10 2021-06-15 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US11048473B2 (en) 2013-06-09 2021-06-29 Apple Inc. Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant
US11070949B2 (en) 2015-05-27 2021-07-20 Apple Inc. Systems and methods for proactively identifying and surfacing relevant content on an electronic device with a touch-sensitive display
US11120372B2 (en) 2011-06-03 2021-09-14 Apple Inc. Performing actions associated with task items that represent tasks to perform
US11126400B2 (en) 2015-09-08 2021-09-21 Apple Inc. Zero latency digital assistant
US11127397B2 (en) 2015-05-27 2021-09-21 Apple Inc. Device voice control
US11133008B2 (en) 2014-05-30 2021-09-28 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US11140099B2 (en) 2019-05-21 2021-10-05 Apple Inc. Providing message response suggestions
US11170166B2 (en) 2018-09-28 2021-11-09 Apple Inc. Neural typographical error modeling via generative adversarial networks
US11169616B2 (en) 2018-05-07 2021-11-09 Apple Inc. Raise to speak
CN113707139A (en) * 2020-09-02 2021-11-26 南宁玄鸟网络科技有限公司 Voice communication and communication service system of artificial intelligent robot
WO2021261664A1 (en) * 2020-06-23 2021-12-30 주식회사 오투오 Tourism service system based on artificial intelligence speech dialogue
US11217251B2 (en) 2019-05-06 2022-01-04 Apple Inc. Spoken notifications
US11227589B2 (en) 2016-06-06 2022-01-18 Apple Inc. Intelligent list reading
US11231904B2 (en) 2015-03-06 2022-01-25 Apple Inc. Reducing response latency of intelligent automated assistants
US11237797B2 (en) 2019-05-31 2022-02-01 Apple Inc. User activity shortcut suggestions
US11269678B2 (en) 2012-05-15 2022-03-08 Apple Inc. Systems and methods for integrating third party services with a digital assistant
US11289073B2 (en) 2019-05-31 2022-03-29 Apple Inc. Device text to speech
US11301477B2 (en) 2017-05-12 2022-04-12 Apple Inc. Feedback analysis of a digital assistant
US11307752B2 (en) 2019-05-06 2022-04-19 Apple Inc. User configurable task triggers
US11314370B2 (en) 2013-12-06 2022-04-26 Apple Inc. Method for extracting salient dialog usage from live data
US11348573B2 (en) 2019-03-18 2022-05-31 Apple Inc. Multimodality in digital assistant systems
US11348582B2 (en) 2008-10-02 2022-05-31 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
US11360641B2 (en) 2019-06-01 2022-06-14 Apple Inc. Increasing the relevance of new available information
US11380310B2 (en) 2017-05-12 2022-07-05 Apple Inc. Low-latency intelligent automated assistant
US11388291B2 (en) 2013-03-14 2022-07-12 Apple Inc. System and method for processing voicemail
US11405466B2 (en) 2017-05-12 2022-08-02 Apple Inc. Synchronization and task delegation of a digital assistant
US11423886B2 (en) 2010-01-18 2022-08-23 Apple Inc. Task flow identification based on user intent
US11423908B2 (en) 2019-05-06 2022-08-23 Apple Inc. Interpreting spoken requests
US11462215B2 (en) 2018-09-28 2022-10-04 Apple Inc. Multi-modal inputs for voice commands
US11467802B2 (en) 2017-05-11 2022-10-11 Apple Inc. Maintaining privacy of personal information
US11468282B2 (en) 2015-05-15 2022-10-11 Apple Inc. Virtual assistant in a communication session
US11475898B2 (en) 2018-10-26 2022-10-18 Apple Inc. Low-latency multi-speaker speech recognition
US11475884B2 (en) 2019-05-06 2022-10-18 Apple Inc. Reducing digital assistant latency when a language is incorrectly determined
US11488406B2 (en) 2019-09-25 2022-11-01 Apple Inc. Text detection using global geometry estimators
US11496600B2 (en) 2019-05-31 2022-11-08 Apple Inc. Remote execution of machine-learned models
US11500672B2 (en) 2015-09-08 2022-11-15 Apple Inc. Distributed personal assistant
US11516537B2 (en) 2014-06-30 2022-11-29 Apple Inc. Intelligent automated assistant for TV user interactions
US11526368B2 (en) 2015-11-06 2022-12-13 Apple Inc. Intelligent automated assistant in a messaging environment
US11532306B2 (en) 2017-05-16 2022-12-20 Apple Inc. Detecting a trigger of a digital assistant
US11580990B2 (en) 2017-05-12 2023-02-14 Apple Inc. User-specific acoustic models
US11599331B2 (en) 2017-05-11 2023-03-07 Apple Inc. Maintaining privacy of personal information
US11638059B2 (en) 2019-01-04 2023-04-25 Apple Inc. Content playback on multiple devices
US11656884B2 (en) 2017-01-09 2023-05-23 Apple Inc. Application integration with a digital assistant
US11657813B2 (en) 2019-05-31 2023-05-23 Apple Inc. Voice identification in digital assistant systems
US11671920B2 (en) 2007-04-03 2023-06-06 Apple Inc. Method and system for operating a multifunction portable electronic device using voice-activation
US11696060B2 (en) 2020-07-21 2023-07-04 Apple Inc. User identification using headphones
US11710482B2 (en) 2018-03-26 2023-07-25 Apple Inc. Natural assistant interaction
US11755276B2 (en) 2020-05-12 2023-09-12 Apple Inc. Reducing description length based on confidence
US11765209B2 (en) 2020-05-11 2023-09-19 Apple Inc. Digital assistant hardware abstraction
US11790914B2 (en) 2019-06-01 2023-10-17 Apple Inc. Methods and user interfaces for voice-based control of electronic devices
US11798547B2 (en) 2013-03-15 2023-10-24 Apple Inc. Voice activated device for use with a voice-based digital assistant
US11809483B2 (en) 2015-09-08 2023-11-07 Apple Inc. Intelligent automated assistant for media search and playback
US11810578B2 (en) 2020-05-11 2023-11-07 Apple Inc. Device arbitration for digital assistant-based intercom systems
US11809783B2 (en) 2016-06-11 2023-11-07 Apple Inc. Intelligent device arbitration and control
US11838734B2 (en) 2020-07-20 2023-12-05 Apple Inc. Multi-device audio adjustment coordination
US11853536B2 (en) 2015-09-08 2023-12-26 Apple Inc. Intelligent automated assistant in a media environment
US11854539B2 (en) 2018-05-07 2023-12-26 Apple Inc. Intelligent automated assistant for delivering content from user experiences
US11914848B2 (en) 2020-05-11 2024-02-27 Apple Inc. Providing relevant data items based on context
US11928604B2 (en) 2005-09-08 2024-03-12 Apple Inc. Method and apparatus for building an intelligent automated assistant

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20020010226A (en) * 2000-07-28 2002-02-04 정명수 Internet Anything Response System
WO2014010879A1 (en) * 2012-07-09 2014-01-16 엘지전자 주식회사 Speech recognition apparatus and method
WO2014088377A1 (en) * 2012-12-07 2014-06-12 삼성전자 주식회사 Voice recognition device and method of controlling same
US20140222436A1 (en) * 2013-02-07 2014-08-07 Apple Inc. Voice trigger for a digital assistant

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20020010226A (en) * 2000-07-28 2002-02-04 정명수 Internet Anything Response System
WO2014010879A1 (en) * 2012-07-09 2014-01-16 엘지전자 주식회사 Speech recognition apparatus and method
WO2014088377A1 (en) * 2012-12-07 2014-06-12 삼성전자 주식회사 Voice recognition device and method of controlling same
US20140222436A1 (en) * 2013-02-07 2014-08-07 Apple Inc. Voice trigger for a digital assistant

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
KOREA CREATIVE CONTENT AGENCY: "This Month's Issue, Trend and Prospect of Speech Recognition Technology", CULTURE TECHNOLOGY(CT) IN-DEPTH STUDY, November 2011 (2011-11-01), Retrieved from the Internet <URL:https://www.kocca.kr/knowledge/publication/ct/ksFiles/afieldfile/2011/12/07/87NEmyIcVWMc.pdf> *

Cited By (139)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11928604B2 (en) 2005-09-08 2024-03-12 Apple Inc. Method and apparatus for building an intelligent automated assistant
US11671920B2 (en) 2007-04-03 2023-06-06 Apple Inc. Method and system for operating a multifunction portable electronic device using voice-activation
US11348582B2 (en) 2008-10-02 2022-05-31 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
US11900936B2 (en) 2008-10-02 2024-02-13 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
US11423886B2 (en) 2010-01-18 2022-08-23 Apple Inc. Task flow identification based on user intent
US10741185B2 (en) 2010-01-18 2020-08-11 Apple Inc. Intelligent automated assistant
US10692504B2 (en) 2010-02-25 2020-06-23 Apple Inc. User profiling for voice input processing
US11120372B2 (en) 2011-06-03 2021-09-14 Apple Inc. Performing actions associated with task items that represent tasks to perform
US11269678B2 (en) 2012-05-15 2022-03-08 Apple Inc. Systems and methods for integrating third party services with a digital assistant
US11321116B2 (en) 2012-05-15 2022-05-03 Apple Inc. Systems and methods for integrating third party services with a digital assistant
US10978090B2 (en) 2013-02-07 2021-04-13 Apple Inc. Voice trigger for a digital assistant
US11862186B2 (en) 2013-02-07 2024-01-02 Apple Inc. Voice trigger for a digital assistant
US11557310B2 (en) 2013-02-07 2023-01-17 Apple Inc. Voice trigger for a digital assistant
US10714117B2 (en) 2013-02-07 2020-07-14 Apple Inc. Voice trigger for a digital assistant
US11636869B2 (en) 2013-02-07 2023-04-25 Apple Inc. Voice trigger for a digital assistant
US11388291B2 (en) 2013-03-14 2022-07-12 Apple Inc. System and method for processing voicemail
US11798547B2 (en) 2013-03-15 2023-10-24 Apple Inc. Voice activated device for use with a voice-based digital assistant
US11727219B2 (en) 2013-06-09 2023-08-15 Apple Inc. System and method for inferring user intent from speech inputs
US11048473B2 (en) 2013-06-09 2021-06-29 Apple Inc. Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant
US10769385B2 (en) 2013-06-09 2020-09-08 Apple Inc. System and method for inferring user intent from speech inputs
US11314370B2 (en) 2013-12-06 2022-04-26 Apple Inc. Method for extracting salient dialog usage from live data
US11257504B2 (en) 2014-05-30 2022-02-22 Apple Inc. Intelligent assistant for home automation
US11699448B2 (en) 2014-05-30 2023-07-11 Apple Inc. Intelligent assistant for home automation
US11133008B2 (en) 2014-05-30 2021-09-28 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US10714095B2 (en) 2014-05-30 2020-07-14 Apple Inc. Intelligent assistant for home automation
US10417344B2 (en) 2014-05-30 2019-09-17 Apple Inc. Exemplar-based natural language processing
US10878809B2 (en) 2014-05-30 2020-12-29 Apple Inc. Multi-command single utterance input method
US11810562B2 (en) 2014-05-30 2023-11-07 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US10657966B2 (en) 2014-05-30 2020-05-19 Apple Inc. Better resolution when referencing to concepts
US11670289B2 (en) 2014-05-30 2023-06-06 Apple Inc. Multi-command single utterance input method
US10699717B2 (en) 2014-05-30 2020-06-30 Apple Inc. Intelligent assistant for home automation
US11516537B2 (en) 2014-06-30 2022-11-29 Apple Inc. Intelligent automated assistant for TV user interactions
US11838579B2 (en) 2014-06-30 2023-12-05 Apple Inc. Intelligent automated assistant for TV user interactions
US10390213B2 (en) 2014-09-30 2019-08-20 Apple Inc. Social reminders
US10438595B2 (en) 2014-09-30 2019-10-08 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US11231904B2 (en) 2015-03-06 2022-01-25 Apple Inc. Reducing response latency of intelligent automated assistants
US11842734B2 (en) 2015-03-08 2023-12-12 Apple Inc. Virtual assistant activation
US10529332B2 (en) 2015-03-08 2020-01-07 Apple Inc. Virtual assistant activation
US11087759B2 (en) 2015-03-08 2021-08-10 Apple Inc. Virtual assistant activation
US10930282B2 (en) 2015-03-08 2021-02-23 Apple Inc. Competing devices responding to voice triggers
US11468282B2 (en) 2015-05-15 2022-10-11 Apple Inc. Virtual assistant in a communication session
US11070949B2 (en) 2015-05-27 2021-07-20 Apple Inc. Systems and methods for proactively identifying and surfacing relevant content on an electronic device with a touch-sensitive display
US11127397B2 (en) 2015-05-27 2021-09-21 Apple Inc. Device voice control
US10681212B2 (en) 2015-06-05 2020-06-09 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US11947873B2 (en) 2015-06-29 2024-04-02 Apple Inc. Virtual assistant for media playback
US11010127B2 (en) 2015-06-29 2021-05-18 Apple Inc. Virtual assistant for media playback
US11853536B2 (en) 2015-09-08 2023-12-26 Apple Inc. Intelligent automated assistant in a media environment
US11500672B2 (en) 2015-09-08 2022-11-15 Apple Inc. Distributed personal assistant
US11550542B2 (en) 2015-09-08 2023-01-10 Apple Inc. Zero latency digital assistant
US11126400B2 (en) 2015-09-08 2021-09-21 Apple Inc. Zero latency digital assistant
US11809483B2 (en) 2015-09-08 2023-11-07 Apple Inc. Intelligent automated assistant for media search and playback
US11954405B2 (en) 2015-09-08 2024-04-09 Apple Inc. Zero latency digital assistant
US11526368B2 (en) 2015-11-06 2022-12-13 Apple Inc. Intelligent automated assistant in a messaging environment
US11809886B2 (en) 2015-11-06 2023-11-07 Apple Inc. Intelligent automated assistant in a messaging environment
US10956666B2 (en) 2015-11-09 2021-03-23 Apple Inc. Unconventional virtual assistant interactions
US11886805B2 (en) 2015-11-09 2024-01-30 Apple Inc. Unconventional virtual assistant interactions
US10942703B2 (en) 2015-12-23 2021-03-09 Apple Inc. Proactive assistance based on dialog communication between devices
US11227589B2 (en) 2016-06-06 2022-01-18 Apple Inc. Intelligent list reading
US11657820B2 (en) 2016-06-10 2023-05-23 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US11037565B2 (en) 2016-06-10 2021-06-15 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US11152002B2 (en) 2016-06-11 2021-10-19 Apple Inc. Application integration with a digital assistant
US10580409B2 (en) 2016-06-11 2020-03-03 Apple Inc. Application integration with a digital assistant
US11809783B2 (en) 2016-06-11 2023-11-07 Apple Inc. Intelligent device arbitration and control
US11749275B2 (en) 2016-06-11 2023-09-05 Apple Inc. Application integration with a digital assistant
US10474753B2 (en) 2016-09-07 2019-11-12 Apple Inc. Language identification using recurrent neural networks
US11656884B2 (en) 2017-01-09 2023-05-23 Apple Inc. Application integration with a digital assistant
US10741181B2 (en) 2017-05-09 2020-08-11 Apple Inc. User interface for correcting recognition errors
US11599331B2 (en) 2017-05-11 2023-03-07 Apple Inc. Maintaining privacy of personal information
US10395654B2 (en) 2017-05-11 2019-08-27 Apple Inc. Text normalization based on a data-driven learning network
US11467802B2 (en) 2017-05-11 2022-10-11 Apple Inc. Maintaining privacy of personal information
US11837237B2 (en) 2017-05-12 2023-12-05 Apple Inc. User-specific acoustic models
US11538469B2 (en) 2017-05-12 2022-12-27 Apple Inc. Low-latency intelligent automated assistant
US11380310B2 (en) 2017-05-12 2022-07-05 Apple Inc. Low-latency intelligent automated assistant
US11580990B2 (en) 2017-05-12 2023-02-14 Apple Inc. User-specific acoustic models
US11405466B2 (en) 2017-05-12 2022-08-02 Apple Inc. Synchronization and task delegation of a digital assistant
US11301477B2 (en) 2017-05-12 2022-04-12 Apple Inc. Feedback analysis of a digital assistant
US11862151B2 (en) 2017-05-12 2024-01-02 Apple Inc. Low-latency intelligent automated assistant
US11675829B2 (en) 2017-05-16 2023-06-13 Apple Inc. Intelligent automated assistant for media exploration
US10909171B2 (en) 2017-05-16 2021-02-02 Apple Inc. Intelligent automated assistant for media exploration
US10311144B2 (en) 2017-05-16 2019-06-04 Apple Inc. Emoji word sense disambiguation
US10748546B2 (en) 2017-05-16 2020-08-18 Apple Inc. Digital assistant services based on device capabilities
US11532306B2 (en) 2017-05-16 2022-12-20 Apple Inc. Detecting a trigger of a digital assistant
WO2019000326A1 (en) * 2017-06-29 2019-01-03 Microsoft Technology Licensing, Llc Generating responses in automated chatting
CN109582763B (en) * 2017-09-27 2023-08-22 韩国电子通信研究院 Answering system and method in moving picture expert group media Internet of things environment
CN109582763A (en) * 2017-09-27 2019-04-05 韩国电子通信研究院 Answering system and method in Motion Picture Experts Group's media environment of internet of things
US10592604B2 (en) 2018-03-12 2020-03-17 Apple Inc. Inverse text normalization for automatic speech recognition
US11710482B2 (en) 2018-03-26 2023-07-25 Apple Inc. Natural assistant interaction
US11907436B2 (en) 2018-05-07 2024-02-20 Apple Inc. Raise to speak
US11900923B2 (en) 2018-05-07 2024-02-13 Apple Inc. Intelligent automated assistant for delivering content from user experiences
US11487364B2 (en) 2018-05-07 2022-11-01 Apple Inc. Raise to speak
US11854539B2 (en) 2018-05-07 2023-12-26 Apple Inc. Intelligent automated assistant for delivering content from user experiences
US11169616B2 (en) 2018-05-07 2021-11-09 Apple Inc. Raise to speak
US10984798B2 (en) 2018-06-01 2021-04-20 Apple Inc. Voice interaction at a primary device to access call functionality of a companion device
US11009970B2 (en) 2018-06-01 2021-05-18 Apple Inc. Attention aware virtual assistant dismissal
US11630525B2 (en) 2018-06-01 2023-04-18 Apple Inc. Attention aware virtual assistant dismissal
US10684703B2 (en) 2018-06-01 2020-06-16 Apple Inc. Attention aware virtual assistant dismissal
US10720160B2 (en) 2018-06-01 2020-07-21 Apple Inc. Voice interaction at a primary device to access call functionality of a companion device
US10892996B2 (en) 2018-06-01 2021-01-12 Apple Inc. Variable latency device coordination
US11360577B2 (en) 2018-06-01 2022-06-14 Apple Inc. Attention aware virtual assistant dismissal
US11431642B2 (en) 2018-06-01 2022-08-30 Apple Inc. Variable latency device coordination
US10496705B1 (en) 2018-06-03 2019-12-03 Apple Inc. Accelerated task performance
US10504518B1 (en) 2018-06-03 2019-12-10 Apple Inc. Accelerated task performance
US10944859B2 (en) 2018-06-03 2021-03-09 Apple Inc. Accelerated task performance
US11010561B2 (en) 2018-09-27 2021-05-18 Apple Inc. Sentiment prediction from textual data
US11462215B2 (en) 2018-09-28 2022-10-04 Apple Inc. Multi-modal inputs for voice commands
US11170166B2 (en) 2018-09-28 2021-11-09 Apple Inc. Neural typographical error modeling via generative adversarial networks
US10839159B2 (en) 2018-09-28 2020-11-17 Apple Inc. Named entity normalization in a spoken dialog system
US11893992B2 (en) 2018-09-28 2024-02-06 Apple Inc. Multi-modal inputs for voice commands
US11475898B2 (en) 2018-10-26 2022-10-18 Apple Inc. Low-latency multi-speaker speech recognition
US11638059B2 (en) 2019-01-04 2023-04-25 Apple Inc. Content playback on multiple devices
US11348573B2 (en) 2019-03-18 2022-05-31 Apple Inc. Multimodality in digital assistant systems
US11783815B2 (en) 2019-03-18 2023-10-10 Apple Inc. Multimodality in digital assistant systems
US11423908B2 (en) 2019-05-06 2022-08-23 Apple Inc. Interpreting spoken requests
US11475884B2 (en) 2019-05-06 2022-10-18 Apple Inc. Reducing digital assistant latency when a language is incorrectly determined
US11307752B2 (en) 2019-05-06 2022-04-19 Apple Inc. User configurable task triggers
US11675491B2 (en) 2019-05-06 2023-06-13 Apple Inc. User configurable task triggers
US11705130B2 (en) 2019-05-06 2023-07-18 Apple Inc. Spoken notifications
US11217251B2 (en) 2019-05-06 2022-01-04 Apple Inc. Spoken notifications
US11888791B2 (en) 2019-05-21 2024-01-30 Apple Inc. Providing message response suggestions
US11140099B2 (en) 2019-05-21 2021-10-05 Apple Inc. Providing message response suggestions
US11237797B2 (en) 2019-05-31 2022-02-01 Apple Inc. User activity shortcut suggestions
US11657813B2 (en) 2019-05-31 2023-05-23 Apple Inc. Voice identification in digital assistant systems
US11289073B2 (en) 2019-05-31 2022-03-29 Apple Inc. Device text to speech
US11496600B2 (en) 2019-05-31 2022-11-08 Apple Inc. Remote execution of machine-learned models
US11360739B2 (en) 2019-05-31 2022-06-14 Apple Inc. User activity shortcut suggestions
US11360641B2 (en) 2019-06-01 2022-06-14 Apple Inc. Increasing the relevance of new available information
US11790914B2 (en) 2019-06-01 2023-10-17 Apple Inc. Methods and user interfaces for voice-based control of electronic devices
US11488406B2 (en) 2019-09-25 2022-11-01 Apple Inc. Text detection using global geometry estimators
US11914848B2 (en) 2020-05-11 2024-02-27 Apple Inc. Providing relevant data items based on context
US11765209B2 (en) 2020-05-11 2023-09-19 Apple Inc. Digital assistant hardware abstraction
US11810578B2 (en) 2020-05-11 2023-11-07 Apple Inc. Device arbitration for digital assistant-based intercom systems
US11924254B2 (en) 2020-05-11 2024-03-05 Apple Inc. Digital assistant hardware abstraction
US11755276B2 (en) 2020-05-12 2023-09-12 Apple Inc. Reducing description length based on confidence
WO2021261664A1 (en) * 2020-06-23 2021-12-30 주식회사 오투오 Tourism service system based on artificial intelligence speech dialogue
US11838734B2 (en) 2020-07-20 2023-12-05 Apple Inc. Multi-device audio adjustment coordination
US11750962B2 (en) 2020-07-21 2023-09-05 Apple Inc. User identification using headphones
US11696060B2 (en) 2020-07-21 2023-07-04 Apple Inc. User identification using headphones
CN113707139A (en) * 2020-09-02 2021-11-26 南宁玄鸟网络科技有限公司 Voice communication and communication service system of artificial intelligent robot
CN113707139B (en) * 2020-09-02 2024-04-09 南宁玄鸟网络科技有限公司 Voice communication and communication service system of artificial intelligent robot

Similar Documents

Publication Publication Date Title
WO2016175354A1 (en) Artificial intelligence conversation device and method
WO2021051506A1 (en) Voice interaction method and apparatus, computer device and storage medium
KR20190095181A (en) Video conference system using artificial intelligence
WO2011074771A2 (en) Apparatus and method for foreign language study
US20140214426A1 (en) System and method for improving voice communication over a network
CN109961792A (en) The method and apparatus of voice for identification
CN110047481A (en) Method for voice recognition and device
JP2012530954A (en) Method and apparatus for improving language communication
CN109712610A (en) The method and apparatus of voice for identification
JP6689953B2 (en) Interpreter service system, interpreter service method, and interpreter service program
CN111063346A (en) Cross-media star emotion accompany interaction system based on machine learning
KR20080114100A (en) Method and apparatus of naturally talking with computer
WO2019142976A1 (en) Display control method, computer-readable recording medium, and computer device for displaying conversation response candidate for user speech input
CN113630309B (en) Robot conversation system, method, device, computer equipment and storage medium
KR20120073557A (en) Speech recognition based intelligent robot system
KR100677435B1 (en) Interactive foreign language studying system and method thereof
KR20220140301A (en) Video learning systems for enable learners to be identified through artificial intelligence and method thereof
KR20220140304A (en) Video learning systems for recognize learners&#39; voice commands
CN110675856A (en) Man-machine conversation method and device for call center
CN112309183A (en) Interactive listening and speaking exercise system suitable for foreign language teaching
CN110519470A (en) A kind of method of speech processing, server and audio access device
KR102577643B1 (en) Online one to one korean lecture platform system and operating server included in the same
KR102364935B1 (en) A method and apparatus for data transmission for improving 5G-based speech recognition response speed
JP6349149B2 (en) Lesson progression system, lesson progression method and lesson progression program
KR102359228B1 (en) Method for customized conversation connection service

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15890795

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 14/02/2018)

122 Ep: pct application non-entry in european phase

Ref document number: 15890795

Country of ref document: EP

Kind code of ref document: A1