WO2019041517A1 - Electronic device, question recognition and confirmation method, and computer-readable storage medium - Google Patents

Electronic device, question recognition and confirmation method, and computer-readable storage medium Download PDF

Info

Publication number
WO2019041517A1
WO2019041517A1 PCT/CN2017/108763 CN2017108763W WO2019041517A1 WO 2019041517 A1 WO2019041517 A1 WO 2019041517A1 CN 2017108763 W CN2017108763 W CN 2017108763W WO 2019041517 A1 WO2019041517 A1 WO 2019041517A1
Authority
WO
WIPO (PCT)
Prior art keywords
question
text
word
probability
feature word
Prior art date
Application number
PCT/CN2017/108763
Other languages
French (fr)
Chinese (zh)
Inventor
王健宗
韩茂琨
肖京
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2019041517A1 publication Critical patent/WO2019041517A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/903Querying
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems

Definitions

  • the present application relates to the field of intelligent voice technologies, and in particular, to an electronic device, a problem recognition confirmation method, and a computer readable storage medium.
  • offline intelligent customer service robots for example The intelligent customer service robot set in the physical office area and/or an online intelligent customer service response system (for example, an intelligent voice response system) serves the customer.
  • the existing schemes for such offline intelligent customer service robots and/or online intelligent customer service response systems are: pre-configure mapping relationship data between standard questions and standard answers; when receiving standard questions raised by customers, according to pre-configured The mapping relationship between the standard question and the standard answer determines the standard answer corresponding to the received standard question and feeds back the determined standard answer to the customer. For existing non-standard questions, this existing solution will be difficult to give feedback.
  • the main purpose of the present application is to provide a method for identifying and identifying problems, which aims to improve the accuracy of the intelligent customer service system for identifying and confirming non-standard problems, thereby improving the accuracy of feedback answers to non-standard questions.
  • a first aspect of the present application provides an electronic device including a memory, a processor, and a memory identification confirmation system operable on the processor, where the problem recognition confirmation system is executed by the processor Implement the following steps:
  • the problem of determining the maximum probability corresponding to the predetermined feature word is determined according to the probability distribution between the feature word and the question, and according to the mapping between the predetermined question and the answer Relationship, the answer corresponding to the question that determines the maximum probability;
  • the determined answer is fed back to the user.
  • a second aspect of the present application provides a method for identifying a problem, the method comprising the steps of:
  • the problem of determining the maximum probability corresponding to the predetermined feature word is determined according to the probability distribution between the feature word and the question, and according to the mapping between the predetermined question and the answer Relationship, the answer corresponding to the question that determines the maximum probability;
  • the determined answer is fed back to the user.
  • a third aspect of the present application provides a computer readable storage medium storing a problem identification confirmation system, the problem identification confirmation system being executable by at least one processor to cause the at least one processor Perform the following steps:
  • the problem of determining the maximum probability corresponding to the predetermined feature word is determined according to the probability distribution between the feature word and the question, and according to the mapping between the predetermined question and the answer Relationship, the answer corresponding to the question that determines the maximum probability;
  • the determined answer is fed back to the user.
  • the technical solution of the present application After the user's question speech is recognized as a question text, the technical solution of the present application performs word segmentation on the question text, and obtains a feature word contained in the word segmentation result that reflects the subject or semantic direction of the user question, and is between the feature word and the question.
  • FIG. 1 is a schematic flowchart of an embodiment of a method for identifying and confirming a problem of the present application
  • FIG. 2 is a schematic flowchart of an embodiment of a method for identifying and confirming a problem of the present application
  • FIG. 3 is a schematic flowchart of determining a probability distribution between a feature word and a question in the problem identification and confirmation method of the present application
  • FIG. 4 is a schematic diagram of an operating environment of a preferred embodiment of the problem identification and confirmation system of the present application.
  • FIG. 5 is a block diagram of a program of an embodiment of the problem identification confirmation system of the present application.
  • FIG. 6 is a program block diagram of a second embodiment of the problem identification and confirmation system of the present application.
  • the present application proposes a method for identifying and identifying problems, which is mainly used for intelligent customer service products such as an intelligent customer service response system or an intelligent customer service robot.
  • FIG. 1 is a schematic flowchart of an embodiment of a method for identifying and confirming a problem of the present application.
  • the method for identifying and identifying the problem includes:
  • Step S1 receiving a question voice sent by the user, performing voice recognition on the received question voice, and generating a question text;
  • the problem recognition confirmation system receives the question voice issued by the user when questioning, identifies the received question voice, and generates the corresponding question text of the recognized question voice.
  • Step S2 performing segmentation processing on the generated problem text according to a predetermined word segmentation rule, and obtaining a word segment corresponding to the problem text;
  • the question recognition confirming system After converting the received question speech recognition into the question text, the question recognition confirming system performs word segmentation processing on the question text according to the predetermined word segmentation rule, and after the word segmentation process, the word segment corresponding to the question text is obtained.
  • the word segmentation includes words and words.
  • the question text may be “Ping An has launched a product of Zunhong Life”, and after the word segmentation, the result is “Ping An”, “Pushing”, “Yes”, “Zhonghong life”, “product”, “?”.
  • Step S3 if the obtained participle contains a predetermined feature word, the problem of determining the maximum probability corresponding to the predetermined feature word is determined according to the probability distribution between the feature word and the question, and according to the predetermined question and answer The mapping relationship between the two, determining the answer corresponding to the question of the maximum probability;
  • the system has preset feature words (for example, “Zhonghong Life”, “Peace”, etc.), the feature words can reflect the subject or semantic direction of the question corresponding to the question text; the system also has predetermined feature words and questions.
  • the probability distribution between each feature word has a probability value corresponding to each pre-existing problem, and the problem text containing each feature word may be the probability of each problem; the system is also set between the preset question and the answer. Mapping relational tables. After obtaining the word segment corresponding to the question text, the system analyzes whether the obtained word segment contains a predetermined feature word; when the analyzed word segment does not contain the predetermined feature word, the user is prompted to re-question or the prompt is unrecognizable. Ask questions and other treatments.
  • the problem of determining the maximum probability corresponding to the predetermined feature word is determined according to the probability distribution between the feature word and the question, that is, the most probable problem
  • the answer corresponding to the question of the maximum probability is obtained according to the mapping relationship between the predetermined question and the answer.
  • step S4 the determined answer is fed back to the user.
  • the system feeds the determined answer to the user by means of voice broadcast or sending to the display device for display or sending to the user's preset terminal.
  • the technical solution of the present invention performs word segmentation on the question text, and obtains a feature word contained in the word segmentation result that reflects the subject or semantic direction of the user question, and is between the feature word and the question.
  • the probability distribution in order to find the problem with the greatest probability (ie the most probable problem), and then determine the answer corresponding to the question of the maximum probability to feed back to the user; because in this technical solution, the feature word can reflect the subject or semantic of the user problem Direction, the corresponding answer found by the problem of the maximum probability corresponding to the feature word. Therefore, compared with the prior art, the similarity between the whole problem and the standard question is compared to obtain the answer corresponding to the most similar question. The accuracy of the answers to the feedback users in this case has improved significantly.
  • the predetermined word segmentation rule is a long word priority word segmentation rule.
  • the long word priority word segmentation rule refers to: for a phrase T1 requiring a word segmentation, starting with the first word A, finding a longest word X1 starting from A from the pre-stored vocabulary, and then culling from T1 X1 has T2 left, and then the same segmentation principle is applied to T2.
  • the result after segmentation is “X1/X2/, ,,,,,”; for example, including “Peace” and “Push” in the pre-stored thesaurus.
  • FIG. 2 is a schematic flowchart of the second embodiment of the method for identifying and confirming the problem of the present application.
  • the solution of the embodiment is replaced by the following steps on the basis of the first embodiment:
  • Step S301 if the obtained word segment contains a predetermined feature word, determining, according to the probability distribution between the feature word and the question, determining the predetermined feature word corresponding to each problem The probability;
  • Step S302 sorting each problem according to the order of probability from large to small, determining a preset number of questions in the prior order as candidate questions, and providing or broadcasting the determined candidate questions to the user for selection;
  • the problems are sorted in descending order according to the obtained probabilities, and then the pre-ordered number in the sorted problem sequence is extracted.
  • the questions (for example, 3 or 4) are used as candidate questions, and the extracted candidate questions are fed back to the user for selection by the user.
  • the manner in which the candidate question is fed back to the user may be: 1. voice broadcast; 2. providing a selection interface, and the candidate question is displayed on the selection interface (for example, generating a question selection interface for the user to select, the selection interface may include a candidate question list, Each candidate question in the list corresponds to an "OK" button, the user can click the button to select a corresponding question);
  • Step S303 after the user selects a question, the answer corresponding to the question is determined according to a mapping relationship between the predetermined question and the answer.
  • the system After the user makes a selection based on the candidate question of the system feedback, the system receives the question selected by the user, and determines the answer corresponding to the question selected by the user according to the mapping relationship between the predetermined question and the answer in the system. .
  • the probability distribution between the feature words and the problem is determined as follows:
  • Step S51 adding a preset number of implicit topics between the feature words and the question;
  • an implicit number of predicted numbers (for example, 50) is added between the feature words and the problem layer as an intermediate layer to constitute a problem selection model; wherein the implicit theme is virtual and there is no real Meaning; each implied topic usually contains multiple feature words, each of which usually contains multiple implied topics.
  • Step S52 acquiring a problem text to be trained, and performing word segmentation processing on the obtained problem texts respectively, to obtain word segments corresponding to the respective problem texts;
  • the problem text to be trained is obtained (the problem text is prepared in advance), and the obtained problem texts are separately subjected to word segmentation processing, thereby obtaining the word segmentation results corresponding to the respective problem texts.
  • Step S53 determining, according to a predetermined mapping relationship between the implicit subject and the feature word, a first quantity of the feature words included in each hidden topic, and respectively determining a second quantity of the hidden topic to which each feature word belongs, according to Corresponding first quantity and second quantity determining a first selection probability of each feature word for each implicit topic;
  • the first number of words is X1
  • the probability of selection of the feature word Y for the hidden subject is: 1/(X1*X2).
  • Step S54 Determine a third quantity of the hidden topic included in each question text according to a predetermined mapping relationship between the implicit topic and the question text, and determine a fourth quantity of the problem text to which each hidden topic belongs, according to Corresponding third and fourth quantities determine a second selection probability of each implicit topic for each question text;
  • the mapping relationship between the predetermined implicit topic and the problem text in the system respectively determine the third quantity of the implicit topic contained in each question text and the fourth quantity of the problem text to which each implicit topic belongs, and then according to the corresponding The third quantity and the fourth quantity respectively obtain a second selection probability of each implicit topic for each question text; for example, the fourth quantity of the problem text to which the implicit theme K belongs is J2, and an implicit subject contained in a question text
  • the third quantity is J1
  • the probability of selection of the implicit subject K for the question text is: 1/(J1*J2).
  • the sequence of the step S54 and the step S53 can be changed.
  • Step S55 Substituting the corresponding first selection probability and the second selection probability into a predetermined calculation formula to calculate, calculating a third selection probability of each feature word for each problem text, and calculating each feature word separately for each problem
  • the third selection probability of the text is the probability distribution between the feature words and the problem.
  • the third selection probability distribution of the feature word to the problem text can be further obtained.
  • the selection probability of each feature word for each problem text is obtained, that is, the probability distribution between the feature word and the problem is obtained.
  • the first selection probability of the feature word Y for the implicit topic K is: 1/(X1*X2)
  • the second selection probability of the implicit topic K for the question text W is: 1/(J1*J2)
  • the feature The probability of selection of the word Y for the question text W is 1/(X1*X2)*(J1*J2).
  • the present application also proposes a problem identification confirmation system.
  • FIG. 4 is a schematic diagram of an operating environment of a preferred embodiment of the problem identification and confirmation system 10 of the present application.
  • the problem recognition confirmation system 10 is installed and operated in the electronic device 1.
  • the electronic device 1 may be a computing device such as a desktop computer, a notebook, a palmtop computer, and a server.
  • the electronic device 1 may include, but is not limited to, a memory 11, a processor 12 and display 13.
  • Figure 4 shows only the electronic device 1 with components 11-13, but it should be understood that not all illustrated components may be implemented, and more or fewer components may be implemented instead.
  • the memory 11 is a computer storage medium, which in some embodiments may be an internal storage unit of the electronic device 1, such as a hard disk or memory of the electronic device 1.
  • the memory 11 may also be an external storage device of the electronic device 1 in other embodiments, such as a plug-in hard disk equipped on the electronic device 1, a smart memory card (SMC), and a secure digital (SD). Card, flash card, etc. Further, the memory 11 may also include both an internal storage unit of the electronic device 1 and an external storage device.
  • the memory 11 is used to store application software and various types of data installed in the electronic device 1, such as program codes of the problem recognition confirmation system 10.
  • the memory 11 can also be used to temporarily store data that has been output or is about to be output.
  • the processor 12 in some embodiments, may be a Central Processing Unit (CPU), microprocessor or other data processing chip for running program code or processing data stored in the memory 11, such as performing problem identification confirmation. System 10 and so on.
  • CPU Central Processing Unit
  • microprocessor or other data processing chip for running program code or processing data stored in the memory 11, such as performing problem identification confirmation. System 10 and so on.
  • the display 13 may be an LED display, a liquid crystal display, a touch-sensitive liquid crystal display, an OLED (Organic Light-Emitting Diode) touch sensor, or the like in some embodiments.
  • the display 13 is for displaying information processed in the electronic device 1 and a user interface for displaying visualization, such as a business customization interface or the like.
  • the components 11-13 of the electronic device 1 communicate with one another via a system bus.
  • FIG. 5 is a program module diagram of an embodiment of the problem identification confirmation system 10 of the present application.
  • the problem identification confirmation system 10 can be divided into one or more modules, one or more modules are stored in the memory 11, and by one or more processors (the processor 12 in this embodiment) Executed to complete the application.
  • the problem identification confirmation system 10 can be divided into an identification module 101, a word segmentation module 102, a determination module 103, and a feedback module 104.
  • a module referred to in this application refers to a series of computer program instruction segments capable of performing a specific function, and is more suitable than the program for describing the execution process of the problem recognition confirmation system 10 in the electronic device 1, wherein:
  • the identification module 101 is configured to receive a question voice sent by the user, perform voice recognition on the received question voice, and generate a question text.
  • the problem recognition confirmation system receives the question voice issued by the user when questioning, identifies the received question voice, and generates the corresponding question text of the recognized question voice.
  • the word segmentation module 102 is configured to perform word segmentation processing on the generated problem text according to a predetermined word segmentation rule to obtain a word segment corresponding to the problem text;
  • the problem recognition confirmation system After converting the received speech recognition into a question text, the problem recognition confirmation system performs word segmentation on the problem text according to a predetermined word segmentation rule, through the word segmentation After the rule, the word segment corresponding to the question text is obtained.
  • the word segmentation includes words and words.
  • the question text may be “Ping An has launched a product of Zunhong Life”, and after the word segmentation, the result is “Ping An”, “Pushing”, “Yes”, “Zhonghong life”, “product”, “?”.
  • a determining module 103 configured to determine a maximum probability corresponding to the predetermined feature word according to a probability distribution between the feature word and the question, after the obtained word segment includes the predetermined feature word, and according to the predetermined problem The mapping relationship with the answer, determining the answer corresponding to the question of the maximum probability;
  • the system has preset feature words (for example, “Zhonghong Life”, “Peace”, etc.), the feature words can reflect the subject or semantic direction of the question corresponding to the question text; the system also has predetermined feature words and questions.
  • the probability distribution between each feature word has a probability value corresponding to each pre-existing problem, and the problem text containing each feature word may be the probability of each problem; the system is also set between the preset question and the answer. Mapping relational tables. After obtaining the word segment corresponding to the question text, the system analyzes whether the obtained word segment contains a predetermined feature word; when the analyzed word segment contains the predetermined feature word, the probability between the feature word and the question is determined.
  • the problem of determining the maximum probability corresponding to the predetermined feature word that is, the most probable problem, after the problem is determined, the problem of the maximum probability is obtained according to the mapping relationship between the predetermined question and the answer.
  • the corresponding answer the determining module 103 also prompts the user to re-question or prompts that the proposed question cannot be recognized after the obtained participle does not contain the predetermined feature word.
  • the feedback module 104 is configured to feed back the determined answer to the user.
  • the system feeds the determined answer to the user by means of voice broadcast or sending to the display device for display or sending to the user's preset terminal.
  • the technical solution of the present invention performs word segmentation on the question text, and obtains a feature word contained in the word segmentation result that reflects the subject or semantic direction of the user question, and is between the feature word and the question.
  • the probability distribution in order to find the problem with the greatest probability (ie the most probable problem), and then determine the answer corresponding to the question of the maximum probability to feed back to the user; because in this technical solution, the feature word can reflect the subject or semantic of the user problem Direction, the corresponding answer found by the problem of the maximum probability corresponding to the feature word. Therefore, compared with the prior art, the similarity between the whole problem and the standard question is compared to obtain the answer corresponding to the most similar question. The accuracy of the answers to the feedback users in this case has improved significantly.
  • the predetermined word segmentation rule is a long word priority word segmentation rule.
  • the long word priority word segmentation rule refers to: for a phrase T1 requiring a word segmentation, starting with the first word A, finding a longest word X1 starting from A from the pre-stored vocabulary, and then culling from T1 X1 has T2 left, and then the same segmentation principle is applied to T2.
  • the result after segmentation is "X1/X2/, ,,,,,”; for example, included in the pre-stored thesaurus
  • the phrase “Ping An Hung Hsing Life Product” is the result of “Peace”/“Push” / " ⁇ " / " ⁇ ” / "product” / "?”.
  • FIG. 6 is a program module diagram of a second embodiment of the problem identification and confirmation system of the present application.
  • the solution of the embodiment is replaced by the following module on the basis of the first embodiment:
  • the first determining sub-module 105 is configured to determine a probability that the predetermined feature word corresponds to each question according to a probability distribution between the feature word and the question, after the obtained word segment includes the predetermined feature word;
  • the second determining sub-module 106 is configured to sort each problem according to a descending order of probabilities, determine a preset number of questions in the prior order as candidate questions, and provide or broadcast the determined candidate questions to the candidate questions. The user makes a selection;
  • the problems are sorted in descending order according to the obtained probabilities, and then the pre-ordered number in the sorted problem sequence is extracted.
  • the questions (for example, 3 or 4) are used as candidate questions, and the extracted candidate questions are fed back to the user for selection by the user.
  • the manner in which the candidate question is fed back to the user may be: 1. voice broadcast; 2. providing a selection interface, and the candidate question is displayed on the selection interface (for example, generating a question selection interface for the user to select, the selection interface may include a candidate question list, Each candidate question in the list corresponds to an "OK" button, the user can click the button to select a corresponding question);
  • the third determining sub-module 107 is configured to determine an answer corresponding to the question according to a mapping relationship between the predetermined question and the answer after the user selects a question.
  • the system After the user makes a selection based on the candidate question of the system feedback, the system receives the question selected by the user, and determines the answer corresponding to the question selected by the user according to the mapping relationship between the predetermined question and the answer in the system. .
  • the probability distribution between the feature words and the problem is determined according to the following steps:
  • an implicit number of predicted numbers (for example, 50) is added between the feature words and the problem layer as an intermediate layer to constitute a problem selection model; wherein the implicit theme is virtual and there is no real Meaning; each implied topic usually contains multiple feature words, each of which usually contains multiple implied topics.
  • the problem text to be trained is obtained (the problem text is prepared in advance), and the obtained problem texts are separately subjected to word segmentation processing, thereby obtaining the word segmentation results corresponding to the respective problem texts.
  • the first number of words is X1
  • the probability of selection of the feature word Y for the hidden subject is: 1/(X1*X2).
  • the mapping relationship between the predetermined implicit topic and the problem text in the system respectively determine the third quantity of the implicit topic contained in each question text and the fourth quantity of the problem text to which each implicit topic belongs, and then according to the corresponding The third quantity and the fourth quantity respectively obtain a second selection probability of each implicit topic for each question text; for example, the fourth quantity of the problem text to which the implicit theme K belongs is J2, and an implicit subject contained in a question text
  • the third quantity is J1
  • the probability of selection of the implicit subject K for the question text is: 1/(J1*J2).
  • the third selection probability is the probability distribution between the feature word and the problem.
  • the third selection probability distribution of the feature word to the problem text can be further obtained.
  • the selection probability of each feature word for each problem text is obtained, that is, the probability distribution between the feature word and the problem is obtained.
  • the first selection probability of the feature word Y for the implicit topic K is: 1/(X1*X2), and the implicit topic K is asked.
  • the second selection probability of the question text W is: 1/(J1*J2), then the selection probability of the feature word Y for the question text W is 1/(X1*X2)*(J1*J2).
  • the present application also provides a computer readable storage medium storing a problem identification confirmation system, the problem identification confirmation system being executable by at least one processor to cause the at least one processor to perform the above-described The problem identification confirmation method described in an embodiment.

Abstract

An electronic device, a question recognition and confirmation method and a-computer-readable storage medium. The method comprises: receiving a question speech pronounced by a user, and carrying out voice recognition on the received question speech to generate a question text (S1); carrying out word segmentation processing on the generated question text according to a predetermined word segmentation rule to obtain segmented words corresponding to the question text (S2); if the obtained segmented words include a predetermined feature word, determining, according to the probability distribution between a feature word and a question, the most likely question corresponding to the predetermined feature word, and determining, according to a predetermined mapping relationship between a question and an answer, an answer corresponding to the most likely question (S3); and feeding the determined answer back to the user (S4). By means of the method, the accuracy of an answer fed back to a user by an intelligent customer service robot and an intelligent customer service response system is improved.

Description

电子装置、问题识别确认方法和计算机可读存储介质Electronic device, problem identification confirmation method, and computer readable storage medium
本申请基于巴黎公约申明享有2017年8月29日递交的申请号为CN 201710754550.6、名称为“电子装置、问题识别确认方法和计算机可读存储介质”中国专利申请的优先权,该中国专利申请的整体内容以参考的方式结合在本申请中。The present application is based on the priority of the Chinese Patent Application entitled "Electronic Device, Problem Identification Confirmation Method, and Computer Readable Storage Medium", filed on August 29, 2017, with the application number of CN 201710754550.6, which is filed on August 29, 2017. The entire content is incorporated herein by reference.
技术领域Technical field
本申请涉及智能语音技术领域,特别涉及一种电子装置、问题识别确认方法和计算机可读存储介质。The present application relates to the field of intelligent voice technologies, and in particular, to an electronic device, a problem recognition confirmation method, and a computer readable storage medium.
背景技术Background technique
目前,为了有效降低客户服务的等待状况、提升服务质量、提高客户服务的便捷性,很多服务型的公司(例如,移动运用商、保险公司、金融机构等)采用了线下智能客服机器人(例如,实体办公区域内设置的智能客服机器人)及/或线上智能客服应答系统(例如,智能语音应答系统)为客户进行服务。这类线下智能客服机器人及/或线上智能客服应答系统通常采用的现有方案是:预先配置标准问题与标准答案的映射关系数据;当接收到客户提出的标准问题后,根据预先配置的标准问题与标准答案的映射关系数据,确定出接收的标准问题对应的标准答案,并将确定出的标准答案反馈给客户。对于用户提出的非标准问题,这种现有方案将难以予以答案反馈。At present, in order to effectively reduce the waiting conditions of customer service, improve service quality, and improve the convenience of customer service, many service companies (for example, mobile operators, insurance companies, financial institutions, etc.) have adopted offline intelligent customer service robots (for example The intelligent customer service robot set in the physical office area and/or an online intelligent customer service response system (for example, an intelligent voice response system) serves the customer. The existing schemes for such offline intelligent customer service robots and/or online intelligent customer service response systems are: pre-configure mapping relationship data between standard questions and standard answers; when receiving standard questions raised by customers, according to pre-configured The mapping relationship between the standard question and the standard answer determines the standard answer corresponding to the received standard question and feeds back the determined standard answer to the customer. For existing non-standard questions, this existing solution will be difficult to give feedback.
虽然,目前市面上存在一种解决非标准问题的改进方案:当无法找到非标准问题对应的答案时,将非标准问题与各个标准问题进行相似度计算,并将最大相似度的标准问题对应的标准答案作为非标准问题对应的答案进行反馈。但是,由于大多数情况下,非标准问题与标准问题之间的相似度都是因为一些与语句含义无关的字(例如,“的”、“吗”)而产生,因此,这种改进方案的准确性很低,经常出错,造成答非所问。Although there is currently an improvement solution on the market to solve non-standard problems: when the answer corresponding to the non-standard question cannot be found, the non-standard question is calculated similarly to each standard question, and the standard question of the maximum similarity is corresponding. The standard answer is fed back as the answer to the non-standard question. However, since in most cases, the similarity between a non-standard problem and a standard question is caused by words that are not related to the meaning of the statement (for example, "", "?"), therefore, the improvement The accuracy is very low, often making mistakes, causing answers.
发明内容Summary of the invention
本申请的主要目的是提供一种问题识别确认方法,旨在提升智能客服系统对非标准问题识别确认的准确性,从而提升针对非标准问题的反馈答案的准确性。The main purpose of the present application is to provide a method for identifying and identifying problems, which aims to improve the accuracy of the intelligent customer service system for identifying and confirming non-standard problems, thereby improving the accuracy of feedback answers to non-standard questions.
本申请第一方面提供一种电子装置,包括存储器、处理器,所述存储器上存储有可在所述处理器上运行的问题识别确认系统,所述问题识别确认系统被所述处理器执行时实现如下步骤: A first aspect of the present application provides an electronic device including a memory, a processor, and a memory identification confirmation system operable on the processor, where the problem recognition confirmation system is executed by the processor Implement the following steps:
接收用户发出的问题语音,对接收的问题语音进行语音识别,生成问题文本;Receiving a question voice sent by the user, performing voice recognition on the received question voice, and generating a question text;
对生成的问题文本按照预先确定的分词规则进行分词处理,获得所述问题文本对应的分词;Performing word segmentation on the generated question text according to a predetermined word segmentation rule to obtain a word segment corresponding to the question text;
若获得的分词中含有预先确定的特征词,则根据特征词与问题之间的概率分布,确定该预先确定的特征词对应的最大概率的问题,并根据预先确定的问题与答案之间的映射关系,确定该最大概率的问题对应的答案;If the obtained participle contains a predetermined feature word, the problem of determining the maximum probability corresponding to the predetermined feature word is determined according to the probability distribution between the feature word and the question, and according to the mapping between the predetermined question and the answer Relationship, the answer corresponding to the question that determines the maximum probability;
将确定的答案反馈给用户。The determined answer is fed back to the user.
本申请第二方面提供一种问题识别确认方法,该方法包括步骤:A second aspect of the present application provides a method for identifying a problem, the method comprising the steps of:
接收用户发出的问题语音,对接收的问题语音进行语音识别,生成问题文本;Receiving a question voice sent by the user, performing voice recognition on the received question voice, and generating a question text;
对生成的问题文本按照预先确定的分词规则进行分词处理,获得所述问题文本对应的分词;Performing word segmentation on the generated question text according to a predetermined word segmentation rule to obtain a word segment corresponding to the question text;
若获得的分词中含有预先确定的特征词,则根据特征词与问题之间的概率分布,确定该预先确定的特征词对应的最大概率的问题,并根据预先确定的问题与答案之间的映射关系,确定该最大概率的问题对应的答案;If the obtained participle contains a predetermined feature word, the problem of determining the maximum probability corresponding to the predetermined feature word is determined according to the probability distribution between the feature word and the question, and according to the mapping between the predetermined question and the answer Relationship, the answer corresponding to the question that determines the maximum probability;
将确定的答案反馈给用户。The determined answer is fed back to the user.
本申请第三方面提供一种计算机可读存储介质,所述计算机可读存储介质存储有问题识别确认系统,所述问题识别确认系统可被至少一个处理器执行,以使所述至少一个处理器执行如下步骤:A third aspect of the present application provides a computer readable storage medium storing a problem identification confirmation system, the problem identification confirmation system being executable by at least one processor to cause the at least one processor Perform the following steps:
接收用户发出的问题语音,对接收的问题语音进行语音识别,生成问题文本;Receiving a question voice sent by the user, performing voice recognition on the received question voice, and generating a question text;
对生成的问题文本按照预先确定的分词规则进行分词处理,获得所述问题文本对应的分词;Performing word segmentation on the generated question text according to a predetermined word segmentation rule to obtain a word segment corresponding to the question text;
若获得的分词中含有预先确定的特征词,则根据特征词与问题之间的概率分布,确定该预先确定的特征词对应的最大概率的问题,并根据预先确定的问题与答案之间的映射关系,确定该最大概率的问题对应的答案;If the obtained participle contains a predetermined feature word, the problem of determining the maximum probability corresponding to the predetermined feature word is determined according to the probability distribution between the feature word and the question, and according to the mapping between the predetermined question and the answer Relationship, the answer corresponding to the question that determines the maximum probability;
将确定的答案反馈给用户。The determined answer is fed back to the user.
本申请技术方案通过将用户的问题语音识别成问题文本后,对问题文本进行分词,获取分词结果中含有的能够反映用户问题的主题或语义方向的特征词,并按特征词与问题之间的概率分布,从而找出最大概率的问题(即最可能的问题),继而确定最大概率的问题对应的答案,以反馈给用户;由于本技术方案中,特征词能够反映用户问题的主题或语义方向,通过特征词对应的最大概率的问题所找到的相应 答案,因此,相较于现有技术采取将整个问题与标准问题进行相似度比较,以获得最相似问题对应的答案的方式而言,本案反馈给用户的答案的准确性显著提高。After the user's question speech is recognized as a question text, the technical solution of the present application performs word segmentation on the question text, and obtains a feature word contained in the word segmentation result that reflects the subject or semantic direction of the user question, and is between the feature word and the question. Probabilistic distribution to find the problem with the greatest probability (ie, the most probable problem), and then determine the answer corresponding to the question with the highest probability to feed back to the user; because in this technical solution, the feature word can reflect the subject or semantic direction of the user problem Corresponding to the problem of the maximum probability corresponding to the feature word The answer, therefore, compared to the prior art, the accuracy of the answer to the user is significantly improved in the way of comparing the similarity of the whole problem with the standard problem to obtain the answer corresponding to the most similar question.
附图说明DRAWINGS
图1为本申请问题识别确认方法一实施例的流程示意图;1 is a schematic flowchart of an embodiment of a method for identifying and confirming a problem of the present application;
图2为本申请问题识别确认方法二实施例的流程示意图;2 is a schematic flowchart of an embodiment of a method for identifying and confirming a problem of the present application;
图3为本申请问题识别确认方法中确定特征词与问题之间的概率分布的流程示意图;FIG. 3 is a schematic flowchart of determining a probability distribution between a feature word and a question in the problem identification and confirmation method of the present application;
图4为本申请问题识别确认系统较佳实施例的运行环境示意图;4 is a schematic diagram of an operating environment of a preferred embodiment of the problem identification and confirmation system of the present application;
图5为本申请问题识别确认系统一实施例的程序模块图;FIG. 5 is a block diagram of a program of an embodiment of the problem identification confirmation system of the present application; FIG.
图6为本申请问题识别确认系统二实施例的程序模块图。FIG. 6 is a program block diagram of a second embodiment of the problem identification and confirmation system of the present application.
具体实施方式Detailed ways
以下结合附图对本申请的原理和特征进行描述,所举实例只用于解释本申请,并非用于限定本申请的范围。The principles and features of the present application are described in the following with reference to the accompanying drawings, which are only used to explain the present application and are not intended to limit the scope of the application.
本申请提出一种问题识别确认方法,主要用于智能客服应答系统或智能客服机器人等智能客服产品。The present application proposes a method for identifying and identifying problems, which is mainly used for intelligent customer service products such as an intelligent customer service response system or an intelligent customer service robot.
如图1所示,图1为本申请问题识别确认方法一实施例的流程示意图。As shown in FIG. 1 , FIG. 1 is a schematic flowchart of an embodiment of a method for identifying and confirming a problem of the present application.
本实施例中,该问题识别确认方法包括:In this embodiment, the method for identifying and identifying the problem includes:
步骤S1,接收用户发出的问题语音,对接收的问题语音进行语音识别,生成问题文本;Step S1: receiving a question voice sent by the user, performing voice recognition on the received question voice, and generating a question text;
当用户向智能客服语音系统或智能客服机器人提问时,问题识别确认系统接收用户提问时发出的问题语音,识别接收到的问题语音并将识别的问题语音生成对应的问题文本。When the user asks a question to the intelligent customer service voice system or the intelligent customer service robot, the problem recognition confirmation system receives the question voice issued by the user when questioning, identifies the received question voice, and generates the corresponding question text of the recognized question voice.
步骤S2,对生成的问题文本按照预先确定的分词规则进行分词处理,获得所述问题文本对应的分词;Step S2, performing segmentation processing on the generated problem text according to a predetermined word segmentation rule, and obtaining a word segment corresponding to the problem text;
在将接收到的问题语音识别转换成问题文本后,问题识别确认系统按照预先确定的分词规则对该问题文本进行分词处理,通过分词处理后,则得到该问题文本对应的分词。本实施例中,所述分词包括字和词,例如:所述问题文本可以是“平安推出了尊宏人生产品吗”,经过分词后的结果为“平安”、“推出”、“了”、“尊宏人生”、“产品”、“吗”。After converting the received question speech recognition into the question text, the question recognition confirming system performs word segmentation processing on the question text according to the predetermined word segmentation rule, and after the word segmentation process, the word segment corresponding to the question text is obtained. In this embodiment, the word segmentation includes words and words. For example, the question text may be “Ping An has launched a product of Zunhong Life”, and after the word segmentation, the result is “Ping An”, “Pushing”, “Yes”, "Zhonghong life", "product", "?".
步骤S3,若获得的分词中含有预先确定的特征词,则根据特征词与问题之间的概率分布,确定该预先确定的特征词对应的最大概率的问题,并根据预先确定的问题与答案之间的映射关系,确定该最大概率的问题对应的答案; Step S3, if the obtained participle contains a predetermined feature word, the problem of determining the maximum probability corresponding to the predetermined feature word is determined according to the probability distribution between the feature word and the question, and according to the predetermined question and answer The mapping relationship between the two, determining the answer corresponding to the question of the maximum probability;
系统中具有预先设定的特征词(例如,“尊宏人生”、“平安”等),特征词能够反映问题文本对应的问题的主题或语义方向;系统中还具有预先确定的特征词与问题之间的概率分布,即每个特征词分别具有与各个预存的问题对应的概率值,含有各个特征词的问题文本可能为各个问题的概率;系统还设置有预设的问题与答案之间的映射关系表。系统在获得所述问题文本对应的分词后,分析获得的分词中是否含有预先确定的特征词;当分析出获得的分词中不含有预先确定的特征词,则提示用户重新提问或者提示无法识别所提问题等处理。当分析出获得的分词中含有预先确定的特征词时,则根据特征词与问题之间的概率分布,确定含有的预先确定的特征词对应的最大概率的问题,即最有可能的问题,在问题确定之后,则根据预先确定的问题与答案之间的映射关系,得到该最大概率的问题所对应的答案。The system has preset feature words (for example, “Zhonghong Life”, “Peace”, etc.), the feature words can reflect the subject or semantic direction of the question corresponding to the question text; the system also has predetermined feature words and questions. The probability distribution between each feature word has a probability value corresponding to each pre-existing problem, and the problem text containing each feature word may be the probability of each problem; the system is also set between the preset question and the answer. Mapping relational tables. After obtaining the word segment corresponding to the question text, the system analyzes whether the obtained word segment contains a predetermined feature word; when the analyzed word segment does not contain the predetermined feature word, the user is prompted to re-question or the prompt is unrecognizable. Ask questions and other treatments. When it is analyzed that the obtained participle contains a predetermined feature word, the problem of determining the maximum probability corresponding to the predetermined feature word is determined according to the probability distribution between the feature word and the question, that is, the most probable problem After the problem is determined, the answer corresponding to the question of the maximum probability is obtained according to the mapping relationship between the predetermined question and the answer.
步骤S4,将确定的答案反馈给用户。In step S4, the determined answer is fed back to the user.
系统在获得确定的答案后,将确定的答案通过语音播报或发送至显示设备显示或发送至用户的预设终端等方式反馈给用户。After obtaining the determined answer, the system feeds the determined answer to the user by means of voice broadcast or sending to the display device for display or sending to the user's preset terminal.
本实施例技术方案通过将用户的问题语音识别成问题文本后,对问题文本进行分词,获取分词结果中含有的能够反映用户问题的主题或语义方向的特征词,并按特征词与问题之间的概率分布,从而找出最大概率的问题(即最可能的问题),继而确定最大概率的问题对应的答案,以反馈给用户;由于本技术方案中,特征词能够反映用户问题的主题或语义方向,通过特征词对应的最大概率的问题所找到的相应答案,因此,相较于现有技术采取将整个问题与标准问题进行相似度比较,以获得最相似问题对应的答案的方式而言,本案反馈用户的答案的准确性显著提高。After the user's question speech is recognized as the question text, the technical solution of the present invention performs word segmentation on the question text, and obtains a feature word contained in the word segmentation result that reflects the subject or semantic direction of the user question, and is between the feature word and the question. The probability distribution, in order to find the problem with the greatest probability (ie the most probable problem), and then determine the answer corresponding to the question of the maximum probability to feed back to the user; because in this technical solution, the feature word can reflect the subject or semantic of the user problem Direction, the corresponding answer found by the problem of the maximum probability corresponding to the feature word. Therefore, compared with the prior art, the similarity between the whole problem and the standard question is compared to obtain the answer corresponding to the most similar question. The accuracy of the answers to the feedback users in this case has improved significantly.
优选地,本实施例中,所述预先确定的分词规则为长词优先分词规则。该长词优先分词规则指的是:对于一个需要分词的短语T1,先从第一个字A开始,从预存的词库找出一个由A起始的最长词语X1,然后从T1中剔除X1剩下T2,再对T2采用相同的切分原理,切分后的结果为“X1/X2/、、、、、、”;例如,在预存的词库中包括“平安”、“推出”、“了”、“尊宏人生”、“产品”、“吗”时,短语“平安推出了尊宏人生产品吗”的切分结果为“平安”/“推出”/“了”/“尊宏人生”/“产品”/“吗”。Preferably, in this embodiment, the predetermined word segmentation rule is a long word priority word segmentation rule. The long word priority word segmentation rule refers to: for a phrase T1 requiring a word segmentation, starting with the first word A, finding a longest word X1 starting from A from the pre-stored vocabulary, and then culling from T1 X1 has T2 left, and then the same segmentation principle is applied to T2. The result after segmentation is “X1/X2/, ,,,,,”; for example, including “Peace” and “Push” in the pre-stored thesaurus. When "Ye", "Zhonghong Life", "Product", "?", the phrase "Ping An has introduced the product of Zunhong Life" is the result of "Peace" / "Push" / "Y" / "Zun" Macro life" / "product" / "?".
如图2所示,图2为本申请问题识别确认方法二实施例的流程示意图,本实施例方案在第一实施例的基础上,将所述步骤S3替换为如下步骤:As shown in FIG. 2, FIG. 2 is a schematic flowchart of the second embodiment of the method for identifying and confirming the problem of the present application. The solution of the embodiment is replaced by the following steps on the basis of the first embodiment:
步骤S301,若获得的分词中含有预先确定的特征词,则根据特征词与问题之间的概率分布,确定该预先确定的特征词对应各个问题 的概率;Step S301, if the obtained word segment contains a predetermined feature word, determining, according to the probability distribution between the feature word and the question, determining the predetermined feature word corresponding to each problem The probability;
当分析出获得的分词中含有预先确定的特征词后,根据系统中预先确定的特征词与问题之间的概率分布,确定得出该获得的分词中含有的预先确定的特征词分别对应各个问题的概率。After analyzing the obtained participle containing the predetermined feature words, according to the probability distribution between the predetermined feature words and the problem in the system, it is determined that the predetermined feature words contained in the obtained word segment respectively correspond to each problem The probability.
步骤S302,按照概率的从大到小的顺序为各个问题进行排序,确定出排序在前的预设数量的问题作为候选问题,并将确定的各个候选问题提供或者播报给用户进行选择;Step S302, sorting each problem according to the order of probability from large to small, determining a preset number of questions in the prior order as candidate questions, and providing or broadcasting the determined candidate questions to the user for selection;
在得出该获得的分词中含有的预先确定的特征词分别对应各个问题的概率后,对各个问题按照得到的概率进行降序排序,再提取排序后的问题序列中的排序在前的预设数量(例如3个、4个)的问题作为候选问题,将提取的候选问题反馈给用户,以供用户进行选择。其中,候选问题反馈给用户的方式可以为:1、语音播报;2、提供选择界面,候选问题显示在选择界面上(例如,生成问题选择界面供用户选择,该选择界面可以包括候选问题列表,所述列表中的每个候选问题对应一个“确定”按钮,用户可以点击所述按钮选择对应的问题);等。After the predetermined feature words contained in the obtained word segment respectively correspond to the probabilities of the respective questions, the problems are sorted in descending order according to the obtained probabilities, and then the pre-ordered number in the sorted problem sequence is extracted. The questions (for example, 3 or 4) are used as candidate questions, and the extracted candidate questions are fed back to the user for selection by the user. The manner in which the candidate question is fed back to the user may be: 1. voice broadcast; 2. providing a selection interface, and the candidate question is displayed on the selection interface (for example, generating a question selection interface for the user to select, the selection interface may include a candidate question list, Each candidate question in the list corresponds to an "OK" button, the user can click the button to select a corresponding question);
步骤S303,在用户选择了一个问题后,根据预先确定的问题与答案之间的映射关系,确定该问题对应的答案。Step S303, after the user selects a question, the answer corresponding to the question is determined according to a mapping relationship between the predetermined question and the answer.
当用户基于系统反馈的候选问题做出选择后,系统接收到用户选择的问题,则根据系统中预先确定的问题与答案之间的映射关系,确定出接收到的用户选择的问题所对应的答案。After the user makes a selection based on the candidate question of the system feedback, the system receives the question selected by the user, and determines the answer corresponding to the question selected by the user according to the mapping relationship between the predetermined question and the answer in the system. .
如图3所示,所述特征词与问题之间的概率分布按照如下步骤确定:As shown in FIG. 3, the probability distribution between the feature words and the problem is determined as follows:
步骤S51,在特征词与问题之间添加预设数量的隐含主题;Step S51, adding a preset number of implicit topics between the feature words and the question;
首先,在特征词与问题这两层之间添加预测数量(例如,50个)的隐含主题,作为中间层,从而构成问题选择模型;其中,所述隐含主题是虚拟的,并没有真实含义;每个隐含主题通常包含多个特征词,每个问题又通常包含多个隐含主题。First, an implicit number of predicted numbers (for example, 50) is added between the feature words and the problem layer as an intermediate layer to constitute a problem selection model; wherein the implicit theme is virtual and there is no real Meaning; each implied topic usually contains multiple feature words, each of which usually contains multiple implied topics.
步骤S52,获取待进行训练的问题文本,并对获取的问题文本分别进行分词处理,得到各个问题文本对应的分词;Step S52: acquiring a problem text to be trained, and performing word segmentation processing on the obtained problem texts respectively, to obtain word segments corresponding to the respective problem texts;
在形成问题选择模型后,获取待进行训练的问题文本(该问题文本为预先准备的),对获取的各个问题文本分别进行分词处理,从而得到各个问题文本对应的分词结果。After the problem selection model is formed, the problem text to be trained is obtained (the problem text is prepared in advance), and the obtained problem texts are separately subjected to word segmentation processing, thereby obtaining the word segmentation results corresponding to the respective problem texts.
步骤S53,根据预先确定的隐含主题与特征词的映射关系,分别确定每个隐含主题含有的特征词的第一数量,分别确定每个特征词所属的隐含主题的第二数量,根据对应的第一数量和第二数量确定每个特征词对各个隐含主题的第一选择概率; Step S53: determining, according to a predetermined mapping relationship between the implicit subject and the feature word, a first quantity of the feature words included in each hidden topic, and respectively determining a second quantity of the hidden topic to which each feature word belongs, according to Corresponding first quantity and second quantity determining a first selection probability of each feature word for each implicit topic;
根据系统中预先确定的隐含主题与特征词的映射关系,分别确定出每个隐含主题中含有的特征词的第一数量及每个特征词所属的隐含主题的第二数量,再根据相应的第一数量与第二数量分别得到每个特征词对各个隐含主题的第一选择概率;例如,特征词Y所属的隐含主题的第二数量为X2,一个隐含主题含有的特征词的第一数量为X1,则该特征词Y对该隐含主题的选择概率为:1/(X1*X2)。Determining, according to a predetermined mapping relationship between the implicit subject and the feature words in the system, the first quantity of the feature words contained in each hidden topic and the second quantity of the hidden topic to which each feature word belongs, and then according to Corresponding first quantity and second quantity respectively obtain a first selection probability of each feature word for each implicit topic; for example, the second quantity of the implicit topic to which the feature word Y belongs is X2, and the feature of an implicit topic The first number of words is X1, and the probability of selection of the feature word Y for the hidden subject is: 1/(X1*X2).
步骤S54,根据预先确定的隐含主题与问题文本的映射关系,分别确定每个问题文本含有的隐含主题的第三数量,分别确定每个隐含主题所属的问题文本的第四数量,根据对应的第三数量和第四数量确定每个隐含主题对各个问题文本的第二选择概率;Step S54: Determine a third quantity of the hidden topic included in each question text according to a predetermined mapping relationship between the implicit topic and the question text, and determine a fourth quantity of the problem text to which each hidden topic belongs, according to Corresponding third and fourth quantities determine a second selection probability of each implicit topic for each question text;
根据系统中预先确定的隐含主题与问题文本的映射关系,分别确定每个问题文本中含有的隐含主题的第三数量及每个隐含主题所属的问题文本的第四数量,再根据相应的第三数量与第四数量分别得到每个隐含主题对各个问题文本的第二选择概率;例如,隐含主题K所属的问题文本的第四数量为J2,一个问题文本含有的隐含主题的第三数量为J1,则该隐含主题K对该问题文本的选择概率为:1/(J1*J2)。本实施例中,所述步骤S54与步骤S53的顺序可调换。According to the mapping relationship between the predetermined implicit topic and the problem text in the system, respectively determine the third quantity of the implicit topic contained in each question text and the fourth quantity of the problem text to which each implicit topic belongs, and then according to the corresponding The third quantity and the fourth quantity respectively obtain a second selection probability of each implicit topic for each question text; for example, the fourth quantity of the problem text to which the implicit theme K belongs is J2, and an implicit subject contained in a question text The third quantity is J1, then the probability of selection of the implicit subject K for the question text is: 1/(J1*J2). In this embodiment, the sequence of the step S54 and the step S53 can be changed.
步骤S55,将对应的第一选择概率和第二选择概率代入预先确定的计算公式进行计算,计算出每个特征词对各个问题文本的第三选择概率,计算出的各个特征词分别对各个问题文本的第三选择概率即为特征词与问题之间的概率分布。Step S55: Substituting the corresponding first selection probability and the second selection probability into a predetermined calculation formula to calculate, calculating a third selection probability of each feature word for each problem text, and calculating each feature word separately for each problem The third selection probability of the text is the probability distribution between the feature words and the problem.
根据特征词对隐含主题的第一选择概率分布,以及隐含主题对问题文本的第二选择概率分布,进一步即可得出特征词对问题文本的第三选择概率分布。具体的,通过将对应的第一选择概率与第二选择概率代入预先确定的计算公式计算,得出各个特征词分别对各个问题文本的选择概率,即得到特征词与问题之间的概率分布。本实施例中,该预先确定的计算公式为:P3=P1*P2,其中,P1代表第一选择概率,P2代表第二选择概率,P3代表第三选择概率。例如,特征词Y对隐含主题K的第一选择概率为:1/(X1*X2),隐含主题K对问题文本W的第二选择概率为:1/(J1*J2),那么特征词Y对问题文本W的选择概率则为1/(X1*X2)*(J1*J2)。According to the first selection probability distribution of the feature word to the implicit topic, and the second selection probability distribution of the implicit topic to the question text, the third selection probability distribution of the feature word to the problem text can be further obtained. Specifically, by substituting the corresponding first selection probability and the second selection probability into a predetermined calculation formula, the selection probability of each feature word for each problem text is obtained, that is, the probability distribution between the feature word and the problem is obtained. In this embodiment, the predetermined calculation formula is: P3=P1*P2, where P1 represents a first selection probability, P2 represents a second selection probability, and P3 represents a third selection probability. For example, the first selection probability of the feature word Y for the implicit topic K is: 1/(X1*X2), and the second selection probability of the implicit topic K for the question text W is: 1/(J1*J2), then the feature The probability of selection of the word Y for the question text W is 1/(X1*X2)*(J1*J2).
本申请还提出一种问题识别确认系统。The present application also proposes a problem identification confirmation system.
请参阅图4,是本申请问题识别确认系统10较佳实施例的运行环境示意图。Please refer to FIG. 4 , which is a schematic diagram of an operating environment of a preferred embodiment of the problem identification and confirmation system 10 of the present application.
在本实施例中,问题识别确认系统10安装并运行于电子装置1中。电子装置1可以是桌上型计算机、笔记本、掌上电脑及服务器等计算设备。该电子装置1可包括,但不仅限于,存储器11、处理器 12及显示器13。图4仅示出了具有组件11-13的电子装置1,但是应理解的是,并不要求实施所有示出的组件,可以替代的实施更多或者更少的组件。In the present embodiment, the problem recognition confirmation system 10 is installed and operated in the electronic device 1. The electronic device 1 may be a computing device such as a desktop computer, a notebook, a palmtop computer, and a server. The electronic device 1 may include, but is not limited to, a memory 11, a processor 12 and display 13. Figure 4 shows only the electronic device 1 with components 11-13, but it should be understood that not all illustrated components may be implemented, and more or fewer components may be implemented instead.
存储器11是一种计算机存储介质,在一些实施例中可以是电子装置1的内部存储单元,例如该电子装置1的硬盘或内存。存储器11在另一些实施例中也可以是电子装置1的外部存储设备,例如电子装置1上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。进一步地,存储器11还可以既包括电子装置1的内部存储单元也包括外部存储设备。存储器11用于存储安装于电子装置1的应用软件及各类数据,例如问题识别确认系统10的程序代码等。存储器11还可以用于暂时地存储已经输出或者将要输出的数据。The memory 11 is a computer storage medium, which in some embodiments may be an internal storage unit of the electronic device 1, such as a hard disk or memory of the electronic device 1. The memory 11 may also be an external storage device of the electronic device 1 in other embodiments, such as a plug-in hard disk equipped on the electronic device 1, a smart memory card (SMC), and a secure digital (SD). Card, flash card, etc. Further, the memory 11 may also include both an internal storage unit of the electronic device 1 and an external storage device. The memory 11 is used to store application software and various types of data installed in the electronic device 1, such as program codes of the problem recognition confirmation system 10. The memory 11 can also be used to temporarily store data that has been output or is about to be output.
处理器12在一些实施例中可以是一中央处理器(Central Processing Unit,CPU),微处理器或其他数据处理芯片,用于运行存储器11中存储的程序代码或处理数据,例如执行问题识别确认系统10等。The processor 12, in some embodiments, may be a Central Processing Unit (CPU), microprocessor or other data processing chip for running program code or processing data stored in the memory 11, such as performing problem identification confirmation. System 10 and so on.
显示器13在一些实施例中可以是LED显示器、液晶显示器、触控式液晶显示器以及OLED(Organic Light-Emitting Diode,有机发光二极管)触摸器等。显示器13用于显示在电子装置1中处理的信息以及用于显示可视化的用户界面,例如业务定制界面等。电子装置1的部件11-13通过系统总线相互通信。The display 13 may be an LED display, a liquid crystal display, a touch-sensitive liquid crystal display, an OLED (Organic Light-Emitting Diode) touch sensor, or the like in some embodiments. The display 13 is for displaying information processed in the electronic device 1 and a user interface for displaying visualization, such as a business customization interface or the like. The components 11-13 of the electronic device 1 communicate with one another via a system bus.
请参阅图5,是本申请问题识别确认系统10一实施例的程序模块图。在本实施例中,问题识别确认系统10可以被分割成一个或多个模块,一个或者多个模块被存储于存储器11中,并由一个或多个处理器(本实施例为处理器12)所执行,以完成本申请。例如,在图5中,问题识别确认系统10可以被分割成识别模块101、分词模块102、确定模块103及反馈模块104。本申请所称的模块是指能够完成特定功能的一系列计算机程序指令段,比程序更适合于描述问题识别确认系统10在电子装置1中的执行过程,其中:Please refer to FIG. 5, which is a program module diagram of an embodiment of the problem identification confirmation system 10 of the present application. In the present embodiment, the problem identification confirmation system 10 can be divided into one or more modules, one or more modules are stored in the memory 11, and by one or more processors (the processor 12 in this embodiment) Executed to complete the application. For example, in FIG. 5, the problem identification confirmation system 10 can be divided into an identification module 101, a word segmentation module 102, a determination module 103, and a feedback module 104. A module referred to in this application refers to a series of computer program instruction segments capable of performing a specific function, and is more suitable than the program for describing the execution process of the problem recognition confirmation system 10 in the electronic device 1, wherein:
识别模块101,用于接收用户发出的问题语音,对接收的问题语音进行语音识别,生成问题文本;The identification module 101 is configured to receive a question voice sent by the user, perform voice recognition on the received question voice, and generate a question text.
当用户向智能客服语音系统或智能客服机器人提问时,问题识别确认系统接收用户提问时发出的问题语音,识别接收到的问题语音并将识别的问题语音生成对应的问题文本。When the user asks a question to the intelligent customer service voice system or the intelligent customer service robot, the problem recognition confirmation system receives the question voice issued by the user when questioning, identifies the received question voice, and generates the corresponding question text of the recognized question voice.
分词模块102,用于对生成的问题文本按照预先确定的分词规则进行分词处理,获得所述问题文本对应的分词;The word segmentation module 102 is configured to perform word segmentation processing on the generated problem text according to a predetermined word segmentation rule to obtain a word segment corresponding to the problem text;
在将接收到的问题语音识别转换成问题文本后,问题识别确认系统按照预先确定的分词规则对该问题文本进行分词处理,通过分词处 理后,则得到该问题文本对应的分词。本实施例中,所述分词包括字和词,例如:所述问题文本可以是“平安推出了尊宏人生产品吗”,经过分词后的结果为“平安”、“推出”、“了”、“尊宏人生”、“产品”、“吗”。After converting the received speech recognition into a question text, the problem recognition confirmation system performs word segmentation on the problem text according to a predetermined word segmentation rule, through the word segmentation After the rule, the word segment corresponding to the question text is obtained. In this embodiment, the word segmentation includes words and words. For example, the question text may be “Ping An has launched a product of Zunhong Life”, and after the word segmentation, the result is “Ping An”, “Pushing”, “Yes”, "Zhonghong life", "product", "?".
确定模块103,用于在获得的分词中含有预先确定的特征词后,根据特征词与问题之间的概率分布,确定该预先确定的特征词对应的最大概率的问题,并根据预先确定的问题与答案之间的映射关系,确定该最大概率的问题对应的答案;a determining module 103, configured to determine a maximum probability corresponding to the predetermined feature word according to a probability distribution between the feature word and the question, after the obtained word segment includes the predetermined feature word, and according to the predetermined problem The mapping relationship with the answer, determining the answer corresponding to the question of the maximum probability;
系统中具有预先设定的特征词(例如,“尊宏人生”、“平安”等),特征词能够反映问题文本对应的问题的主题或语义方向;系统中还具有预先确定的特征词与问题之间的概率分布,即每个特征词分别具有与各个预存的问题对应的概率值,含有各个特征词的问题文本可能为各个问题的概率;系统还设置有预设的问题与答案之间的映射关系表。系统在获得所述问题文本对应的分词后,分析获得的分词中是否含有预先确定的特征词;当分析出获得的分词中含有预先确定的特征词时,则根据特征词与问题之间的概率分布,确定含有的预先确定的特征词对应的最大概率的问题,即最有可能的问题,在问题确定之后,则根据预先确定的问题与答案之间的映射关系,得到该最大概率的问题所对应的答案。另外,确定模块103还在分析出获得的分词中不含有预先确定的特征词后,提示用户重新提问或者提示无法识别所提问题等处理。The system has preset feature words (for example, “Zhonghong Life”, “Peace”, etc.), the feature words can reflect the subject or semantic direction of the question corresponding to the question text; the system also has predetermined feature words and questions. The probability distribution between each feature word has a probability value corresponding to each pre-existing problem, and the problem text containing each feature word may be the probability of each problem; the system is also set between the preset question and the answer. Mapping relational tables. After obtaining the word segment corresponding to the question text, the system analyzes whether the obtained word segment contains a predetermined feature word; when the analyzed word segment contains the predetermined feature word, the probability between the feature word and the question is determined. Distribution, the problem of determining the maximum probability corresponding to the predetermined feature word, that is, the most probable problem, after the problem is determined, the problem of the maximum probability is obtained according to the mapping relationship between the predetermined question and the answer. The corresponding answer. In addition, the determining module 103 also prompts the user to re-question or prompts that the proposed question cannot be recognized after the obtained participle does not contain the predetermined feature word.
反馈模块104,用于将确定的答案反馈给用户。The feedback module 104 is configured to feed back the determined answer to the user.
系统在获得确定的答案后,将确定的答案通过语音播报或发送至显示设备显示或发送至用户的预设终端等方式反馈给用户。After obtaining the determined answer, the system feeds the determined answer to the user by means of voice broadcast or sending to the display device for display or sending to the user's preset terminal.
本实施例技术方案通过将用户的问题语音识别成问题文本后,对问题文本进行分词,获取分词结果中含有的能够反映用户问题的主题或语义方向的特征词,并按特征词与问题之间的概率分布,从而找出最大概率的问题(即最可能的问题),继而确定最大概率的问题对应的答案,以反馈给用户;由于本技术方案中,特征词能够反映用户问题的主题或语义方向,通过特征词对应的最大概率的问题所找到的相应答案,因此,相较于现有技术采取将整个问题与标准问题进行相似度比较,以获得最相似问题对应的答案的方式而言,本案反馈用户的答案的准确性显著提高。After the user's question speech is recognized as the question text, the technical solution of the present invention performs word segmentation on the question text, and obtains a feature word contained in the word segmentation result that reflects the subject or semantic direction of the user question, and is between the feature word and the question. The probability distribution, in order to find the problem with the greatest probability (ie the most probable problem), and then determine the answer corresponding to the question of the maximum probability to feed back to the user; because in this technical solution, the feature word can reflect the subject or semantic of the user problem Direction, the corresponding answer found by the problem of the maximum probability corresponding to the feature word. Therefore, compared with the prior art, the similarity between the whole problem and the standard question is compared to obtain the answer corresponding to the most similar question. The accuracy of the answers to the feedback users in this case has improved significantly.
优选地,本实施例中,所述预先确定的分词规则为长词优先分词规则。该长词优先分词规则指的是:对于一个需要分词的短语T1,先从第一个字A开始,从预存的词库找出一个由A起始的最长词语X1,然后从T1中剔除X1剩下T2,再对T2采用相同的切分原理,切分后的结果为“X1/X2/、、、、、、”;例如,在预存的词库中包括 “平安”、“推出”、“了”、“尊宏人生”、“产品”、“吗”时,短语“平安推出了尊宏人生产品吗”的切分结果为“平安”/“推出”/“了”/“尊宏人生”/“产品”/“吗”。Preferably, in this embodiment, the predetermined word segmentation rule is a long word priority word segmentation rule. The long word priority word segmentation rule refers to: for a phrase T1 requiring a word segmentation, starting with the first word A, finding a longest word X1 starting from A from the pre-stored vocabulary, and then culling from T1 X1 has T2 left, and then the same segmentation principle is applied to T2. The result after segmentation is "X1/X2/, ,,,,,"; for example, included in the pre-stored thesaurus When “Peace”, “Publish”, “Yes”, “Zhonghong Life”, “Product”, “?”, the phrase “Ping An Hung Hsing Life Product” is the result of “Peace”/“Push” / "了" / "尊宏人生" / "product" / "?".
如图6所示,图6为本申请问题识别确认系统二实施例的程序模块图,本实施例方案在第一实施例的基础上,将所述确定模块103替换为如下模块:As shown in FIG. 6, FIG. 6 is a program module diagram of a second embodiment of the problem identification and confirmation system of the present application. The solution of the embodiment is replaced by the following module on the basis of the first embodiment:
第一确定子模块105,用于在获得的分词中含有预先确定的特征词后,根据特征词与问题之间的概率分布,确定该预先确定的特征词对应各个问题的概率;The first determining sub-module 105 is configured to determine a probability that the predetermined feature word corresponds to each question according to a probability distribution between the feature word and the question, after the obtained word segment includes the predetermined feature word;
当分析出获得的分词中含有预先确定的特征词后,根据系统中预先确定的特征词与问题之间的概率分布,确定得出该获得的分词中含有的预先确定的特征词分别对应各个问题的概率。After analyzing the obtained participle containing the predetermined feature words, according to the probability distribution between the predetermined feature words and the problem in the system, it is determined that the predetermined feature words contained in the obtained word segment respectively correspond to each problem The probability.
第二确定子模块106,用于按照概率的从大到小的顺序为各个问题进行排序,确定出排序在前的预设数量的问题作为候选问题,并将确定的各个候选问题提供或者播报给用户进行选择;The second determining sub-module 106 is configured to sort each problem according to a descending order of probabilities, determine a preset number of questions in the prior order as candidate questions, and provide or broadcast the determined candidate questions to the candidate questions. The user makes a selection;
在得出该获得的分词中含有的预先确定的特征词分别对应各个问题的概率后,对各个问题按照得到的概率进行降序排序,再提取排序后的问题序列中的排序在前的预设数量(例如3个、4个)的问题作为候选问题,将提取的候选问题反馈给用户,以供用户进行选择。其中,候选问题反馈给用户的方式可以为:1、语音播报;2、提供选择界面,候选问题显示在选择界面上(例如,生成问题选择界面供用户选择,该选择界面可以包括候选问题列表,所述列表中的每个候选问题对应一个“确定”按钮,用户可以点击所述按钮选择对应的问题);等。After the predetermined feature words contained in the obtained word segment respectively correspond to the probabilities of the respective questions, the problems are sorted in descending order according to the obtained probabilities, and then the pre-ordered number in the sorted problem sequence is extracted. The questions (for example, 3 or 4) are used as candidate questions, and the extracted candidate questions are fed back to the user for selection by the user. The manner in which the candidate question is fed back to the user may be: 1. voice broadcast; 2. providing a selection interface, and the candidate question is displayed on the selection interface (for example, generating a question selection interface for the user to select, the selection interface may include a candidate question list, Each candidate question in the list corresponds to an "OK" button, the user can click the button to select a corresponding question);
第三确定子模块107,用于在用户选择了一个问题后,根据预先确定的问题与答案之间的映射关系,确定该问题对应的答案。The third determining sub-module 107 is configured to determine an answer corresponding to the question according to a mapping relationship between the predetermined question and the answer after the user selects a question.
当用户基于系统反馈的候选问题做出选择后,系统接收到用户选择的问题,则根据系统中预先确定的问题与答案之间的映射关系,确定出接收到的用户选择的问题所对应的答案。After the user makes a selection based on the candidate question of the system feedback, the system receives the question selected by the user, and determines the answer corresponding to the question selected by the user according to the mapping relationship between the predetermined question and the answer in the system. .
优选地,本实施例中,所述特征词与问题之间的概率分布按照如下步骤确定:Preferably, in this embodiment, the probability distribution between the feature words and the problem is determined according to the following steps:
1、在特征词与问题之间添加预设数量的隐含主题;1. Add a preset number of implied topics between feature words and questions;
首先,在特征词与问题这两层之间添加预测数量(例如,50个)的隐含主题,作为中间层,从而构成问题选择模型;其中,所述隐含主题是虚拟的,并没有真实含义;每个隐含主题通常包含多个特征词,每个问题又通常包含多个隐含主题。 First, an implicit number of predicted numbers (for example, 50) is added between the feature words and the problem layer as an intermediate layer to constitute a problem selection model; wherein the implicit theme is virtual and there is no real Meaning; each implied topic usually contains multiple feature words, each of which usually contains multiple implied topics.
2、获取待进行训练的问题文本,并对获取的问题文本分别进行分词处理,得到各个问题文本对应的分词;2. Obtaining the problem text to be trained, and separately performing word segmentation on the obtained problem text, and obtaining the word segment corresponding to each question text;
在形成问题选择模型后,获取待进行训练的问题文本(该问题文本为预先准备的),对获取的各个问题文本分别进行分词处理,从而得到各个问题文本对应的分词结果。After the problem selection model is formed, the problem text to be trained is obtained (the problem text is prepared in advance), and the obtained problem texts are separately subjected to word segmentation processing, thereby obtaining the word segmentation results corresponding to the respective problem texts.
3、根据预先确定的隐含主题与特征词的映射关系,分别确定每个隐含主题含有的特征词的第一数量,分别确定每个特征词所属的隐含主题的第二数量,根据对应的第一数量和第二数量确定每个特征词对各个隐含主题的第一选择概率;3. According to the mapping relationship between the predetermined implicit topic and the feature word, respectively determine the first quantity of the feature words contained in each hidden topic, and determine the second quantity of the hidden topic to which each feature word belongs, according to the corresponding The first quantity and the second quantity determine a first selection probability of each feature word for each implicit topic;
根据系统中预先确定的隐含主题与特征词的映射关系,分别确定出每个隐含主题中含有的特征词的第一数量及每个特征词所属的隐含主题的第二数量,再根据相应的第一数量与第二数量分别得到每个特征词对各个隐含主题的第一选择概率;例如,特征词Y所属的隐含主题的第二数量为X2,一个隐含主题含有的特征词的第一数量为X1,则该特征词Y对该隐含主题的选择概率为:1/(X1*X2)。Determining, according to a predetermined mapping relationship between the implicit subject and the feature words in the system, the first quantity of the feature words contained in each hidden topic and the second quantity of the hidden topic to which each feature word belongs, and then according to Corresponding first quantity and second quantity respectively obtain a first selection probability of each feature word for each implicit topic; for example, the second quantity of the implicit topic to which the feature word Y belongs is X2, and the feature of an implicit topic The first number of words is X1, and the probability of selection of the feature word Y for the hidden subject is: 1/(X1*X2).
4、根据预先确定的隐含主题与问题文本的映射关系,分别确定每个问题文本含有的隐含主题的第三数量,分别确定每个隐含主题所属的问题文本的第四数量,根据对应的第三数量和第四数量确定每个隐含主题对各个问题文本的第二选择概率;4. According to the mapping relationship between the predetermined implicit topic and the problem text, respectively determine the third quantity of the implicit topic contained in each question text, and determine the fourth quantity of the problem text to which each hidden topic belongs, according to the corresponding The third quantity and the fourth quantity determine a second selection probability of each implicit topic for each question text;
根据系统中预先确定的隐含主题与问题文本的映射关系,分别确定每个问题文本中含有的隐含主题的第三数量及每个隐含主题所属的问题文本的第四数量,再根据相应的第三数量与第四数量分别得到每个隐含主题对各个问题文本的第二选择概率;例如,隐含主题K所属的问题文本的第四数量为J2,一个问题文本含有的隐含主题的第三数量为J1,则该隐含主题K对该问题文本的选择概率为:1/(J1*J2)。According to the mapping relationship between the predetermined implicit topic and the problem text in the system, respectively determine the third quantity of the implicit topic contained in each question text and the fourth quantity of the problem text to which each implicit topic belongs, and then according to the corresponding The third quantity and the fourth quantity respectively obtain a second selection probability of each implicit topic for each question text; for example, the fourth quantity of the problem text to which the implicit theme K belongs is J2, and an implicit subject contained in a question text The third quantity is J1, then the probability of selection of the implicit subject K for the question text is: 1/(J1*J2).
5、将对应的第一选择概率和第二选择概率代入预先确定的计算公式进行计算,计算出每个特征词对各个问题文本的第三选择概率,计算出的各个特征词分别对各个问题文本的第三选择概率即为特征词与问题之间的概率分布。5. Substituting the corresponding first selection probability and the second selection probability into a predetermined calculation formula to calculate, and calculating a third selection probability of each feature word for each problem text, and calculating each feature word separately for each problem text The third selection probability is the probability distribution between the feature word and the problem.
根据特征词对隐含主题的第一选择概率分布,以及隐含主题对问题文本的第二选择概率分布,进一步即可得出特征词对问题文本的第三选择概率分布。具体的,通过将对应的第一选择概率与第二选择概率代入预先确定的计算公式计算,得出各个特征词分别对各个问题文本的选择概率,即得到特征词与问题之间的概率分布。本实施例中,该预先确定的计算公式为:P3=P1*P2,其中,P1代表第一选择概率,P2代表第二选择概率,P3代表第三选择概率。例如,特征词Y对隐含主题K的第一选择概率为:1/(X1*X2),隐含主题K对问 题文本W的第二选择概率为:1/(J1*J2),那么特征词Y对问题文本W的选择概率则为1/(X1*X2)*(J1*J2)。According to the first selection probability distribution of the feature word to the implicit topic, and the second selection probability distribution of the implicit topic to the question text, the third selection probability distribution of the feature word to the problem text can be further obtained. Specifically, by substituting the corresponding first selection probability and the second selection probability into a predetermined calculation formula, the selection probability of each feature word for each problem text is obtained, that is, the probability distribution between the feature word and the problem is obtained. In this embodiment, the predetermined calculation formula is: P3=P1*P2, where P1 represents a first selection probability, P2 represents a second selection probability, and P3 represents a third selection probability. For example, the first selection probability of the feature word Y for the implicit topic K is: 1/(X1*X2), and the implicit topic K is asked. The second selection probability of the question text W is: 1/(J1*J2), then the selection probability of the feature word Y for the question text W is 1/(X1*X2)*(J1*J2).
本申请还提出一种计算机可读存储介质,该计算机可读存储介质存储有问题识别确认系统,所述问题识别确认系统可被至少一个处理器执行,以使所述至少一个处理器执行上述任一实施例中所述的问题识别确认方法。The present application also provides a computer readable storage medium storing a problem identification confirmation system, the problem identification confirmation system being executable by at least one processor to cause the at least one processor to perform the above-described The problem identification confirmation method described in an embodiment.
以上所述仅为本发明的优选实施例,并非因此限制本发明的专利范围,凡是在本发明的发明构思下,利用本发明说明书及附图内容所作的等效结构变换,或直接/间接运用在其他相关的技术领域均包括在本发明的专利保护范围内。 The above description is only a preferred embodiment of the present invention, and is not intended to limit the scope of the invention, and the equivalent structural transformation, or direct/indirect use, of the present invention and the contents of the drawings are used in the inventive concept of the present invention. It is included in the scope of the patent protection of the present invention in other related technical fields.

Claims (20)

  1. 一种电子装置,其特征在于,所述电子装置包括存储器、处理器,所述存储器上存储有可在所述处理器上运行的问题识别确认系统,所述问题识别确认系统被所述处理器执行时实现如下步骤:An electronic device, comprising: a memory, a processor, wherein the memory stores a problem identification confirmation system operable on the processor, the problem identification confirmation system being the processor The following steps are implemented during execution:
    S1、接收用户发出的问题语音,对接收的问题语音进行语音识别,生成问题文本;S1: receiving a question voice sent by the user, performing voice recognition on the received question voice, and generating a question text;
    S2、对生成的问题文本按照预先确定的分词规则进行分词处理,获得所述问题文本对应的分词;S2. Performing word segmentation on the generated problem text according to a predetermined word segmentation rule to obtain a word segment corresponding to the question text;
    S3、若获得的分词中含有预先确定的特征词,则根据特征词与问题之间的概率分布,确定该预先确定的特征词对应的最大概率的问题,并根据预先确定的问题与答案之间的映射关系,确定该最大概率的问题对应的答案;S3. If the obtained participle contains a predetermined feature word, the problem of determining the maximum probability corresponding to the predetermined feature word is determined according to the probability distribution between the feature word and the question, and according to the predetermined question and answer Mapping relationship, determining the answer corresponding to the question of the maximum probability;
    S4、将确定的答案反馈给用户。S4. The determined answer is fed back to the user.
  2. 如权利要求1所述的电子装置,其特征在于,所述特征词与问题之间的概率分布按照如下步骤确定:The electronic device according to claim 1, wherein the probability distribution between the feature word and the question is determined as follows:
    在特征词与问题之间添加预设数量的隐含主题;Add a preset number of implied topics between feature words and questions;
    获取待进行训练的问题文本,并对获取的问题文本分别进行分词处理,得到各个问题文本对应的分词;Obtaining the problem text to be trained, and separately performing word segmentation on the obtained problem text, and obtaining the word segment corresponding to each question text;
    根据预先确定的隐含主题与特征词的映射关系,分别确定每个隐含主题含有的特征词的第一数量,分别确定每个特征词所属的隐含主题的第二数量,根据对应的第一数量和第二数量确定每个特征词对各个隐含主题的第一选择概率;Determining, according to a predetermined mapping relationship between the implicit subject and the feature word, a first quantity of the feature words contained in each hidden topic, and respectively determining a second quantity of the hidden subject to which each feature word belongs, according to the corresponding number A quantity and a second quantity determine a first selection probability of each feature word for each implicit topic;
    根据预先确定的隐含主题与问题文本的映射关系,分别确定每个问题文本含有的隐含主题的第三数量,分别确定每个隐含主题所属的问题文本的第四数量,根据对应的第三数量和第四数量确定每个隐含主题对各个问题文本的第二选择概率;Determining, according to a predetermined mapping relationship between the implicit topic and the problem text, respectively determining a third quantity of the implicit topic contained in each question text, and respectively determining a fourth quantity of the problem text to which each implicit topic belongs, according to the corresponding number The third quantity and the fourth quantity determine a second selection probability of each implicit topic for each question text;
    将对应的第一选择概率和第二选择概率代入预先确定的计算公式进行计算,计算出每个特征词对各个问题文本的第三选择概率,计算出的各个特征词分别对各个问题文本的第三选择概率即为特征词与问题之间的概率分布。Substituting the corresponding first selection probability and the second selection probability into a predetermined calculation formula to calculate, calculating a third selection probability of each feature word for each problem text, and calculating each feature word separately for each question text The three-choice probability is the probability distribution between the feature word and the problem.
  3. 如权利要求2所述的电子装置,其特征在于,所述预先确定的计算公式为:The electronic device according to claim 2, wherein said predetermined calculation formula is:
    P3=P1*P2,其中,P1代表第一选择概率,P2代表第二选择概率,P3代表第三选择概率。P3 = P1 * P2, where P1 represents the first selection probability, P2 represents the second selection probability, and P3 represents the third selection probability.
  4. 如权利要求1所述的电子装置,其特征在于,所述步骤S3替 换为如下步骤:The electronic device according to claim 1, wherein said step S3 is replaced Change to the following steps:
    若获得的分词中含有预先确定的特征词,则根据特征词与问题之间的概率分布,确定该预先确定的特征词对应各个问题的概率;If the obtained participle contains a predetermined feature word, determining a probability that the predetermined feature word corresponds to each question according to a probability distribution between the feature word and the question;
    按照概率的从大到小的顺序为各个问题进行排序,确定出排序在前的预设数量的问题作为候选问题,并将确定的各个候选问题提供或者播报给用户进行选择;Sorting each problem according to the order of probability, determining the pre-sorted number of questions as candidate questions, and providing or broadcasting the determined candidate questions to the user for selection;
    在用户选择了一个问题后,根据预先确定的问题与答案之间的映射关系,确定该问题对应的答案。After the user selects a question, the answer corresponding to the question is determined according to the mapping relationship between the predetermined question and the answer.
  5. 如权利要求4所述的电子装置,其特征在于,所述特征词与问题之间的概率分布按照如下步骤确定:The electronic device according to claim 4, wherein the probability distribution between the feature word and the question is determined as follows:
    在特征词与问题之间添加预设数量的隐含主题;Add a preset number of implied topics between feature words and questions;
    获取待进行训练的问题文本,并对获取的问题文本分别进行分词处理,得到各个问题文本对应的分词;Obtaining the problem text to be trained, and separately performing word segmentation on the obtained problem text, and obtaining the word segment corresponding to each question text;
    根据预先确定的隐含主题与特征词的映射关系,分别确定每个隐含主题含有的特征词的第一数量,分别确定每个特征词所属的隐含主题的第二数量,根据对应的第一数量和第二数量确定每个特征词对各个隐含主题的第一选择概率;Determining, according to a predetermined mapping relationship between the implicit subject and the feature word, a first quantity of the feature words contained in each hidden topic, and respectively determining a second quantity of the hidden subject to which each feature word belongs, according to the corresponding number A quantity and a second quantity determine a first selection probability of each feature word for each implicit topic;
    根据预先确定的隐含主题与问题文本的映射关系,分别确定每个问题文本含有的隐含主题的第三数量,分别确定每个隐含主题所属的问题文本的第四数量,根据对应的第三数量和第四数量确定每个隐含主题对各个问题文本的第二选择概率;Determining, according to a predetermined mapping relationship between the implicit topic and the problem text, respectively determining a third quantity of the implicit topic contained in each question text, and respectively determining a fourth quantity of the problem text to which each implicit topic belongs, according to the corresponding number The third quantity and the fourth quantity determine a second selection probability of each implicit topic for each question text;
    将对应的第一选择概率和第二选择概率代入预先确定的计算公式进行计算,计算出每个特征词对各个问题文本的第三选择概率,计算出的各个特征词分别对各个问题文本的第三选择概率即为特征词与问题之间的概率分布。Substituting the corresponding first selection probability and the second selection probability into a predetermined calculation formula to calculate, calculating a third selection probability of each feature word for each problem text, and calculating each feature word separately for each question text The three-choice probability is the probability distribution between the feature word and the problem.
  6. 如权利要求5所述的电子装置,其特征在于,所述预先确定的计算公式为:The electronic device according to claim 5, wherein said predetermined calculation formula is:
    P3=P1*P2,其中,P1代表第一选择概率,P2代表第二选择概率,P3代表第三选择概率。P3 = P1 * P2, where P1 represents the first selection probability, P2 represents the second selection probability, and P3 represents the third selection probability.
  7. 如权利要求1所述的电子装置,其特征在于,所述预先确定的分词规则为长词优先分词规则。The electronic device according to claim 1, wherein said predetermined word segmentation rule is a long word priority word segmentation rule.
  8. 如权利要求7所述的电子装置,其特征在于,所述特征词与问题之间的概率分布按照如下步骤确定:The electronic device according to claim 7, wherein the probability distribution between the feature word and the question is determined as follows:
    在特征词与问题之间添加预设数量的隐含主题; Add a preset number of implied topics between feature words and questions;
    获取待进行训练的问题文本,并对获取的问题文本分别进行分词处理,得到各个问题文本对应的分词;Obtaining the problem text to be trained, and separately performing word segmentation on the obtained problem text, and obtaining the word segment corresponding to each question text;
    根据预先确定的隐含主题与特征词的映射关系,分别确定每个隐含主题含有的特征词的第一数量,分别确定每个特征词所属的隐含主题的第二数量,根据对应的第一数量和第二数量确定每个特征词对各个隐含主题的第一选择概率;Determining, according to a predetermined mapping relationship between the implicit subject and the feature word, a first quantity of the feature words contained in each hidden topic, and respectively determining a second quantity of the hidden subject to which each feature word belongs, according to the corresponding number A quantity and a second quantity determine a first selection probability of each feature word for each implicit topic;
    根据预先确定的隐含主题与问题文本的映射关系,分别确定每个问题文本含有的隐含主题的第三数量,分别确定每个隐含主题所属的问题文本的第四数量,根据对应的第三数量和第四数量确定每个隐含主题对各个问题文本的第二选择概率;Determining, according to a predetermined mapping relationship between the implicit topic and the problem text, respectively determining a third quantity of the implicit topic contained in each question text, and respectively determining a fourth quantity of the problem text to which each implicit topic belongs, according to the corresponding number The third quantity and the fourth quantity determine a second selection probability of each implicit topic for each question text;
    将对应的第一选择概率和第二选择概率代入预先确定的计算公式进行计算,计算出每个特征词对各个问题文本的第三选择概率,计算出的各个特征词分别对各个问题文本的第三选择概率即为特征词与问题之间的概率分布。Substituting the corresponding first selection probability and the second selection probability into a predetermined calculation formula to calculate, calculating a third selection probability of each feature word for each problem text, and calculating each feature word separately for each question text The three-choice probability is the probability distribution between the feature word and the problem.
  9. 如权利要求8所述的电子装置,其特征在于,所述预先确定的计算公式为:The electronic device according to claim 8, wherein said predetermined calculation formula is:
    P3=P1*P2,其中,P1代表第一选择概率,P2代表第二选择概率,P3代表第三选择概率。P3 = P1 * P2, where P1 represents the first selection probability, P2 represents the second selection probability, and P3 represents the third selection probability.
  10. 一种问题识别确认方法,其特征在于,该方法包括步骤:A method for identifying and identifying a problem, characterized in that the method comprises the steps of:
    A1、接收用户发出的问题语音,对接收的问题语音进行语音识别,生成问题文本;A1. Receiving a question voice sent by the user, performing voice recognition on the received question voice, and generating a question text;
    A2、对生成的问题文本按照预先确定的分词规则进行分词处理,获得所述问题文本对应的分词;A2. Performing word segmentation on the generated problem text according to a predetermined word segmentation rule to obtain a word segment corresponding to the question text;
    A3、若获得的分词中含有预先确定的特征词,则根据特征词与问题之间的概率分布,确定该预先确定的特征词对应的最大概率的问题,并根据预先确定的问题与答案之间的映射关系,确定该最大概率的问题对应的答案;A3. If the obtained participle contains a predetermined feature word, the problem of determining the maximum probability corresponding to the predetermined feature word is determined according to the probability distribution between the feature word and the question, and according to the predetermined question and answer Mapping relationship, determining the answer corresponding to the question of the maximum probability;
    A4、将确定的答案反馈给用户。A4. The determined answer is fed back to the user.
  11. 如权利要求10所述的问题识别确认方法,其特征在于,所述特征词与问题之间的概率分布按照如下步骤确定:The problem recognition confirmation method according to claim 10, wherein the probability distribution between the feature word and the question is determined as follows:
    在特征词与问题之间添加预设数量的隐含主题;Add a preset number of implied topics between feature words and questions;
    获取待进行训练的问题文本,并对获取的问题文本分别进行分词处理,得到各个问题文本对应的分词;Obtaining the problem text to be trained, and separately performing word segmentation on the obtained problem text, and obtaining the word segment corresponding to each question text;
    根据预先确定的隐含主题与特征词的映射关系,分别确定每个隐含主题含有的特征词的第一数量,分别确定每个特征词所属的隐含主 题的第二数量,根据对应的第一数量和第二数量确定每个特征词对各个隐含主题的第一选择概率;Determining the first quantity of the feature words contained in each hidden topic according to the predetermined mapping relationship between the implicit topic and the feature words, respectively determining the implicit master to which each feature word belongs a second quantity of the problem, determining, according to the corresponding first quantity and the second quantity, a first selection probability of each feature word for each implied topic;
    根据预先确定的隐含主题与问题文本的映射关系,分别确定每个问题文本含有的隐含主题的第三数量,分别确定每个隐含主题所属的问题文本的第四数量,根据对应的第三数量和第四数量确定每个隐含主题对各个问题文本的第二选择概率;Determining, according to a predetermined mapping relationship between the implicit topic and the problem text, respectively determining a third quantity of the implicit topic contained in each question text, and respectively determining a fourth quantity of the problem text to which each implicit topic belongs, according to the corresponding number The third quantity and the fourth quantity determine a second selection probability of each implicit topic for each question text;
    将对应的第一选择概率和第二选择概率代入预先确定的计算公式进行计算,计算出每个特征词对各个问题文本的第三选择概率,计算出的各个特征词分别对各个问题文本的第三选择概率即为特征词与问题之间的概率分布。Substituting the corresponding first selection probability and the second selection probability into a predetermined calculation formula to calculate, calculating a third selection probability of each feature word for each problem text, and calculating each feature word separately for each question text The three-choice probability is the probability distribution between the feature word and the problem.
  12. 如权利要求11所述的问题识别确认方法,其特征在于,所述预先确定的计算公式为:The problem recognition confirmation method according to claim 11, wherein said predetermined calculation formula is:
    P3=P1*P2,其中,P1代表第一选择概率,P2代表第二选择概率,P3代表第三选择概率。P3 = P1 * P2, where P1 represents the first selection probability, P2 represents the second selection probability, and P3 represents the third selection probability.
  13. 如权利要求10所述的问题识别确认方法,其特征在于,所述步骤A3替换为如下步骤:The problem identification confirmation method according to claim 10, wherein the step A3 is replaced by the following steps:
    若获得的分词中含有预先确定的特征词,则根据特征词与问题之间的概率分布,确定该预先确定的特征词对应各个问题的概率;If the obtained participle contains a predetermined feature word, determining a probability that the predetermined feature word corresponds to each question according to a probability distribution between the feature word and the question;
    按照概率的从大到小的顺序为各个问题进行排序,确定出排序在前的预设数量的问题作为候选问题,并将确定的各个候选问题提供或者播报给用户进行选择;Sorting each problem according to the order of probability, determining the pre-sorted number of questions as candidate questions, and providing or broadcasting the determined candidate questions to the user for selection;
    在用户选择了一个问题后,根据预先确定的问题与答案之间的映射关系,确定该问题对应的答案。After the user selects a question, the answer corresponding to the question is determined according to the mapping relationship between the predetermined question and the answer.
  14. 如权利要求13所述的问题识别确认方法,其特征在于,所述特征词与问题之间的概率分布按照如下步骤确定:The problem recognition confirmation method according to claim 13, wherein the probability distribution between the feature word and the question is determined as follows:
    在特征词与问题之间添加预设数量的隐含主题;Add a preset number of implied topics between feature words and questions;
    获取待进行训练的问题文本,并对获取的问题文本分别进行分词处理,得到各个问题文本对应的分词;Obtaining the problem text to be trained, and separately performing word segmentation on the obtained problem text, and obtaining the word segment corresponding to each question text;
    根据预先确定的隐含主题与特征词的映射关系,分别确定每个隐含主题含有的特征词的第一数量,分别确定每个特征词所属的隐含主题的第二数量,根据对应的第一数量和第二数量确定每个特征词对各个隐含主题的第一选择概率;Determining, according to a predetermined mapping relationship between the implicit subject and the feature word, a first quantity of the feature words contained in each hidden topic, and respectively determining a second quantity of the hidden subject to which each feature word belongs, according to the corresponding number A quantity and a second quantity determine a first selection probability of each feature word for each implicit topic;
    根据预先确定的隐含主题与问题文本的映射关系,分别确定每个问题文本含有的隐含主题的第三数量,分别确定每个隐含主题所属的问题文本的第四数量,根据对应的第三数量和第四数量确定每个隐含 主题对各个问题文本的第二选择概率;Determining, according to a predetermined mapping relationship between the implicit topic and the problem text, respectively determining a third quantity of the implicit topic contained in each question text, and respectively determining a fourth quantity of the problem text to which each implicit topic belongs, according to the corresponding number Three quantities and a fourth quantity determine each implied The second selection probability of the subject's text for each question;
    将对应的第一选择概率和第二选择概率代入预先确定的计算公式进行计算,计算出每个特征词对各个问题文本的第三选择概率,计算出的各个特征词分别对各个问题文本的第三选择概率即为特征词与问题之间的概率分布。Substituting the corresponding first selection probability and the second selection probability into a predetermined calculation formula to calculate, calculating a third selection probability of each feature word for each problem text, and calculating each feature word separately for each question text The three-choice probability is the probability distribution between the feature word and the problem.
  15. 如权利要求14所述的问题识别确认方法,其特征在于,所述预先确定的计算公式为:The problem recognition confirmation method according to claim 14, wherein said predetermined calculation formula is:
    P3=P1*P2,其中,P1代表第一选择概率,P2代表第二选择概率,P3代表第三选择概率。P3 = P1 * P2, where P1 represents the first selection probability, P2 represents the second selection probability, and P3 represents the third selection probability.
  16. 如权利要求10所述的问题识别确认方法,其特征在于,所述预先确定的分词规则为长词优先分词规则。The problem recognition confirmation method according to claim 10, wherein said predetermined word segmentation rule is a long word priority word segmentation rule.
  17. 如权利要求16所述的问题识别确认方法,其特征在于,所述特征词与问题之间的概率分布按照如下步骤确定:The problem recognition confirmation method according to claim 16, wherein the probability distribution between the feature word and the question is determined as follows:
    在特征词与问题之间添加预设数量的隐含主题;Add a preset number of implied topics between feature words and questions;
    获取待进行训练的问题文本,并对获取的问题文本分别进行分词处理,得到各个问题文本对应的分词;Obtaining the problem text to be trained, and separately performing word segmentation on the obtained problem text, and obtaining the word segment corresponding to each question text;
    根据预先确定的隐含主题与特征词的映射关系,分别确定每个隐含主题含有的特征词的第一数量,分别确定每个特征词所属的隐含主题的第二数量,根据对应的第一数量和第二数量确定每个特征词对各个隐含主题的第一选择概率;Determining, according to a predetermined mapping relationship between the implicit subject and the feature word, a first quantity of the feature words contained in each hidden topic, and respectively determining a second quantity of the hidden subject to which each feature word belongs, according to the corresponding number A quantity and a second quantity determine a first selection probability of each feature word for each implicit topic;
    根据预先确定的隐含主题与问题文本的映射关系,分别确定每个问题文本含有的隐含主题的第三数量,分别确定每个隐含主题所属的问题文本的第四数量,根据对应的第三数量和第四数量确定每个隐含主题对各个问题文本的第二选择概率;Determining, according to a predetermined mapping relationship between the implicit topic and the problem text, respectively determining a third quantity of the implicit topic contained in each question text, and respectively determining a fourth quantity of the problem text to which each implicit topic belongs, according to the corresponding number The third quantity and the fourth quantity determine a second selection probability of each implicit topic for each question text;
    将对应的第一选择概率和第二选择概率代入预先确定的计算公式进行计算,计算出每个特征词对各个问题文本的第三选择概率,计算出的各个特征词分别对各个问题文本的第三选择概率即为特征词与问题之间的概率分布。Substituting the corresponding first selection probability and the second selection probability into a predetermined calculation formula to calculate, calculating a third selection probability of each feature word for each problem text, and calculating each feature word separately for each question text The three-choice probability is the probability distribution between the feature word and the problem.
  18. 如权利要求17所述的问题识别确认方法,其特征在于,所述预先确定的计算公式为:The problem recognition confirmation method according to claim 17, wherein said predetermined calculation formula is:
    P3=P1*P2,其中,P1代表第一选择概率,P2代表第二选择概率,P3代表第三选择概率。P3 = P1 * P2, where P1 represents the first selection probability, P2 represents the second selection probability, and P3 represents the third selection probability.
  19. 一种计算机可读存储介质,其特征在于,所述计算机可读存 储介质存储有问题识别确认系统,所述问题识别确认系统可被至少一个处理器执行,以使所述至少一个处理器执行如下步骤:A computer readable storage medium, wherein the computer readable storage The storage medium stores a problem identification confirmation system that is executable by at least one processor to cause the at least one processor to perform the following steps:
    B1、接收用户发出的问题语音,对接收的问题语音进行语音识别,生成问题文本;B1. Receiving a question voice sent by the user, performing voice recognition on the received question voice, and generating a question text;
    B2、对生成的问题文本按照预先确定的分词规则进行分词处理,获得所述问题文本对应的分词;B2. Performing word segmentation on the generated problem text according to a predetermined word segmentation rule to obtain a word segment corresponding to the question text;
    B3、若获得的分词中含有预先确定的特征词,则根据特征词与问题之间的概率分布,确定该预先确定的特征词对应的最大概率的问题,并根据预先确定的问题与答案之间的映射关系,确定该最大概率的问题对应的答案;B3. If the obtained participle contains a predetermined feature word, the problem of determining the maximum probability corresponding to the predetermined feature word is determined according to the probability distribution between the feature word and the question, and according to the predetermined question and answer Mapping relationship, determining the answer corresponding to the question of the maximum probability;
    B4、将确定的答案反馈给用户。B4. The determined answer is fed back to the user.
  20. 如权利要求19所述的计算机可读存储介质,其特征在于,所述步骤B3替换为如下步骤:The computer readable storage medium of claim 19, wherein said step B3 is replaced by the following steps:
    若获得的分词中含有预先确定的特征词,则根据特征词与问题之间的概率分布,确定该预先确定的特征词对应各个问题的概率;If the obtained participle contains a predetermined feature word, determining a probability that the predetermined feature word corresponds to each question according to a probability distribution between the feature word and the question;
    按照概率的从大到小的顺序为各个问题进行排序,确定出排序在前的预设数量的问题作为候选问题,并将确定的各个候选问题提供或者播报给用户进行选择;Sorting each problem according to the order of probability, determining the pre-sorted number of questions as candidate questions, and providing or broadcasting the determined candidate questions to the user for selection;
    在用户选择了一个问题后,根据预先确定的问题与答案之间的映射关系,确定该问题对应的答案。 After the user selects a question, the answer corresponding to the question is determined according to the mapping relationship between the predetermined question and the answer.
PCT/CN2017/108763 2017-08-29 2017-10-31 Electronic device, question recognition and confirmation method, and computer-readable storage medium WO2019041517A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201710754550.6 2017-08-29
CN201710754550.6A CN107644012B (en) 2017-08-29 2017-08-29 Electronic device, problem identification confirmation method and computer readable storage medium

Publications (1)

Publication Number Publication Date
WO2019041517A1 true WO2019041517A1 (en) 2019-03-07

Family

ID=61110870

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/108763 WO2019041517A1 (en) 2017-08-29 2017-10-31 Electronic device, question recognition and confirmation method, and computer-readable storage medium

Country Status (2)

Country Link
CN (1) CN107644012B (en)
WO (1) WO2019041517A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110211704A (en) * 2019-05-05 2019-09-06 平安科技(深圳)有限公司 The engine method and server of matter of opening
CN110968669A (en) * 2019-11-30 2020-04-07 南京森林警察学院 Intelligent video analysis police test question classification recommendation method
CN111881694A (en) * 2020-08-05 2020-11-03 科大讯飞股份有限公司 Chapter point detection method, device, equipment and storage medium

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108388558B (en) * 2018-02-07 2022-04-19 平安普惠企业管理有限公司 Question matching method and device, customer service robot and storage medium
CN108595519A (en) * 2018-03-26 2018-09-28 平安科技(深圳)有限公司 Focus incident sorting technique, device and storage medium
CN109697228A (en) * 2018-12-13 2019-04-30 平安科技(深圳)有限公司 Intelligent answer method, apparatus, computer equipment and storage medium
CN109829048B (en) * 2019-01-23 2023-06-23 平安科技(深圳)有限公司 Electronic device, interview assisting method, and computer-readable storage medium
CN110414006B (en) * 2019-07-31 2023-09-08 京东方科技集团股份有限公司 Text theme labeling method and device, electronic equipment and storage medium
CN112399019A (en) * 2020-09-16 2021-02-23 中国农业银行股份有限公司河北省分行 Intelligent outbound call method, terminal equipment and readable storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120101807A1 (en) * 2010-10-25 2012-04-26 Electronics And Telecommunications Research Institute Question type and domain identifying apparatus and method
US20140316768A1 (en) * 2012-12-14 2014-10-23 Pramod Khandekar Systems and methods for natural language processing
CN104424290A (en) * 2013-09-02 2015-03-18 佳能株式会社 Voice based question-answering system and method for interactive voice system
CN104657346A (en) * 2015-01-15 2015-05-27 深圳市前海安测信息技术有限公司 Question matching system and question matching system in intelligent interaction system
CN105955976A (en) * 2016-04-15 2016-09-21 中国工商银行股份有限公司 Automatic answering system and method
CN106601237A (en) * 2016-12-29 2017-04-26 上海智臻智能网络科技股份有限公司 Interactive voice response system and voice recognition method thereof

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8990400B2 (en) * 2007-08-17 2015-03-24 The Invention Science Fund I, Llc Facilitating communications among message recipients
US8434066B2 (en) * 2008-06-12 2013-04-30 Oracle International Corporation Enabling a business entity to validate support for specific activities in a CRM system before migration from earlier versions
CN102622413A (en) * 2012-02-17 2012-08-01 清华大学 Method and device for answering natural language questions
US9916298B2 (en) * 2014-09-03 2018-03-13 International Business Machines Corporation Management of content tailoring by services
CN105094641A (en) * 2015-07-15 2015-11-25 百度在线网络技术(北京)有限公司 Question-answer interaction method and device based on voice
CN106445905B (en) * 2015-08-04 2019-05-21 阿里巴巴集团控股有限公司 Question and answer data processing, automatic question-answering method and device
CN106847279A (en) * 2017-01-10 2017-06-13 西安电子科技大学 Man-machine interaction method based on robot operating system ROS
CN106550156A (en) * 2017-01-23 2017-03-29 苏州咖啦魔哆信息技术有限公司 A kind of artificial intelligence's customer service system and its implementation based on speech recognition

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120101807A1 (en) * 2010-10-25 2012-04-26 Electronics And Telecommunications Research Institute Question type and domain identifying apparatus and method
US20140316768A1 (en) * 2012-12-14 2014-10-23 Pramod Khandekar Systems and methods for natural language processing
CN104424290A (en) * 2013-09-02 2015-03-18 佳能株式会社 Voice based question-answering system and method for interactive voice system
CN104657346A (en) * 2015-01-15 2015-05-27 深圳市前海安测信息技术有限公司 Question matching system and question matching system in intelligent interaction system
CN105955976A (en) * 2016-04-15 2016-09-21 中国工商银行股份有限公司 Automatic answering system and method
CN106601237A (en) * 2016-12-29 2017-04-26 上海智臻智能网络科技股份有限公司 Interactive voice response system and voice recognition method thereof

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110211704A (en) * 2019-05-05 2019-09-06 平安科技(深圳)有限公司 The engine method and server of matter of opening
CN110968669A (en) * 2019-11-30 2020-04-07 南京森林警察学院 Intelligent video analysis police test question classification recommendation method
CN111881694A (en) * 2020-08-05 2020-11-03 科大讯飞股份有限公司 Chapter point detection method, device, equipment and storage medium

Also Published As

Publication number Publication date
CN107644012A (en) 2018-01-30
CN107644012B (en) 2019-03-01

Similar Documents

Publication Publication Date Title
WO2019041517A1 (en) Electronic device, question recognition and confirmation method, and computer-readable storage medium
CN108491433B (en) Chat response method, electronic device and storage medium
US11380305B2 (en) System and method for using a question and answer engine
US7822699B2 (en) Adaptive semantic reasoning engine
WO2019153607A1 (en) Intelligent response method, electronic device and storage medium
US9575936B2 (en) Word cloud display
JP5171962B2 (en) Text classification with knowledge transfer from heterogeneous datasets
CN107004212B (en) Modeling actions, results, and goal realizations from social media and other digital tracks
CN111368043A (en) Event question-answering method, device, equipment and storage medium based on artificial intelligence
US10643601B2 (en) Detection mechanism for automated dialog systems
CN104834651B (en) Method and device for providing high-frequency question answers
EP3680802A1 (en) System and method for using a question and answer engine
US10248716B2 (en) Real-time guidance for content collection
US20200074475A1 (en) Intelligent system enabling automated scenario-based responses in customer service
WO2021042529A1 (en) Article abstract automatic generation method, device, and computer-readable storage medium
US20210141865A1 (en) Machine learning based tenant-specific chatbots for performing actions in a multi-tenant system
WO2019041528A1 (en) Method, electronic apparatus, and computer readable storage medium for determining polarity of news sentiment
CN112989050B (en) Form classification method, device, equipment and storage medium
US20230334075A1 (en) Search platform for unstructured interaction summaries
CN111400340A (en) Natural language processing method and device, computer equipment and storage medium
CN112989003B (en) Intention recognition method, device, processing equipment and medium
CN116048463A (en) Intelligent recommendation method and device for content of demand item based on label management
CN108959327B (en) Service processing method, device and computer readable storage medium
TW201935229A (en) Method and system for operating a virtual assistant
CN112562734B (en) Voice interaction method and device based on voice detection

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17923147

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC , EPO FORM 1205A DATED 28.09.2020.

122 Ep: pct application non-entry in european phase

Ref document number: 17923147

Country of ref document: EP

Kind code of ref document: A1