WO2020098249A1 - Electronic device, response conversation technique recommendation method and computer readable storage medium - Google Patents
Electronic device, response conversation technique recommendation method and computer readable storage medium Download PDFInfo
- Publication number
- WO2020098249A1 WO2020098249A1 PCT/CN2019/088635 CN2019088635W WO2020098249A1 WO 2020098249 A1 WO2020098249 A1 WO 2020098249A1 CN 2019088635 W CN2019088635 W CN 2019088635W WO 2020098249 A1 WO2020098249 A1 WO 2020098249A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- preset
- voice
- feature vector
- audio feature
- model
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 71
- 230000004044 response Effects 0.000 title abstract description 15
- 239000013598 vector Substances 0.000 claims abstract description 132
- 230000008451 emotion Effects 0.000 claims abstract description 126
- 239000003795 chemical substances by application Substances 0.000 claims description 141
- 238000012549 training Methods 0.000 claims description 56
- 238000012795 verification Methods 0.000 claims description 22
- 238000000605 extraction Methods 0.000 claims description 21
- 230000002159 abnormal effect Effects 0.000 claims description 18
- 230000008569 process Effects 0.000 claims description 14
- 239000000284 extract Substances 0.000 claims description 12
- 230000002452 interceptive effect Effects 0.000 claims description 4
- 238000005516 engineering process Methods 0.000 abstract description 6
- 206010022998 Irritability Diseases 0.000 description 7
- 238000010586 diagram Methods 0.000 description 6
- 230000002996 emotional effect Effects 0.000 description 4
- 230000008859 change Effects 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 239000004973 liquid crystal related substance Substances 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- 238000012546 transfer Methods 0.000 description 2
- 208000027534 Emotional disease Diseases 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/50—Centralised arrangements for answering calls; Centralised arrangements for recording messages for absent or busy subscribers ; Centralised arrangements for recording messages
- H04M3/51—Centralised call answering arrangements requiring operator intervention, e.g. call or contact centers for telemarketing
- H04M3/5183—Call or contact centers with computer-telephony arrangements
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/63—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for estimating an emotional state
Definitions
- the present application relates to the field of call center customer service systems, and in particular, to an electronic device, a method for responding to conversations, and a computer-readable storage medium.
- the main purpose of the present application is to provide an electronic device, a method for recommending communication technology, and a computer-readable storage medium, designed to help agents deal with customers and improve customer complaints and customer churn.
- the electronic device includes a memory and a processor, and the memory stores a recommended program that can be run on the processor.
- the following procedure is implemented when the technical recommendation program is executed by the processor:
- A1 After receiving the incoming customer line, obtain the voice streams of the incoming customer and agent in real time;
- D1 input the recognized first speech content and the first sentiment classification into a pre-trained dialogue-recommended recommendation model for analysis, to obtain a recommended dialogue-reporting technique, and put the recommended dialogue-reported technique in real time Sent to the terminal of the agent.
- the second aspect of the present application also proposes a method for recommending conversational techniques, including the following steps:
- A2 After receiving the incoming customer line, obtain the voice streams of the incoming customer and agent in real time;
- the first audio feature vector is input into a preset voice emotion analysis model for analysis, and the first speech content corresponding to the first audio feature vector and the first emotion classification are identified;
- the recognized first speech content and the first emotion classification are input into a pre-trained conversational dialogue recommendation model for analysis to obtain a recommended conversational dialogue, and the recommended conversational dialogue is real-time Sent to the terminal of the agent.
- a third aspect of the present application also provides a computer-readable storage medium that stores a computer-speaking recommendation program that can be executed by at least one processor so that the at least one A processor performs the following steps:
- the voice stream of the incoming client and the agent is obtained in real time when the customer talks to the agent after entering the line, and the first audio feature vector corresponding to the voice segment input by the customer is extracted from the obtained voice stream.
- the first audio feature vector is input into a preset speech emotion analysis model for analysis and recognition, the first speech content and the first emotion classification corresponding to the first audio feature vector are identified, and finally the pre-trained response model is used for recognition
- the first voice content and the first sentiment classification are analyzed, and the recommended response should be sent to the agent ’s terminal in real time for reference by the agent to help the agent deal with the customer; thus, the customer ’s own emotions are effectively improved.
- the agent does not know how to deal with the customer's situation, reducing customer complaints and customer churn.
- FIG. 1 is a schematic flowchart of an embodiment of a method for recommending a conversational application
- FIG. 2 is a schematic flowchart of the second embodiment of the application-recommended method for application of dialogue
- FIG. 3 is a schematic flow chart of three embodiments of the recommended method of application of dialogue technology
- FIG. 4 is a schematic diagram of the operating environment of the preferred embodiment of the application recommended procedure of the application
- FIG. 5 is a program module diagram of an embodiment of the application-recommended procedure of the application.
- FIG. 6 is a program module diagram of the second embodiment of the recommended procedure for application of dialogue technology
- FIG. 7 is a program module diagram of three embodiments of the application-recommended application procedure.
- This application proposes a recommended method of dialogue technology, which is mainly used in the call center customer service system.
- FIG. 1 is a schematic flowchart of an embodiment of a method for recommending a conversational application in this application.
- the method for recommending conversational techniques includes:
- Step S10 After receiving the incoming customer line, obtain the voice streams of the incoming customer and the agent in real time.
- the call center customer service system obtains in real time the voice stream of the incoming customer and agent (that is, the interactive audio stream between the customer and the agent) through the telephone voice platform.
- Step S20 Perform feature extraction on the voice stream to extract the first audio feature vector of the first voice segment in the voice stream, where the first voice segment corresponds to the voice input by the customer in the voice stream segment.
- the call center customer service system performs feature extraction on the currently acquired voice stream, and extracts the first audio feature vector corresponding to the voice segment input by the customer in the voice stream.
- the first audio feature vector may be one or more of the following audio features: energy feature, pronunciation frame feature, pitch frequency feature, formant feature, harmonic noise ratio feature, and Mel cepstrum coefficient feature.
- step S30 the first audio feature vector is input into a preset voice emotion analysis model for analysis, and the first voice content and the first emotion classification corresponding to the first audio feature vector are identified.
- the call center customer service system has a preset voice emotion analysis model. After extracting the first audio feature vector from the currently acquired voice stream, the call center customer service system inputs the first audio feature vector into the preset voice emotion analysis Perform analysis and recognition in the model to identify the first speech content corresponding to the first audio feature vector (that is, the text content corresponding to the speech segment input by the customer in the acquired speech stream) and the first emotion classification (that is, the customer's Emotion classification).
- the preset voice sentiment analysis model preferably adopts a sentiment analysis model including a DNN-HMM acoustic model, an n-gram language model and a wfst weighted graph, and the first audio feature vector is preferably a melody cepstrum coefficient feature vector
- the preset voice sentiment analysis model may also be other sentiment analysis models.
- the first emotion classification includes: satisfaction, calm, irritability, and anger, etc.
- Step S40 the recognized first speech content and the first emotion classification are input into a pre-trained conversational dialogue recommendation model for analysis, to obtain a recommended conversational dialogue, and the recommended conversational dialogue Send to the agent's terminal in real time.
- the customer service system in the call center also has a pre-trained recommendation model for conversational skills.
- the first speech feature and the first emotion classification corresponding to the first audio feature vector are identified through a preset speech emotion analysis model, the The first speech content and the first sentiment classification are input into the recommended model for analysis, and the recommended response technique is obtained, and the recommended response technique is sent to the terminal of the agent for display in real time for Agent reference, help agents deal with customers.
- the technical solution of this embodiment by acquiring the voice streams of the incoming client and the agent in real time after the customer enters the line, and extracting the first audio feature vector corresponding to the voice segment input by the customer from the acquired voice stream, the The first audio feature vector is input into a preset speech emotion analysis model for analysis and recognition, and the first speech content and the first emotion classification corresponding to the first audio feature vector are identified, and finally a pre-trained response model is used.
- the recognized first speech content and the first sentiment classification are analyzed, and the recommended response should be sent to the agent's terminal in real time for reference by the agent to help the agent deal with the customer; in this way, the customer's own When emotional issues conflict with the agent, the agent does not know how to deal with the customer's situation, reducing customer complaints and customer churn.
- FIG. 2 is a schematic flowchart of a second embodiment of a method for recommending conversational application in this application.
- the method for recommending conversational skills further includes:
- Step S50 Perform feature extraction on the voice stream to extract a second audio feature vector of a second voice segment in the voice stream, where the second voice segment corresponds to the voice input by the agent in the voice stream segment.
- the call center customer service system performs feature extraction on the currently acquired voice stream to extract the second audio feature vector corresponding to the voice segment input by the agent in the voice stream.
- the second audio feature vector may include one or more of the following audio features: energy feature, pronunciation frame feature, pitch frequency feature, formant feature, harmonic-to-noise ratio feature, and Mel cepstral coefficient feature .
- step S60 the second audio feature vector is input into a preset voice emotion analysis model for analysis, and a second emotion classification corresponding to the second audio feature vector is identified.
- the call center customer service system After extracting the second audio feature vector from the currently acquired voice stream, the call center customer service system inputs the second audio feature vector into a preset voice sentiment analysis model for analysis and recognition to identify the first audio
- the second emotion classification corresponding to the feature vector ie, the emotion classification of the agent.
- the second emotion classification includes: satisfaction, calm, irritability, and anger, etc.
- Step S70 if the second emotion classification is a preset abnormal emotion classification, then preset preset reminder information is sent to the terminal of the agent.
- the call center customer service system When it is found that the emotion classification of the agent (ie, the second emotion classification) is an abnormal emotion classification (for example, irritability, anger, and other emotional classifications that are not positive emotions), the call center customer service system will send a preset to the terminal of the agent
- the first reminder message is to remind the agent to pay attention to emotions and adjust the service attitude in time.
- the first reminder message is, for example: "Your service attitude is detected to be negative, please pay attention to adjust the service attitude", and so on.
- This embodiment recognizes the emotion classification of the agent in real time, detects the change of the agent's emotion, and prompts the agent to adjust the emotion and state in real time when the agent's emotion is abnormal (that is, deteriorates), so as to better ensure the service quality of the agent to the customer To increase customer satisfaction.
- FIG. 3 is a schematic flowchart of three embodiments of the application-recommended method of application.
- the preset voice sentiment analysis model also recognizes the second voice content corresponding to the second audio feature vector (that is, the agent in the acquired voice stream The text content corresponding to the input voice segment); the recommended dialogue method recommendation method after the step S60, further includes:
- Step S01 When the second emotion classification is a preset abnormal emotion classification, analyze whether the second speech content contains a preset sensitive word.
- a sensitive thesaurus is set in the customer service system of the call center (the sensitive word library includes many sensitive words).
- the call center customer service system is based on Sensitive word library, analyze whether the second speech content (that is, the text content corresponding to the speech segment input by the agent in the acquired speech stream) contains preset sensitive words (for example, impolite and uncivilized words).
- Step S02 if the preset sensitive words are included, analyze whether the number of times the preset sensitive words appear in the second speech content is greater than the first threshold;
- the agent's word use is further judged by analyzing the number of times the preset sensitive words appear in the second voice content Improper severity, according to the severity of the corresponding treatment. Specifically, the number of occurrences of the preset sensitive word is compared with a first threshold (for example, 3 times).
- Step S03 If the number of times the preset sensitive words appear in the second voice content is less than or equal to the first threshold, send preset second reminder information to the terminal of the agent.
- the call center customer service system determines that the use of the agent's words is not particularly serious, and then sends the agent Terminal sends a preset second reminder message to remind the agent to pay attention to the words and not to use sensitive words.
- the call center customer service system can also send the preset sensitive words appearing in the second voice content of the agent to the terminal of the agent for highlighting.
- the second reminder message is, for example, "Please pay attention to words, and prohibit the use of sensitive words", etc.
- Step S04 if the number of times the preset sensitive words appear in the second voice content is greater than the first threshold, send preset third reminder information to the superior management terminal of the agent.
- the call center customer service system determines that the agent's words are used improperly, which may be due to the agent arguing with the customer, etc. Abnormal situation, at this time, the call center customer service system sends the preset third reminder message to the superior management terminal of the agent (the terminal of the superior manager) to remind the superior leader or manager of the agent to pay special attention to the agent's call happening.
- the call center customer service system can also transfer the call voice of the agent and the customer to the superior management terminal of the agent in real time, so that the superior leader or manager can directly monitor the voice process of the call to quarrel with the customer at the agent Timely handling.
- the third reminder message is, for example: "There is a serious problem with the agent's speech, please deal with it in time", etc.
- the training process of the recommended model should include:
- the customer service system of the call center will record every call that a customer enters and save it in the call recording database.
- Each recorded data is usually marked with a service tag of the customer's satisfaction for the call service provided by the agent.
- These recording data are all recording data whose service tag is satisfactory; extract audio feature vectors for each recording data obtained to obtain each The first audio feature vector corresponding to the first speech segment (voice segment input by the customer) in the recorded data and the second audio feature vector corresponding to the second speech segment (voice segment input by the agent).
- the audio feature vector may include one or more of the following audio features: energy feature, pronunciation frame number feature, pitch frequency feature, formant feature, harmonic noise ratio feature, and Mel cepstrum coefficient feature.
- a preset voice sentiment analysis model is used to analyze and recognize the first audio feature vector and the second audio feature vector corresponding to each recording data, and the first speech content and the first speech feature corresponding to each first audio feature vector are identified.
- An emotion classification, and recognizing the second speech content corresponding to each second audio feature vector; the first speech content, the first emotion classification, and the second speech content corresponding to each recording data form a sample, so that A preset number of samples were obtained.
- the preset speech sentiment analysis model preferably uses a sentiment analysis model including a DNN-HMM acoustic model, an n-gram language model and a wfst weighted graph, and the first audio feature vector and the second audio feature vector are preferably Mel cepstrum coefficients Feature vector; of course, the preset voice sentiment analysis model may also be other sentiment analysis models.
- the first emotion classification includes: satisfaction, calm, irritability, and anger, etc.
- the first percentage of the preset number of samples is used as the training set, and the second percentage is used as the verification set.
- the sum of the first percentage and the second percentage is less than 100%.
- a first percentage for example, 70%
- a second percentage for example, 25% to 30%
- the samples of the training set are used to train the pre-trained recommendation model. After the training is completed, the samples in the verification set are used to verify the completed training model.
- the following criteria are used to test the accuracy of the recommended dialogue model:
- the recommended dialogue model predicts a sample of the verification set after the corresponding dialogue, if the content of the corresponding dialogue exceeds N% ( For example, the content of 95%) is the same as the second speech content of the sample, and it is determined that the prediction is accurate.
- a preset threshold for example, 97%) of the prediction accuracy rate is preset in the system, which is used to check the training effect of the recommended model of the application. If the prediction accuracy rate is greater than the preset threshold, then the application The training of the dialogue recommended model has reached the preset standard, and then the model training ends.
- the prediction accuracy rate is less than or equal to the preset threshold, it means that the training of the recommended model should not reach the preset standard. It may be that the number of samples in the training set is insufficient or the number of samples in the verification set is insufficient. In this case, increase the preset number (ie, increase the number of samples, for example, increase the fixed number each time or increase the random number each time), and then, on this basis, re-execute the above steps S1-S4 , And so on, until the requirement of step S5 is reached, then the model training is ended.
- increase the preset number ie, increase the number of samples, for example, increase the fixed number each time or increase the random number each time
- this application also proposes a recommended procedure for dialogue.
- FIG. 4 is a schematic diagram of the operating environment of the preferred embodiment of the application recommended procedure 10 of the present application.
- the conversational recommendation program 10 should be installed and run in the electronic device 1.
- the electronic device 1 may be a computing device such as a desktop computer, a notebook, a palmtop computer, and a server.
- the electronic device 1 may include, but is not limited to, a memory 11, a processor 12, and a display 13.
- FIG. 4 only shows the electronic device 1 having the components 11-13, but it should be understood that it is not required to implement all the components shown, and more or fewer components may be implemented instead.
- the memory 11 may be an internal storage unit of the electronic device 1 in some embodiments, such as a hard disk or a memory of the electronic device 1. In other embodiments, the memory 11 may also be an external storage device of the electronic device 1, such as a plug-in hard disk equipped on the electronic device 1, a smart memory card (Smart, Media, Card, SMC), and a secure digital (SD) Cards, flash cards, etc. Further, the memory 11 may also include both an internal storage unit of the electronic device 1 and an external storage device.
- the memory 11 is used to store application software and various types of data installed in the electronic device 1, for example, program codes of the program 10 recommended for communication. The memory 11 can also be used to temporarily store data that has been or will be output.
- the processor 12 may be a central processing unit (CPU), microprocessor, or other data processing chip, which is used to run the program code or process data stored in the memory 11, for example, to execute the interactive technology Recommended program 10 etc.
- CPU central processing unit
- microprocessor or other data processing chip, which is used to run the program code or process data stored in the memory 11, for example, to execute the interactive technology Recommended program 10 etc.
- the display 13 may be an LED display, a liquid crystal display, a touch-sensitive liquid crystal display, an OLED (Organic Light-Emitting Diode, organic light emitting diode) touch device, or the like.
- the display 13 is used to display information processed in the electronic device 1 and to display a visual user interface.
- the components 11-13 of the electronic device 1 communicate with each other through a program bus.
- FIG. 5 is a program block diagram of an embodiment of the application recommended program 10 of this application.
- the application recommendation program 10 can be divided into one or more modules, and the one or more modules are stored in the memory 11 and are processed by one or more processors (in this embodiment, the processor 12 ) To complete this application.
- the conversational skill recommendation program 10 may be divided into a real-time acquisition module 101, a first extraction module 102, a first recognition module 103, and a recommendation module 104.
- the module referred to in this application refers to a series of computer program instruction segments capable of performing specific functions, and is more suitable than the program for describing the execution process of the recommended program 10 in the electronic device 1, wherein:
- the real-time acquiring module 101 is configured to acquire the voice streams of the incoming customer and the agent in real time after receiving the incoming customer.
- the call center customer service system obtains in real time the voice stream of the incoming customer and the agent (that is, the interactive audio stream between the customer and the agent) through the telephone voice platform.
- the first extraction module 102 is configured to perform feature extraction on the voice stream to extract a first audio feature vector of a first voice segment in the voice stream, where the first voice segment corresponds to the voice stream Speech segments entered by customers.
- the call center customer service system performs feature extraction on the currently acquired voice stream, and extracts the first audio feature vector corresponding to the voice segment input by the customer in the voice stream.
- the first audio feature vector may be one or more of the following audio features: energy feature, pronunciation frame feature, pitch frequency feature, formant feature, harmonic noise ratio feature, and Mel cepstrum coefficient feature.
- the first recognition module 103 is configured to input the first audio feature vector into a preset voice emotion analysis model for analysis, and recognize the first voice content and the first emotion classification corresponding to the first audio feature vector.
- the call center customer service system has a preset voice emotion analysis model. After extracting the first audio feature vector from the currently acquired voice stream, the call center customer service system inputs the first audio feature vector into the preset voice emotion analysis Perform analysis and recognition in the model to identify the first speech content corresponding to the first audio feature vector (that is, the text content corresponding to the speech segment input by the customer in the acquired speech stream) and the first emotion classification (that is, the customer's Emotion classification).
- the preset voice sentiment analysis model preferably adopts a sentiment analysis model including a DNN-HMM acoustic model, an n-gram language model and a wfst weighted graph, and the first audio feature vector is preferably a melody cepstrum coefficient feature vector
- the preset voice sentiment analysis model may also be other sentiment analysis models.
- the first emotion classification includes: satisfaction, calm, irritability, and anger, etc.
- the recommendation module 104 is configured to input the recognized first speech content and the first emotion classification into a pre-trained dialogue-recommendation recommendation model for analysis, so as to obtain a recommended dialogue-reaction technique, and convert the recommended Dialogue is sent to the agent's terminal in real time.
- the customer service system in the call center also has a pre-trained recommendation model for conversational skills.
- the first speech feature and the first emotion classification corresponding to the first audio feature vector are identified through a preset speech emotion analysis model, the The first speech content and the first sentiment classification are input into the recommended model for analysis, and the recommended response technique is obtained, and the recommended response technique is sent to the terminal of the agent for display in real time for Agent reference, help agents deal with customers.
- the technical solution of this embodiment by acquiring the voice streams of the incoming client and the agent in real time after the customer enters the line, and extracting the first audio feature vector corresponding to the voice segment input by the customer from the acquired voice stream, the The first audio feature vector is input into a preset speech emotion analysis model for analysis and recognition, and the first speech content and the first emotion classification corresponding to the first audio feature vector are identified, and finally a pre-trained response model is used.
- the recognized first speech content and the first sentiment classification are analyzed, and the recommended response should be sent to the agent's terminal in real time for reference by the agent to help the agent deal with the customer; in this way, the customer's own When emotional issues conflict with the agent, the agent does not know how to deal with the customer's situation, reducing customer complaints and customer churn.
- the conversational skill recommendation program further includes a second extraction module 105, a second recognition module 106 and a reminder module 107.
- the second extraction module 105 is configured to perform feature extraction on the voice stream to extract a second audio feature vector of a second voice segment in the voice stream, where the second voice segment corresponds to the voice stream Voice segment entered by the agent.
- the call center customer service system performs feature extraction on the currently acquired voice stream to extract the second audio feature vector corresponding to the voice segment input by the agent in the voice stream.
- the second audio feature vector may include one or more of the following audio features: energy feature, pronunciation frame feature, pitch frequency feature, formant feature, harmonic-to-noise ratio feature, and Mel cepstral coefficient feature .
- the second recognition module 106 is configured to input the second audio feature vector into a preset voice emotion analysis model for analysis, and recognize a second emotion classification corresponding to the second audio feature vector.
- the call center customer service system After extracting the second audio feature vector from the currently acquired voice stream, the call center customer service system inputs the second audio feature vector into a preset voice sentiment analysis model for analysis and recognition to identify the first audio
- the second emotion classification corresponding to the feature vector ie, the emotion classification of the agent.
- the second emotion classification includes: satisfaction, calm, irritability, and anger, etc.
- the reminder module 107 is configured to send preset first reminder information to the terminal of the agent when the second emotion category is a preset abnormal emotion category.
- the call center customer service system When it is found that the emotion classification of the agent (ie, the second emotion classification) is an abnormal emotion classification (for example, irritability, anger, and other emotional classifications with inactive emotions), the call center customer service system will send the preset
- the first reminder message is to remind the agent to pay attention to emotions and adjust the service attitude in time.
- the first reminder message is, for example: "Your service attitude is detected to be negative, please pay attention to adjust the service attitude", and so on.
- This embodiment recognizes the emotion classification of the agent in real time, detects the change of the agent's emotion, and prompts the agent to adjust the emotion and state in real time when the agent's emotion is abnormal (that is, deteriorates), so as to better ensure the service quality of the agent to the customer To increase customer satisfaction.
- the second recognition module 106 is further configured to input the second audio feature vector into a preset voice sentiment analysis model for analysis, and identify the second corresponding to the second audio feature vector Voice content (that is, the text content corresponding to the voice segment input by the agent in the acquired voice stream); the dialogue-speaking recommendation program further includes a first analysis module 108 and a second analysis module 109. among them,
- the first analysis module 108 is configured to analyze whether the second speech content contains preset sensitive words when determining that the second emotion classification is a preset abnormal emotion classification.
- a sensitive thesaurus is set in the customer service system of the call center (the sensitive word library includes many sensitive words).
- the call center customer service system is based on Sensitive word library, analyze whether the second speech content (that is, the text content corresponding to the speech segment input by the agent in the acquired speech stream) contains preset sensitive words (for example, impolite and uncivilized words).
- the second analysis module 109 is configured to, when determining that the second voice content contains preset sensitive words, analyze whether the number of times the preset sensitive words appear in the second voice content is greater than a first threshold;
- the agent's word use is further determined by analyzing the number of times the preset sensitive words appear in the second voice content Improper severity, according to the severity of the corresponding treatment. Specifically, the number of occurrences of the preset sensitive word is compared with a first threshold (for example, 3 times).
- the reminder module 107 is further configured to send preset second reminder information to the terminal of the agent when it is determined that the number of preset sensitive words appearing in the second voice content is less than or equal to the first threshold.
- the call center customer service system determines that the use of the agent's words is not particularly serious, and then sends the agent Terminal sends a preset second reminder message to remind the agent to pay attention to the words and not to use sensitive words.
- the call center customer service system can also send the preset sensitive words appearing in the second voice content of the agent to the terminal of the agent for highlighting.
- the second reminder message is, for example, "Please pay attention to words, and prohibit the use of sensitive words", etc.
- the reminder module 107 is further configured to send preset third reminder information to the superior management terminal of the agent when it is determined that the number of times the preset sensitive words appear in the second voice content is greater than the first threshold.
- the call center customer service system determines that the agent's words are used improperly, which may be due to the agent arguing with the customer. Abnormal situation, at this time, the call center customer service system sends the preset third reminder message to the superior management terminal of the agent (the terminal of the superior manager) to remind the superior leader or manager of the agent to pay special attention to the agent's call happening.
- the call center customer service system can also transfer the call voice of the agent and the customer to the superior management terminal of the agent in real time, so that the superior leader or manager can directly monitor the voice process of the call to quarrel with the customer at the agent Timely handling.
- the third reminder message is, for example: "There is a serious problem with the agent's speech, please deal with it in time", etc.
- the present application also proposes a computer-readable storage medium, the computer-readable storage medium storing a recommended program for conversational skills, the recommended program for conversational skills can be executed by at least one processor, so that the at least A processor executes the method for recommending interaction in any of the above embodiments.
Landscapes
- Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Health & Medical Sciences (AREA)
- Psychiatry (AREA)
- Marketing (AREA)
- Child & Adolescent Psychology (AREA)
- General Health & Medical Sciences (AREA)
- Hospice & Palliative Care (AREA)
- Business, Economics & Management (AREA)
- Computational Linguistics (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Telephonic Communication Services (AREA)
Abstract
The present application relates to semantic recognition technology, and disclosed thereby is a response conversation technique recommendation method. The method comprises: after receiving the incoming line of a customer, acquiring a voice stream of the incoming line customer and an operator in real time; extracting a first audio feature vector of a first voice segment in the voice stream, wherein the first voice segment corresponds to a voice segment inputted by the customer in the voice stream; inputting the first audio feature vector into a preset voice emotion analysis model for analysis, and recognizing first voice content and a first emotion classification corresponding to the first audio feature vector; inputting the recognized first voice content and the first emotion classification into a pre-trained response conversation technique recommendation model for analysis to obtain a recommended response conversation technique, and sending the recommended response conversation technique to a terminal of the operator in real time. Also disclosed by the present application are an electronic device and a computer-readable storage medium. The technical solution of the present application effectively ameliorates the condition of operators not knowing how to respond to customers, which reduces customer complaints and customer attrition.
Description
优先权申明Priority declaration
本申请基于巴黎公约申明享有2018年11月12日递交的申请号为CN 201811340705.2、名称为“电子装置、应对话术推荐方法和计算机可读存储介质”中国专利申请的优先权,该中国专利申请的整体内容以参考的方式结合在本申请中。This application is based on the declaration of the Paris Convention and enjoys the priority of the Chinese patent application with the application number CN201811340705.2 and the name "Electronic device, recommended method and computer readable storage medium" submitted on November 12, 2018. The entire content of is incorporated into this application by reference.
本申请涉及电话中心客服系统领域,特别涉及一种电子装置、应对话术推荐方法和计算机可读存储介质。The present application relates to the field of call center customer service systems, and in particular, to an electronic device, a method for responding to conversations, and a computer-readable storage medium.
目前,电话中心客服系统在分配坐席为进线客户服务后,坐席在为客户服务的过程中,部分客户可能会因自身情绪问题而与坐席产生冲突,这种情形下,很多经验不足的坐席不知如何有效应对客户,造成客户投诉或客户流失等后果。At present, in the call center customer service system, after assigning agents to serve incoming customers, some agents may conflict with agents due to their own emotional problems in the process of serving customers. In this case, many inexperienced agents do not know How to effectively deal with customers, resulting in customer complaints or customer churn and other consequences.
发明内容Summary of the invention
本申请的主要目的是提供一种电子装置、应对话术推荐方法和计算机可读存储介质,旨在帮助坐席应对客户,改善客户投诉和客户流失的情况。The main purpose of the present application is to provide an electronic device, a method for recommending communication technology, and a computer-readable storage medium, designed to help agents deal with customers and improve customer complaints and customer churn.
为实现上述目的,本申请第一方面提出的电子装置,所述电子装置包括存储器和处理器,所述存储器上存储有可在所述处理器上运行的应对话术推荐程序,所述应对话术推荐程序被所述处理器执行时实现如下步骤:In order to achieve the above object, the electronic device proposed in the first aspect of the present application, the electronic device includes a memory and a processor, and the memory stores a recommended program that can be run on the processor. The following procedure is implemented when the technical recommendation program is executed by the processor:
A1、在接收到客户进线后,实时获取所述进线客户与坐席的语音流;A1. After receiving the incoming customer line, obtain the voice streams of the incoming customer and agent in real time;
B1、对所述语音流进行特征提取,提取出所述语音流中的第一语音片段的第一音频特征向量,其中,所述第一语音片段对应所述语音流中的客户输入的语音段;B1. Perform feature extraction on the voice stream to extract a first audio feature vector of a first voice segment in the voice stream, where the first voice segment corresponds to a voice segment input by a customer in the voice stream ;
C1、将所述第一音频特征向量输入预设的语音情绪分析模型中进行分析,识别出所述第一音频特征向量对应的第一语音内容和第一情绪分类;C1. Input the first audio feature vector into a preset voice emotion analysis model for analysis, and identify the first voice content and the first emotion classification corresponding to the first audio feature vector;
D1、将识别出的所述第一语音内容和所述第一情绪分类输入预先训练好的应对话术推荐模型中进行分析,以得到推荐的应对话术,将所述推荐的应对话术实时发送至所述坐席的终端。D1, input the recognized first speech content and the first sentiment classification into a pre-trained dialogue-recommended recommendation model for analysis, to obtain a recommended dialogue-reporting technique, and put the recommended dialogue-reported technique in real time Sent to the terminal of the agent.
本申请第二方面还提出一种应对话术推荐方法,包括以下步骤:The second aspect of the present application also proposes a method for recommending conversational techniques, including the following steps:
A2、在接收到客户进线后,实时获取所述进线客户与坐席的语音流;A2. After receiving the incoming customer line, obtain the voice streams of the incoming customer and agent in real time;
B2、对所述语音流进行特征提取,提取出所述语音流中的第一语音片段的第一音频特征向量,其中,所述第一语音片段对应所述语音流中的客户输入的语音段;B2. Perform feature extraction on the voice stream to extract a first audio feature vector of a first voice segment in the voice stream, where the first voice segment corresponds to a voice segment input by a customer in the voice stream ;
C2、将所述第一音频特征向量输入预设的语音情绪分析模型中进行分析,识别出所述第一音频特征向量对应的第一语音内容和所述第一情绪分类;C2. The first audio feature vector is input into a preset voice emotion analysis model for analysis, and the first speech content corresponding to the first audio feature vector and the first emotion classification are identified;
D2、将识别出的所述第一语音内容和所述第一情绪分类输入预先训练好的应对话术推荐模型中进行分析,以得到推荐的应对话术,将所述推荐的应对话术实时发送至所述坐席的终端。D2. The recognized first speech content and the first emotion classification are input into a pre-trained conversational dialogue recommendation model for analysis to obtain a recommended conversational dialogue, and the recommended conversational dialogue is real-time Sent to the terminal of the agent.
本申请第三方面还提出一种计算机可读存储介质,所述计算机可读存储介质存储有应对话术推荐程序,所述应对话术推荐程序可被至少一个处理器执行,以使所述至少一个处理器执行如下步骤:A third aspect of the present application also provides a computer-readable storage medium that stores a computer-speaking recommendation program that can be executed by at least one processor so that the at least one A processor performs the following steps:
在接收到客户进线后,实时获取所述进线客户与坐席的语音流;After receiving the incoming customer line, obtain the voice streams of the incoming customer and agent in real time;
对所述语音流进行特征提取,提取出所述语音流中的第一语音片段的第一音频特征向量,其中,所述第一语音片段对应所述语音流中的客户输入的语音段;Performing feature extraction on the voice stream to extract a first audio feature vector of a first voice segment in the voice stream, where the first voice segment corresponds to a voice segment input by a customer in the voice stream;
将所述第一音频特征向量输入预设的语音情绪分析模型中进行分析,识别出所述第一音频特征向量对应的第一语音内容和第一情绪分类;Input the first audio feature vector into a preset voice emotion analysis model for analysis, and identify the first voice content and the first emotion classification corresponding to the first audio feature vector;
将识别出的所述第一语音内容和所述第一情绪分类输入预先训练好的应对话术推荐模型中进行分析,以得到推荐的应对话术,将所述推荐的应对话术实时发送至所述坐席的终端。Input the recognized first speech content and the first sentiment classification into a pre-trained conversational recommendation model for analysis to obtain a recommended conversational conversation, and send the recommended conversational conversation in real time to The terminal of the agent.
本申请技术方案,通过在客户进线后与坐席通话时,实时获取进线客户与坐席的语音流,从获取的语音流中提取出客户输入的语音段对应的第一音频特征向量,将所述第一音频特征向量输入预设的语音情绪分析模型中进行分析识别,识别出第一音频特征向量对应的第一语音内容和第一情绪分类,最终利用预先训练好的应对话术模型对识别出的第一语音内容和第一情绪分类进行分析,得出推荐的应对话术实时发送给所述坐席的终端,以供坐席参考,帮助坐席应对客户;如此,有效的改善了客户因自身情绪问题而与坐席产生冲突时,坐席不知如何应对客户的情况,减少了客户投诉和客户流失。In the technical solution of the present application, the voice stream of the incoming client and the agent is obtained in real time when the customer talks to the agent after entering the line, and the first audio feature vector corresponding to the voice segment input by the customer is extracted from the obtained voice stream. The first audio feature vector is input into a preset speech emotion analysis model for analysis and recognition, the first speech content and the first emotion classification corresponding to the first audio feature vector are identified, and finally the pre-trained response model is used for recognition The first voice content and the first sentiment classification are analyzed, and the recommended response should be sent to the agent ’s terminal in real time for reference by the agent to help the agent deal with the customer; thus, the customer ’s own emotions are effectively improved. When the problem conflicts with the agent, the agent does not know how to deal with the customer's situation, reducing customer complaints and customer churn.
为了更清楚地说明本申请实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而 易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图示出的结构获得其他的附图。In order to more clearly explain the embodiments of the present application or the technical solutions in the prior art, the following will briefly introduce the drawings required in the embodiments or the description of the prior art. Obviously, the drawings in the following description are only These are some embodiments of the present application. For those of ordinary skill in the art, without paying any creative work, other drawings can be obtained according to the structures shown in these drawings.
图1为本申请应对话术推荐方法一实施例的流程示意图;FIG. 1 is a schematic flowchart of an embodiment of a method for recommending a conversational application;
图2为本申请应对话术推荐方法二实施例的流程示意图;FIG. 2 is a schematic flowchart of the second embodiment of the application-recommended method for application of dialogue;
图3为本申请应对话术推荐方法三实施例的流程示意图;FIG. 3 is a schematic flow chart of three embodiments of the recommended method of application of dialogue technology;
图4为本申请应对话术推荐程序较佳实施例的运行环境示意图;FIG. 4 is a schematic diagram of the operating environment of the preferred embodiment of the application recommended procedure of the application;
图5为本申请应对话术推荐程序一实施例的程序模块图;FIG. 5 is a program module diagram of an embodiment of the application-recommended procedure of the application;
图6为本申请应对话术推荐程序二实施例的程序模块图;FIG. 6 is a program module diagram of the second embodiment of the recommended procedure for application of dialogue technology;
图7为本申请应对话术推荐程序三实施例的程序模块图。FIG. 7 is a program module diagram of three embodiments of the application-recommended application procedure.
本申请目的的实现、功能特点及优点将结合实施例,参照附图做进一步说明。The implementation, functional characteristics and advantages of the present application will be further described in conjunction with the embodiments and with reference to the drawings.
以下结合附图对本申请的原理和特征进行描述,所举实例只用于解释本申请,并非用于限定本申请的范围。The following describes the principles and features of the present application with reference to the drawings. The examples given are only used to explain the present application, not to limit the scope of the present application.
本申请提出一种应对话术推荐方法,主要应用于电话中心客服系统中。This application proposes a recommended method of dialogue technology, which is mainly used in the call center customer service system.
如图1所示,图1为本申请应对话术推荐方法一实施例的流程示意图。As shown in FIG. 1, FIG. 1 is a schematic flowchart of an embodiment of a method for recommending a conversational application in this application.
本实施例中,该应对话术推荐方法包括:In this embodiment, the method for recommending conversational techniques includes:
步骤S10,在接收到客户进线后,实时获取所述进线客户与坐席的语音流。Step S10: After receiving the incoming customer line, obtain the voice streams of the incoming customer and the agent in real time.
当有客户进线并与坐席接通后,电话中心客服系统通过电话语音平台实时获取所述进线客户与坐席的语音流(即客户与坐席的交互音频流)。When a customer enters the line and connects with the agent, the call center customer service system obtains in real time the voice stream of the incoming customer and agent (that is, the interactive audio stream between the customer and the agent) through the telephone voice platform.
步骤S20,对所述语音流进行特征提取,提取出所述语音流中的第一语音片段的第一音频特征向量,其中,所述第一语音片段对应所述语音流中的客户输入的语音段。Step S20: Perform feature extraction on the voice stream to extract the first audio feature vector of the first voice segment in the voice stream, where the first voice segment corresponds to the voice input by the customer in the voice stream segment.
电话中心客服系统对当前获取的语音流进行特征提取,提取出该语音流中的客户输入的语音段对应的第一音频特征向量。所述第一音频特征向量可以为包括以下几种音频特征中的一种或多种:能量特征、发音帧数特征、基音频率特征、共振峰特征、谐波噪声比特征以及梅尔倒谱系数特征。The call center customer service system performs feature extraction on the currently acquired voice stream, and extracts the first audio feature vector corresponding to the voice segment input by the customer in the voice stream. The first audio feature vector may be one or more of the following audio features: energy feature, pronunciation frame feature, pitch frequency feature, formant feature, harmonic noise ratio feature, and Mel cepstrum coefficient feature.
步骤S30,将所述第一音频特征向量输入预设的语音情绪分析模型中进行分析,识别出所述第一音频特征向量对应的第一语音内容和第一情绪分类。In step S30, the first audio feature vector is input into a preset voice emotion analysis model for analysis, and the first voice content and the first emotion classification corresponding to the first audio feature vector are identified.
电话中心客服系统中具有预设的语音情绪分析模型,在从当前获取的语音流中提取出第一音频特征向量后,电话中心客服系统将所述第一音频特征向量输入预设的语音情绪分析模型中进行分析识别,以识别出所述第一音频特征向量所对应的第一语音内容(即获取的语音流中客户输入的语音段所对应的文字内容)和第一情绪分类(即客户的情绪分类)。本实施例中,所述预设的语音情绪分析模型优选采用包括DNN-HMM声学模型、n-gram语言模型和wfst加权图的情绪分析模型,第一音频特征向量优选梅尔倒谱系数特征向量;当然,所述预设的语音情绪分析模型也可以是其它情绪分析模型。所述第一情绪分类包括:满意类、平静类、烦躁类以及生气类,等。The call center customer service system has a preset voice emotion analysis model. After extracting the first audio feature vector from the currently acquired voice stream, the call center customer service system inputs the first audio feature vector into the preset voice emotion analysis Perform analysis and recognition in the model to identify the first speech content corresponding to the first audio feature vector (that is, the text content corresponding to the speech segment input by the customer in the acquired speech stream) and the first emotion classification (that is, the customer's Emotion classification). In this embodiment, the preset voice sentiment analysis model preferably adopts a sentiment analysis model including a DNN-HMM acoustic model, an n-gram language model and a wfst weighted graph, and the first audio feature vector is preferably a melody cepstrum coefficient feature vector Of course, the preset voice sentiment analysis model may also be other sentiment analysis models. The first emotion classification includes: satisfaction, calm, irritability, and anger, etc.
步骤S40,将识别出的所述第一语音内容和所述第一情绪分类输入预先训练好的应对话术推荐模型中进行分析,以得到推荐的应对话术,将所述推荐的应对话术实时发送至所述坐席的终端。Step S40, the recognized first speech content and the first emotion classification are input into a pre-trained conversational dialogue recommendation model for analysis, to obtain a recommended conversational dialogue, and the recommended conversational dialogue Send to the agent's terminal in real time.
电话中心客服系统中还具有预先训练好的应对话术推荐模型,在通过预设的语音情绪分析模型识别出所述第一音频特征向量对应的第一语音内容和第一情绪分类时,将识别出的第一语音内容和第一情绪分类输入该应对话术推荐模型中分析,得出推荐的应对话术,并实时将该推荐的应对话术发送至所述坐席的终端进行展示,以供坐席参考,帮助坐席应对客户。The customer service system in the call center also has a pre-trained recommendation model for conversational skills. When the first speech feature and the first emotion classification corresponding to the first audio feature vector are identified through a preset speech emotion analysis model, the The first speech content and the first sentiment classification are input into the recommended model for analysis, and the recommended response technique is obtained, and the recommended response technique is sent to the terminal of the agent for display in real time for Agent reference, help agents deal with customers.
本实施例技术方案,通过在客户进线后与坐席通话时,实时获取进线客户与坐席的语音流,从获取的语音流中提取出客户输入的语音段对应的第一音频特征向量,将所述第一音频特征向量输入预设的语音情绪分析模型中进行分析识别,识别出第一音频特征向量对应的第一语音内容和第一情绪分类,最终利用预先训练好的应对话术模型对识别出的第一语音内容和第一情绪分类进行分析,得出推荐的应对话术实时发送给所述坐席的终端,以供坐席参考,帮助坐席应对客户;如此,有效的改善了客户因自身情绪问题而与坐席产生冲突时,坐席不知如何应对客户的情况,减少了客户投诉和客户流失。In the technical solution of this embodiment, by acquiring the voice streams of the incoming client and the agent in real time after the customer enters the line, and extracting the first audio feature vector corresponding to the voice segment input by the customer from the acquired voice stream, the The first audio feature vector is input into a preset speech emotion analysis model for analysis and recognition, and the first speech content and the first emotion classification corresponding to the first audio feature vector are identified, and finally a pre-trained response model is used The recognized first speech content and the first sentiment classification are analyzed, and the recommended response should be sent to the agent's terminal in real time for reference by the agent to help the agent deal with the customer; in this way, the customer's own When emotional issues conflict with the agent, the agent does not know how to deal with the customer's situation, reducing customer complaints and customer churn.
如图2所示,图2为本申请应对话术推荐方法二实施例的流程示意图。As shown in FIG. 2, FIG. 2 is a schematic flowchart of a second embodiment of a method for recommending conversational application in this application.
本实施例中,所述应对话术推荐方法在所述步骤S10之后,还包括:In this embodiment, after the step S10, the method for recommending conversational skills further includes:
步骤S50,对所述语音流进行特征提取,提取出所述语音流中的第二语音片段的第二音频特征向量,其中,所述第二语音片段对应所述语音流中的坐席输入的语音段。Step S50: Perform feature extraction on the voice stream to extract a second audio feature vector of a second voice segment in the voice stream, where the second voice segment corresponds to the voice input by the agent in the voice stream segment.
电话中心客服系统对当前获取的语音流进行特征提取,提取出该语音流中的坐席输入的语音段对应的第二音频特征向量。所述第二音 频特征向量可以包括以下几种音频特征中的一种或多种:能量特征、发音帧数特征、基音频率特征、共振峰特征、谐波噪声比特征以及梅尔倒谱系数特征。The call center customer service system performs feature extraction on the currently acquired voice stream to extract the second audio feature vector corresponding to the voice segment input by the agent in the voice stream. The second audio feature vector may include one or more of the following audio features: energy feature, pronunciation frame feature, pitch frequency feature, formant feature, harmonic-to-noise ratio feature, and Mel cepstral coefficient feature .
步骤S60,将所述第二音频特征向量输入预设的语音情绪分析模型中进行分析,识别出所述第二音频特征向量对应的第二情绪分类。In step S60, the second audio feature vector is input into a preset voice emotion analysis model for analysis, and a second emotion classification corresponding to the second audio feature vector is identified.
在从当前获取的语音流中提取出第二音频特征向量后,电话中心客服系统将所述第二音频特征向量输入预设的语音情绪分析模型中进行分析识别,以识别出所述第一音频特征向量所对应的第二情绪分类(即坐席的情绪分类)。所述第二情绪分类包括:满意类、平静类、烦躁类以及生气类,等。After extracting the second audio feature vector from the currently acquired voice stream, the call center customer service system inputs the second audio feature vector into a preset voice sentiment analysis model for analysis and recognition to identify the first audio The second emotion classification corresponding to the feature vector (ie, the emotion classification of the agent). The second emotion classification includes: satisfaction, calm, irritability, and anger, etc.
步骤S70,若所述第二情绪分类为预设的异常情绪分类,则向所述坐席的终端发送预设的第一提醒信息。Step S70, if the second emotion classification is a preset abnormal emotion classification, then preset preset reminder information is sent to the terminal of the agent.
当发现坐席的情绪分类(即第二情绪分类)为异常情绪分类(例如,烦躁类、生气类,等情绪不积极的情绪分类)时,电话中心客服系统则会向坐席的终端发送预设的第一提醒信息,以提醒坐席注意情绪,及时调整好服务态度。所述第一提醒信息例如为:“检测到你的服务态度消极,请注意调整服务态度”,等等。When it is found that the emotion classification of the agent (ie, the second emotion classification) is an abnormal emotion classification (for example, irritability, anger, and other emotional classifications that are not positive emotions), the call center customer service system will send a preset to the terminal of the agent The first reminder message is to remind the agent to pay attention to emotions and adjust the service attitude in time. The first reminder message is, for example: "Your service attitude is detected to be negative, please pay attention to adjust the service attitude", and so on.
本实施例通过实时识别坐席的情绪分类,侦测坐席的情绪变化,在坐席情绪发生异常(即变差)时,实时提醒坐席注意调整情绪和状态,从而更好的保证坐席对客户的服务质量,提升客户的满意度。This embodiment recognizes the emotion classification of the agent in real time, detects the change of the agent's emotion, and prompts the agent to adjust the emotion and state in real time when the agent's emotion is abnormal (that is, deteriorates), so as to better ensure the service quality of the agent to the customer To increase customer satisfaction.
如图3所示,图3为本申请应对话术推荐方法三实施例的流程示意图。As shown in FIG. 3, FIG. 3 is a schematic flowchart of three embodiments of the application-recommended method of application.
本实施例的应对话术推荐方法,在所述步骤S60中,所述预设的语音情绪分析模型还识别出所述第二音频特征向量对应的第二语音内容(即获取的语音流中坐席输入的语音段所对应的文字内容);所述应对话术推荐方法在所述步骤S60之后,还包括:In the method for recommending conversational skills in this embodiment, in step S60, the preset voice sentiment analysis model also recognizes the second voice content corresponding to the second audio feature vector (that is, the agent in the acquired voice stream The text content corresponding to the input voice segment); the recommended dialogue method recommendation method after the step S60, further includes:
步骤S01,在所述第二情绪分类为预设的异常情绪分类时,分析所述第二语音内容中是否包含预设的敏感词。Step S01: When the second emotion classification is a preset abnormal emotion classification, analyze whether the second speech content contains a preset sensitive word.
电话中心客服系统中设置了敏感词库(敏感词库中包括很多敏感词),当识别出的第二情绪分类(即坐席的情绪分类)位预设的异常情绪分类时,电话中心客服系统根据敏感词库,分析第二语音内容(即获取的语音流中坐席输入的语音段所对应的文字内容)中是否包含预设的敏感词(例如,不礼貌、不文明的词汇)。A sensitive thesaurus is set in the customer service system of the call center (the sensitive word library includes many sensitive words). When the identified second emotion classification (that is, the emotion classification of the agent) is a preset abnormal emotion classification, the call center customer service system is based on Sensitive word library, analyze whether the second speech content (that is, the text content corresponding to the speech segment input by the agent in the acquired speech stream) contains preset sensitive words (for example, impolite and uncivilized words).
步骤S02,若包含预设的敏感词,则分析所述第二语音内容中出现预设的敏感词的次数否大于第一阈值;Step S02: if the preset sensitive words are included, analyze whether the number of times the preset sensitive words appear in the second speech content is greater than the first threshold;
若确定所述第二语音内容中包含预设的敏感词,则说明当前坐席的言辞使用不当,此时进一步通过分析所述第二语音内容中出现预设 的敏感词的次数来判断坐席言辞使用不当的严重程度,根据严重程度进行相应的处理。具体为将出现预设的敏感词的次数与第一阈值(例如3次)比较。If it is determined that the second voice content contains preset sensitive words, it means that the current agent's words are used improperly. At this time, the agent's word use is further judged by analyzing the number of times the preset sensitive words appear in the second voice content Improper severity, according to the severity of the corresponding treatment. Specifically, the number of occurrences of the preset sensitive word is compared with a first threshold (for example, 3 times).
步骤S03,若所述第二语音内容中出现预设的敏感词的次数小于等于所述第一阈值,则向所述坐席的终端发送预设的第二提醒信息。Step S03: If the number of times the preset sensitive words appear in the second voice content is less than or equal to the first threshold, send preset second reminder information to the terminal of the agent.
当确定第二语音内容中出现预设的敏感词的次数小于等于所述第一阈值时,电话中心客服系统判定为所述坐席的言辞使用不当的情况不是特别严重,此时则向所述坐席的终端发送预设的第二提醒信息,以提醒坐席注意言辞,不要使用敏感词。同时,电话中心客服系统还可将坐席的第二语音内容中出现的预设的敏感词发送到坐席的终端突出显示。所述第二提醒消息例如为:“请注意言辞谨慎,禁止使用敏感词”,等。When it is determined that the number of occurrences of the preset sensitive words in the second voice content is less than or equal to the first threshold, the call center customer service system determines that the use of the agent's words is not particularly serious, and then sends the agent Terminal sends a preset second reminder message to remind the agent to pay attention to the words and not to use sensitive words. At the same time, the call center customer service system can also send the preset sensitive words appearing in the second voice content of the agent to the terminal of the agent for highlighting. The second reminder message is, for example, "Please pay attention to words, and prohibit the use of sensitive words", etc.
步骤S04,若所述第二语音内容中出现预设的敏感词的次数大于所述第一阈值,则向所述坐席的上级管理终端发送预设的第三提醒信息。Step S04, if the number of times the preset sensitive words appear in the second voice content is greater than the first threshold, send preset third reminder information to the superior management terminal of the agent.
当确定第二语音内容中出现预设的敏感词的次数大于所述第一阈值时,电话中心客服系统判定为所述坐席的言辞使用不当的情况非常严重,有可能是出现坐席与客户争吵等异常情况,此时电话中心客服系统则向所述坐席的上级管理终端(上级管理人员的终端)发送预设的第三提醒信息,以提醒该坐席的上级领导或管理人员特别注意该坐席的通话情况。同时,电话中心客服系统也可以将该坐席与客户的通话语音实时转接到所述坐席的上级管理终端,让上级领导或管理人员直接监听到该次通话语音过程,以在坐席与客户发生争吵时及时处理。所述第三提醒消息例如为:“该坐席言辞出现严重问题,请及时处理”,等。When it is determined that the number of occurrences of the preset sensitive words in the second voice content is greater than the first threshold, the call center customer service system determines that the agent's words are used improperly, which may be due to the agent arguing with the customer, etc. Abnormal situation, at this time, the call center customer service system sends the preset third reminder message to the superior management terminal of the agent (the terminal of the superior manager) to remind the superior leader or manager of the agent to pay special attention to the agent's call Happening. At the same time, the call center customer service system can also transfer the call voice of the agent and the customer to the superior management terminal of the agent in real time, so that the superior leader or manager can directly monitor the voice process of the call to quarrel with the customer at the agent Timely handling. The third reminder message is, for example: "There is a serious problem with the agent's speech, please deal with it in time", etc.
本实施例中,所述应对话术推荐模型的训练过程包括:In this embodiment, the training process of the recommended model should include:
S1、从通话录音数据库中获取预设数量的服务标签为满意的录音数据,提取每一则录音数据中的第一语音片段的第一音频特征向量和第二语音片段的第二音频特征向量,其中,所述第一语音片段对应所述录音数据中的客户输入的语音段,所述第二语音片段对应所述录音数据中的坐席输入的语音段。S1. Obtain a preset number of service tags from the call recording database as satisfactory recording data, and extract the first audio feature vector of the first speech segment and the second audio feature vector of the second speech segment in each recording data, Wherein, the first voice segment corresponds to a voice segment input by a customer in the recording data, and the second voice segment corresponds to a voice segment input by an agent in the recording data.
电话中心客服系统会对每一次客户进线的通话进行录音并保存到通话录音数据库中,每则录音数据通常都标记有客户针对坐席提供的通话服务所反馈的满意度的服务标签。先从通话录音数据库中获取预设数量(例如10000个)的录音数据,这些录音数据均是服务标签为满意的录音数据;对获取的每一则录音数据进行音频特征向量提取,得到每一则录音数据中的第一语音片段(客户输入的语音段)对 应的第一音频特征向量和第二语音片段(坐席输入的语音段)对应的第二音频特征向量。所述音频特征向量可以为包括以下几种音频特征中的一种或多种:能量特征、发音帧数特征、基音频率特征、共振峰特征、谐波噪声比特征以及梅尔倒谱系数特征。The customer service system of the call center will record every call that a customer enters and save it in the call recording database. Each recorded data is usually marked with a service tag of the customer's satisfaction for the call service provided by the agent. First obtain a preset number (for example, 10,000) of recording data from the call recording database. These recording data are all recording data whose service tag is satisfactory; extract audio feature vectors for each recording data obtained to obtain each The first audio feature vector corresponding to the first speech segment (voice segment input by the customer) in the recorded data and the second audio feature vector corresponding to the second speech segment (voice segment input by the agent). The audio feature vector may include one or more of the following audio features: energy feature, pronunciation frame number feature, pitch frequency feature, formant feature, harmonic noise ratio feature, and Mel cepstrum coefficient feature.
S2、采用预设的语音情绪分析模型中分别识别出所述第一音频特征向量对应的第一语音内容和第一情绪分类,以及所述第二音频特征向量对应的第二语音内容,将每一则录音数据对应的第一语音内容、第一情绪分类和第二语音内容作为一个样本,得到预设数量的样本。S2. Identify the first speech content and the first sentiment category corresponding to the first audio feature vector and the second speech content corresponding to the second audio feature vector using a preset speech emotion analysis model. The first voice content, the first emotion classification, and the second voice content corresponding to the recorded data are used as a sample to obtain a preset number of samples.
然后,采用预设的语音情绪分析模型对每一则录音数据对应的第一音频特征向量和第二音频特征向量进行分析识别,识别得到每一个第一音频特征向量对应的第一语音内容和第一情绪分类,以及识别得到每一个第二音频特征向量对应的第二语音内容;,将每一则录音数据对应的第一语音内容、第一情绪分类和第二语音内容构成一个样本,这样就得到了预设数量的样本。所述预设的语音情绪分析模型优选采用包括DNN-HMM声学模型、n-gram语言模型和wfst加权图的情绪分析模型,第一音频特征向量和第二音频特征向量优选为梅尔倒谱系数特征向量;当然,所述预设的语音情绪分析模型也可以是其它情绪分析模型。所述第一情绪分类包括:满意类、平静类、烦躁类以及生气类,等。Then, a preset voice sentiment analysis model is used to analyze and recognize the first audio feature vector and the second audio feature vector corresponding to each recording data, and the first speech content and the first speech feature corresponding to each first audio feature vector are identified. An emotion classification, and recognizing the second speech content corresponding to each second audio feature vector; the first speech content, the first emotion classification, and the second speech content corresponding to each recording data form a sample, so that A preset number of samples were obtained. The preset speech sentiment analysis model preferably uses a sentiment analysis model including a DNN-HMM acoustic model, an n-gram language model and a wfst weighted graph, and the first audio feature vector and the second audio feature vector are preferably Mel cepstrum coefficients Feature vector; of course, the preset voice sentiment analysis model may also be other sentiment analysis models. The first emotion classification includes: satisfaction, calm, irritability, and anger, etc.
S3、将预设数量的样本的第一百分比作为训练集,第二百分比作为验证集,第一百分比和第二百分比之和小于百分之百。S3. The first percentage of the preset number of samples is used as the training set, and the second percentage is used as the verification set. The sum of the first percentage and the second percentage is less than 100%.
从得到的预设数量的样本的第一百分比(例如70%)的样本作为训练集,第二百分比(例如25%~30%)的样本作为验证集。From the obtained preset number of samples, a first percentage (for example, 70%) of the samples is used as the training set, and a second percentage (for example, 25% to 30%) of the samples is used as the verification set.
S4、利用所述训练集中的样本对预设的应对话术推荐模型进行训练,并在训练结束后,利用验证集中的样本对所述应对话术推荐模型进行验证;S4. Use the samples in the training set to train the preset recommended conversational model, and after the training, use the samples in the verification set to verify the recommended conversational model;
利用训练集的样本对预设的应对话术推荐模型进行训练,训练结束后,再利用验证集中的样本对训练完成的应对话术推荐模型进行验证,检验应对话术推荐模型的训练效果。本实施例中,采用以下标准来检验应对话术推荐模型的准确率:应对话术推荐模型针对验证集中的一个样本预测得到的应对话术后,若该应对话术的内容中超过N%(例如95%)的内容与该样本的第二语音内容相同,则判定为预测准确。The samples of the training set are used to train the pre-trained recommendation model. After the training is completed, the samples in the verification set are used to verify the completed training model. In this embodiment, the following criteria are used to test the accuracy of the recommended dialogue model: The recommended dialogue model predicts a sample of the verification set after the corresponding dialogue, if the content of the corresponding dialogue exceeds N% ( For example, the content of 95%) is the same as the second speech content of the sample, and it is determined that the prediction is accurate.
S5、若所述预测准确率大于预设阈值,则模型训练结束。S5. If the prediction accuracy rate is greater than a preset threshold, the model training ends.
系统中预先设置了预测准确率的预设阈值(例如97%),用于对所述应对话术推荐模型的训练效果进行检验,若预测准确率大于所述预设阈值,那么说明所述应对话术推荐模型的训练达到了预设标准,此时则结束模型训练。A preset threshold (for example, 97%) of the prediction accuracy rate is preset in the system, which is used to check the training effect of the recommended model of the application. If the prediction accuracy rate is greater than the preset threshold, then the application The training of the dialogue recommended model has reached the preset standard, and then the model training ends.
S6、若所述预设准确率小于或等于所述预设阈值,则增大所述预设数量的值,并重复执行步骤S1至S4。S6. If the preset accuracy rate is less than or equal to the preset threshold, increase the preset number of values, and repeat steps S1 to S4.
若是预测准确率小于或等于所述预设阈值,那么说明所述应对话术推荐模型的训练还没有达到了预设标准,可能是训练集的样本数量不够或验证集的样本数量不够,所以,在这种情况时,则增大所述预设数量(即增大了样本数量,例如,每次增加固定数量或每次增加随机数量),然后在这基础上,重新执行上述步骤S1-S4,如此循环执行,直至达到了步骤S5的要求,则结束模型训练。If the prediction accuracy rate is less than or equal to the preset threshold, it means that the training of the recommended model should not reach the preset standard. It may be that the number of samples in the training set is insufficient or the number of samples in the verification set is insufficient. In this case, increase the preset number (ie, increase the number of samples, for example, increase the fixed number each time or increase the random number each time), and then, on this basis, re-execute the above steps S1-S4 , And so on, until the requirement of step S5 is reached, then the model training is ended.
此外,本申请还提出一种应对话术推荐程序。In addition, this application also proposes a recommended procedure for dialogue.
请参阅图4,是本申请应对话术推荐程序10较佳实施例的运行环境示意图。Please refer to FIG. 4, which is a schematic diagram of the operating environment of the preferred embodiment of the application recommended procedure 10 of the present application.
在本实施例中,应对话术推荐程序10安装并运行于电子装置1中。电子装置1可以是桌上型计算机、笔记本、掌上电脑及服务器等计算设备。该电子装置1可包括,但不仅限于,存储器11、处理器12及显示器13。图4仅示出了具有组件11-13的电子装置1,但是应理解的是,并不要求实施所有示出的组件,可以替代的实施更多或者更少的组件。In this embodiment, the conversational recommendation program 10 should be installed and run in the electronic device 1. The electronic device 1 may be a computing device such as a desktop computer, a notebook, a palmtop computer, and a server. The electronic device 1 may include, but is not limited to, a memory 11, a processor 12, and a display 13. FIG. 4 only shows the electronic device 1 having the components 11-13, but it should be understood that it is not required to implement all the components shown, and more or fewer components may be implemented instead.
存储器11在一些实施例中可以是电子装置1的内部存储单元,例如该电子装置1的硬盘或内存。存储器11在另一些实施例中也可以是电子装置1的外部存储设备,例如电子装置1上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。进一步地,存储器11还可以既包括电子装置1的内部存储单元也包括外部存储设备。存储器11用于存储安装于电子装置1的应用软件及各类数据,例如应对话术推荐程序10的程序代码等。存储器11还可以用于暂时地存储已经输出或者将要输出的数据。The memory 11 may be an internal storage unit of the electronic device 1 in some embodiments, such as a hard disk or a memory of the electronic device 1. In other embodiments, the memory 11 may also be an external storage device of the electronic device 1, such as a plug-in hard disk equipped on the electronic device 1, a smart memory card (Smart, Media, Card, SMC), and a secure digital (SD) Cards, flash cards, etc. Further, the memory 11 may also include both an internal storage unit of the electronic device 1 and an external storage device. The memory 11 is used to store application software and various types of data installed in the electronic device 1, for example, program codes of the program 10 recommended for communication. The memory 11 can also be used to temporarily store data that has been or will be output.
处理器12在一些实施例中可以是一中央处理器(Central Processing Unit,CPU),微处理器或其他数据处理芯片,用于运行存储器11中存储的程序代码或处理数据,例如执行应对话术推荐程序10等。In some embodiments, the processor 12 may be a central processing unit (CPU), microprocessor, or other data processing chip, which is used to run the program code or process data stored in the memory 11, for example, to execute the interactive technology Recommended program 10 etc.
显示器13在一些实施例中可以是LED显示器、液晶显示器、触控式液晶显示器以及OLED(Organic Light-Emitting Diode,有机发光二极管)触摸器等。显示器13用于显示在电子装置1中处理的信息以及用于显示可视化的用户界面。电子装置1的部件11-13通过程序总线相互通信。In some embodiments, the display 13 may be an LED display, a liquid crystal display, a touch-sensitive liquid crystal display, an OLED (Organic Light-Emitting Diode, organic light emitting diode) touch device, or the like. The display 13 is used to display information processed in the electronic device 1 and to display a visual user interface. The components 11-13 of the electronic device 1 communicate with each other through a program bus.
请参阅图5,是本申请应对话术推荐程序10一实施例的程序模块图。在本实施例中,应对话术推荐程序10可以被分割成一个或多 个模块,一个或者多个模块被存储于存储器11中,并由一个或多个处理器(本实施例为处理器12)所执行,以完成本申请。例如,在图5中,应对话术推荐程序10可以被分割成实时获取模块101、第一提取模块102、第一识别模块103及推荐模块104。本申请所称的模块是指能够完成特定功能的一系列计算机程序指令段,比程序更适合于描述应对话术推荐程序10在电子装置1中的执行过程,其中:Please refer to FIG. 5, which is a program block diagram of an embodiment of the application recommended program 10 of this application. In this embodiment, the application recommendation program 10 can be divided into one or more modules, and the one or more modules are stored in the memory 11 and are processed by one or more processors (in this embodiment, the processor 12 ) To complete this application. For example, in FIG. 5, the conversational skill recommendation program 10 may be divided into a real-time acquisition module 101, a first extraction module 102, a first recognition module 103, and a recommendation module 104. The module referred to in this application refers to a series of computer program instruction segments capable of performing specific functions, and is more suitable than the program for describing the execution process of the recommended program 10 in the electronic device 1, wherein:
实时获取模块101,用于在接收到客户进线后,实时获取所述进线客户与坐席的语音流。The real-time acquiring module 101 is configured to acquire the voice streams of the incoming customer and the agent in real time after receiving the incoming customer.
当有客户进线并与坐席接通后,电话中心客服系统通过电话语音平台实时获取所述进线客户与坐席的语音流(即客户与坐席的交互音频流)。When a customer enters the line and connects with the agent, the call center customer service system obtains in real time the voice stream of the incoming customer and the agent (that is, the interactive audio stream between the customer and the agent) through the telephone voice platform.
第一提取模块102,用于对所述语音流进行特征提取,提取出所述语音流中的第一语音片段的第一音频特征向量,其中,所述第一语音片段对应所述语音流中的客户输入的语音段。The first extraction module 102 is configured to perform feature extraction on the voice stream to extract a first audio feature vector of a first voice segment in the voice stream, where the first voice segment corresponds to the voice stream Speech segments entered by customers.
电话中心客服系统对当前获取的语音流进行特征提取,提取出该语音流中的客户输入的语音段对应的第一音频特征向量。所述第一音频特征向量可以为包括以下几种音频特征中的一种或多种:能量特征、发音帧数特征、基音频率特征、共振峰特征、谐波噪声比特征以及梅尔倒谱系数特征。The call center customer service system performs feature extraction on the currently acquired voice stream, and extracts the first audio feature vector corresponding to the voice segment input by the customer in the voice stream. The first audio feature vector may be one or more of the following audio features: energy feature, pronunciation frame feature, pitch frequency feature, formant feature, harmonic noise ratio feature, and Mel cepstrum coefficient feature.
第一识别模块103,用于将所述第一音频特征向量输入预设的语音情绪分析模型中进行分析,识别出所述第一音频特征向量对应的第一语音内容和第一情绪分类。The first recognition module 103 is configured to input the first audio feature vector into a preset voice emotion analysis model for analysis, and recognize the first voice content and the first emotion classification corresponding to the first audio feature vector.
电话中心客服系统中具有预设的语音情绪分析模型,在从当前获取的语音流中提取出第一音频特征向量后,电话中心客服系统将所述第一音频特征向量输入预设的语音情绪分析模型中进行分析识别,以识别出所述第一音频特征向量所对应的第一语音内容(即获取的语音流中客户输入的语音段所对应的文字内容)和第一情绪分类(即客户的情绪分类)。本实施例中,所述预设的语音情绪分析模型优选采用包括DNN-HMM声学模型、n-gram语言模型和wfst加权图的情绪分析模型,第一音频特征向量优选梅尔倒谱系数特征向量;当然,所述预设的语音情绪分析模型也可以是其它情绪分析模型。所述第一情绪分类包括:满意类、平静类、烦躁类以及生气类,等。The call center customer service system has a preset voice emotion analysis model. After extracting the first audio feature vector from the currently acquired voice stream, the call center customer service system inputs the first audio feature vector into the preset voice emotion analysis Perform analysis and recognition in the model to identify the first speech content corresponding to the first audio feature vector (that is, the text content corresponding to the speech segment input by the customer in the acquired speech stream) and the first emotion classification (that is, the customer's Emotion classification). In this embodiment, the preset voice sentiment analysis model preferably adopts a sentiment analysis model including a DNN-HMM acoustic model, an n-gram language model and a wfst weighted graph, and the first audio feature vector is preferably a melody cepstrum coefficient feature vector Of course, the preset voice sentiment analysis model may also be other sentiment analysis models. The first emotion classification includes: satisfaction, calm, irritability, and anger, etc.
推荐模块104,用于将识别出的所述第一语音内容和所述第一情绪分类输入预先训练好的应对话术推荐模型中进行分析,以得到推荐的应对话术,将所述推荐的应对话术实时发送至所述坐席的终端。The recommendation module 104 is configured to input the recognized first speech content and the first emotion classification into a pre-trained dialogue-recommendation recommendation model for analysis, so as to obtain a recommended dialogue-reaction technique, and convert the recommended Dialogue is sent to the agent's terminal in real time.
电话中心客服系统中还具有预先训练好的应对话术推荐模型,在通过预设的语音情绪分析模型识别出所述第一音频特征向量对应的第一语音内容和第一情绪分类时,将识别出的第一语音内容和第一情 绪分类输入该应对话术推荐模型中分析,得出推荐的应对话术,并实时将该推荐的应对话术发送至所述坐席的终端进行展示,以供坐席参考,帮助坐席应对客户。The customer service system in the call center also has a pre-trained recommendation model for conversational skills. When the first speech feature and the first emotion classification corresponding to the first audio feature vector are identified through a preset speech emotion analysis model, the The first speech content and the first sentiment classification are input into the recommended model for analysis, and the recommended response technique is obtained, and the recommended response technique is sent to the terminal of the agent for display in real time for Agent reference, help agents deal with customers.
本实施例技术方案,通过在客户进线后与坐席通话时,实时获取进线客户与坐席的语音流,从获取的语音流中提取出客户输入的语音段对应的第一音频特征向量,将所述第一音频特征向量输入预设的语音情绪分析模型中进行分析识别,识别出第一音频特征向量对应的第一语音内容和第一情绪分类,最终利用预先训练好的应对话术模型对识别出的第一语音内容和第一情绪分类进行分析,得出推荐的应对话术实时发送给所述坐席的终端,以供坐席参考,帮助坐席应对客户;如此,有效的改善了客户因自身情绪问题而与坐席产生冲突时,坐席不知如何应对客户的情况,减少了客户投诉和客户流失。In the technical solution of this embodiment, by acquiring the voice streams of the incoming client and the agent in real time after the customer enters the line, and extracting the first audio feature vector corresponding to the voice segment input by the customer from the acquired voice stream, the The first audio feature vector is input into a preset speech emotion analysis model for analysis and recognition, and the first speech content and the first emotion classification corresponding to the first audio feature vector are identified, and finally a pre-trained response model is used The recognized first speech content and the first sentiment classification are analyzed, and the recommended response should be sent to the agent's terminal in real time for reference by the agent to help the agent deal with the customer; in this way, the customer's own When emotional issues conflict with the agent, the agent does not know how to deal with the customer's situation, reducing customer complaints and customer churn.
本实施例中,所述应对话术推荐模型的训练方法参照上述应对话术推荐方法的实施例中的描述,在此不赘述。In this embodiment, for the training method of the conversational recommendation model, refer to the description in the above embodiment of the conversational recommendation method, and details are not described herein.
参照图6,本实施例中,所述应对话术推荐程序还包括第二提取模块105、第二识别模块106和提醒模块107。Referring to FIG. 6, in this embodiment, the conversational skill recommendation program further includes a second extraction module 105, a second recognition module 106 and a reminder module 107.
第二提取模块105,用于对所述语音流进行特征提取,提取出所述语音流中的第二语音片段的第二音频特征向量,其中,所述第二语音片段对应所述语音流中的坐席输入的语音段。The second extraction module 105 is configured to perform feature extraction on the voice stream to extract a second audio feature vector of a second voice segment in the voice stream, where the second voice segment corresponds to the voice stream Voice segment entered by the agent.
电话中心客服系统对当前获取的语音流进行特征提取,提取出该语音流中的坐席输入的语音段对应的第二音频特征向量。所述第二音频特征向量可以包括以下几种音频特征中的一种或多种:能量特征、发音帧数特征、基音频率特征、共振峰特征、谐波噪声比特征以及梅尔倒谱系数特征。The call center customer service system performs feature extraction on the currently acquired voice stream to extract the second audio feature vector corresponding to the voice segment input by the agent in the voice stream. The second audio feature vector may include one or more of the following audio features: energy feature, pronunciation frame feature, pitch frequency feature, formant feature, harmonic-to-noise ratio feature, and Mel cepstral coefficient feature .
第二识别模块106,用于将所述第二音频特征向量输入预设的语音情绪分析模型中进行分析,识别出所述第二音频特征向量对应的第二情绪分类。The second recognition module 106 is configured to input the second audio feature vector into a preset voice emotion analysis model for analysis, and recognize a second emotion classification corresponding to the second audio feature vector.
在从当前获取的语音流中提取出第二音频特征向量后,电话中心客服系统将所述第二音频特征向量输入预设的语音情绪分析模型中进行分析识别,以识别出所述第一音频特征向量所对应的第二情绪分类(即坐席的情绪分类)。所述第二情绪分类包括:满意类、平静类、烦躁类以及生气类,等。After extracting the second audio feature vector from the currently acquired voice stream, the call center customer service system inputs the second audio feature vector into a preset voice sentiment analysis model for analysis and recognition to identify the first audio The second emotion classification corresponding to the feature vector (ie, the emotion classification of the agent). The second emotion classification includes: satisfaction, calm, irritability, and anger, etc.
提醒模块107,用于在所述第二情绪分类为预设的异常情绪分类时,向所述坐席的终端发送预设的第一提醒信息。The reminder module 107 is configured to send preset first reminder information to the terminal of the agent when the second emotion category is a preset abnormal emotion category.
当发现坐席的情绪分类(即第二情绪分类)为异常情绪分类(例如,烦躁类、生气类,等情绪不积极的情绪分类)时,电话中心客服系统则会向坐席的终端发送预设的第一提醒信息,以提醒坐席注意情 绪,及时调整好服务态度。所述第一提醒信息例如为:“检测到你的服务态度消极,请注意调整服务态度”,等等。When it is found that the emotion classification of the agent (ie, the second emotion classification) is an abnormal emotion classification (for example, irritability, anger, and other emotional classifications with inactive emotions), the call center customer service system will send the preset The first reminder message is to remind the agent to pay attention to emotions and adjust the service attitude in time. The first reminder message is, for example: "Your service attitude is detected to be negative, please pay attention to adjust the service attitude", and so on.
本实施例通过实时识别坐席的情绪分类,侦测坐席的情绪变化,在坐席情绪发生异常(即变差)时,实时提醒坐席注意调整情绪和状态,从而更好的保证坐席对客户的服务质量,提升客户的满意度。This embodiment recognizes the emotion classification of the agent in real time, detects the change of the agent's emotion, and prompts the agent to adjust the emotion and state in real time when the agent's emotion is abnormal (that is, deteriorates), so as to better ensure the service quality of the agent to the customer To increase customer satisfaction.
参照图7,本实施例中,第二识别模块106还用于将所述第二音频特征向量输入预设的语音情绪分析模型中进行分析,识别出所述第二音频特征向量对应的第二语音内容(即获取的语音流中坐席输入的语音段所对应的文字内容);所述应对话术推荐程序还包括第一分析模块108和第二分析模块109。其中,Referring to FIG. 7, in this embodiment, the second recognition module 106 is further configured to input the second audio feature vector into a preset voice sentiment analysis model for analysis, and identify the second corresponding to the second audio feature vector Voice content (that is, the text content corresponding to the voice segment input by the agent in the acquired voice stream); the dialogue-speaking recommendation program further includes a first analysis module 108 and a second analysis module 109. among them,
第一分析模块108,用于在确定所述第二情绪分类为预设的异常情绪分类时,分析所述第二语音内容中是否包含预设的敏感词。The first analysis module 108 is configured to analyze whether the second speech content contains preset sensitive words when determining that the second emotion classification is a preset abnormal emotion classification.
电话中心客服系统中设置了敏感词库(敏感词库中包括很多敏感词),当识别出的第二情绪分类(即坐席的情绪分类)位预设的异常情绪分类时,电话中心客服系统根据敏感词库,分析第二语音内容(即获取的语音流中坐席输入的语音段所对应的文字内容)中是否包含预设的敏感词(例如,不礼貌、不文明的词汇)。A sensitive thesaurus is set in the customer service system of the call center (the sensitive word library includes many sensitive words). When the identified second emotion classification (that is, the emotion classification of the agent) is a preset abnormal emotion classification, the call center customer service system is based on Sensitive word library, analyze whether the second speech content (that is, the text content corresponding to the speech segment input by the agent in the acquired speech stream) contains preset sensitive words (for example, impolite and uncivilized words).
第二分析模块109,用于在确定所述第二语音内容中包含预设的敏感词时,分析所述第二语音内容中出现预设的敏感词的次数否大于第一阈值;The second analysis module 109 is configured to, when determining that the second voice content contains preset sensitive words, analyze whether the number of times the preset sensitive words appear in the second voice content is greater than a first threshold;
若确定所述第二语音内容中包含预设的敏感词,则说明当前坐席的言辞使用不当,此时进一步通过分析所述第二语音内容中出现预设的敏感词的次数来判断坐席言辞使用不当的严重程度,根据严重程度进行相应的处理。具体为将出现预设的敏感词的次数与第一阈值(例如3次)比较。If it is determined that the second voice content contains preset sensitive words, it means that the current agent's words are used improperly. At this time, the agent's word use is further determined by analyzing the number of times the preset sensitive words appear in the second voice content Improper severity, according to the severity of the corresponding treatment. Specifically, the number of occurrences of the preset sensitive word is compared with a first threshold (for example, 3 times).
所述提醒模块107还用于在确定所述第二语音内容中出现预设的敏感词的次数小于等于所述第一阈值时,向所述坐席的终端发送预设的第二提醒信息。The reminder module 107 is further configured to send preset second reminder information to the terminal of the agent when it is determined that the number of preset sensitive words appearing in the second voice content is less than or equal to the first threshold.
当确定第二语音内容中出现预设的敏感词的次数小于等于所述第一阈值时,电话中心客服系统判定为所述坐席的言辞使用不当的情况不是特别严重,此时则向所述坐席的终端发送预设的第二提醒信息,以提醒坐席注意言辞,不要使用敏感词。同时,电话中心客服系统还可将坐席的第二语音内容中出现的预设的敏感词发送到坐席的终端突出显示。所述第二提醒消息例如为:“请注意言辞谨慎,禁止使用敏感词”,等。When it is determined that the number of occurrences of the preset sensitive words in the second voice content is less than or equal to the first threshold, the call center customer service system determines that the use of the agent's words is not particularly serious, and then sends the agent Terminal sends a preset second reminder message to remind the agent to pay attention to the words and not to use sensitive words. At the same time, the call center customer service system can also send the preset sensitive words appearing in the second voice content of the agent to the terminal of the agent for highlighting. The second reminder message is, for example, "Please pay attention to words, and prohibit the use of sensitive words", etc.
所述提醒模块107还用于在确定所述第二语音内容中出现预设的敏感词的次数大于所述第一阈值时,向所述坐席的上级管理终端发 送预设的第三提醒信息。The reminder module 107 is further configured to send preset third reminder information to the superior management terminal of the agent when it is determined that the number of times the preset sensitive words appear in the second voice content is greater than the first threshold.
当确定第二语音内容中出现预设的敏感词的次数大于所述第一阈值时,电话中心客服系统判定为所述坐席的言辞使用不当的情况非常严重,有可能是出现坐席与客户争吵等异常情况,此时电话中心客服系统则向所述坐席的上级管理终端(上级管理人员的终端)发送预设的第三提醒信息,以提醒该坐席的上级领导或管理人员特别注意该坐席的通话情况。同时,电话中心客服系统也可以将该坐席与客户的通话语音实时转接到所述坐席的上级管理终端,让上级领导或管理人员直接监听到该次通话语音过程,以在坐席与客户发生争吵时及时处理。所述第三提醒消息例如为:“该坐席言辞出现严重问题,请及时处理”,等。When it is determined that the number of occurrences of the preset sensitive words in the second voice content is greater than the first threshold, the call center customer service system determines that the agent's words are used improperly, which may be due to the agent arguing with the customer. Abnormal situation, at this time, the call center customer service system sends the preset third reminder message to the superior management terminal of the agent (the terminal of the superior manager) to remind the superior leader or manager of the agent to pay special attention to the agent's call Happening. At the same time, the call center customer service system can also transfer the call voice of the agent and the customer to the superior management terminal of the agent in real time, so that the superior leader or manager can directly monitor the voice process of the call to quarrel with the customer at the agent Timely handling. The third reminder message is, for example: "There is a serious problem with the agent's speech, please deal with it in time", etc.
进一步地,本申请还提出一种计算机可读存储介质,所述计算机可读存储介质存储有应对话术推荐程序,所述应对话术推荐程序可被至少一个处理器执行,以使所述至少一个处理器执行上述任一实施例中的应对话术推荐方法。Further, the present application also proposes a computer-readable storage medium, the computer-readable storage medium storing a recommended program for conversational skills, the recommended program for conversational skills can be executed by at least one processor, so that the at least A processor executes the method for recommending interaction in any of the above embodiments.
以上所述仅为本申请的优选实施例,并非因此限制本申请的专利范围,凡是在本申请的发明构思下,利用本申请说明书及附图内容所作的等效结构变换,或直接/间接运用在其他相关的技术领域均包括在本申请的专利保护范围内。The above are only the preferred embodiments of the present application, and therefore do not limit the patent scope of the present application. Any equivalent structural transformation or direct / indirect use of the content of the description and drawings of the present application under the inventive concept of the present application All other related technical fields are included in the patent protection scope of this application.
Claims (20)
- 一种电子装置,其特征在于,所述电子装置包括存储器和处理器,所述存储器上存储有可在所述处理器上运行的应对话术推荐程序,所述应对话术推荐程序被所述处理器执行时实现如下步骤:An electronic device, characterized in that the electronic device includes a memory and a processor, and the memory stores a recommended interactive program that can run on the processor, and the recommended recommended program is The processor implements the following steps during execution:A1、在接收到客户进线后,实时获取所述进线客户与坐席的语音流;A1. After receiving the incoming customer line, obtain the voice streams of the incoming customer and agent in real time;B1、对所述语音流进行特征提取,提取出所述语音流中的第一语音片段的第一音频特征向量,其中,所述第一语音片段对应所述语音流中的客户输入的语音段;B1. Perform feature extraction on the voice stream to extract a first audio feature vector of a first voice segment in the voice stream, where the first voice segment corresponds to a voice segment input by a customer in the voice stream ;C1、将所述第一音频特征向量输入预设的语音情绪分析模型中进行分析,识别出所述第一音频特征向量对应的第一语音内容和第一情绪分类;C1. Input the first audio feature vector into a preset voice emotion analysis model for analysis, and identify the first voice content and the first emotion classification corresponding to the first audio feature vector;D1、将识别出的所述第一语音内容和所述第一情绪分类输入预先训练好的应对话术推荐模型中进行分析,以得到推荐的应对话术,将所述推荐的应对话术实时发送至所述坐席的终端。D1, input the recognized first speech content and the first sentiment classification into a pre-trained dialogue-recommended recommendation model for analysis, to obtain a recommended dialogue-reporting technique, and put the recommended dialogue-reported technique in real time Sent to the terminal of the agent.
- 如权利要求1所述的电子装置,其特征在于,在所述步骤A1之后,所述应对话术推荐程序被所述处理器执行时,还实现如下步骤:The electronic device according to claim 1, characterized in that, after the step A1, when the recommended program for conversational techniques is executed by the processor, the following steps are further implemented:E1、对所述语音流进行特征提取,提取出所述语音流中的第二语音片段的第二音频特征向量,其中,所述第二语音片段对应所述语音流中的坐席输入的语音段;E1: Perform feature extraction on the voice stream to extract a second audio feature vector of a second voice segment in the voice stream, where the second voice segment corresponds to a voice segment input by an agent in the voice stream ;F1、将所述第二音频特征向量输入预设的语音情绪分析模型中进行分析,识别出所述第二音频特征向量对应的第二情绪分类;F1. Input the second audio feature vector into a preset voice emotion analysis model for analysis, and identify a second emotion classification corresponding to the second audio feature vector;G1、若所述第二情绪分类为预设的异常情绪分类,则向所述坐席的终端发送预设的第一提醒信息。G1. If the second emotion classification is a preset abnormal emotion classification, send preset first reminder information to the terminal of the agent.
- 如权利要求2所述的电子装置,其特征在于,所述预设的语音情绪分析模型还识别出所述第二音频特征向量对应的第二语音内容,在所述步骤F1之后,所述应对话术推荐程序被所述处理器执行时,还实现如下步骤:The electronic device according to claim 2, wherein the preset voice sentiment analysis model also recognizes the second voice content corresponding to the second audio feature vector. After the step F1, the application When the dialogue recommendation program is executed by the processor, the following steps are also implemented:在所述第二情绪分类为预设的异常情绪分类时,分析所述第二语音内容中是否包含预设的敏感词;When the second emotion classification is a preset abnormal emotion classification, analyze whether the second speech content contains a preset sensitive word;若包含预设的敏感词,且所述第二语音内容中出现预设的敏感词的次数小于等于第一阈值,则向所述坐席的终端发送预设的第二提醒信息;If the preset sensitive words are included, and the number of times the preset sensitive words appear in the second voice content is less than or equal to the first threshold, then send preset second reminder information to the terminal of the agent;若包含预设的敏感词,且所述第二语音内容中出现预设的敏感词的次数大于第一阈值,则向所述坐席的上级管理终端发送预设的第三提醒信息。If the preset sensitive words are included, and the number of times the preset sensitive words appear in the second voice content is greater than the first threshold, the preset third reminder information is sent to the superior management terminal of the agent.
- 如权利要求1所述的电子装置,其特征在于,所述预设的语 音情绪分析模型包括DNN-HMM声学模型、n-gram语言模型和wfst加权图的情绪分析模型。The electronic device of claim 1, wherein the preset voice sentiment analysis model includes a DNN-HMM acoustic model, an n-gram language model, and a wfst weighted graph sentiment analysis model.
- 如权利要求1所述的电子装置,其特征在于,所述应对话术推荐模型的训练过程包括:The electronic device according to claim 1, characterized in that the training process of the recommended model should include:S1、从通话录音数据库中获取预设数量的服务标签为满意的录音数据,提取每一则录音数据中的第一语音片段的第一音频特征向量和第二语音片段的第二音频特征向量,其中,所述第一语音片段对应所述录音数据中的客户输入的语音段,所述第二语音片段对应所述录音数据中的坐席输入的语音段;S1. Obtain a preset number of service tags from the call recording database as satisfactory recording data, and extract the first audio feature vector of the first speech segment and the second audio feature vector of the second speech segment in each recording data, Wherein, the first voice segment corresponds to a voice segment input by a customer in the recording data, and the second voice segment corresponds to a voice segment input by an agent in the recording data;S2、采用预设的语音情绪分析模型中分别识别出所述第一音频特征向量对应的第一语音内容和第一情绪分类,以及所述第二音频特征向量对应的第二语音内容,将每一则录音数据对应的第一语音内容、第一情绪分类和第二语音内容作为一个样本,得到预设数量的样本;S2. Identify the first speech content and the first sentiment category corresponding to the first audio feature vector and the second speech content corresponding to the second audio feature vector using a preset speech emotion analysis model. The first voice content, the first emotion classification and the second voice content corresponding to the recorded data are used as a sample to obtain a preset number of samples;S3、将预设数量的样本的第一百分比作为训练集,第二百分比作为验证集,第一百分比和第二百分比之和小于百分之百;S3. The first percentage of the preset number of samples is used as the training set, the second percentage is used as the verification set, and the sum of the first percentage and the second percentage is less than 100%;S4、利用所述训练集中的样本对预设的应对话术推荐模型进行训练,并在训练结束后,利用验证集中的样本对所述应对话术推荐模型进行验证;S4. Use the samples in the training set to train the preset recommended conversational model, and after the training, use the samples in the verification set to verify the recommended conversational model;S5、若所述预测准确率大于预设阈值,则模型训练结束;S5. If the prediction accuracy rate is greater than a preset threshold, the model training ends;S6、若所述预设准确率小于或等于所述预设阈值,则增大所述预设数量的值,并重复执行步骤S1至S4。S6. If the preset accuracy rate is less than or equal to the preset threshold, increase the preset number of values, and repeat steps S1 to S4.
- 如权利要求2所述的电子装置,其特征在于,所述应对话术推荐模型的训练过程包括:The electronic device according to claim 2, characterized in that the training process of the recommended model should include:S1、从通话录音数据库中获取预设数量的服务标签为满意的录音数据,提取每一则录音数据中的第一语音片段的第一音频特征向量和第二语音片段的第二音频特征向量,其中,所述第一语音片段对应所述录音数据中的客户输入的语音段,所述第二语音片段对应所述录音数据中的坐席输入的语音段;S1. Obtain a preset number of service tags from the call recording database as satisfactory recording data, and extract the first audio feature vector of the first speech segment and the second audio feature vector of the second speech segment in each recording data, Wherein, the first voice segment corresponds to a voice segment input by a customer in the recording data, and the second voice segment corresponds to a voice segment input by an agent in the recording data;S2、采用预设的语音情绪分析模型中分别识别出所述第一音频特征向量对应的第一语音内容和第一情绪分类,以及所述第二音频特征向量对应的第二语音内容,将每一则录音数据对应的第一语音内容、第一情绪分类和第二语音内容作为一个样本,得到预设数量的样本;S2. Identify the first speech content and the first sentiment category corresponding to the first audio feature vector and the second speech content corresponding to the second audio feature vector using a preset speech emotion analysis model. The first voice content, the first emotion classification and the second voice content corresponding to the recorded data are used as a sample to obtain a preset number of samples;S3、将预设数量的样本的第一百分比作为训练集,第二百分比作为验证集,第一百分比和第二百分比之和小于百分之百;S3. The first percentage of the preset number of samples is used as the training set, the second percentage is used as the verification set, and the sum of the first percentage and the second percentage is less than 100%;S4、利用所述训练集中的样本对预设的应对话术推荐模型进行训练,并在训练结束后,利用验证集中的样本对所述应对话术推荐模型进行验证;S4. Use the samples in the training set to train the preset recommended conversational model, and after the training, use the samples in the verification set to verify the recommended conversational model;S5、若所述预测准确率大于预设阈值,则模型训练结束;S5. If the prediction accuracy rate is greater than a preset threshold, the model training ends;S6、若所述预设准确率小于或等于所述预设阈值,则增大所述预设数量的值,并重复执行步骤S1至S4。S6. If the preset accuracy rate is less than or equal to the preset threshold, increase the preset number of values, and repeat steps S1 to S4.
- 如权利要求3所述的电子装置,其特征在于,所述应对话术推荐模型的训练过程包括:The electronic device according to claim 3, characterized in that the training process of the recommended model should include:S1、从通话录音数据库中获取预设数量的服务标签为满意的录音数据,提取每一则录音数据中的第一语音片段的第一音频特征向量和第二语音片段的第二音频特征向量,其中,所述第一语音片段对应所述录音数据中的客户输入的语音段,所述第二语音片段对应所述录音数据中的坐席输入的语音段;S1. Obtain a preset number of service tags from the call recording database as satisfactory recording data, and extract the first audio feature vector of the first speech segment and the second audio feature vector of the second speech segment in each recording data, Wherein, the first voice segment corresponds to a voice segment input by a customer in the recording data, and the second voice segment corresponds to a voice segment input by an agent in the recording data;S2、采用预设的语音情绪分析模型中分别识别出所述第一音频特征向量对应的第一语音内容和第一情绪分类,以及所述第二音频特征向量对应的第二语音内容,将每一则录音数据对应的第一语音内容、第一情绪分类和第二语音内容作为一个样本,得到预设数量的样本;S2. Identify the first speech content and the first sentiment category corresponding to the first audio feature vector and the second speech content corresponding to the second audio feature vector using a preset speech emotion analysis model. The first voice content, the first emotion classification and the second voice content corresponding to the recorded data are used as a sample to obtain a preset number of samples;S3、将预设数量的样本的第一百分比作为训练集,第二百分比作为验证集,第一百分比和第二百分比之和小于百分之百;S3. The first percentage of the preset number of samples is used as the training set, the second percentage is used as the verification set, and the sum of the first percentage and the second percentage is less than 100%;S4、利用所述训练集中的样本对预设的应对话术推荐模型进行训练,并在训练结束后,利用验证集中的样本对所述应对话术推荐模型进行验证;S4. Use the samples in the training set to train the preset recommended conversational model, and after the training, use the samples in the verification set to verify the recommended conversational model;S5、若所述预测准确率大于预设阈值,则模型训练结束;S5. If the prediction accuracy rate is greater than a preset threshold, the model training ends;S6、若所述预设准确率小于或等于所述预设阈值,则增大所述预设数量的值,并重复执行步骤S1至S4。S6. If the preset accuracy rate is less than or equal to the preset threshold, increase the preset number of values, and repeat steps S1 to S4.
- 一种应对话术推荐方法,其特征在于,包括以下步骤:A method for recommending dialogue technique is characterized by the following steps:A2、在接收到客户进线后,实时获取所述进线客户与坐席的语音流;A2. After receiving the incoming customer line, obtain the voice streams of the incoming customer and agent in real time;B2、对所述语音流进行特征提取,提取出所述语音流中的第一语音片段的第一音频特征向量,其中,所述第一语音片段对应所述语音流中的客户输入的语音段;B2. Perform feature extraction on the voice stream to extract a first audio feature vector of a first voice segment in the voice stream, where the first voice segment corresponds to a voice segment input by a customer in the voice stream ;C2、将所述第一音频特征向量输入预设的语音情绪分析模型中进行分析,识别出所述第一音频特征向量对应的第一语音内容和第一情绪分类;C2. The first audio feature vector is input into a preset voice emotion analysis model for analysis, and the first voice content and the first emotion classification corresponding to the first audio feature vector are identified;D2、将识别出的所述第一语音内容和所述第一情绪分类输入预先训练好的应对话术推荐模型中进行分析,以得到推荐的应对话术,将所述推荐的应对话术实时发送至所述坐席的终端。D2. The recognized first speech content and the first emotion classification are input into a pre-trained conversational dialogue recommendation model for analysis to obtain a recommended conversational dialogue, and the recommended conversational dialogue is real-time Sent to the terminal of the agent.
- 如权利要求8所述的应对话术推荐方法,其特征在于,在所述步骤A2之后,所述应对话术推荐方法还包括:The method for recommending conversational skills according to claim 8, wherein after the step A2, the method for recommending conversational skills further comprises:E2、对所述语音流进行特征提取,提取出所述语音流中的第二语音片段的第二音频特征向量,其中,所述第二语音片段对应所述语音流中的坐席输入的语音段;E2. Perform feature extraction on the voice stream to extract a second audio feature vector of a second voice segment in the voice stream, where the second voice segment corresponds to a voice segment input by an agent in the voice stream ;F2、将所述第二音频特征向量输入预设的语音情绪分析模型中进行分析,识别出所述第二音频特征向量对应的第二情绪分类;F2. Input the second audio feature vector into a preset voice emotion analysis model for analysis, and identify a second emotion classification corresponding to the second audio feature vector;G2、若所述第二情绪分类为预设的异常情绪分类,则向所述坐席的终端发送预设的第一提醒信息。G2. If the second emotion classification is a preset abnormal emotion classification, send preset first reminder information to the terminal of the agent.
- 如权利要求9所述的应对话术推荐方法,其特征在于,所述预设的语音情绪分析模型还识别出所述第二音频特征向量对应的第二语音内容,在所述步骤F2之后,所述应对话术推荐方法还包括:The method according to claim 9, wherein the preset speech emotion analysis model also recognizes the second speech content corresponding to the second audio feature vector, after step F2, The method for recommending conversational skills also includes:在所述第二情绪分类为预设的异常情绪分类时,分析所述第二语音内容中是否包含预设的敏感词;When the second emotion classification is a preset abnormal emotion classification, analyze whether the second speech content contains a preset sensitive word;若包含预设的敏感词,且所述第二语音内容中出现预设的敏感词的次数小于等于第一阈值,则向所述坐席的终端发送预设的第二提醒信息;If the preset sensitive words are included, and the number of times the preset sensitive words appear in the second voice content is less than or equal to the first threshold, then send preset second reminder information to the terminal of the agent;若包含预设的敏感词,且所述第二语音内容中出现预设的敏感词的次数大于第一阈值,则向所述坐席的上级管理终端发送预设的第三提醒信息。If the preset sensitive words are included, and the number of times the preset sensitive words appear in the second voice content is greater than the first threshold, the preset third reminder information is sent to the superior management terminal of the agent.
- 如权利要求8所述的应对话术推荐方法,其特征在于,所述预设的语音情绪分析模型包括DNN-HMM声学模型、n-gram语言模型和wfst加权图的情绪分析模型。The method according to claim 8, wherein the preset speech sentiment analysis model includes a DNN-HMM acoustic model, an n-gram language model and a wfst weighted graph sentiment analysis model.
- 如权利要求8所述的应对话术推荐方法,其特征在于,所述应对话术推荐模型的训练过程包括:The method for recommending conversational skills according to claim 8, wherein the training process of the recommendation model for conversational skills includes:S1、从通话录音数据库中获取预设数量的服务标签为满意的录音数据,提取每一则录音数据中的第一语音片段的第一音频特征向量和第二语音片段的第二音频特征向量,其中,所述第一语音片段对应所述录音数据中的客户输入的语音段,所述第二语音片段对应所述录音数据中的坐席输入的语音段;S1. Obtain a preset number of service tags from the call recording database as satisfactory recording data, and extract the first audio feature vector of the first speech segment and the second audio feature vector of the second speech segment in each recording data, Wherein, the first voice segment corresponds to a voice segment input by a customer in the recording data, and the second voice segment corresponds to a voice segment input by an agent in the recording data;S2、采用预设的语音情绪分析模型中分别识别出所述第一音频特征向量对应的第一语音内容和第一情绪分类,以及所述第二音频特征向量对应的第二语音内容,将每一则录音数据对应的第一语音内容、第一情绪分类和第二语音内容作为一个样本,得到预设数量的样本;S2. Identify the first speech content and the first sentiment category corresponding to the first audio feature vector and the second speech content corresponding to the second audio feature vector using a preset speech emotion analysis model. The first voice content, the first emotion classification and the second voice content corresponding to the recorded data are used as a sample to obtain a preset number of samples;S3、将预设数量的样本的第一百分比作为训练集,第二百分比作为验证集,第一百分比和第二百分比之和小于百分之百;S3. The first percentage of the preset number of samples is used as the training set, the second percentage is used as the verification set, and the sum of the first percentage and the second percentage is less than 100%;S4、利用所述训练集中的样本对预设的应对话术推荐模型进行训练,并在训练结束后,利用验证集中的样本对所述应对话术推荐模型进行验证;S4. Use the samples in the training set to train the preset recommended conversational model, and after the training, use the samples in the verification set to verify the recommended conversational model;S5、若所述预测准确率大于预设阈值,则模型训练结束;S5. If the prediction accuracy rate is greater than a preset threshold, the model training ends;S6、若所述预设准确率小于或等于所述预设阈值,则增大所述预设数量的值,并重复执行步骤S1至S4。S6. If the preset accuracy rate is less than or equal to the preset threshold, increase the preset number of values, and repeat steps S1 to S4.
- 如权利要求9所述的应对话术推荐方法,其特征在于,所述 应对话术推荐模型的训练过程包括:The method for recommending conversational skills according to claim 9, wherein the training process of the recommendation model for conversational skills includes:S1、从通话录音数据库中获取预设数量的服务标签为满意的录音数据,提取每一则录音数据中的第一语音片段的第一音频特征向量和第二语音片段的第二音频特征向量,其中,所述第一语音片段对应所述录音数据中的客户输入的语音段,所述第二语音片段对应所述录音数据中的坐席输入的语音段;S1. Obtain a preset number of service tags from the call recording database as satisfactory recording data, and extract the first audio feature vector of the first speech segment and the second audio feature vector of the second speech segment in each recording data, Wherein, the first voice segment corresponds to a voice segment input by a customer in the recording data, and the second voice segment corresponds to a voice segment input by an agent in the recording data;S2、采用预设的语音情绪分析模型中分别识别出所述第一音频特征向量对应的第一语音内容和第一情绪分类,以及所述第二音频特征向量对应的第二语音内容,将每一则录音数据对应的第一语音内容、第一情绪分类和第二语音内容作为一个样本,得到预设数量的样本;S2. Identify the first speech content and the first sentiment category corresponding to the first audio feature vector and the second speech content corresponding to the second audio feature vector using a preset speech emotion analysis model. The first voice content, the first emotion classification and the second voice content corresponding to the recorded data are used as a sample to obtain a preset number of samples;S3、将预设数量的样本的第一百分比作为训练集,第二百分比作为验证集,第一百分比和第二百分比之和小于百分之百;S3. The first percentage of the preset number of samples is used as the training set, the second percentage is used as the verification set, and the sum of the first percentage and the second percentage is less than 100%;S4、利用所述训练集中的样本对预设的应对话术推荐模型进行训练,并在训练结束后,利用验证集中的样本对所述应对话术推荐模型进行验证;S4. Use the samples in the training set to train the preset recommended conversational model, and after the training, use the samples in the verification set to verify the recommended conversational model;S5、若所述预测准确率大于预设阈值,则模型训练结束;S5. If the prediction accuracy rate is greater than a preset threshold, the model training ends;S6、若所述预设准确率小于或等于所述预设阈值,则增大所述预设数量的值,并重复执行步骤S1至S4。S6. If the preset accuracy rate is less than or equal to the preset threshold, increase the preset number of values, and repeat steps S1 to S4.
- 如权利要求10所述的应对话术推荐方法,其特征在于,所述应对话术推荐模型的训练过程包括:The method for recommending conversational skills according to claim 10, wherein the training process of the recommendation model for conversational skills includes:S1、从通话录音数据库中获取预设数量的服务标签为满意的录音数据,提取每一则录音数据中的第一语音片段的第一音频特征向量和第二语音片段的第二音频特征向量,其中,所述第一语音片段对应所述录音数据中的客户输入的语音段,所述第二语音片段对应所述录音数据中的坐席输入的语音段;S1. Obtain a preset number of service tags from the call recording database as satisfactory recording data, and extract the first audio feature vector of the first speech segment and the second audio feature vector of the second speech segment in each recording data, Wherein, the first voice segment corresponds to a voice segment input by a customer in the recording data, and the second voice segment corresponds to a voice segment input by an agent in the recording data;S2、采用预设的语音情绪分析模型中分别识别出所述第一音频特征向量对应的第一语音内容和第一情绪分类,以及所述第二音频特征向量对应的第二语音内容,将每一则录音数据对应的第一语音内容、第一情绪分类和第二语音内容作为一个样本,得到预设数量的样本;S2. Identify the first speech content and the first sentiment category corresponding to the first audio feature vector and the second speech content corresponding to the second audio feature vector using a preset speech emotion analysis model. The first voice content, the first emotion classification and the second voice content corresponding to the recorded data are used as a sample to obtain a preset number of samples;S3、将预设数量的样本的第一百分比作为训练集,第二百分比作为验证集,第一百分比和第二百分比之和小于百分之百;S3. The first percentage of the preset number of samples is used as the training set, the second percentage is used as the verification set, and the sum of the first percentage and the second percentage is less than 100%;S4、利用所述训练集中的样本对预设的应对话术推荐模型进行训练,并在训练结束后,利用验证集中的样本对所述应对话术推荐模型进行验证;S4. Use the samples in the training set to train the preset recommended conversational model, and after the training, use the samples in the verification set to verify the recommended conversational model;S5、若所述预测准确率大于预设阈值,则模型训练结束;S5. If the prediction accuracy rate is greater than a preset threshold, the model training ends;S6、若所述预设准确率小于或等于所述预设阈值,则增大所述预设数量的值,并重复执行步骤S1至S4。S6. If the preset accuracy rate is less than or equal to the preset threshold, increase the preset number of values, and repeat steps S1 to S4.
- 一种计算机可读存储介质,其特征在于,所述计算机可读存 储介质存储有应对话术推荐程序,所述应对话术推荐程序可被至少一个处理器执行,以使所述至少一个处理器执行如下步骤:A computer-readable storage medium, characterized in that the computer-readable storage medium stores an application-recommendation program, and the application-recommendation program can be executed by at least one processor so that the at least one processor Perform the following steps:在接收到客户进线后,实时获取所述进线客户与坐席的语音流;After receiving the incoming customer line, obtain the voice streams of the incoming customer and agent in real time;对所述语音流进行特征提取,提取出所述语音流中的第一语音片段的第一音频特征向量,其中,所述第一语音片段对应所述语音流中的客户输入的语音段;Performing feature extraction on the voice stream to extract a first audio feature vector of a first voice segment in the voice stream, where the first voice segment corresponds to a voice segment input by a customer in the voice stream;将所述第一音频特征向量输入预设的语音情绪分析模型中进行分析,识别出所述第一音频特征向量对应的第一语音内容和第一情绪分类;Input the first audio feature vector into a preset voice emotion analysis model for analysis, and identify the first voice content and the first emotion classification corresponding to the first audio feature vector;将识别出的所述第一语音内容和所述第一情绪分类输入预先训练好的应对话术推荐模型中进行分析,以得到推荐的应对话术,将所述推荐的应对话术实时发送至所述坐席的终端。Input the recognized first speech content and the first sentiment classification into a pre-trained conversational recommendation model for analysis to obtain a recommended conversational conversation, and send the recommended conversational conversation in real time to The terminal of the agent.
- 如权利要求15所述的计算机可读存储介质,其特征在于,在实时获取所述进线客户与坐席的语音流之后,所述应对话术推荐程序被所述处理器执行时,还实现如下步骤:The computer-readable storage medium according to claim 15, wherein after acquiring the voice streams of the incoming client and the agent in real time, when the conversational recommendation program is executed by the processor, it is also implemented as follows step:对所述语音流进行特征提取,提取出所述语音流中的第二语音片段的第二音频特征向量,其中,所述第二语音片段对应所述语音流中的坐席输入的语音段;Performing feature extraction on the voice stream to extract a second audio feature vector of a second voice segment in the voice stream, where the second voice segment corresponds to a voice segment input by an agent in the voice stream;将所述第二音频特征向量输入预设的语音情绪分析模型中进行分析,识别出所述第二音频特征向量对应的第二情绪分类;Input the second audio feature vector into a preset voice emotion analysis model for analysis, and identify a second emotion classification corresponding to the second audio feature vector;若所述第二情绪分类为预设的异常情绪分类,则向所述坐席的终端发送预设的第一提醒信息。If the second emotion classification is a preset abnormal emotion classification, preset preset reminder information is sent to the terminal of the agent.
- 如权利要求16所述的计算机可读存储介质,其特征在于,所述预设的语音情绪分析模型还识别出所述第二音频特征向量对应的第二语音内容,在所述将所述第二音频特征向量输入预设的语音情绪分析模型中进行分析,识别出所述第二音频特征向量对应的第二情绪分类的步骤之后,所述应对话术推荐程序被所述处理器执行时,还实现如下步骤:The computer-readable storage medium of claim 16, wherein the preset voice sentiment analysis model also recognizes the second voice content corresponding to the second audio feature vector, and the Two audio feature vectors are input into a preset voice sentiment analysis model for analysis, and after the step of recognizing the second sentiment classification corresponding to the second audio feature vector is performed, when the recommended conversational program is executed by the processor The following steps are also achieved:在所述第二情绪分类为预设的异常情绪分类时,分析所述第二语音内容中是否包含预设的敏感词;When the second emotion classification is a preset abnormal emotion classification, analyze whether the second speech content contains a preset sensitive word;若包含预设的敏感词,且所述第二语音内容中出现预设的敏感词的次数小于等于第一阈值,则向所述坐席的终端发送预设的第二提醒信息;If the preset sensitive words are included, and the number of times the preset sensitive words appear in the second voice content is less than or equal to the first threshold, then send preset second reminder information to the terminal of the agent;若包含预设的敏感词,且所述第二语音内容中出现预设的敏感词的次数大于第一阈值,则向所述坐席的上级管理终端发送预设的第三提醒信息。If the preset sensitive words are included, and the number of times the preset sensitive words appear in the second voice content is greater than the first threshold, the preset third reminder information is sent to the superior management terminal of the agent.
- 如权利要求15所述的计算机可读存储介质,其特征在于,所述预设的语音情绪分析模型包括DNN-HMM声学模型、n-gram语 言模型和wfst加权图的情绪分析模型。The computer-readable storage medium of claim 15, wherein the preset speech sentiment analysis model includes a DNN-HMM acoustic model, an n-gram language model, and a wfst weighted graph sentiment analysis model.
- 如权利要求15所述的计算机可读存储介质,其特征在于,所述应对话术推荐模型的训练过程包括:The computer-readable storage medium according to claim 15, wherein the training process of the recommended model should include:S1、从通话录音数据库中获取预设数量的服务标签为满意的录音数据,提取每一则录音数据中的第一语音片段的第一音频特征向量和第二语音片段的第二音频特征向量,其中,所述第一语音片段对应所述录音数据中的客户输入的语音段,所述第二语音片段对应所述录音数据中的坐席输入的语音段;S1. Obtain a preset number of service tags from the call recording database as satisfactory recording data, and extract the first audio feature vector of the first speech segment and the second audio feature vector of the second speech segment in each recording data, Wherein, the first voice segment corresponds to a voice segment input by a customer in the recording data, and the second voice segment corresponds to a voice segment input by an agent in the recording data;S2、采用预设的语音情绪分析模型中分别识别出所述第一音频特征向量对应的第一语音内容和第一情绪分类,以及所述第二音频特征向量对应的第二语音内容,将每一则录音数据对应的第一语音内容、第一情绪分类和第二语音内容作为一个样本,得到预设数量的样本;S2. Identify the first speech content and the first sentiment category corresponding to the first audio feature vector and the second speech content corresponding to the second audio feature vector using a preset speech emotion analysis model. The first voice content, the first emotion classification and the second voice content corresponding to the recorded data are used as a sample to obtain a preset number of samples;S3、将预设数量的样本的第一百分比作为训练集,第二百分比作为验证集,第一百分比和第二百分比之和小于百分之百;S3. The first percentage of the preset number of samples is used as the training set, the second percentage is used as the verification set, and the sum of the first percentage and the second percentage is less than 100%;S4、利用所述训练集中的样本对预设的应对话术推荐模型进行训练,并在训练结束后,利用验证集中的样本对所述应对话术推荐模型进行验证;S4. Use the samples in the training set to train the preset recommended conversational model, and after the training, use the samples in the verification set to verify the recommended conversational model;S5、若所述预测准确率大于预设阈值,则模型训练结束;S5. If the prediction accuracy rate is greater than a preset threshold, the model training ends;S6、若所述预设准确率小于或等于所述预设阈值,则增大所述预设数量的值,并重复执行步骤S1至S4。S6. If the preset accuracy rate is less than or equal to the preset threshold, increase the preset number of values, and repeat steps S1 to S4.
- 如权利要求16所述的计算机可读存储介质,其特征在于,所述应对话术推荐模型的训练过程包括:The computer-readable storage medium according to claim 16, wherein the training process of the recommended model should include:S1、从通话录音数据库中获取预设数量的服务标签为满意的录音数据,提取每一则录音数据中的第一语音片段的第一音频特征向量和第二语音片段的第二音频特征向量,其中,所述第一语音片段对应所述录音数据中的客户输入的语音段,所述第二语音片段对应所述录音数据中的坐席输入的语音段;S1. Obtain a preset number of service tags from the call recording database as satisfactory recording data, and extract the first audio feature vector of the first speech segment and the second audio feature vector of the second speech segment in each recording data, Wherein, the first voice segment corresponds to a voice segment input by a customer in the recording data, and the second voice segment corresponds to a voice segment input by an agent in the recording data;S2、采用预设的语音情绪分析模型中分别识别出所述第一音频特征向量对应的第一语音内容和第一情绪分类,以及所述第二音频特征向量对应的第二语音内容,将每一则录音数据对应的第一语音内容、第一情绪分类和第二语音内容作为一个样本,得到预设数量的样本;S2. Identify the first speech content and the first sentiment category corresponding to the first audio feature vector and the second speech content corresponding to the second audio feature vector using a preset speech emotion analysis model. The first voice content, the first emotion classification and the second voice content corresponding to the recorded data are used as a sample to obtain a preset number of samples;S3、将预设数量的样本的第一百分比作为训练集,第二百分比作为验证集,第一百分比和第二百分比之和小于百分之百;S3. The first percentage of the preset number of samples is used as the training set, the second percentage is used as the verification set, and the sum of the first percentage and the second percentage is less than 100%;S4、利用所述训练集中的样本对预设的应对话术推荐模型进行训练,并在训练结束后,利用验证集中的样本对所述应对话术推荐模型进行验证;S4. Use the samples in the training set to train the preset recommended conversational model, and after the training, use the samples in the verification set to verify the recommended conversational model;S5、若所述预测准确率大于预设阈值,则模型训练结束;S5. If the prediction accuracy rate is greater than a preset threshold, the model training ends;S6、若所述预设准确率小于或等于所述预设阈值,则增大所述预 设数量的值,并重复执行步骤S1至S4。S6. If the preset accuracy rate is less than or equal to the preset threshold, increase the value of the preset number and repeat steps S1 to S4.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811340705.2A CN109587360B (en) | 2018-11-12 | 2018-11-12 | Electronic device, method for coping with tactical recommendation, and computer-readable storage medium |
CN201811340705.2 | 2018-11-12 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020098249A1 true WO2020098249A1 (en) | 2020-05-22 |
Family
ID=65922254
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/088635 WO2020098249A1 (en) | 2018-11-12 | 2019-05-27 | Electronic device, response conversation technique recommendation method and computer readable storage medium |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN109587360B (en) |
WO (1) | WO2020098249A1 (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112507094A (en) * | 2020-12-11 | 2021-03-16 | 润联软件系统(深圳)有限公司 | Customer service robot dialogue method based on reinforcement learning and related components thereof |
CN113312468A (en) * | 2021-07-30 | 2021-08-27 | 平安科技(深圳)有限公司 | Conversation mode-based conversation recommendation method, device, equipment and medium |
CN113688221A (en) * | 2021-09-08 | 2021-11-23 | 中国平安人寿保险股份有限公司 | Model-based dialect recommendation method and device, computer equipment and storage medium |
Families Citing this family (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109587360B (en) * | 2018-11-12 | 2021-07-13 | 平安科技(深圳)有限公司 | Electronic device, method for coping with tactical recommendation, and computer-readable storage medium |
CN110444229A (en) * | 2019-06-17 | 2019-11-12 | 深圳壹账通智能科技有限公司 | Communication service method, device, computer equipment and storage medium based on speech recognition |
CN110472007A (en) * | 2019-07-04 | 2019-11-19 | 深圳追一科技有限公司 | Information-pushing method, device, equipment and storage medium |
CN110535749B (en) * | 2019-07-09 | 2023-04-25 | 中国平安财产保险股份有限公司 | Dialogue pushing method and device, electronic equipment and storage medium |
CN110364183A (en) * | 2019-07-09 | 2019-10-22 | 深圳壹账通智能科技有限公司 | Method, apparatus, computer equipment and the storage medium of voice quality inspection |
CN110491372A (en) * | 2019-07-22 | 2019-11-22 | 平安科技(深圳)有限公司 | A kind of feedback information generating method, device, storage medium and smart machine |
CN110459210A (en) * | 2019-07-30 | 2019-11-15 | 平安科技(深圳)有限公司 | Answering method, device, equipment and storage medium based on speech analysis |
CN110473549A (en) * | 2019-08-21 | 2019-11-19 | 北京智合大方科技有限公司 | A kind of voice dialogue analysis system, method and storage medium |
CN110717733A (en) * | 2019-10-09 | 2020-01-21 | 浙江百应科技有限公司 | Enterprise work mobile phone intelligent management technology solution |
CN110689878B (en) * | 2019-10-11 | 2020-07-28 | 浙江百应科技有限公司 | Intelligent voice conversation intention recognition method based on X L Net |
CN110990543A (en) * | 2019-10-18 | 2020-04-10 | 平安科技(深圳)有限公司 | Intelligent conversation generation method and device, computer equipment and computer storage medium |
CN110556110A (en) * | 2019-10-24 | 2019-12-10 | 北京九狐时代智能科技有限公司 | Voice processing method and device, intelligent terminal and storage medium |
CN111008586A (en) * | 2019-11-29 | 2020-04-14 | 上海能塔智能科技有限公司 | Data processing method, device, equipment and storage medium for passenger car conflict detection |
CN111128189A (en) * | 2019-12-30 | 2020-05-08 | 秒针信息技术有限公司 | Warning information prompting method and device |
CN110933239A (en) * | 2019-12-30 | 2020-03-27 | 秒针信息技术有限公司 | Method and apparatus for detecting dialect |
WO2021139737A1 (en) * | 2020-01-08 | 2021-07-15 | 北京嘀嘀无限科技发展有限公司 | Method and system for man-machine interaction |
CN111429157A (en) * | 2020-03-27 | 2020-07-17 | 上海东普信息科技有限公司 | Method, device and equipment for evaluating and processing complaint work order and storage medium |
CN112488750A (en) * | 2020-11-27 | 2021-03-12 | 上海容大数字技术有限公司 | Intelligent recommendation and renewal system for insurance scene |
CN112992187B (en) * | 2021-02-26 | 2023-04-18 | 平安科技(深圳)有限公司 | Context-based voice emotion detection method, device, equipment and storage medium |
CN112687293B (en) * | 2021-03-22 | 2021-06-22 | 北京孵家科技股份有限公司 | Intelligent agent training method and system based on machine learning and data mining |
CN116886819B (en) * | 2023-08-07 | 2024-02-02 | 云南电网有限责任公司 | Multi-dimensional telephone traffic data monitoring method, device and storage medium |
CN117711444B (en) * | 2024-02-05 | 2024-04-23 | 新励成教育科技股份有限公司 | Interaction method, device, equipment and storage medium based on talent expression |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104538043A (en) * | 2015-01-16 | 2015-04-22 | 北京邮电大学 | Real-time emotion reminder for call |
US9105042B2 (en) * | 2013-02-07 | 2015-08-11 | Verizon Patent And Licensing Inc. | Customer sentiment analysis using recorded conversation |
CN107705807A (en) * | 2017-08-24 | 2018-02-16 | 平安科技(深圳)有限公司 | Voice quality detecting method, device, equipment and storage medium based on Emotion identification |
CN108564968A (en) * | 2018-04-26 | 2018-09-21 | 广州势必可赢网络科技有限公司 | Method and device for evaluating customer service |
CN109587360A (en) * | 2018-11-12 | 2019-04-05 | 平安科技(深圳)有限公司 | Electronic device should talk with art recommended method and computer readable storage medium |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103701999B (en) * | 2012-09-27 | 2017-03-01 | 中国电信股份有限公司 | The monitoring method of call center's voice call and system |
CN104113643B (en) * | 2014-06-27 | 2017-01-04 | 国家电网公司 | A kind of Customer Service Center's Field Monitoring System and method |
CN104616666B (en) * | 2015-03-03 | 2018-05-25 | 广东小天才科技有限公司 | Method and device for improving conversation communication effect based on voice analysis |
US10025775B2 (en) * | 2015-09-04 | 2018-07-17 | Conduent Business Services, Llc | Emotion, mood and personality inference in real-time environments |
CN105261362B (en) * | 2015-09-07 | 2019-07-05 | 科大讯飞股份有限公司 | A kind of call voice monitoring method and system |
CN106548788B (en) * | 2015-09-23 | 2020-01-07 | 中国移动通信集团山东有限公司 | Intelligent emotion determining method and system |
CN106024015A (en) * | 2016-06-14 | 2016-10-12 | 上海航动科技有限公司 | Call center agent monitoring method and system |
CN108122552B (en) * | 2017-12-15 | 2021-10-15 | 上海智臻智能网络科技股份有限公司 | Voice emotion recognition method and device |
CN108259686A (en) * | 2017-12-28 | 2018-07-06 | 合肥凯捷技术有限公司 | A kind of customer service system based on speech analysis |
CN108197115B (en) * | 2018-01-26 | 2022-04-22 | 上海智臻智能网络科技股份有限公司 | Intelligent interaction method and device, computer equipment and computer readable storage medium |
-
2018
- 2018-11-12 CN CN201811340705.2A patent/CN109587360B/en active Active
-
2019
- 2019-05-27 WO PCT/CN2019/088635 patent/WO2020098249A1/en active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9105042B2 (en) * | 2013-02-07 | 2015-08-11 | Verizon Patent And Licensing Inc. | Customer sentiment analysis using recorded conversation |
CN104538043A (en) * | 2015-01-16 | 2015-04-22 | 北京邮电大学 | Real-time emotion reminder for call |
CN107705807A (en) * | 2017-08-24 | 2018-02-16 | 平安科技(深圳)有限公司 | Voice quality detecting method, device, equipment and storage medium based on Emotion identification |
CN108564968A (en) * | 2018-04-26 | 2018-09-21 | 广州势必可赢网络科技有限公司 | Method and device for evaluating customer service |
CN109587360A (en) * | 2018-11-12 | 2019-04-05 | 平安科技(深圳)有限公司 | Electronic device should talk with art recommended method and computer readable storage medium |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112507094A (en) * | 2020-12-11 | 2021-03-16 | 润联软件系统(深圳)有限公司 | Customer service robot dialogue method based on reinforcement learning and related components thereof |
CN113312468A (en) * | 2021-07-30 | 2021-08-27 | 平安科技(深圳)有限公司 | Conversation mode-based conversation recommendation method, device, equipment and medium |
CN113312468B (en) * | 2021-07-30 | 2021-11-05 | 平安科技(深圳)有限公司 | Conversation mode-based conversation recommendation method, device, equipment and medium |
CN113688221A (en) * | 2021-09-08 | 2021-11-23 | 中国平安人寿保险股份有限公司 | Model-based dialect recommendation method and device, computer equipment and storage medium |
CN113688221B (en) * | 2021-09-08 | 2023-07-25 | 中国平安人寿保险股份有限公司 | Model-based conversation recommendation method, device, computer equipment and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN109587360A (en) | 2019-04-05 |
CN109587360B (en) | 2021-07-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2020098249A1 (en) | Electronic device, response conversation technique recommendation method and computer readable storage medium | |
US11315366B2 (en) | Conference recording method and data processing device employing the same | |
US11908477B2 (en) | Automatic extraction of conversation highlights | |
JP2020030408A (en) | Method, apparatus, device and medium for identifying key phrase in audio | |
US8285539B2 (en) | Extracting tokens in a natural language understanding application | |
CN112966082B (en) | Audio quality inspection method, device, equipment and storage medium | |
JP5025353B2 (en) | Dialog processing apparatus, dialog processing method, and computer program | |
US20170200167A1 (en) | Funnel Analysis | |
WO2021012649A1 (en) | Method and device for expanding question and answer sample | |
CN111177350A (en) | Method, device and system for forming dialect of intelligent voice robot | |
CN110738998A (en) | Voice-based personal credit evaluation method, device, terminal and storage medium | |
CN111540353A (en) | Semantic understanding method, device, equipment and storage medium | |
CN112632245A (en) | Intelligent customer service distribution method and device, computer equipment and storage medium | |
CN113486170B (en) | Natural language processing method, device, equipment and medium based on man-machine interaction | |
CN113239204A (en) | Text classification method and device, electronic equipment and computer-readable storage medium | |
CN113240510A (en) | Abnormal user prediction method, device, equipment and storage medium | |
CN113314150A (en) | Emotion recognition method and device based on voice data and storage medium | |
CN113064983A (en) | Semantic detection method and device, computer equipment and storage medium | |
US11064075B2 (en) | System for processing voice responses using a natural language processing engine | |
CN112908315A (en) | Question-answer intention judgment method based on voice characteristics and voice recognition | |
CN116541494A (en) | Model training method, device, equipment and medium for replying information | |
CN115878768A (en) | NLP-based vehicle insurance service call-back clue recommendation method and related equipment thereof | |
CN114925159A (en) | User emotion analysis model training method and device, electronic equipment and storage medium | |
CN114265917A (en) | Work order category identification method, model training method, device and equipment | |
WO2023272833A1 (en) | Data detection method, apparatus and device and readable storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19885521 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19885521 Country of ref document: EP Kind code of ref document: A1 |