CN105810205A - Speech processing method and device - Google Patents

Speech processing method and device Download PDF

Info

Publication number
CN105810205A
CN105810205A CN201410838240.9A CN201410838240A CN105810205A CN 105810205 A CN105810205 A CN 105810205A CN 201410838240 A CN201410838240 A CN 201410838240A CN 105810205 A CN105810205 A CN 105810205A
Authority
CN
China
Prior art keywords
hot line
speech data
speech
line speech
user
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201410838240.9A
Other languages
Chinese (zh)
Inventor
王朝民
冯俊兰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China Mobile Communications Group Co Ltd
Original Assignee
China Mobile Communications Group Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China Mobile Communications Group Co Ltd filed Critical China Mobile Communications Group Co Ltd
Priority to CN201410838240.9A priority Critical patent/CN105810205A/en
Publication of CN105810205A publication Critical patent/CN105810205A/en
Pending legal-status Critical Current

Links

Landscapes

  • Telephonic Communication Services (AREA)

Abstract

The invention provides a speech processing method and device. The speech processing method comprises the steps of processing hot speech data to acquire text information of the hot speech data and user speech information involved in the hot speech data; and acquiring an evaluation result of the hot speech data according to the text information and the user speech information. According to the scheme, the user satisfaction of customer service is evaluated in combination with the text information and the speech information in the speech data, so that the user satisfaction of customer service records can be evaluated more comprehensively and more meticulously, the object scale of the satisfaction evaluation is enlarged, meanwhile, the manpower cost is reduced, and a technical support is better provided for customer service hotline operation.

Description

A kind of method of speech processing and device
Technical field
The present invention relates to speech recognition and voice processing technology field, particularly to a kind of method of speech processing and device.
Background technology
Along with network, communication, computer technology development, enterprise presents the feature of electronization, long-range, virtualization, networking, and more Xian Shang enterprise emerges in multitude.And communication between client and enterprise and dialogue, the exchange and conmmunication of the remotely means such as Network Based, phone are also developed into by aspectant consulting, negotiation.In this context, become, based on the client service center (call center) of phone, the important channel that enterprise is mutual with user.Client service center every day, all facing to substantial amounts of telephone voice service, processes the diversified demand for services of client, including pre-sales consulting, purchase, after sale, complaint etc..In the process of telephone service, customer service needs the service object of the different emotion of reply, and makes suitable reaction.It can be said that the image that client service center is enterprise is represented, the service quality of client service center directly affects user's loyalty to enterprise, therefore by improving the service quality of customer service and improving customer satisfaction, loyalty has become the important public relations direction of enterprise.Additionally, by accuracy customer demand solution, promote customer service efficiency;And by reasonably managing customer service team, the service quality etc. of accurate evaluation employee, all become the direction that client service center needs constantly to inquire into, grope and study.
In prior art, adopting by speech recognition engine is content of text by hot line recording data " translation ", then pass through natural language understanding technology and obtain the satisfaction correlation analysis of text, thus judging the user satisfaction situation of this recording of taking on the telephone, but, existing customer service hot line recording customer satisfaction evaluation method is owing to using speech recognition engine to be identified rear text, process this text message thus obtaining user satisfaction information, actually lost the voice satisfaction information in recording originally, the unsatisfied situation of a lot of users does not have complete reaction out, have impact on satisfaction evaluation accuracy.
Known technology:
1) speech recognition technology: speech recognition technology, it is also referred to as automatic speech recognition (AutomaticSpeechRecognition, ASR), its target is to be computer-readable input by the vocabulary Content Transformation in the voice of the mankind, such as button, binary coding or character string, different from Speaker Identification and speaker verification, the latter attempts identifying or confirm to send the speaker of voice but not the vocabulary content that wherein comprises.
2) speech parameter fractional analysis: speech parameter is analyzed after including direct Waveform Analysis Method and speech parameterization extraction by the method for speech analysis, and current main flow is Parametric Analysis.Popular speech parameter has the parametrization that speech channel is composed, such as linear predictor coefficient (LPC (LinearPredictiveCoding, linear predictive coding) coefficient), MFCC cepstrum (MelFrequencyCepstrumCoefficient and MFCC) and line spectrum pair (LSP) etc.;Voice sound source parameter, as fundamental frequency, aperiodic composition etc.;Also has the linguistics content that higher level analysis obtains, such as the analysis of sound mother stock, tone, intonation;Also have deeper emotion information and speaker's individual information etc..By the analysis of these speech parameters, we can comprehensively understand the information that voice carries more, and more directly effectively processes voice.
3) speech emotion recognition: contain abundant information in the natural-sounding of people, including semantic information, speaker information and emotion information etc..Speech emotion recognition is through extracting the emotion correlated characteristic in voice signal, identifies the emotion information of speaker by setting up category of model, and the model being generally adopted and method have gauss hybrid models, principal component analysis, support vector machine etc..
Summary of the invention
The technical problem to be solved in the present invention is to provide a kind of method of speech processing and device, only user satisfaction information is obtained according to the text that hot line recording data is corresponding in order to solve existing customer service hot line recording customer satisfaction evaluation method, actually lost the voice satisfaction information in recording originally, the unsatisfied situation of a lot of users does not have complete reaction out, the problem that have impact on satisfaction evaluation accuracy.
In order to solve above-mentioned technical problem, the embodiment of the present invention provides a kind of method of speech processing, including:
Hot line speech data is processed, obtains the user speech information related in the text message of described hot line speech data and hot line speech data;
According to described text message and user speech information, obtain the evaluation result of described hot line speech data.
Further, described hot line speech data being processed, the step obtaining the user speech information related in the text message of described hot line speech data and hot line speech data includes:
Described hot line speech data is carried out voice recognition processing, it is thus achieved that the text message of hot line speech data;
Described hot line speech data is carried out cutting process, it is thus achieved that the user speech information related in hot line speech data.
Further, described described hot line speech data is carried out voice recognition processing, it is thus achieved that the step of the text message of hot line speech data includes:
By speech recognition engine, described hot line speech data is carried out voice recognition processing, it is thus achieved that the text message of hot line speech data.
Further, described described hot line speech data is carried out cutting process, it is thus achieved that the step of the user speech information related in hot line speech data includes:
By voice partitioning algorithm, described hot line speech data is carried out phonetic segmentation process, it is thus achieved that the user speech information related in hot line speech data.
Further, described according to described text message and user speech information, the step of the evaluation result obtaining described hot line speech data includes:
The text feature relevant to user satisfaction and emotion is extracted from described text message;
Extracting from described user speech information and speech emotional, abstract characteristics that customer satisfaction evaluation is relevant, as voice assemblage characteristic, described abstract characteristics includes: temporal signatures, frequency domain character, short-time characteristic, long time feature and session features;
It is combined described text feature and voice assemblage characteristic processing, obtains the evaluation result of described hot line speech data.
Further, described described text feature and voice assemblage characteristic being combined processes, and the step obtaining described hot line speech data evaluation result includes:
Determine the Feature Words weight in text feature and voice assemblage characteristic respectively according to presetting, be combined processing, using the combined treatment result evaluation result as described hot line speech data.
Further, if described hot line speech data includes mark degrees of data, then hot line speech data is processed, after obtaining the user speech information related in the text message of described hot line speech data and hot line speech data, according to described mark degrees of data and described text message and user speech information, obtain the evaluation result of described hot line speech data.
The embodiment of the present invention provides a kind of voice processing apparatus, including:
First acquisition module, for hot line speech data is processed, obtains the user speech information related in the text message of described hot line speech data and hot line speech data;
Second acquisition module, for according to described text message and user speech information, obtaining the evaluation result of described hot line speech data.
Further, described first acquisition module, including:
Voice recognition unit, for carrying out voice recognition processing to described hot line speech data, it is thus achieved that the text message of hot line speech data;
Phonetic segmentation unit, for carrying out cutting process to described hot line speech data, it is thus achieved that the user speech information related in hot line speech data.
Further, described second acquisition module, including:
Text character extraction unit, for extracting the text feature relevant to user satisfaction and emotion from described text message;
Speech feature extraction unit, for extracting from described user speech information and speech emotional, abstract characteristics that customer satisfaction evaluation is relevant, as voice assemblage characteristic, described abstract characteristics includes: temporal signatures, frequency domain character, short-time characteristic, long time feature and session features;
Processing unit, for being combined described text feature and voice assemblage characteristic processing, obtains the evaluation result of described hot line speech data.
The invention has the beneficial effects as follows:
Such scheme, in conjunction with the text message in speech data and voice messaging, the user satisfaction of customer service is evaluated, the user satisfaction situation of evaluation customer service recording that can be more comprehensively more careful, expand the object scale of satisfaction evaluation, save human cost simultaneously, better provide technical support for customer service hot line operation.
Accompanying drawing explanation
Fig. 1 represents the overview flow chart of the described method of speech processing of the embodiment of the present invention;
Fig. 2 represents the detail flowchart of step 20 in Fig. 1;
Fig. 3 represents the module diagram of the described voice processing apparatus of the embodiment of the present invention;
Fig. 4 represents the composition structural representation of described second acquisition module of the embodiment of the present invention;
Fig. 5 degree of being satisfied with evaluation realize schematic flow sheet.
Detailed description of the invention
For making the object, technical solutions and advantages of the present invention clearly, describe the present invention below in conjunction with the accompanying drawings and the specific embodiments.
The present invention is directed to existing customer service hot line recording customer satisfaction evaluation method owing to using speech recognition engine to be identified rear text, process this text message thus obtaining user satisfaction information, actually lost the voice satisfaction information in recording originally, the unsatisfied situation of a lot of users does not have complete reaction out, the problem that have impact on satisfaction evaluation accuracy, it is provided that a kind of method of speech processing and device.
As it is shown in figure 1, the embodiment of the present invention provides a kind of method of speech processing, including:
Step 10, processes hot line speech data, obtains the user speech information related in the text message of described hot line speech data and hot line speech data;
Step 20, according to described text message and user speech information, obtains the evaluation result of described hot line speech data.
Such scheme, by the text message in extraction hot line voice and voice messaging, it is combined the two processing, jointly obtain the evaluation result of user satisfaction, this kind of method, it is possible to the more comprehensively more careful user satisfaction situation evaluating customer service recording, expands the object scale of satisfaction evaluation, save human cost simultaneously, better provide technical support for customer service hot line operation.
In another embodiment of the present invention, described step 10 includes:
Step 11, carries out voice recognition processing to described hot line speech data, it is thus achieved that the text message of hot line speech data;
Step 12, carries out cutting process to described hot line speech data, it is thus achieved that the user speech information related in hot line speech data.
It should be noted that, the corresponding text identification in hot line speech data is gone out by described step 11 mainly through speech recognition engine, forming the text after identifying, this text only comprises voice content information in hot line speech data;Described step 12 mainly through voice cutting techniques by the dialogue both sides' phonetic segmentation in hot line voice, obtain the voice segments collection of user, customer service represent voice segments collection and relevant dialog information, after completing step 10, just obtain the voice segments collection of user and voice content corresponding to described user, voice segments collection that customer service represents and described customer service represent corresponding voice content.
It should be noted that, the information obtained in described step 10, it it is the follow-up basis that described hot line speech data is evaluated, it follows that be the step being evaluated according to information acquired in described step 10, as shown in Figure 2, in further embodiment of this invention, described step 20 particularly as follows:
Step 21, extracts the text feature relevant to user satisfaction and emotion from described text message;
Step 22, extracts and speech emotional, abstract characteristics that customer satisfaction evaluation is relevant from described user speech information, and as voice assemblage characteristic, described abstract characteristics includes: temporal signatures, frequency domain character, short-time characteristic, long time feature and session features;
Step 23, is combined described text feature and voice assemblage characteristic processing, obtains the evaluation result of described hot line speech data.
Alternatively, described step 21 concrete implementation mode is: extract the text feature relevant to user satisfaction and emotion by the correlation technique of natural language processing, including emphasis vocabulary and assemblage characteristic etc., described emphasis vocabulary and assemblage characteristic are formulated according to practical situation;Described step 22 concrete implementation mode is: extract the abstract characteristics in voice signal by technology such as relevant treatment, Fourier transformation, homomorphic filterings, including temporal signatures and frequency domain character, include short-time characteristic (with speech frame for analytic unit) from feature span, long time feature (with phrase or sentence for analytic unit) and session features (with the context of dialogue for analytic unit), the present invention selects combine with speech emotional, feature that customer satisfaction evaluation is relevant to be analyzed.
After obtaining text feature and voice assemblage characteristic, it is analyzed described text feature and voice assemblage characteristic processing, just the evaluation result that hot line speech data is corresponding can be obtained, therefore, in further embodiment of this invention, described step 23 concrete implementation mode is: determine the Feature Words weight in text feature and voice assemblage characteristic respectively according to presetting, and is combined processing, using the combined treatment result evaluation result as described hot line speech data.
The realization of described step 23 is described as follows that (setting rule can be: unsatisfied emotion of particular words is more strong, and the weight of the text message of its correspondence is more big;Passing judgment on of the weight of voice messaging is regular identical with text message), in each hot line voice, only one of which Feature Words illustrates:
Example 1: when the default Feature Words that need to analyze is for " being unsatisfied with ", when including the word of " being unsatisfied with " in described hot line speech data being detected, because this word emotional color is stronger, the weight of the text message that this hot line speech data is corresponding is 10, and the information in the voice dialogue corresponding to word obtaining described " being unsatisfied with " of analyzing is " being not be unsatisfied with ", now, the weight of described default Feature Words correspondence voice messaging is 1, the numerical value that the weight of described text message is corresponding with the weight of voice messaging is overlapped or subtracts each other, result computing obtained is as the evaluation result of described hot line speech data, to be superposed to example in the present invention, therefore features described above word is the evaluation result of " being unsatisfied with " is 11.
Example 2: when the default Feature Words that need to analyze is for " call ", when including the word of " call " in described hot line speech data being detected, because this word belongs to neutral word, do not comprise emotional color, the weight of the text message that this hot line speech data is corresponding is 0, and the information in the voice dialogue corresponding to word obtaining described " call " of analyzing is " call response is very slow ", emotional color with user, now, the weight of described default Feature Words correspondence voice messaging is 3, the numerical value that the weight of described text message is corresponding with the weight of voice messaging is overlapped, the value obtained is 3, therefore the evaluation result that " call " Feature Words described in this hot line speech data is corresponding is 3.
In actual applications; evaluation to each hot line voice; multiple Feature Words would generally be selected to carry out overall merit; when obtaining evaluation result; can first obtain the evaluation result of each Feature Words respectively; then the evaluation result of each Feature Words is carried out corresponding computing again and (can first acquire the shared proportion of the emotion of each Feature Words; then undertaken being multiplied the computing of then superposition respectively by the proportion of the weight of these Feature Words and correspondence), finally obtain the comprehensive evaluation result of this hot line speech data.
It is above to hot line speech data does not mark degrees of data that (described mark degrees of data is after call completes, the actual evaluating data provided of user) the explanation that carries out of situation, in another embodiment of the present invention, if described hot line speech data includes mark degrees of data, then hot line speech data is processed, after obtaining the user speech information related in the text message of described hot line speech data and hot line speech data, according to described mark degrees of data and described text message and user speech information, obtain the evaluation result of described hot line speech data.
Should be noted that, when containing mark degrees of data, direct basis mark degrees of data can obtain evaluation result, first can also analyze described text message according to said method and user speech information obtains rough evaluation result, then according to the proportion that mark degrees of data accounts for is 80%, the data of rough evaluation result account for the proportion of 20%, carry out comprehensive computing, finally obtain the evaluation result of described hot line speech data.
Should be noted that, said method is used to obtain the customer service satisfaction evaluation result of hot line speech data, the user satisfaction situation of evaluation customer service recording that can be more comprehensively more careful, expand the object scale of satisfaction evaluation, save human cost simultaneously, better provide technical support for customer service hot line operation.
As it is shown on figure 3, the embodiment of the present invention also provides for a kind of voice processing apparatus, including:
First acquisition module 100, for hot line speech data is processed, obtains the user speech information related in the text message of described hot line speech data and hot line speech data;
Second acquisition module 200, for according to described text message and user speech information, obtaining the evaluation result of described hot line speech data.
Alternatively, described first acquisition module 100, including:
Voice recognition unit, for carrying out voice recognition processing to described hot line speech data, it is thus achieved that the text message of hot line speech data;
Phonetic segmentation unit, for carrying out cutting process to described hot line speech data, it is thus achieved that the user speech information related in hot line speech data.
Alternatively, as shown in Figure 4, described second acquisition module 200, including:
Text character extraction unit 201, for extracting the text feature relevant to user satisfaction and emotion from described text message;
Speech feature extraction unit 202, for extracting from described user speech information and speech emotional, abstract characteristics that customer satisfaction evaluation is relevant, as voice assemblage characteristic, described abstract characteristics includes: temporal signatures, frequency domain character, short-time characteristic, long time feature and session features;
Processing unit 203, for being combined described text feature and voice assemblage characteristic processing, obtains the evaluation result of described hot line speech data.
It should be noted that this device embodiment is the device corresponding with said method, all implementations of said method, all suitable in this device embodiment, also can reach technique effect same as mentioned above.
The actual process that realizes of the present invention is: is first the training stage of evaluation model, by repeatedly simulation training, acquires the good evaluation model of evaluation effect, is then according to the evaluation of existing hot line recording data to obtaining of this evaluation model.As shown in Figure 5, in the training stage: first the hot line recording data containing satisfaction labeled data is carried out speaker's cutting and acquire speech data, and speech recognition obtains text data, then speech data is carried out speech feature extraction, text data is carried out Text character extraction, last according to said method foundation phonetic feature and text feature, omnibearing parsing user satisfaction feature is evaluated the training of model;Evaluation phase: after the evaluation model obtaining optimum, determines, by this evaluation model identification or classification, the user satisfaction confidence level that wall scroll hot line is recorded, thus automatically providing user satisfaction marking, just can obtain satisfaction evaluation result with this.
In actual applications, training end adopts news to fly the identification engine of offer to identify that wall scroll 10086 customer service recording is identified rear text, then ties up (namely 13 kinds) satisfaction related text feature by text analyzing and Text character extraction to 13;Totally 45 dimension (namely 45 kinds) the satisfaction related voice feature of 10 classes in hot line recording is extracted by various voice tools of increasing income;Combine 13 Balakrishnan eigens and 45 dimension phonetic features and existing satisfaction labeled data trains a SVM (SupportVectorMachine, support vector machine) grader;
Identify that a new hot line customer service recording is extracted 13 dimension demand characteristics by speech recognition engine and Text Feature Extraction instrument by end, extract 45 dimension phonetic features of demand again through voice tool;Then features described above is obtained classification results by SVM classifier, namely satisfaction evaluation result, complete automatic satisfaction evaluation.
Above-described is the preferred embodiment of the present invention; should be understood that the ordinary person for the art; can also making some improvements and modifications under without departing from principle premise of the present invention, these improvements and modifications are also in protection scope of the present invention.

Claims (10)

1. a method of speech processing, it is characterised in that including:
Hot line speech data is processed, obtains the user speech information related in the text message of described hot line speech data and hot line speech data;
According to described text message and user speech information, obtain the evaluation result of described hot line speech data.
2. method of speech processing according to claim 1, it is characterised in that described hot line speech data is processed, the step obtaining the user speech information related in the text message of described hot line speech data and hot line speech data includes:
Described hot line speech data is carried out voice recognition processing, it is thus achieved that the text message of hot line speech data;
Described hot line speech data is carried out cutting process, it is thus achieved that the user speech information related in hot line speech data.
3. method of speech processing according to claim 2, it is characterised in that described described hot line speech data is carried out voice recognition processing, it is thus achieved that the step of the text message of hot line speech data includes:
By speech recognition engine, described hot line speech data is carried out voice recognition processing, it is thus achieved that the text message of hot line speech data.
4. method of speech processing according to claim 2, it is characterised in that described described hot line speech data is carried out cutting process, it is thus achieved that the step of the user speech information related in hot line speech data includes:
By voice partitioning algorithm, described hot line speech data is carried out phonetic segmentation process, it is thus achieved that the user speech information related in hot line speech data.
5. method of speech processing according to claim 1, it is characterised in that described according to described text message and user speech information, the step of the evaluation result obtaining described hot line speech data includes:
The text feature relevant to user satisfaction and emotion is extracted from described text message;
Extracting from described user speech information and speech emotional, abstract characteristics that customer satisfaction evaluation is relevant, as voice assemblage characteristic, described abstract characteristics includes: temporal signatures, frequency domain character, short-time characteristic, long time feature and session features;
It is combined described text feature and voice assemblage characteristic processing, obtains the evaluation result of described hot line speech data.
6. method of speech processing according to claim 5, it is characterised in that described described text feature and voice assemblage characteristic being combined processes, and the step obtaining described hot line speech data evaluation result includes:
Determine the Feature Words weight in text feature and voice assemblage characteristic respectively according to presetting, be combined processing, using the combined treatment result evaluation result as described hot line speech data.
7. method of speech processing according to claim 1, it is characterized in that, if described hot line speech data includes mark degrees of data, then hot line speech data is processed, after obtaining the user speech information related in the text message of described hot line speech data and hot line speech data, according to described mark degrees of data and described text message and user speech information, obtain the evaluation result of described hot line speech data.
8. a voice processing apparatus, it is characterised in that including:
First acquisition module, for hot line speech data is processed, obtains the user speech information related in the text message of described hot line speech data and hot line speech data;
Second acquisition module, for according to described text message and user speech information, obtaining the evaluation result of described hot line speech data.
9. voice processing apparatus according to claim 8, it is characterised in that described first acquisition module, including:
Voice recognition unit, for carrying out voice recognition processing to described hot line speech data, it is thus achieved that the text message of hot line speech data;
Phonetic segmentation unit, for carrying out cutting process to described hot line speech data, it is thus achieved that the user speech information related in hot line speech data.
10. voice processing apparatus according to claim 8, it is characterised in that described second acquisition module, including:
Text character extraction unit, for extracting the text feature relevant to user satisfaction and emotion from described text message;
Speech feature extraction unit, for extracting from described user speech information and speech emotional, abstract characteristics that customer satisfaction evaluation is relevant, as voice assemblage characteristic, described abstract characteristics includes: temporal signatures, frequency domain character, short-time characteristic, long time feature and session features;
Processing unit, for being combined described text feature and voice assemblage characteristic processing, obtains the evaluation result of described hot line speech data.
CN201410838240.9A 2014-12-29 2014-12-29 Speech processing method and device Pending CN105810205A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410838240.9A CN105810205A (en) 2014-12-29 2014-12-29 Speech processing method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410838240.9A CN105810205A (en) 2014-12-29 2014-12-29 Speech processing method and device

Publications (1)

Publication Number Publication Date
CN105810205A true CN105810205A (en) 2016-07-27

Family

ID=56979889

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410838240.9A Pending CN105810205A (en) 2014-12-29 2014-12-29 Speech processing method and device

Country Status (1)

Country Link
CN (1) CN105810205A (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017133165A1 (en) * 2016-02-01 2017-08-10 百度在线网络技术(北京)有限公司 Method, apparatus and device for automatic evaluation of satisfaction and computer storage medium
CN107293309A (en) * 2017-05-19 2017-10-24 四川新网银行股份有限公司 A kind of method that lifting public sentiment monitoring efficiency is analyzed based on customer anger
CN107731219A (en) * 2017-09-06 2018-02-23 百度在线网络技术(北京)有限公司 Phonetic synthesis processing method, device and equipment
CN107767038A (en) * 2017-10-01 2018-03-06 上海量科电子科技有限公司 voice-based payment evaluation method, client and system
CN108021565A (en) * 2016-11-01 2018-05-11 中国移动通信有限公司研究院 A kind of analysis method and device of the user satisfaction based on linguistic level
CN108735200A (en) * 2018-06-27 2018-11-02 北京灵伴即时智能科技有限公司 A kind of speaker's automatic marking method
CN108962281A (en) * 2018-08-15 2018-12-07 三星电子(中国)研发中心 A kind of evaluation of language expression and householder method and device
CN109064789A (en) * 2018-08-17 2018-12-21 重庆第二师范学院 A kind of adjoint cerebral palsy speaks with a lisp supplementary controlled system and method, assistor
CN110062117A (en) * 2019-04-08 2019-07-26 商客通尚景科技(上海)股份有限公司 A kind of sonic detection and method for early warning
CN110782916A (en) * 2019-09-30 2020-02-11 北京淇瑀信息科技有限公司 Multi-modal complaint recognition method, device and system

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101364408A (en) * 2008-10-07 2009-02-11 西安成峰科技有限公司 Sound image combined monitoring method and system
CN102623009A (en) * 2012-03-02 2012-08-01 安徽科大讯飞信息技术股份有限公司 Abnormal emotion automatic detection and extraction method and system on basis of short-time analysis
CN102655003A (en) * 2012-03-21 2012-09-05 北京航空航天大学 Method for recognizing emotion points of Chinese pronunciation based on sound-track modulating signals MFCC (Mel Frequency Cepstrum Coefficient)
CN103811009A (en) * 2014-03-13 2014-05-21 华东理工大学 Smart phone customer service system based on speech analysis
CN104050965A (en) * 2013-09-02 2014-09-17 广东外语外贸大学 English phonetic pronunciation quality evaluation system with emotion recognition function and method thereof

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101364408A (en) * 2008-10-07 2009-02-11 西安成峰科技有限公司 Sound image combined monitoring method and system
CN102623009A (en) * 2012-03-02 2012-08-01 安徽科大讯飞信息技术股份有限公司 Abnormal emotion automatic detection and extraction method and system on basis of short-time analysis
CN102655003A (en) * 2012-03-21 2012-09-05 北京航空航天大学 Method for recognizing emotion points of Chinese pronunciation based on sound-track modulating signals MFCC (Mel Frequency Cepstrum Coefficient)
CN104050965A (en) * 2013-09-02 2014-09-17 广东外语外贸大学 English phonetic pronunciation quality evaluation system with emotion recognition function and method thereof
CN103811009A (en) * 2014-03-13 2014-05-21 华东理工大学 Smart phone customer service system based on speech analysis

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017133165A1 (en) * 2016-02-01 2017-08-10 百度在线网络技术(北京)有限公司 Method, apparatus and device for automatic evaluation of satisfaction and computer storage medium
CN108021565A (en) * 2016-11-01 2018-05-11 中国移动通信有限公司研究院 A kind of analysis method and device of the user satisfaction based on linguistic level
CN108021565B (en) * 2016-11-01 2021-09-10 中国移动通信有限公司研究院 User satisfaction analysis method and device based on conversation
CN107293309A (en) * 2017-05-19 2017-10-24 四川新网银行股份有限公司 A kind of method that lifting public sentiment monitoring efficiency is analyzed based on customer anger
CN107731219B (en) * 2017-09-06 2021-07-20 百度在线网络技术(北京)有限公司 Speech synthesis processing method, device and equipment
CN107731219A (en) * 2017-09-06 2018-02-23 百度在线网络技术(北京)有限公司 Phonetic synthesis processing method, device and equipment
CN107767038A (en) * 2017-10-01 2018-03-06 上海量科电子科技有限公司 voice-based payment evaluation method, client and system
CN108735200A (en) * 2018-06-27 2018-11-02 北京灵伴即时智能科技有限公司 A kind of speaker's automatic marking method
CN108962281A (en) * 2018-08-15 2018-12-07 三星电子(中国)研发中心 A kind of evaluation of language expression and householder method and device
CN108962281B (en) * 2018-08-15 2021-05-07 三星电子(中国)研发中心 Language expression evaluation and auxiliary method and device
CN109064789A (en) * 2018-08-17 2018-12-21 重庆第二师范学院 A kind of adjoint cerebral palsy speaks with a lisp supplementary controlled system and method, assistor
CN110062117A (en) * 2019-04-08 2019-07-26 商客通尚景科技(上海)股份有限公司 A kind of sonic detection and method for early warning
CN110782916A (en) * 2019-09-30 2020-02-11 北京淇瑀信息科技有限公司 Multi-modal complaint recognition method, device and system
CN110782916B (en) * 2019-09-30 2023-09-05 北京淇瑀信息科技有限公司 Multi-mode complaint identification method, device and system

Similar Documents

Publication Publication Date Title
CN105810205A (en) Speech processing method and device
CN112804400B (en) Customer service call voice quality inspection method and device, electronic equipment and storage medium
CN109599093B (en) Intelligent quality inspection keyword detection method, device and equipment and readable storage medium
CN109256150B (en) Speech emotion recognition system and method based on machine learning
CN108737667B (en) Voice quality inspection method and device, computer equipment and storage medium
WO2021128741A1 (en) Voice emotion fluctuation analysis method and apparatus, and computer device and storage medium
Kelly et al. Deep neural network based forensic automatic speaker recognition in VOCALISE using x-vectors
CN108388638B (en) Semantic parsing method, device, equipment and storage medium
CN107657017A (en) Method and apparatus for providing voice service
CN107154257A (en) Customer service quality evaluating method and system based on customer voice emotion
CN107886951B (en) Voice detection method, device and equipment
CN111193834B (en) Man-machine interaction method and device based on user sound characteristic analysis and electronic equipment
CN103811009A (en) Smart phone customer service system based on speech analysis
CN107767881B (en) Method and device for acquiring satisfaction degree of voice information
CN108735200A (en) A kind of speaker's automatic marking method
Gold et al. Issues and opportunities: The application of the numerical likelihood ratio framework to forensic speaker comparison
CN106782517A (en) A kind of speech audio keyword filter method and device
CN112133277A (en) Sample generation method and device
CN106356054A (en) Method and system for collecting information of agricultural products based on voice recognition
Płaza et al. Call transcription methodology for contact center systems
CN112349266A (en) Voice editing method and related equipment
CN114267340A (en) Method, device, storage medium and equipment for evaluating service quality of 4S shop
CN108962281B (en) Language expression evaluation and auxiliary method and device
Johar Paralinguistic profiling using speech recognition
CN113990288B (en) Method for automatically generating and deploying voice synthesis model by voice customer service

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20160727