CN113645364B - Intelligent voice outbound method for power dispatching - Google Patents
Intelligent voice outbound method for power dispatching Download PDFInfo
- Publication number
- CN113645364B CN113645364B CN202110683217.7A CN202110683217A CN113645364B CN 113645364 B CN113645364 B CN 113645364B CN 202110683217 A CN202110683217 A CN 202110683217A CN 113645364 B CN113645364 B CN 113645364B
- Authority
- CN
- China
- Prior art keywords
- outbound
- voice
- service
- character string
- intelligent
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 32
- 230000003993 interaction Effects 0.000 claims abstract description 20
- 230000008451 emotion Effects 0.000 claims abstract description 16
- 230000008569 process Effects 0.000 claims abstract description 13
- 230000000295 complement effect Effects 0.000 claims abstract description 9
- 238000012423 maintenance Methods 0.000 claims abstract description 8
- 238000007781 pre-processing Methods 0.000 claims abstract description 8
- 230000006870 function Effects 0.000 claims abstract description 7
- 230000008909 emotion recognition Effects 0.000 claims description 12
- 230000002452 interceptive effect Effects 0.000 claims description 11
- 238000013528 artificial neural network Methods 0.000 claims description 8
- 125000004122 cyclic group Chemical group 0.000 claims description 5
- 238000013507 mapping Methods 0.000 claims description 5
- 210000001503 joint Anatomy 0.000 claims description 4
- 230000007812 deficiency Effects 0.000 claims description 3
- 238000000605 extraction Methods 0.000 claims description 3
- 238000007689 inspection Methods 0.000 claims description 3
- 230000000306 recurrent effect Effects 0.000 claims description 3
- 238000012546 transfer Methods 0.000 claims description 3
- 238000005457 optimization Methods 0.000 abstract description 4
- 238000013473 artificial intelligence Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000012544 monitoring process Methods 0.000 description 2
- 230000004075 alteration Effects 0.000 description 1
- 238000012076 audiometry Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000000977 initiatory effect Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000007619 statistical method Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/50—Centralised arrangements for answering calls; Centralised arrangements for recording messages for absent or busy subscribers ; Centralised arrangements for recording messages
- H04M3/51—Centralised call answering arrangements requiring operator intervention, e.g. call or contact centers for telemarketing
- H04M3/523—Centralised call answering arrangements requiring operator intervention, e.g. call or contact centers for telemarketing with call distribution or queueing
- H04M3/5232—Call distribution algorithms
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/06—Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
- G10L15/063—Training
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/27—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique
- G10L25/30—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique using neural networks
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/60—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for measuring the quality of voice signals
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/63—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for estimating an emotional state
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y04—INFORMATION OR COMMUNICATION TECHNOLOGIES HAVING AN IMPACT ON OTHER TECHNOLOGY AREAS
- Y04S—SYSTEMS INTEGRATING TECHNOLOGIES RELATED TO POWER NETWORK OPERATION, COMMUNICATION OR INFORMATION TECHNOLOGIES FOR IMPROVING THE ELECTRICAL POWER GENERATION, TRANSMISSION, DISTRIBUTION, MANAGEMENT OR USAGE, i.e. SMART GRIDS
- Y04S10/00—Systems supporting electrical power generation, transmission or distribution
- Y04S10/50—Systems or methods supporting the power network operation or management, involving a certain degree of interaction with the load-side end user applications
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Multimedia (AREA)
- Acoustics & Sound (AREA)
- Human Computer Interaction (AREA)
- Computational Linguistics (AREA)
- Signal Processing (AREA)
- Artificial Intelligence (AREA)
- Hospice & Palliative Care (AREA)
- Psychiatry (AREA)
- Business, Economics & Management (AREA)
- General Health & Medical Sciences (AREA)
- Child & Adolescent Psychology (AREA)
- Quality & Reliability (AREA)
- Marketing (AREA)
- Evolutionary Computation (AREA)
- Telephonic Communication Services (AREA)
Abstract
The application provides an intelligent voice outbound method facing power dispatching, which comprises the following steps of preprocessing voice information: collecting voice matching content to form a voice signal, identifying service appeal of a user in real time, and preprocessing the voice signal to obtain a text character string S; judging whether VIP users, people with weak self-help ability, emergency repeated incoming calls and the like are required, if so, manual and intelligent voice interaction understanding is needed, multiple-round interaction understanding functions are supported, text character strings S automatically fill grooves and ask backwards in the interaction process, and users are guided to complement information S n Decoding according to the semantic code C, the set emotion representation Con (Q) and the set theme representation Sof (Q) to obtain a target sequence Y; and checking the quality of the outbound result, and checking the service completion condition with an outbound service system. The method can effectively and accurately connect the service completion condition with the outbound service system, provide continuous optimization service of the telephone operation model for power dispatching, and save maintenance cost of the telephone operation model for power dispatching.
Description
Technical Field
The application relates to the field of voice service, in particular to an intelligent voice outbound method facing power dispatching.
Background
The intelligent voice-based navigation system is intelligent, and has important significance for intelligent upgrading of dispatcher business operations such as ground dispatching operation, fault handling, signal monitoring and the like in current power grid dispatching. The development of artificial intelligence technology is mature day by day, and the system has strong universality, standardized, automatic and modularized industrial mass production characteristics and has the capability of deeply enabling power dispatching. The application of the artificial intelligence technology depends on the continuous maintenance of manpower, the good intelligent outbound service effect can be achieved only by the continuous tuning of the conversation model by the organization professional, and the personnel investment of most institutions in this aspect is difficult to ensure, so that the intelligent outbound service effect is affected. Traditional customer service system excessively relies on the manpower mode, needs to reduce power dispatching input cost, provides the continuous optimization service of conversation model for power dispatching, saves the conversation model maintenance cost of power dispatching.
Disclosure of Invention
The embodiment of the application provides an intelligent voice outbound method for power dispatching, which can effectively and accurately dock the service completion condition with an outbound service system, provide continuous optimization service of a telephone model for power dispatching, and save the maintenance cost of the telephone model for power dispatching.
The method specifically comprises the following steps:
step one, preprocessing voice information: collecting voice matching content to form a voice signal, identifying service appeal of a user in real time, and preprocessing the voice signal to obtain a text character string S;
step two, intelligent prejudging and transferring are carried out manually, and whether VIP users, people with weak self-help ability, emergency repeated incoming calls and the like are judged according to the text character string S in the step one, if so, the step eight is directly transferred, and if not, the text character string S is transferred to the step three;
step four, intelligent voice interactive understanding, supporting a multi-round interactive understanding function, automatically filling grooves and asking back in the text character string S in the interactive process, and guiding a user to complement information S n The context information is associated to accurately acquire a user demand character string Q;
mapping the source sequence intelligent voice interaction understanding model U into a semantic vector C, using a cyclic neural network as a decoder, and decoding to obtain a target sequence Y according to the semantic code C, the set emotion representation Con (Q) and the set theme representation Sof (Q);
step six, the quality inspection of the outbound result, check the service completion situation with the outbound business system, whether to finish effectively, transfer the result into step seven and step eight;
and step seven, model retraining is carried out by updating the model, the result of the step six is fed back to the step five in real time, the intelligent outbound model is updated, model parameters are trained, and the model is updated.
Optionally, the fourth step includes:
the text character string S automatically fills the slot for back inquiry in the interaction process, guides the user to complement the information Sn, and correlates the context information to accurately acquire the character string Q required by the user,
Q={S,S 1 ,...,S n };
wherein Q is a user demand string, S 1 ,...,S n The method is that intelligent voice interaction understanding complement each information character string;
the intelligent voice interaction understanding model U comprises voice emotion recognition and semantic emotion recognition;
U=∑Con(Q)+Sof(Q);
wherein Con (Q) is speech emotion recognition; judging the emotion of the customer according to the threshold value of the speed and the severity of the language of the customer; sof (Q) is semantic emotion recognition; judging the dissatisfaction of the client according to the keywords of the sentences of the client, such as the keywords of the dirty words, complaints and the like;
if the emotion of the customer continuously fluctuates, the demand understanding information U is transmitted to the fifth step.
Optionally, the fifth step includes:
mapping the source sequence intelligent voice interaction understanding model U into a semantic vector C, using a cyclic neural network (Recurrent Neural Network, RNN) as a decoder, and decoding to obtain a target sequence Y according to the semantic code C, the set emotion representation Con (Q) and the set theme representation Sof (Q);
C=f(Q 1 ,Q 2 ,...,Q M );
the target sequence Y supports the butt joint of third party service systems such as a dispatching system and the like, realizes the extraction of key information in the interaction process, and is pushed to the third party service system through an interface, and the outbound result condition is fed back to the step six.
Optionally, the method includes:
and step eight, artificial service deficiency.
The beneficial effects are that:
the method can effectively and accurately connect the service completion condition with the outbound service system, provide continuous optimization service of the telephone operation model for power dispatching, and save maintenance cost of the telephone operation model for power dispatching.
Drawings
In order to more clearly illustrate the technical solutions of the present application, the drawings that are needed in the description of the embodiments will be briefly described below, it being obvious that the drawings in the following description are only some embodiments of the present application, and that other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
Fig. 1 is a flow chart of an intelligent voice outbound method for power dispatching.
Detailed Description
In order to make the structure and advantages of the present application more apparent, the structure of the present application will be further described with reference to the accompanying drawings.
Referring to fig. 1, the application discloses an intelligent voice outbound method facing power dispatching, which comprises the following steps:
step one, preprocessing voice information: collecting voice matching content to form a voice signal, identifying service appeal of a user in real time, and preprocessing the voice signal to obtain a text character string S;
step two, intelligent prejudging and transferring are carried out manually, and whether VIP users, people with weak self-help ability, emergency repeated incoming calls and the like are judged according to the text character string S in the step one, if so, the step eight is directly transferred, and if not, the text character string S is transferred to the step three;
step three, intelligent prejudging and switching keys, wherein step two does not need to switch to manual service, and the step four is further judged whether to switch to key navigation service or not according to database matching, knowledge point matching, speaking matching, keyword matching and the like when the maintenance task is configured in the text character string S by combining number source and interface source information;
step four, intelligent voice interactive understanding, supporting a multi-round interactive understanding function, automatically filling grooves and asking back in the text character string S in the interactive process, and guiding a user to complement information S n And the context information is associated to accurately acquire the user demand character string Q,
Q={S,S 1 ,...,S n };
wherein Q is a user demand string, S 1 ,...,S n Is intelligent voice interactive understanding complement information character strings.
The intelligent voice interaction understanding model U comprises voice emotion recognition and semantic emotion recognition.
U=∑Con(Q)+Sof(Q);
Wherein Con (Q) is speech emotion recognition. And judging the emotion of the customer according to the threshold value of the voice intensity and the voice speed of the customer. Sof (Q) is semantic emotion recognition. Judging the dissatisfaction of the client according to the keywords of the sentences of the client, such as the keywords of the dirty words, complaints and the like
The intelligent speech interaction understanding model U is identified as pleasant, general, discontent, anger. And identifying the emotion state of the client according to the emotion service strategy model, and reminding intervention pacifying in real time. If the emotion of the customer continuously fluctuates, a team leader is warned in time, and meanwhile, the customer emotion is properly strategically compensated and calmed according to the anger emotion service strategy of the customer, so that the complaint risk is reduced. Transmitting the demand understanding information U into a fifth step;
step five, calculating an intelligent outbound model,
mapping the source sequence intelligent voice interaction understanding model U into a semantic vector C, and decoding to obtain a target sequence Y by using a cyclic neural network (Recurrent Neural Network, RNN) as a decoder according to the semantic code C, the set emotion representation Con (Q) and the set theme representation Sof (Q).
C=f(Q 1 ,Q 2 ,...,Q M );
The target sequence Y supports the butt joint of third party service systems such as a dispatching system and the like, realizes the extraction of key information (such as addresses, names, application numbers and the like) in the interaction process, and pushes the key information into the third party service system through an interface, and feeds back the outbound result condition to the step six;
step six, the quality inspection of the outbound result, check the service completion situation with the outbound business system, whether to finish effectively, transfer the result into step seven and step eight;
step seven, model retraining is carried out by updating the model, the result of the step six is fed back to the step five in real time, the intelligent outbound model is updated, model parameters are trained, and the model is updated;
and step eight, artificial service deficiency.
In the application, an intelligent outbound system task monitoring module is constructed, so that the data such as the task progress, call rate and the like of voice outbound can be monitored, and a perfect task report is provided; supporting the butt joint with a third party service system, thereby acquiring outbound numbers, key information and the like from the third party system, perfecting outbound tasks and automatically initiating outbound; the system maintenance function is provided, and is mainly used for maintaining a voice library, an outbound display number, an interface, knowledge points, a speaking operation and keywords in task configuration; supporting a third party application system to call an outbound system interface so as to acquire an outbound result (success or failure); supporting the function of calling voice interruption; the system supports the functions of outbound data recording and statistical analysis, and comprises the statistics of indexes such as audio audiometry of historical conversations, total number of calls recorded, number of calls, total call duration, average call duration, completion amount, success rate and the like. Intelligent outbound interactive dialogue system, voice and semantic recognition rate is not lower than 95%: supporting the user interaction process of intelligent outbound and response is beneficial to improving the service efficiency and enabling the customer service system to be intelligent and efficient.
It should be noted that in this document, terms such as "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
Although embodiments of the present application have been shown and described, it will be understood by those skilled in the art that various changes, modifications, substitutions and alterations can be made therein without departing from the principles and spirit of the application, the scope of which is defined in the appended claims and their equivalents.
Claims (3)
1. The intelligent voice outbound method for power dispatching is characterized by comprising the following steps of:
step one, preprocessing voice information: collecting voice matching content to form a voice signal, identifying service appeal of a user in real time, and preprocessing the voice signal to obtain a text character string S;
step two, intelligent prejudging and transferring are carried out manually, and whether VIP users, people with weak self-help ability, emergency repeated incoming calls and the like are judged according to the text character string S in the step one, if so, the step eight is directly transferred, and if not, the text character string S is transferred to the step three;
step three, intelligent prejudging and switching keys, wherein step two does not need to switch to manual service, and the step four is further judged whether to switch to key navigation service or not according to database matching, knowledge point matching, speaking matching, keyword matching and the like when the maintenance task is configured in the text character string S by combining number source and interface source information;
step four, intelligent voice interactive understanding, supporting a multi-round interactive understanding function, automatically filling grooves and asking back in the text character string S in the interactive process, and guiding a user to complement information S n The context information is associated to accurately acquire a user demand character string Q;
mapping the source sequence intelligent voice interaction understanding model U into a semantic vector C, using a cyclic neural network as a decoder, and decoding to obtain a target sequence Y according to the semantic vector C, the set emotion representation Con (Q) and the set theme representation Sof (Q);
step six, the quality inspection of the outbound result, check the service completion situation with the outbound business system, whether to finish effectively, transfer the result into step seven and step eight;
step seven, model retraining is carried out by updating the model, the result of the step six is fed back to the step five in real time, the intelligent outbound model is updated, model parameters are trained, and the model is updated;
and step eight, artificial service deficiency.
2. The intelligent voice outbound method for power scheduling according to claim 1, wherein the fourth step comprises:
the text character string S automatically fills the slot for back inquiry in the interaction process, guides the user to complement the information Sn, and correlates the context information to accurately acquire the character string Q required by the user,
Q={S,S 1 ,...,S n };
wherein Q is a user demand string, S 1 ,...,S n The method is that intelligent voice interaction understanding complement each information character string;
the intelligent voice interaction understanding model U comprises voice emotion recognition and semantic emotion recognition;
U=∑Con(Q)+Sof(Q);
wherein Con (Q) is speech emotion recognition; judging the emotion of the customer according to the threshold value of the speed and the severity of the language of the customer; sof (Q) is semantic emotion recognition; judging the dissatisfaction of the client according to the keywords of the sentences of the client, such as the keywords of the dirty words, complaints and the like;
if the emotion of the customer continuously fluctuates, the demand understanding information U is transmitted to the fifth step.
3. The intelligent voice outbound method for power scheduling according to claim 1, wherein the fifth step comprises:
mapping the source sequence intelligent voice interaction understanding model U into a semantic vector C, using a cyclic neural network (Recurrent Neural Network, RNN) as a decoder, and decoding to obtain a target sequence Y according to the semantic vector C and a set emotion representation Con (Q) and a theme representation Sof (Q);
C=f(Q 1 ,Q 2 ,...,Q M );
the target sequence Y supports the butt joint of third party service systems such as a dispatching system and the like, realizes the extraction of key information in the interaction process, and is pushed to the third party service system through an interface, and the outbound result condition is fed back to the step six.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110683217.7A CN113645364B (en) | 2021-06-21 | 2021-06-21 | Intelligent voice outbound method for power dispatching |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110683217.7A CN113645364B (en) | 2021-06-21 | 2021-06-21 | Intelligent voice outbound method for power dispatching |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113645364A CN113645364A (en) | 2021-11-12 |
CN113645364B true CN113645364B (en) | 2023-08-22 |
Family
ID=78415993
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110683217.7A Active CN113645364B (en) | 2021-06-21 | 2021-06-21 | Intelligent voice outbound method for power dispatching |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113645364B (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116030811B (en) * | 2023-03-22 | 2023-06-30 | 广州小鹏汽车科技有限公司 | Voice interaction method, vehicle and computer readable storage medium |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104125354A (en) * | 2014-06-27 | 2014-10-29 | 国家电网公司 | Method and system for regulating and controlling voice service based on call requirements |
CN108305642A (en) * | 2017-06-30 | 2018-07-20 | 腾讯科技(深圳)有限公司 | The determination method and apparatus of emotion information |
CN110769124A (en) * | 2019-10-30 | 2020-02-07 | 国网江苏省电力有限公司镇江供电分公司 | Electric power marketing customer communication system |
CN111127031A (en) * | 2018-10-11 | 2020-05-08 | 上海智臻智能网络科技股份有限公司 | Intelligent customer service system interaction method, intelligent customer service robot and storage medium |
CN111192060A (en) * | 2019-12-23 | 2020-05-22 | 广州供电局有限公司 | Electric power IT service-based full-channel self-service response implementation method |
WO2020135194A1 (en) * | 2018-12-26 | 2020-07-02 | 深圳Tcl新技术有限公司 | Emotion engine technology-based voice interaction method, smart terminal, and storage medium |
CN111541821A (en) * | 2020-07-10 | 2020-08-14 | 北京灵伴即时智能科技有限公司 | Telephone customer service system and telephone customer service flexible switching method |
CN112185389A (en) * | 2020-09-22 | 2021-01-05 | 北京小米松果电子有限公司 | Voice generation method and device, storage medium and electronic equipment |
CN112581958A (en) * | 2020-12-07 | 2021-03-30 | 中国南方电网有限责任公司 | Short voice intelligent navigation method applied to electric power field |
CN112804400A (en) * | 2020-12-31 | 2021-05-14 | 中国工商银行股份有限公司 | Customer service call voice quality inspection method and device, electronic equipment and storage medium |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10692492B2 (en) * | 2017-09-29 | 2020-06-23 | Intel IP Corporation | Techniques for client-side speech domain detection using gyroscopic data and a system using the same |
US11164572B2 (en) * | 2018-05-01 | 2021-11-02 | Dell Products, L.P. | Intelligent assistance across voice services |
-
2021
- 2021-06-21 CN CN202110683217.7A patent/CN113645364B/en active Active
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104125354A (en) * | 2014-06-27 | 2014-10-29 | 国家电网公司 | Method and system for regulating and controlling voice service based on call requirements |
CN108305642A (en) * | 2017-06-30 | 2018-07-20 | 腾讯科技(深圳)有限公司 | The determination method and apparatus of emotion information |
CN111127031A (en) * | 2018-10-11 | 2020-05-08 | 上海智臻智能网络科技股份有限公司 | Intelligent customer service system interaction method, intelligent customer service robot and storage medium |
WO2020135194A1 (en) * | 2018-12-26 | 2020-07-02 | 深圳Tcl新技术有限公司 | Emotion engine technology-based voice interaction method, smart terminal, and storage medium |
CN110769124A (en) * | 2019-10-30 | 2020-02-07 | 国网江苏省电力有限公司镇江供电分公司 | Electric power marketing customer communication system |
CN111192060A (en) * | 2019-12-23 | 2020-05-22 | 广州供电局有限公司 | Electric power IT service-based full-channel self-service response implementation method |
CN111541821A (en) * | 2020-07-10 | 2020-08-14 | 北京灵伴即时智能科技有限公司 | Telephone customer service system and telephone customer service flexible switching method |
CN112185389A (en) * | 2020-09-22 | 2021-01-05 | 北京小米松果电子有限公司 | Voice generation method and device, storage medium and electronic equipment |
CN112581958A (en) * | 2020-12-07 | 2021-03-30 | 中国南方电网有限责任公司 | Short voice intelligent navigation method applied to electric power field |
CN112804400A (en) * | 2020-12-31 | 2021-05-14 | 中国工商银行股份有限公司 | Customer service call voice quality inspection method and device, electronic equipment and storage medium |
Non-Patent Citations (1)
Title |
---|
智能机器人在电网培训服务中的落地实践;王佳培;叶碧;李康宁;;电力设备管理(02);全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN113645364A (en) | 2021-11-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CA2886609C (en) | Computer-implemented call center architecture and method for optimizing customer experience through in-band expert intervention | |
US11632463B2 (en) | Automated systems and methods for natural language processing with speaker intention inference | |
CN108156331A (en) | A kind of intelligence auto dialing outgoing call system and method | |
US7280651B2 (en) | Method and system for performing automated telemarketing | |
CN107331383A (en) | One kind is based on artificial intelligence telephone outbound system and its implementation | |
CN110263144A (en) | A kind of answer acquisition methods and device | |
CN104809197A (en) | On-line question and answer method based on intelligent robot | |
CN111739519A (en) | Dialogue management processing method, device, equipment and medium based on voice recognition | |
WO2012012247A1 (en) | Efficiently reducing transcription error using hybrid voice transcription | |
CN111489765A (en) | Telephone traffic service quality inspection method based on intelligent voice technology | |
CN110392168A (en) | Call processing method, device, server, storage medium and system | |
CN112885348A (en) | AI-combined intelligent voice electric marketing method | |
CN102868836A (en) | Real person talk skill system for call center and realization method thereof | |
KR20190097461A (en) | Person-artificial intelligence collaboration Call Center System and service method | |
CN109859032A (en) | Funds on account collection method, apparatus, equipment and storage medium based on intelligent sound | |
CN113645364B (en) | Intelligent voice outbound method for power dispatching | |
CN110705309A (en) | Service quality evaluation method and system | |
CN113011159A (en) | Artificial seat monitoring method and device, electronic equipment and storage medium | |
CN114328867A (en) | Intelligent interruption method and device in man-machine conversation | |
CN116340482A (en) | Multi-skill customer service auxiliary product based on enterprise WeChat combined with NLP engine | |
CN112150694A (en) | Intelligent voice electric charge payment urging system and method | |
CN113784006A (en) | AI outbound based intelligent evaluation method for communication operator | |
CN114040056A (en) | Method, device and storage medium for automatically checking user information | |
CN112929502A (en) | Voice recognition method and system based on electric power customer service | |
CN113472958A (en) | Method, system, electronic device and storage medium for receiving branch telephone in centralized mode |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |