CN112399019A - Intelligent outbound call method, terminal equipment and readable storage medium - Google Patents
Intelligent outbound call method, terminal equipment and readable storage medium Download PDFInfo
- Publication number
- CN112399019A CN112399019A CN202010974415.4A CN202010974415A CN112399019A CN 112399019 A CN112399019 A CN 112399019A CN 202010974415 A CN202010974415 A CN 202010974415A CN 112399019 A CN112399019 A CN 112399019A
- Authority
- CN
- China
- Prior art keywords
- text
- voice
- outbound
- reply content
- converting
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 52
- 238000006243 chemical reaction Methods 0.000 claims abstract description 13
- 238000004590 computer program Methods 0.000 claims description 20
- 238000012545 processing Methods 0.000 claims description 11
- 230000011218 segmentation Effects 0.000 claims description 5
- 230000008030 elimination Effects 0.000 claims description 4
- 238000003379 elimination reaction Methods 0.000 claims description 4
- 230000009467 reduction Effects 0.000 claims description 4
- 238000004891 communication Methods 0.000 abstract description 12
- 238000005516 engineering process Methods 0.000 abstract description 4
- 230000008569 process Effects 0.000 description 15
- 230000006870 function Effects 0.000 description 6
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 230000015572 biosynthetic process Effects 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000003786 synthesis reaction Methods 0.000 description 2
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 238000009499 grossing Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000012805 post-processing Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/487—Arrangements for providing information services, e.g. recorded voice services or time announcements
- H04M3/493—Interactive information services, e.g. directory enquiries ; Arrangements therefor, e.g. interactive voice response [IVR] systems or voice portals
- H04M3/4936—Speech interaction details
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/18—Speech classification or search using natural language modelling
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/20—Speech recognition techniques specially adapted for robustness in adverse environments, e.g. in noise, of stress induced speech
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/50—Centralised arrangements for answering calls; Centralised arrangements for recording messages for absent or busy subscribers ; Centralised arrangements for recording messages
- H04M3/527—Centralised call answering arrangements not requiring operator intervention
Landscapes
- Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Artificial Intelligence (AREA)
- Telephonic Communication Services (AREA)
Abstract
The present application belongs to the field of communications technologies, and in particular, to an intelligent outbound method, a terminal device, and a readable storage medium. Wherein, the method comprises the following steps: acquiring an outbound voice; converting the outbound voice into a corresponding text; selecting corresponding reply content according to the text; and generating a corresponding reply voice according to the reply content. The intelligent outbound method, the terminal device and the readable storage medium provided by the embodiment of the application realize the intelligent selection of the reply content by the computer through the text conversion and the text semantic understanding of the outbound voice sent by the client. In order to be more suitable for telephone communication and provide telephone service for clients, after the reply content is selected, the reply content is further converted into voice, so that the telephone service for the clients can be completed instead of manual customer service, and the labor cost of service industries such as communication, finance and the like is favorably reduced.
Description
Technical Field
The present application belongs to the field of communications technologies, and in particular, to an intelligent outbound method, a terminal device, and a readable storage medium.
Background
Modern service industries, such as the communications service industry and the financial service industry, often require telephone service transactions to be provided to customers. Traditional telephone business handling needs enterprises to configure a large amount of manual customer service, answers incoming calls of clients and handles business, and labor cost is high.
Disclosure of Invention
In view of this, embodiments of the present application provide an intelligent outbound method, a terminal device, and a readable storage medium, so as to solve the problem that the labor cost of the current service industry is high.
According to a first aspect, an embodiment of the present application provides an intelligent outbound method, including: acquiring an outbound voice; converting the outbound voice into a corresponding text; selecting corresponding reply content according to the text; and generating corresponding reply voice according to the reply content.
With reference to the first aspect, in some embodiments of the present application, the step of converting the outbound voice into corresponding text includes: determining the initial position and the end position of pronunciation according to the audio stream of the outbound voice; extracting target audio data according to the starting position and the ending position; and decoding the target audio data to generate a corresponding text.
With reference to the first aspect, in some embodiments of the present application, after the step of decoding the target audio data to generate a corresponding text, the step of converting the outbound voice into the corresponding text further includes: and performing text smoothness, punctuation prediction and text segmentation on the text.
With reference to the first aspect, in some embodiments of the present application, before the step of determining a start position and an end position of a pronunciation according to an audio stream of the outbound voice, the step of converting the outbound voice into a corresponding text further includes: and carrying out noise reduction and reverberation elimination processing on the outbound voice.
With reference to the first aspect, in some embodiments of the present application, the step of selecting a corresponding reply content according to the text includes: matching a corresponding intention scene for the text according to a preset grammar rule; and converting the text into corresponding structured data according to the intention scene, and determining reply content according to the structured data.
With reference to the first aspect, in some embodiments of the present application, the step of selecting a corresponding reply content according to the text includes: extracting text content according to a preset fixed slot position; and converting the text into corresponding structured data according to the text content of the fixed slot, and determining reply content according to the structured data.
With reference to the first aspect, in some embodiments of the present application, the step of generating a corresponding reply voice according to the reply content includes: converting the reply content into a corresponding rhyme sequence; and generating corresponding reply voice according to the rhyme sequence.
According to a second aspect, an embodiment of the present application provides a terminal device, including: the input unit is used for acquiring the outbound voice; the text conversion unit is used for converting the outbound voice into a corresponding text; the text understanding unit is used for selecting corresponding reply content according to the text; and the voice playing unit is used for generating corresponding reply voice according to the reply content.
According to a third aspect, an embodiment of the present application provides a terminal device, which includes a memory, a processor, and a computer program stored in the memory and executable on the processor, where the processor implements the steps of the method according to the first aspect or any embodiment of the first aspect when executing the computer program.
According to a fourth aspect, embodiments of the present application provide a computer-readable storage medium, which stores a computer program, and the computer program, when executed by a processor, implements the steps of the method according to the first aspect or any embodiment of the first aspect.
The intelligent outbound method, the terminal device and the readable storage medium provided by the embodiment of the application realize the intelligent selection of the reply content by the computer through the text conversion and the text semantic understanding of the outbound voice sent by the client. In order to be more suitable for telephone communication and provide telephone service for clients, after the reply content is selected, the reply content is further converted into voice, so that the telephone service for the clients can be completed instead of manual customer service, and the labor cost of service industries such as communication, finance and the like is favorably reduced.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, and it is obvious that the drawings in the following description are some embodiments of the present invention, and other drawings can be obtained by those skilled in the art without creative efforts.
Fig. 1 is a flowchart of a specific example of an intelligent outbound method provided by an embodiment of the present application;
fig. 2 is a flowchart of another specific example of an intelligent outbound method provided by an embodiment of the present application;
fig. 3 is a schematic structural diagram of a specific example of a terminal device provided in an embodiment of the present application;
fig. 4 is a schematic structural diagram of another specific example of the terminal device provided in the embodiment of the present application.
Detailed Description
In the following description, for purposes of explanation and not limitation, specific details are set forth, such as particular system structures, techniques, etc. in order to provide a thorough understanding of the embodiments of the present application. It will be apparent, however, to one skilled in the art that the present application may be practiced in other embodiments that depart from these specific details. In other instances, detailed descriptions of well-known systems, devices, circuits, and methods are omitted so as not to obscure the description of the present application with unnecessary detail.
In order to explain the technical solution described in the present application, the following description will be given by way of specific examples.
The embodiment of the application provides an intelligent outbound method, as shown in fig. 1, the method may include the following steps:
step S101: and acquiring the outbound voice.
When a client transacts communication or financial services such as price package change, credit card payment by stages and the like, the client often selects telephone transaction, namely telephone service transaction is performed by dialing the telephone of the client of an enterprise. The service handling requirement proposed by the client after the call is connected is the outbound voice mentioned in the embodiment of the application.
Step S102: the outbound voice is converted to corresponding text.
In a specific embodiment, as shown in fig. 2, the process of step S102 can be implemented by the following several sub-steps:
step S1021: based on the audio stream of the outbound speech, the start and end positions of the utterance are determined.
Step S1022: target audio data is extracted based on the start position and the end position.
Step S1023: and decoding the target audio data to generate a corresponding text.
In practical applications, in order to improve the recognition effect of the audio, the following sub-steps may be added before step S1021:
step S1020: and carrying out noise reduction and reverberation elimination processing on the outbound voice.
After the noise reduction and reverberation elimination processing is performed on the input voice, the endpoint detection can be performed on the input audio stream, the starting position and the ending position of the speaking are determined, and the identification processing is synchronously performed. In step S1023, in order to improve the user response speed and ensure the recognition effect, two-pass decoding may be adopted in the recognition process. Specifically, the target audio data can be decoded once by using acoustic fusion models such as DFCNN and BilSTM and an nGram language model; and then, carrying out two-time decoding processing on the data obtained by the one-time decoding by utilizing the domain language model and the RNN language model.
In order to improve the readability of the text, post-processing such as punctuation prediction, text smoothness, text segmentation and the like can be performed on the data obtained after the two-pass decoding, and finally the text is generated. As shown in fig. 2, step S1024 may be added after step S1023 to perform text smoothing, punctuation prediction and text segmentation on the text.
Step S103: and selecting corresponding reply content according to the text.
After obtaining the text corresponding to the outbound voice, semantic understanding of the text is required. Semantic understanding (NLP) refers to converting a natural language into computer-readable structured data, and can achieve the purpose of semantic understanding by matching corresponding intention scenes through grammar rules or capturing text contents of fixed slots.
Specifically, the process of step S103 may be implemented by the following several sub-steps:
step S1031: and matching the corresponding intention scene for the text according to a preset grammar rule.
Step S1032: and converting the text into corresponding structured data according to the intention scene, and determining reply content according to the structured data.
In a specific embodiment, the following substeps may also be used to implement the process of step S103, in place of step S1031 and step S1032:
step S1031': and extracting text content according to the preset fixed slot position.
Step S1032': and converting the text into corresponding structured data according to the text content of the fixed slot position, and determining reply content according to the structured data.
Step S104: and generating a corresponding reply voice according to the reply content.
In a specific embodiment, as shown in fig. 2, the process of step S104 can be implemented by the following several sub-steps:
step S1041: and converting the reply content into a corresponding rhyme sequence.
Step S1042: and generating corresponding reply voice according to the rhyme sequence.
In step S104, the text is converted into sound, and the sound corresponding to the anchor is synthesized according to the selected anchor, and the specific principle is as follows:
speech synthesis can be regarded as an artificial intelligence system. In order to synthesize a high-quality language, besides relying on various rules, including semantic rules, lexical rules, and phonetic rules, the content of the text must be well understood, which also relates to the understanding of natural language. The text-to-speech conversion process is to convert the text sequence into a rhyme sequence and then generate a speech waveform according to the rhyme sequence. Wherein, step S1041 involves linguistic processing, such as word segmentation, word-to-speech conversion, etc., and a set of valid prosody control rules; step S1042 requires advanced speech synthesis technology, and can synthesize a high-quality speech stream in real time as required. Generally speaking, a text-to-speech conversion system requires a complex conversion procedure from text sequences to phoneme sequences, that is, the text-to-speech conversion system not only needs digital signal processing technology, but also needs to be supported by a great deal of linguistic knowledge.
According to the intelligent outbound method provided by the embodiment of the application, the text conversion and the text semantic understanding are carried out on the outbound voice sent by the client, so that the reply content can be intelligently selected by the computer. In order to be more suitable for telephone communication and provide telephone service for clients, after the reply content is selected, the reply content is further converted into voice, so that the telephone service for the clients can be completed instead of manual customer service, and the labor cost of service industries such as communication, finance and the like is favorably reduced.
It should be understood that, the sequence numbers of the steps in the foregoing embodiments do not imply an execution sequence, and the execution sequence of each process should be determined by its function and inherent logic, and should not constitute any limitation to the implementation process of the embodiments of the present application.
An embodiment of the present application further provides a terminal device, as shown in fig. 3, where the terminal device may include: an input unit 301, a text conversion unit 302, a text understanding unit 303, and a voice playing unit 304.
Specifically, the input unit 301 is configured to obtain an outbound voice; the corresponding implementation process can be referred to the record of step S101 in the above method embodiment.
The text conversion unit 302 is configured to convert the outbound voice into a corresponding text; the corresponding implementation process can be referred to the description of step S102 in the above method embodiment.
The text understanding unit 303 is configured to select corresponding reply content according to the text; the corresponding implementation process can be referred to the description of step S103 in the above method embodiment.
The voice playing unit 304 is configured to generate a corresponding reply voice according to the reply content; the corresponding implementation process can be referred to the record of step S104 in the above method embodiment.
Fig. 4 is a schematic diagram of another terminal device provided in an embodiment of the present application. As shown in fig. 4, the terminal device 400 of this embodiment includes: a processor 401, a memory 402 and a computer program 403, such as an intelligent call-out program, stored in said memory 402 and executable on said processor 401. The processor 401, when executing the computer program 403, implements the steps in the above-described various embodiments of the intelligent outbound method, such as the steps S101 to S104 shown in fig. 1. Alternatively, the processor 401, when executing the computer program 403, implements the functions of the modules/units in the device embodiments described above, such as the functions of the input unit 301, the text conversion unit 302, the text understanding unit 303, and the speech playing unit 304 shown in fig. 3.
The computer program 403 may be partitioned into one or more modules/units that are stored in the memory 402 and executed by the processor 401 to accomplish the present application. The one or more modules/units may be a series of computer program instruction segments capable of performing specific functions, which are used for describing the execution process of the computer program 403 in the terminal device 400. For example, the computer program 403 may be partitioned into a synchronization module, a summarization module, an acquisition module, a return module (a module in a virtual device).
The terminal device 400 may be a desktop computer, a notebook, a palm computer, a cloud server, or other computing devices. The terminal device may include, but is not limited to, a processor 401, a memory 402. Those skilled in the art will appreciate that fig. 4 is merely an example of a terminal device 400 and does not constitute a limitation of terminal device 400 and may include more or fewer components than shown, or some components may be combined, or different components, e.g., the terminal device may also include input-output devices, network access devices, buses, etc.
The Processor 401 may be a Central Processing Unit (CPU), other general purpose Processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), an off-the-shelf Programmable Gate Array (FPGA) or other Programmable logic device, discrete Gate or transistor logic, discrete hardware components, etc. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
The storage 402 may be an internal storage unit of the terminal device 400, such as a hard disk or a memory of the terminal device 400. The memory 402 may also be an external storage device of the terminal device 400, such as a plug-in hard disk, a Smart Media Card (SMC), a Secure Digital (SD) Card, a Flash memory Card (Flash Card), and the like, which are provided on the terminal device 400. Further, the memory 402 may also include both an internal storage unit and an external storage device of the terminal device 400. The memory 402 is used for storing the computer programs and other programs and data required by the terminal device. The memory 402 may also be used to temporarily store data that has been output or is to be output.
It will be apparent to those skilled in the art that, for convenience and brevity of description, only the above-mentioned division of the functional units and modules is illustrated, and in practical applications, the above-mentioned function distribution may be performed by different functional units and modules according to needs, that is, the internal structure of the apparatus is divided into different functional units or modules to perform all or part of the above-mentioned functions. Each functional unit and module in the embodiments may be integrated in one processing unit, or each unit may exist alone physically, or two or more units are integrated in one unit, and the integrated unit may be implemented in a form of hardware, or in a form of software functional unit. In addition, specific names of the functional units and modules are only for convenience of distinguishing from each other, and are not used for limiting the protection scope of the present application. The specific working processes of the units and modules in the system may refer to the corresponding processes in the foregoing method embodiments, and are not described herein again.
In the above embodiments, the descriptions of the respective embodiments have respective emphasis, and reference may be made to the related descriptions of other embodiments for parts that are not described or illustrated in a certain embodiment.
Those of ordinary skill in the art will appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware or combinations of computer software and electronic hardware. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the implementation. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present application.
In the embodiments provided in the present application, it should be understood that the disclosed apparatus/terminal device and method may be implemented in other ways. For example, the above-described embodiments of the apparatus/terminal device are merely illustrative, and for example, the division of the modules or units is only one logical division, and there may be other divisions when actually implemented, for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection through some interfaces, devices or units, and may be in an electrical, mechanical or other form.
The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In addition, functional units in the embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit.
The integrated modules/units, if implemented in the form of software functional units and sold or used as separate products, may be stored in a computer readable storage medium. Based on such understanding, all or part of the flow in the method of the embodiments described above can be realized by a computer program, which can be stored in a computer-readable storage medium and can realize the steps of the embodiments of the methods described above when the computer program is executed by a processor. . Wherein the computer program comprises computer program code, which may be in the form of source code, object code, an executable file or some intermediate form, etc. The computer-readable medium may include: any entity or device capable of carrying the computer program code, recording medium, usb disk, removable hard disk, magnetic disk, optical disk, computer Memory, Read-Only Memory (ROM), Random Access Memory (RAM), electrical carrier wave signals, telecommunications signals, software distribution medium, and the like. It should be noted that the computer readable medium may contain content that is subject to appropriate increase or decrease as required by legislation and patent practice in jurisdictions, for example, in some jurisdictions, computer readable media does not include electrical carrier signals and telecommunications signals as is required by legislation and patent practice.
The above-mentioned embodiments are only used for illustrating the technical solutions of the present application, and not for limiting the same; although the present application has been described in detail with reference to the foregoing embodiments, it should be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; such modifications and substitutions do not substantially depart from the spirit and scope of the embodiments of the present application and are intended to be included within the scope of the present application.
Claims (10)
1. An intelligent outbound method, comprising:
acquiring an outbound voice;
converting the outbound voice into a corresponding text;
selecting corresponding reply content according to the text;
and generating corresponding reply voice according to the reply content.
2. The intelligent outbound method of claim 1 wherein said step of converting said outbound voice to corresponding text comprises:
determining the initial position and the end position of pronunciation according to the audio stream of the outbound voice;
extracting target audio data according to the starting position and the ending position;
and decoding the target audio data to generate a corresponding text.
3. The intelligent outbound method of claim 2 wherein said step of converting said outbound speech to corresponding text after said step of decoding said target audio data to generate corresponding text further comprises:
and performing text smoothness, punctuation prediction and text segmentation on the text.
4. The intelligent outbound method of claim 2 wherein said step of converting said outbound voice to corresponding text prior to said step of determining a start location and an end location of a pronunciation based on an audio stream of said outbound voice further comprises:
and carrying out noise reduction and reverberation elimination processing on the outbound voice.
5. The intelligent callout method of claim 1, wherein said step of selecting a corresponding reply content based on said text comprises:
matching a corresponding intention scene for the text according to a preset grammar rule;
and converting the text into corresponding structured data according to the intention scene, and determining reply content according to the structured data.
6. The intelligent callout method of claim 1, wherein said step of selecting a corresponding reply content based on said text comprises:
extracting text content according to a preset fixed slot position;
and converting the text into corresponding structured data according to the text content of the fixed slot, and determining reply content according to the structured data.
7. The intelligent outbound method of claim 1 wherein said step of generating a corresponding reply voice based on said reply content comprises:
converting the reply content into a corresponding rhyme sequence;
and generating corresponding reply voice according to the rhyme sequence.
8. A terminal device, comprising:
the input unit is used for acquiring the outbound voice;
the text conversion unit is used for converting the outbound voice into a corresponding text;
the text understanding unit is used for selecting corresponding reply content according to the text;
and the voice playing unit is used for generating corresponding reply voice according to the reply content.
9. A terminal device comprising a memory, a processor and a computer program stored in the memory and executable on the processor, characterized in that the processor implements the steps of the method according to any of claims 1 to 7 when executing the computer program.
10. A computer-readable storage medium, in which a computer program is stored which, when being executed by a processor, carries out the steps of the method according to any one of claims 1 to 7.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010974415.4A CN112399019A (en) | 2020-09-16 | 2020-09-16 | Intelligent outbound call method, terminal equipment and readable storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010974415.4A CN112399019A (en) | 2020-09-16 | 2020-09-16 | Intelligent outbound call method, terminal equipment and readable storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
CN112399019A true CN112399019A (en) | 2021-02-23 |
Family
ID=74595814
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010974415.4A Pending CN112399019A (en) | 2020-09-16 | 2020-09-16 | Intelligent outbound call method, terminal equipment and readable storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112399019A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114679515A (en) * | 2022-05-30 | 2022-06-28 | 杭州一知智能科技有限公司 | Method, device, equipment and storage medium for judging connection time point of outbound system |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102369568A (en) * | 2009-02-03 | 2012-03-07 | 索夫特赫斯公司 | Systems and methods for interactively accessing hosted services using voice communications |
CN105427858A (en) * | 2015-11-06 | 2016-03-23 | 科大讯飞股份有限公司 | Method and system for achieving automatic voice classification |
CN106941619A (en) * | 2017-03-16 | 2017-07-11 | 百度在线网络技术(北京)有限公司 | Program prompting method, device and system based on artificial intelligence |
CN107562760A (en) * | 2016-06-30 | 2018-01-09 | 科大讯飞股份有限公司 | A kind of voice data processing method and device |
CN107644012A (en) * | 2017-08-29 | 2018-01-30 | 平安科技(深圳)有限公司 | Electronic installation, problem identification confirmation method and computer-readable recording medium |
CN108170859A (en) * | 2018-01-22 | 2018-06-15 | 北京百度网讯科技有限公司 | Method, apparatus, storage medium and the terminal device of speech polling |
CN108170792A (en) * | 2017-12-27 | 2018-06-15 | 北京百度网讯科技有限公司 | Question and answer bootstrap technique, device and computer equipment based on artificial intelligence |
CN108418744A (en) * | 2018-02-06 | 2018-08-17 | 国网浙江省电力有限公司信息通信分公司 | A kind of electronics seat system for promoting electrical power services quality |
CN110909137A (en) * | 2019-10-12 | 2020-03-24 | 平安科技(深圳)有限公司 | Information pushing method and device based on man-machine interaction and computer equipment |
CN111026842A (en) * | 2019-11-29 | 2020-04-17 | 微民保险代理有限公司 | Natural language processing method, natural language processing device and intelligent question-answering system |
-
2020
- 2020-09-16 CN CN202010974415.4A patent/CN112399019A/en active Pending
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102369568A (en) * | 2009-02-03 | 2012-03-07 | 索夫特赫斯公司 | Systems and methods for interactively accessing hosted services using voice communications |
CN105427858A (en) * | 2015-11-06 | 2016-03-23 | 科大讯飞股份有限公司 | Method and system for achieving automatic voice classification |
CN107562760A (en) * | 2016-06-30 | 2018-01-09 | 科大讯飞股份有限公司 | A kind of voice data processing method and device |
CN106941619A (en) * | 2017-03-16 | 2017-07-11 | 百度在线网络技术(北京)有限公司 | Program prompting method, device and system based on artificial intelligence |
CN107644012A (en) * | 2017-08-29 | 2018-01-30 | 平安科技(深圳)有限公司 | Electronic installation, problem identification confirmation method and computer-readable recording medium |
CN108170792A (en) * | 2017-12-27 | 2018-06-15 | 北京百度网讯科技有限公司 | Question and answer bootstrap technique, device and computer equipment based on artificial intelligence |
CN108170859A (en) * | 2018-01-22 | 2018-06-15 | 北京百度网讯科技有限公司 | Method, apparatus, storage medium and the terminal device of speech polling |
CN108418744A (en) * | 2018-02-06 | 2018-08-17 | 国网浙江省电力有限公司信息通信分公司 | A kind of electronics seat system for promoting electrical power services quality |
CN110909137A (en) * | 2019-10-12 | 2020-03-24 | 平安科技(深圳)有限公司 | Information pushing method and device based on man-machine interaction and computer equipment |
CN111026842A (en) * | 2019-11-29 | 2020-04-17 | 微民保险代理有限公司 | Natural language processing method, natural language processing device and intelligent question-answering system |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114679515A (en) * | 2022-05-30 | 2022-06-28 | 杭州一知智能科技有限公司 | Method, device, equipment and storage medium for judging connection time point of outbound system |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110136744B (en) | Audio fingerprint generation method, equipment and storage medium | |
CN107481717B (en) | Acoustic model training method and system | |
CN108615525B (en) | Voice recognition method and device | |
TWI711967B (en) | Method, device and equipment for determining broadcast voice | |
CN108492818B (en) | Text-to-speech conversion method and device and computer equipment | |
CN107705782B (en) | Method and device for determining phoneme pronunciation duration | |
CN110415701A (en) | The recognition methods of lip reading and its device | |
WO2023020262A1 (en) | Integrating dialog history into end-to-end spoken language understanding systems | |
US11562735B1 (en) | Multi-modal spoken language understanding systems | |
CN112397051B (en) | Voice recognition method and device and terminal equipment | |
CN112466314A (en) | Emotion voice data conversion method and device, computer equipment and storage medium | |
CN113178201A (en) | Unsupervised voice conversion method, unsupervised voice conversion device, unsupervised voice conversion equipment and unsupervised voice conversion medium | |
CN111508466A (en) | Text processing method, device and equipment and computer readable storage medium | |
CN111696521A (en) | Method for training speech clone model, readable storage medium and speech clone method | |
CN112037776A (en) | Voice recognition method, voice recognition device and terminal equipment | |
CN114495977B (en) | Speech translation and model training method, device, electronic equipment and storage medium | |
JP7329393B2 (en) | Audio signal processing device, audio signal processing method, audio signal processing program, learning device, learning method and learning program | |
CN113327576B (en) | Speech synthesis method, device, equipment and storage medium | |
CN109147146B (en) | Voice number taking method and terminal equipment | |
CN112399019A (en) | Intelligent outbound call method, terminal equipment and readable storage medium | |
CN113327575A (en) | Speech synthesis method, device, computer equipment and storage medium | |
US20230206895A1 (en) | Speech synthesis method, device and computer-readable storage medium | |
CN113808572B (en) | Speech synthesis method, speech synthesis device, electronic equipment and storage medium | |
CN112908339B (en) | Conference link positioning method and device, positioning equipment and readable storage medium | |
CN114783428A (en) | Voice translation method, voice translation device, voice translation model training method, voice translation model training device, voice translation equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20210223 |