CN112399019A - Intelligent outbound call method, terminal equipment and readable storage medium - Google Patents

Intelligent outbound call method, terminal equipment and readable storage medium Download PDF

Info

Publication number
CN112399019A
CN112399019A CN202010974415.4A CN202010974415A CN112399019A CN 112399019 A CN112399019 A CN 112399019A CN 202010974415 A CN202010974415 A CN 202010974415A CN 112399019 A CN112399019 A CN 112399019A
Authority
CN
China
Prior art keywords
text
voice
outbound
reply content
converting
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010974415.4A
Other languages
Chinese (zh)
Inventor
胡宁
马晓萌
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Agricultural Bank of China Hebei Branch
Original Assignee
Agricultural Bank of China Hebei Branch
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Agricultural Bank of China Hebei Branch filed Critical Agricultural Bank of China Hebei Branch
Priority to CN202010974415.4A priority Critical patent/CN112399019A/en
Publication of CN112399019A publication Critical patent/CN112399019A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/487Arrangements for providing information services, e.g. recorded voice services or time announcements
    • H04M3/493Interactive information services, e.g. directory enquiries ; Arrangements therefor, e.g. interactive voice response [IVR] systems or voice portals
    • H04M3/4936Speech interaction details
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/18Speech classification or search using natural language modelling
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/20Speech recognition techniques specially adapted for robustness in adverse environments, e.g. in noise, of stress induced speech
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/50Centralised arrangements for answering calls; Centralised arrangements for recording messages for absent or busy subscribers ; Centralised arrangements for recording messages
    • H04M3/527Centralised call answering arrangements not requiring operator intervention

Landscapes

  • Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Artificial Intelligence (AREA)
  • Telephonic Communication Services (AREA)

Abstract

The present application belongs to the field of communications technologies, and in particular, to an intelligent outbound method, a terminal device, and a readable storage medium. Wherein, the method comprises the following steps: acquiring an outbound voice; converting the outbound voice into a corresponding text; selecting corresponding reply content according to the text; and generating a corresponding reply voice according to the reply content. The intelligent outbound method, the terminal device and the readable storage medium provided by the embodiment of the application realize the intelligent selection of the reply content by the computer through the text conversion and the text semantic understanding of the outbound voice sent by the client. In order to be more suitable for telephone communication and provide telephone service for clients, after the reply content is selected, the reply content is further converted into voice, so that the telephone service for the clients can be completed instead of manual customer service, and the labor cost of service industries such as communication, finance and the like is favorably reduced.

Description

Intelligent outbound call method, terminal equipment and readable storage medium
Technical Field
The present application belongs to the field of communications technologies, and in particular, to an intelligent outbound method, a terminal device, and a readable storage medium.
Background
Modern service industries, such as the communications service industry and the financial service industry, often require telephone service transactions to be provided to customers. Traditional telephone business handling needs enterprises to configure a large amount of manual customer service, answers incoming calls of clients and handles business, and labor cost is high.
Disclosure of Invention
In view of this, embodiments of the present application provide an intelligent outbound method, a terminal device, and a readable storage medium, so as to solve the problem that the labor cost of the current service industry is high.
According to a first aspect, an embodiment of the present application provides an intelligent outbound method, including: acquiring an outbound voice; converting the outbound voice into a corresponding text; selecting corresponding reply content according to the text; and generating corresponding reply voice according to the reply content.
With reference to the first aspect, in some embodiments of the present application, the step of converting the outbound voice into corresponding text includes: determining the initial position and the end position of pronunciation according to the audio stream of the outbound voice; extracting target audio data according to the starting position and the ending position; and decoding the target audio data to generate a corresponding text.
With reference to the first aspect, in some embodiments of the present application, after the step of decoding the target audio data to generate a corresponding text, the step of converting the outbound voice into the corresponding text further includes: and performing text smoothness, punctuation prediction and text segmentation on the text.
With reference to the first aspect, in some embodiments of the present application, before the step of determining a start position and an end position of a pronunciation according to an audio stream of the outbound voice, the step of converting the outbound voice into a corresponding text further includes: and carrying out noise reduction and reverberation elimination processing on the outbound voice.
With reference to the first aspect, in some embodiments of the present application, the step of selecting a corresponding reply content according to the text includes: matching a corresponding intention scene for the text according to a preset grammar rule; and converting the text into corresponding structured data according to the intention scene, and determining reply content according to the structured data.
With reference to the first aspect, in some embodiments of the present application, the step of selecting a corresponding reply content according to the text includes: extracting text content according to a preset fixed slot position; and converting the text into corresponding structured data according to the text content of the fixed slot, and determining reply content according to the structured data.
With reference to the first aspect, in some embodiments of the present application, the step of generating a corresponding reply voice according to the reply content includes: converting the reply content into a corresponding rhyme sequence; and generating corresponding reply voice according to the rhyme sequence.
According to a second aspect, an embodiment of the present application provides a terminal device, including: the input unit is used for acquiring the outbound voice; the text conversion unit is used for converting the outbound voice into a corresponding text; the text understanding unit is used for selecting corresponding reply content according to the text; and the voice playing unit is used for generating corresponding reply voice according to the reply content.
According to a third aspect, an embodiment of the present application provides a terminal device, which includes a memory, a processor, and a computer program stored in the memory and executable on the processor, where the processor implements the steps of the method according to the first aspect or any embodiment of the first aspect when executing the computer program.
According to a fourth aspect, embodiments of the present application provide a computer-readable storage medium, which stores a computer program, and the computer program, when executed by a processor, implements the steps of the method according to the first aspect or any embodiment of the first aspect.
The intelligent outbound method, the terminal device and the readable storage medium provided by the embodiment of the application realize the intelligent selection of the reply content by the computer through the text conversion and the text semantic understanding of the outbound voice sent by the client. In order to be more suitable for telephone communication and provide telephone service for clients, after the reply content is selected, the reply content is further converted into voice, so that the telephone service for the clients can be completed instead of manual customer service, and the labor cost of service industries such as communication, finance and the like is favorably reduced.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, and it is obvious that the drawings in the following description are some embodiments of the present invention, and other drawings can be obtained by those skilled in the art without creative efforts.
Fig. 1 is a flowchart of a specific example of an intelligent outbound method provided by an embodiment of the present application;
fig. 2 is a flowchart of another specific example of an intelligent outbound method provided by an embodiment of the present application;
fig. 3 is a schematic structural diagram of a specific example of a terminal device provided in an embodiment of the present application;
fig. 4 is a schematic structural diagram of another specific example of the terminal device provided in the embodiment of the present application.
Detailed Description
In the following description, for purposes of explanation and not limitation, specific details are set forth, such as particular system structures, techniques, etc. in order to provide a thorough understanding of the embodiments of the present application. It will be apparent, however, to one skilled in the art that the present application may be practiced in other embodiments that depart from these specific details. In other instances, detailed descriptions of well-known systems, devices, circuits, and methods are omitted so as not to obscure the description of the present application with unnecessary detail.
In order to explain the technical solution described in the present application, the following description will be given by way of specific examples.
The embodiment of the application provides an intelligent outbound method, as shown in fig. 1, the method may include the following steps:
step S101: and acquiring the outbound voice.
When a client transacts communication or financial services such as price package change, credit card payment by stages and the like, the client often selects telephone transaction, namely telephone service transaction is performed by dialing the telephone of the client of an enterprise. The service handling requirement proposed by the client after the call is connected is the outbound voice mentioned in the embodiment of the application.
Step S102: the outbound voice is converted to corresponding text.
In a specific embodiment, as shown in fig. 2, the process of step S102 can be implemented by the following several sub-steps:
step S1021: based on the audio stream of the outbound speech, the start and end positions of the utterance are determined.
Step S1022: target audio data is extracted based on the start position and the end position.
Step S1023: and decoding the target audio data to generate a corresponding text.
In practical applications, in order to improve the recognition effect of the audio, the following sub-steps may be added before step S1021:
step S1020: and carrying out noise reduction and reverberation elimination processing on the outbound voice.
After the noise reduction and reverberation elimination processing is performed on the input voice, the endpoint detection can be performed on the input audio stream, the starting position and the ending position of the speaking are determined, and the identification processing is synchronously performed. In step S1023, in order to improve the user response speed and ensure the recognition effect, two-pass decoding may be adopted in the recognition process. Specifically, the target audio data can be decoded once by using acoustic fusion models such as DFCNN and BilSTM and an nGram language model; and then, carrying out two-time decoding processing on the data obtained by the one-time decoding by utilizing the domain language model and the RNN language model.
In order to improve the readability of the text, post-processing such as punctuation prediction, text smoothness, text segmentation and the like can be performed on the data obtained after the two-pass decoding, and finally the text is generated. As shown in fig. 2, step S1024 may be added after step S1023 to perform text smoothing, punctuation prediction and text segmentation on the text.
Step S103: and selecting corresponding reply content according to the text.
After obtaining the text corresponding to the outbound voice, semantic understanding of the text is required. Semantic understanding (NLP) refers to converting a natural language into computer-readable structured data, and can achieve the purpose of semantic understanding by matching corresponding intention scenes through grammar rules or capturing text contents of fixed slots.
Specifically, the process of step S103 may be implemented by the following several sub-steps:
step S1031: and matching the corresponding intention scene for the text according to a preset grammar rule.
Step S1032: and converting the text into corresponding structured data according to the intention scene, and determining reply content according to the structured data.
In a specific embodiment, the following substeps may also be used to implement the process of step S103, in place of step S1031 and step S1032:
step S1031': and extracting text content according to the preset fixed slot position.
Step S1032': and converting the text into corresponding structured data according to the text content of the fixed slot position, and determining reply content according to the structured data.
Step S104: and generating a corresponding reply voice according to the reply content.
In a specific embodiment, as shown in fig. 2, the process of step S104 can be implemented by the following several sub-steps:
step S1041: and converting the reply content into a corresponding rhyme sequence.
Step S1042: and generating corresponding reply voice according to the rhyme sequence.
In step S104, the text is converted into sound, and the sound corresponding to the anchor is synthesized according to the selected anchor, and the specific principle is as follows:
speech synthesis can be regarded as an artificial intelligence system. In order to synthesize a high-quality language, besides relying on various rules, including semantic rules, lexical rules, and phonetic rules, the content of the text must be well understood, which also relates to the understanding of natural language. The text-to-speech conversion process is to convert the text sequence into a rhyme sequence and then generate a speech waveform according to the rhyme sequence. Wherein, step S1041 involves linguistic processing, such as word segmentation, word-to-speech conversion, etc., and a set of valid prosody control rules; step S1042 requires advanced speech synthesis technology, and can synthesize a high-quality speech stream in real time as required. Generally speaking, a text-to-speech conversion system requires a complex conversion procedure from text sequences to phoneme sequences, that is, the text-to-speech conversion system not only needs digital signal processing technology, but also needs to be supported by a great deal of linguistic knowledge.
According to the intelligent outbound method provided by the embodiment of the application, the text conversion and the text semantic understanding are carried out on the outbound voice sent by the client, so that the reply content can be intelligently selected by the computer. In order to be more suitable for telephone communication and provide telephone service for clients, after the reply content is selected, the reply content is further converted into voice, so that the telephone service for the clients can be completed instead of manual customer service, and the labor cost of service industries such as communication, finance and the like is favorably reduced.
It should be understood that, the sequence numbers of the steps in the foregoing embodiments do not imply an execution sequence, and the execution sequence of each process should be determined by its function and inherent logic, and should not constitute any limitation to the implementation process of the embodiments of the present application.
An embodiment of the present application further provides a terminal device, as shown in fig. 3, where the terminal device may include: an input unit 301, a text conversion unit 302, a text understanding unit 303, and a voice playing unit 304.
Specifically, the input unit 301 is configured to obtain an outbound voice; the corresponding implementation process can be referred to the record of step S101 in the above method embodiment.
The text conversion unit 302 is configured to convert the outbound voice into a corresponding text; the corresponding implementation process can be referred to the description of step S102 in the above method embodiment.
The text understanding unit 303 is configured to select corresponding reply content according to the text; the corresponding implementation process can be referred to the description of step S103 in the above method embodiment.
The voice playing unit 304 is configured to generate a corresponding reply voice according to the reply content; the corresponding implementation process can be referred to the record of step S104 in the above method embodiment.
Fig. 4 is a schematic diagram of another terminal device provided in an embodiment of the present application. As shown in fig. 4, the terminal device 400 of this embodiment includes: a processor 401, a memory 402 and a computer program 403, such as an intelligent call-out program, stored in said memory 402 and executable on said processor 401. The processor 401, when executing the computer program 403, implements the steps in the above-described various embodiments of the intelligent outbound method, such as the steps S101 to S104 shown in fig. 1. Alternatively, the processor 401, when executing the computer program 403, implements the functions of the modules/units in the device embodiments described above, such as the functions of the input unit 301, the text conversion unit 302, the text understanding unit 303, and the speech playing unit 304 shown in fig. 3.
The computer program 403 may be partitioned into one or more modules/units that are stored in the memory 402 and executed by the processor 401 to accomplish the present application. The one or more modules/units may be a series of computer program instruction segments capable of performing specific functions, which are used for describing the execution process of the computer program 403 in the terminal device 400. For example, the computer program 403 may be partitioned into a synchronization module, a summarization module, an acquisition module, a return module (a module in a virtual device).
The terminal device 400 may be a desktop computer, a notebook, a palm computer, a cloud server, or other computing devices. The terminal device may include, but is not limited to, a processor 401, a memory 402. Those skilled in the art will appreciate that fig. 4 is merely an example of a terminal device 400 and does not constitute a limitation of terminal device 400 and may include more or fewer components than shown, or some components may be combined, or different components, e.g., the terminal device may also include input-output devices, network access devices, buses, etc.
The Processor 401 may be a Central Processing Unit (CPU), other general purpose Processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), an off-the-shelf Programmable Gate Array (FPGA) or other Programmable logic device, discrete Gate or transistor logic, discrete hardware components, etc. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
The storage 402 may be an internal storage unit of the terminal device 400, such as a hard disk or a memory of the terminal device 400. The memory 402 may also be an external storage device of the terminal device 400, such as a plug-in hard disk, a Smart Media Card (SMC), a Secure Digital (SD) Card, a Flash memory Card (Flash Card), and the like, which are provided on the terminal device 400. Further, the memory 402 may also include both an internal storage unit and an external storage device of the terminal device 400. The memory 402 is used for storing the computer programs and other programs and data required by the terminal device. The memory 402 may also be used to temporarily store data that has been output or is to be output.
It will be apparent to those skilled in the art that, for convenience and brevity of description, only the above-mentioned division of the functional units and modules is illustrated, and in practical applications, the above-mentioned function distribution may be performed by different functional units and modules according to needs, that is, the internal structure of the apparatus is divided into different functional units or modules to perform all or part of the above-mentioned functions. Each functional unit and module in the embodiments may be integrated in one processing unit, or each unit may exist alone physically, or two or more units are integrated in one unit, and the integrated unit may be implemented in a form of hardware, or in a form of software functional unit. In addition, specific names of the functional units and modules are only for convenience of distinguishing from each other, and are not used for limiting the protection scope of the present application. The specific working processes of the units and modules in the system may refer to the corresponding processes in the foregoing method embodiments, and are not described herein again.
In the above embodiments, the descriptions of the respective embodiments have respective emphasis, and reference may be made to the related descriptions of other embodiments for parts that are not described or illustrated in a certain embodiment.
Those of ordinary skill in the art will appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware or combinations of computer software and electronic hardware. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the implementation. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present application.
In the embodiments provided in the present application, it should be understood that the disclosed apparatus/terminal device and method may be implemented in other ways. For example, the above-described embodiments of the apparatus/terminal device are merely illustrative, and for example, the division of the modules or units is only one logical division, and there may be other divisions when actually implemented, for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection through some interfaces, devices or units, and may be in an electrical, mechanical or other form.
The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In addition, functional units in the embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit.
The integrated modules/units, if implemented in the form of software functional units and sold or used as separate products, may be stored in a computer readable storage medium. Based on such understanding, all or part of the flow in the method of the embodiments described above can be realized by a computer program, which can be stored in a computer-readable storage medium and can realize the steps of the embodiments of the methods described above when the computer program is executed by a processor. . Wherein the computer program comprises computer program code, which may be in the form of source code, object code, an executable file or some intermediate form, etc. The computer-readable medium may include: any entity or device capable of carrying the computer program code, recording medium, usb disk, removable hard disk, magnetic disk, optical disk, computer Memory, Read-Only Memory (ROM), Random Access Memory (RAM), electrical carrier wave signals, telecommunications signals, software distribution medium, and the like. It should be noted that the computer readable medium may contain content that is subject to appropriate increase or decrease as required by legislation and patent practice in jurisdictions, for example, in some jurisdictions, computer readable media does not include electrical carrier signals and telecommunications signals as is required by legislation and patent practice.
The above-mentioned embodiments are only used for illustrating the technical solutions of the present application, and not for limiting the same; although the present application has been described in detail with reference to the foregoing embodiments, it should be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; such modifications and substitutions do not substantially depart from the spirit and scope of the embodiments of the present application and are intended to be included within the scope of the present application.

Claims (10)

1. An intelligent outbound method, comprising:
acquiring an outbound voice;
converting the outbound voice into a corresponding text;
selecting corresponding reply content according to the text;
and generating corresponding reply voice according to the reply content.
2. The intelligent outbound method of claim 1 wherein said step of converting said outbound voice to corresponding text comprises:
determining the initial position and the end position of pronunciation according to the audio stream of the outbound voice;
extracting target audio data according to the starting position and the ending position;
and decoding the target audio data to generate a corresponding text.
3. The intelligent outbound method of claim 2 wherein said step of converting said outbound speech to corresponding text after said step of decoding said target audio data to generate corresponding text further comprises:
and performing text smoothness, punctuation prediction and text segmentation on the text.
4. The intelligent outbound method of claim 2 wherein said step of converting said outbound voice to corresponding text prior to said step of determining a start location and an end location of a pronunciation based on an audio stream of said outbound voice further comprises:
and carrying out noise reduction and reverberation elimination processing on the outbound voice.
5. The intelligent callout method of claim 1, wherein said step of selecting a corresponding reply content based on said text comprises:
matching a corresponding intention scene for the text according to a preset grammar rule;
and converting the text into corresponding structured data according to the intention scene, and determining reply content according to the structured data.
6. The intelligent callout method of claim 1, wherein said step of selecting a corresponding reply content based on said text comprises:
extracting text content according to a preset fixed slot position;
and converting the text into corresponding structured data according to the text content of the fixed slot, and determining reply content according to the structured data.
7. The intelligent outbound method of claim 1 wherein said step of generating a corresponding reply voice based on said reply content comprises:
converting the reply content into a corresponding rhyme sequence;
and generating corresponding reply voice according to the rhyme sequence.
8. A terminal device, comprising:
the input unit is used for acquiring the outbound voice;
the text conversion unit is used for converting the outbound voice into a corresponding text;
the text understanding unit is used for selecting corresponding reply content according to the text;
and the voice playing unit is used for generating corresponding reply voice according to the reply content.
9. A terminal device comprising a memory, a processor and a computer program stored in the memory and executable on the processor, characterized in that the processor implements the steps of the method according to any of claims 1 to 7 when executing the computer program.
10. A computer-readable storage medium, in which a computer program is stored which, when being executed by a processor, carries out the steps of the method according to any one of claims 1 to 7.
CN202010974415.4A 2020-09-16 2020-09-16 Intelligent outbound call method, terminal equipment and readable storage medium Pending CN112399019A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010974415.4A CN112399019A (en) 2020-09-16 2020-09-16 Intelligent outbound call method, terminal equipment and readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010974415.4A CN112399019A (en) 2020-09-16 2020-09-16 Intelligent outbound call method, terminal equipment and readable storage medium

Publications (1)

Publication Number Publication Date
CN112399019A true CN112399019A (en) 2021-02-23

Family

ID=74595814

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010974415.4A Pending CN112399019A (en) 2020-09-16 2020-09-16 Intelligent outbound call method, terminal equipment and readable storage medium

Country Status (1)

Country Link
CN (1) CN112399019A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114679515A (en) * 2022-05-30 2022-06-28 杭州一知智能科技有限公司 Method, device, equipment and storage medium for judging connection time point of outbound system

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102369568A (en) * 2009-02-03 2012-03-07 索夫特赫斯公司 Systems and methods for interactively accessing hosted services using voice communications
CN105427858A (en) * 2015-11-06 2016-03-23 科大讯飞股份有限公司 Method and system for achieving automatic voice classification
CN106941619A (en) * 2017-03-16 2017-07-11 百度在线网络技术(北京)有限公司 Program prompting method, device and system based on artificial intelligence
CN107562760A (en) * 2016-06-30 2018-01-09 科大讯飞股份有限公司 A kind of voice data processing method and device
CN107644012A (en) * 2017-08-29 2018-01-30 平安科技(深圳)有限公司 Electronic installation, problem identification confirmation method and computer-readable recording medium
CN108170859A (en) * 2018-01-22 2018-06-15 北京百度网讯科技有限公司 Method, apparatus, storage medium and the terminal device of speech polling
CN108170792A (en) * 2017-12-27 2018-06-15 北京百度网讯科技有限公司 Question and answer bootstrap technique, device and computer equipment based on artificial intelligence
CN108418744A (en) * 2018-02-06 2018-08-17 国网浙江省电力有限公司信息通信分公司 A kind of electronics seat system for promoting electrical power services quality
CN110909137A (en) * 2019-10-12 2020-03-24 平安科技(深圳)有限公司 Information pushing method and device based on man-machine interaction and computer equipment
CN111026842A (en) * 2019-11-29 2020-04-17 微民保险代理有限公司 Natural language processing method, natural language processing device and intelligent question-answering system

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102369568A (en) * 2009-02-03 2012-03-07 索夫特赫斯公司 Systems and methods for interactively accessing hosted services using voice communications
CN105427858A (en) * 2015-11-06 2016-03-23 科大讯飞股份有限公司 Method and system for achieving automatic voice classification
CN107562760A (en) * 2016-06-30 2018-01-09 科大讯飞股份有限公司 A kind of voice data processing method and device
CN106941619A (en) * 2017-03-16 2017-07-11 百度在线网络技术(北京)有限公司 Program prompting method, device and system based on artificial intelligence
CN107644012A (en) * 2017-08-29 2018-01-30 平安科技(深圳)有限公司 Electronic installation, problem identification confirmation method and computer-readable recording medium
CN108170792A (en) * 2017-12-27 2018-06-15 北京百度网讯科技有限公司 Question and answer bootstrap technique, device and computer equipment based on artificial intelligence
CN108170859A (en) * 2018-01-22 2018-06-15 北京百度网讯科技有限公司 Method, apparatus, storage medium and the terminal device of speech polling
CN108418744A (en) * 2018-02-06 2018-08-17 国网浙江省电力有限公司信息通信分公司 A kind of electronics seat system for promoting electrical power services quality
CN110909137A (en) * 2019-10-12 2020-03-24 平安科技(深圳)有限公司 Information pushing method and device based on man-machine interaction and computer equipment
CN111026842A (en) * 2019-11-29 2020-04-17 微民保险代理有限公司 Natural language processing method, natural language processing device and intelligent question-answering system

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114679515A (en) * 2022-05-30 2022-06-28 杭州一知智能科技有限公司 Method, device, equipment and storage medium for judging connection time point of outbound system

Similar Documents

Publication Publication Date Title
CN110136744B (en) Audio fingerprint generation method, equipment and storage medium
CN107481717B (en) Acoustic model training method and system
CN108615525B (en) Voice recognition method and device
TWI711967B (en) Method, device and equipment for determining broadcast voice
CN108492818B (en) Text-to-speech conversion method and device and computer equipment
CN107705782B (en) Method and device for determining phoneme pronunciation duration
CN110415701A (en) The recognition methods of lip reading and its device
WO2023020262A1 (en) Integrating dialog history into end-to-end spoken language understanding systems
US11562735B1 (en) Multi-modal spoken language understanding systems
CN112397051B (en) Voice recognition method and device and terminal equipment
CN112466314A (en) Emotion voice data conversion method and device, computer equipment and storage medium
CN113178201A (en) Unsupervised voice conversion method, unsupervised voice conversion device, unsupervised voice conversion equipment and unsupervised voice conversion medium
CN111508466A (en) Text processing method, device and equipment and computer readable storage medium
CN111696521A (en) Method for training speech clone model, readable storage medium and speech clone method
CN112037776A (en) Voice recognition method, voice recognition device and terminal equipment
CN114495977B (en) Speech translation and model training method, device, electronic equipment and storage medium
JP7329393B2 (en) Audio signal processing device, audio signal processing method, audio signal processing program, learning device, learning method and learning program
CN113327576B (en) Speech synthesis method, device, equipment and storage medium
CN109147146B (en) Voice number taking method and terminal equipment
CN112399019A (en) Intelligent outbound call method, terminal equipment and readable storage medium
CN113327575A (en) Speech synthesis method, device, computer equipment and storage medium
US20230206895A1 (en) Speech synthesis method, device and computer-readable storage medium
CN113808572B (en) Speech synthesis method, speech synthesis device, electronic equipment and storage medium
CN112908339B (en) Conference link positioning method and device, positioning equipment and readable storage medium
CN114783428A (en) Voice translation method, voice translation device, voice translation model training method, voice translation model training device, voice translation equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20210223