CN112967717B - Fuzzy matching training method for English speech translation with high accuracy - Google Patents

Fuzzy matching training method for English speech translation with high accuracy Download PDF

Info

Publication number
CN112967717B
CN112967717B CN202110223114.2A CN202110223114A CN112967717B CN 112967717 B CN112967717 B CN 112967717B CN 202110223114 A CN202110223114 A CN 202110223114A CN 112967717 B CN112967717 B CN 112967717B
Authority
CN
China
Prior art keywords
voice
programmable device
fuzzy
matching
online
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110223114.2A
Other languages
Chinese (zh)
Other versions
CN112967717A (en
Inventor
王晓靖
张敏
李琦
丁桂芝
牛明敏
张晨曦
郭晓斌
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhengzhou Railway Vocational and Technical College
Original Assignee
Zhengzhou Railway Vocational and Technical College
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhengzhou Railway Vocational and Technical College filed Critical Zhengzhou Railway Vocational and Technical College
Priority to CN202110223114.2A priority Critical patent/CN112967717B/en
Publication of CN112967717A publication Critical patent/CN112967717A/en
Application granted granted Critical
Publication of CN112967717B publication Critical patent/CN112967717B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/40Processing or translation of natural language
    • G06F40/58Use of machine translation, e.g. for multi-lingual retrieval, for server-side translation for client devices or for real-time translation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/12Computing arrangements based on biological models using genetic models
    • G06N3/126Evolutionary algorithms, e.g. genetic algorithms or genetic programming
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/28Constructional details of speech recognition systems
    • G10L15/30Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
    • G10L25/24Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters the extracted parameters being the cepstrum
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/27Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique
    • G10L25/33Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique using fuzzy logic
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/225Feedback of the input speech
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Abstract

The invention discloses a fuzzy matching method for English speech translation with high accuracy, which comprises the following steps: s1, acquiring voice information and converting the voice information into a digital voice signal; s2, carrying out voice detection on the digital voice signal by adopting a fuzzy sound image rule to obtain corresponding matching parameters, and obtaining a reconfiguration data stream of the programmable device by adopting genetic algorithm operation according to the matching parameters; s3, reconfiguring the programmable device by adopting the reconfiguration data stream; s4, the online voice receiving module can search online voice data by taking the matching parameters as guidance, and roughly sort the online voice data; inputting the retrieved online voice data into a programmable device for fuzzy matching; s5, finding out the most matched online voice data output.

Description

Fuzzy matching training method for English speech translation with high accuracy
Technical Field
The invention relates to the field of English translation, in particular to a fuzzy matching method for English speech translation with high accuracy.
Background
Along with the development of internet technology, the intellectualization of English translation is also continuously advancing, such as a computer-aided translation system, text translation and voice translation are realized one by one, and particularly, on-line voice translation can be realized by handheld equipment such as a mobile phone APP.
In the prior art, the voice is generally converted into the text, then the similarity matching is carried out on the text information and sentences in a translation library by adopting an algorithm, and the result with the highest similarity is used as output. As disclosed in CN201710532235.9, an information retrieval technology is adopted to construct an index for a large-scale translation memory, rough selection and carefully selecting strategies are adopted, namely, firstly, a matched subset is obtained from the index library according to input sentences to be translated, then, final translation output is obtained by using a fuzzy matching method of linear combination of semantic vector similarity and editing distance of sentences, and finally, manually edited translations and source language sentence segments thereof are returned to the translation memory for incremental updating.
The translation efficiency is not high, the text recognition rate is also related to whether English pronunciation is accurate, text information needs to be checked again after text matching to obtain a translation result, meaning is understood again according to the text information, and communication efficiency is affected.
Disclosure of Invention
The invention provides the following technical scheme for solving the problems in the background technology: a fuzzy matching method for English speech translation with high accuracy comprises the following specific steps:
s1, acquiring voice information and converting the voice information into a digital voice signal;
s2, carrying out voice detection on the digital voice signal by adopting a fuzzy sound image rule to obtain corresponding matching parameters, and obtaining a reconfiguration data stream of the programmable device by adopting genetic algorithm operation according to the matching parameters;
s3, reconfiguring the programmable device by adopting the reconfiguration data stream;
s4, the online voice receiving module can search online voice data by taking the matching parameters as guidance, and roughly sort the online voice data; inputting the retrieved online voice data into a programmable device for fuzzy matching;
s5, finding out the most matched online voice data output.
The fuzzy matching algorithm in the step S4 is as follows: and taking a plurality of matching parameters as a fuzzy rule R, taking digitized voice as an input quantity X, and determining an output U by the output of a plurality of fuzzy rules when the input quantity X activates the plurality of fuzzy rules R.
The genetic algorithm in the step S2 is specifically as follows:
s21, randomly generating chromosome individuals;
s22, calculating the fitness value of the individual;
s23, randomly performing mutation operation on the individuals to generate offspring individuals;
and S24, executing selection operation, if the fitness value of the child individual is higher than that of the individual, copying the child individual to the next generation, otherwise copying the child individual to the next generation with a small probability, and the like until the termination condition is met.
The fuzzy sound image rule in the step S2 includes a description of a sound image feature implemented by a rule formed by a mel cepstrum coefficient analysis method, a short-time energy and short-time average zero-crossing rate statistical method, a formant extraction method obtained based on spectrum analysis and other methods, and outputs of the rule, such as the mel cepstrum coefficient, the short-time energy, the short-time average zero-crossing rate, the formant and the like, are used as matching parameters.
A configurable logic module (LUT) in a programmable device is used as a chromosome, an optimal chromosome is found out through a genetic algorithm, and a reconfiguration data stream is generated and configured on the programmable device, so that a circuit is reconfigured to realize that the circuit is changed according to the change of input voice so as to more effectively match voice data.
The invention also provides a translation fuzzy matching system of English voice, which can translate more accurately, wherein one technical scheme is as follows: a translation fuzzy matching system of English speech comprises a processor, a memory, a programmable device, an online speech receiving module, a speech acquisition module and a display playing module;
the processor and the programmable device are respectively connected with the voice acquisition module; the processor is electrically connected with the memory, the programmable device, the online voice receiving module, the voice acquisition module and the display playing module respectively;
the online voice receiving module and the memory are connected with the programmable device;
the voice acquisition module acquires voice information and converts the voice information into a digital voice signal.
The processor receives the digital voice signal, carries out voice detection on the digital voice signal by adopting a fuzzy sound image rule to obtain corresponding matching parameters, adopts genetic algorithm operation to obtain reconfiguration data flow of the programmable device according to the matching parameters, stores the reconfiguration data flow into the memory, and controls the memory and the programmable device to reconfigure the programmable device.
The matching parameters are also used for retrieving online voice data through an online voice receiving module, and the online voice receiving module can take the matching parameters as a guide to retrieve the online voice data and roughly sort the online voice data; and inputting the retrieved online voice data into a programmable device for fuzzy matching.
The programmable device is configured to realize a fuzzy matching algorithm, wherein a plurality of matching parameters are used as fuzzy rules R, digital voice is used as input quantity X, and when the input quantity X activates a plurality of fuzzy rules R, the output U is determined by the output of the fuzzy rules.
The programmable device takes digital voice signals and online voice data as input, and finally finds out the most matched online voice data to output to a processor for combination and text translation to a display and play module for playing and displaying through calculation of a fuzzy matching algorithm.
According to the characteristics of the received voice information, the matching parameters are obtained, the optimal matching strategy is obtained by using a genetic algorithm, a reconfiguration data stream for configuring a programmable device is generated, the matching of on-line voice data and an input digital voice signal is facilitated, finally, the matching of the voice signal is directly carried out by adopting a fuzzy matching algorithm, and the voice and text output translated after the accurate matching data is obtained according to a fuzzy rule. Therefore, the best matched voice translation information can be obtained according to the characteristics of voice input, and voice output can be realized rapidly.
Drawings
FIG. 1 is a block diagram of a translation fuzzy matching system.
FIG. 2 is a schematic diagram of a translation fuzzy matching system.
Fig. 3 is a step diagram of a translation fuzzy matching system for english speech.
FIG. 4 is a diagram of a genetic algorithm implementation step.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
Example 1
As shown in fig. 3, the present invention provides a specific embodiment: a fuzzy matching method for English speech translation with high accuracy comprises the following specific steps:
s1, acquiring voice information and converting the voice information into a digital voice signal;
s2, carrying out voice detection on the digital voice signal by adopting a fuzzy sound image rule to obtain corresponding matching parameters, and obtaining a reconfiguration data stream of the programmable device by adopting genetic algorithm operation according to the matching parameters;
s3, reconfiguring the programmable device by adopting the reconfiguration data stream;
s4, the online voice receiving module can search online voice data by taking the matching parameters as guidance, and roughly sort the online voice data; inputting the retrieved online voice data into a programmable device for fuzzy matching;
s5, finding out the most matched online voice data output.
The fuzzy matching algorithm in the S4 is as follows: and taking a plurality of matching parameters as a fuzzy rule R, taking digitized voice as an input quantity X, and determining an output U by the output of a plurality of fuzzy rules when the input quantity X activates the plurality of fuzzy rules R.
The genetic algorithm in S2 is specifically as follows:
s21, randomly generating chromosome individuals;
s22, calculating the fitness value of the individual;
s23, randomly performing mutation operation on the individuals to generate offspring individuals;
and S24, executing selection operation, if the fitness value of the child individual is higher than that of the individual, copying the child individual to the next generation, otherwise copying the child individual to the next generation with a small probability, and the like until the termination condition is met.
The fuzzy sound image rule in the S2 comprises sound image feature description realized by the rule formed by the methods of the mel cepstrum coefficient analysis, the short-time energy and short-time average zero-crossing rate statistical method, the formant extraction method obtained based on spectrum analysis and the like, and the output of the rule such as the mel cepstrum coefficient, the short-time energy, the short-time average zero-crossing rate, the formant and the like is used as a matching parameter.
The configurable logic module (LUT) in the programmable device 3 is used as a chromosome, the optimal chromosome is found out through a genetic algorithm, and a reconfiguration data stream is generated and configured on the programmable device 3, so that a circuit is reconfigured to realize that the circuit is changed according to the change of input voice so as to more effectively match voice data.
Example 2
As shown in fig. 1, a specific embodiment of the present invention is a translation fuzzy matching system for english language, which includes a processor 1, a memory 2, a programmable device 3, an online speech receiving module 4, a speech acquiring module 5, and a display playing module 6.
The processor 1 and the programmable device 3 are respectively connected with the voice acquisition module 5; the processor 1 is electrically connected with the memory 2, the programmable device 3, the online voice receiving module 4, the voice obtaining module 5 and the display playing module 6 respectively.
The online voice receiving module 4 and the memory 2 are connected with the programmable device 3;
the voice acquisition module 5 acquires voice information and converts the voice information into a digital voice signal;
the processor 1 receives the digital voice signal, performs voice detection on the digital voice signal by adopting a fuzzy sound image rule to obtain corresponding matching parameters, adopts genetic algorithm operation to obtain a reconfiguration data stream of the programmable device 3 according to the matching parameters, and stores the reconfiguration data stream into the memory 2, and the processor 1 controls the memory 2 and the programmable device 3 to reconfigure the programmable device 3.
The matching parameters are also used for retrieving online voice data through the online voice receiving module 4, and the online voice receiving module 4 can take the matching parameters as guidance to retrieve online voice data and roughly sort the online voice data; the retrieved online voice data is input to the programmable device 3 for fuzzy matching.
The programmable device 3 is configured to implement a fuzzy matching algorithm, taking a plurality of matching parameters as a fuzzy rule R and digitized voice as an input quantity X, and when the input quantity X activates a plurality of fuzzy rules R, the output U is determined by the output of the plurality of fuzzy rules.
The programmable device 3 takes digital voice signals and online voice data as input, and finally finds out the most matched online voice data to output to the processor 1 for combination and text translation to the display and play module 6 for playing and displaying after calculation of a fuzzy matching algorithm.
As shown in fig. 2, as a preferred embodiment, the programmable device 3 may be implemented by using an FPGA3', and the processor 1 can control the configuration data flow stored in the memory to complete the configuration after powering up the FPGA3' chip when starting up the program; the reconfiguration data stream can also be regenerated in the running process of the processor 1 to reconfigure the FPGA3' so as to perform personalized translation on different english voices.
As a preferred embodiment, the genetic algorithm may employ a trend compact genetic algorithm;
as a preferred embodiment, the fuzzy sound image rule includes a feature description of a sound image implemented by a rule composed of a mel-frequency cepstrum coefficient analysis method, a short-time energy and short-time average zero-crossing rate statistical method, a formant extraction method obtained based on spectrum analysis and the like, and outputs of the rule such as the mel-frequency cepstrum coefficient, the short-time energy, the short-time average zero-crossing rate, the formant and the like are used as matching parameters.
As a preferred embodiment, the processor 1 may be implemented by an STM32 series single-chip microcomputer 1' or a DSP.
As a preferred embodiment, the online voice receiving module 4 may remotely obtain an online voice feature library in a server, and implement retrieval of online voice features according to matching parameters and a manner of converting digital voice signals into text; the search results are sequenced and then sent to the programmable device 3 for fuzzy matching; the Websocket server 4' can be built in cooperation with the processor 1.
According to the embodiment, the matching parameters are obtained according to the characteristics of the received voice information, the optimal matching strategy is obtained by using the genetic algorithm, the reconfiguration data stream for configuring the programmable device 3 is generated, the on-line voice data and the input digital voice signals are conveniently matched, finally the voice signals are directly matched by using the fuzzy matching algorithm, and the accurate matching data is obtained according to the fuzzy rule, and then the translated voice and text are output. Therefore, the best matched voice translation information can be obtained according to the characteristics of voice input, and voice output can be realized rapidly.
Although embodiments of the present invention have been shown and described, it will be understood by those skilled in the art that various changes, modifications, substitutions and alterations can be made therein without departing from the principles and spirit of the invention, the scope of which is defined in the appended claims and their equivalents.

Claims (1)

1. A fuzzy matching method for English speech translation with high accuracy is characterized in that: the matching system adopted by the matching method comprises a processor, a memory, a programmable device, an online voice receiving module, a voice acquisition module and a display playing module; the processor and the programmable device are respectively connected with the voice acquisition module; the processor is electrically connected with the memory, the programmable device, the online voice receiving module, the voice acquisition module and the display playing module respectively; the online voice receiving module and the memory are connected with the programmable device;
the fuzzy matching method comprises the following steps:
s1, the voice acquisition module acquires voice information and converts the voice information into a digital voice signal;
s2, carrying out voice detection on the digital voice signal by adopting a fuzzy sound image rule to obtain corresponding matching parameters, and obtaining a reconfiguration data stream of the programmable device by adopting genetic algorithm operation according to the matching parameters; the reconfiguration data flow of the programmable device obtained by adopting genetic algorithm operation according to the matching parameters is specifically as follows: the configurable logic module in the programmable device is used as a chromosome, the optimal chromosome is found out through a genetic algorithm, and a reconfiguration data stream is generated and configured on the programmable device, so that a circuit is reconfigured to realize that the circuit is changed according to the change of input voice so as to more effectively match voice data; the fuzzy sound image rule comprises a sound image feature description realized by a rule composed of a mel cepstrum coefficient analysis method, a short-time energy and short-time average zero-crossing rate statistical method and a formant extraction method obtained based on spectrum analysis, and the output of the rule such as the mel cepstrum coefficient, the short-time energy, the short-time average zero-crossing rate and the formant are used as matching parameters;
the genetic algorithm comprises the following specific implementation steps:
s21, randomly generating chromosome individuals;
s22, calculating the fitness value of the individual;
s23, randomly performing mutation operation on the individuals to generate offspring individuals;
s24, executing selection operation, if the fitness value of the child individual is higher than that of the individual, copying the child individual to the next generation, otherwise copying the child individual to the next generation with a small probability, and then repeating the steps until the termination condition is met;
s3, adopting the reconfiguration data flow to reconfigure the programmable device to realize the change of the programmable device according to the change of the input voice so as to match the voice data;
s4, the online voice receiving module takes the matching parameters as guidance to retrieve online voice data, and roughly sorts the online voice data; inputting the retrieved online voice data into a programmable device for fuzzy matching; the programmable device is configured to realize a fuzzy matching algorithm, a plurality of matching parameters are used as fuzzy rules R, a digital voice signal is used as an input quantity X, and when the input quantity X activates a plurality of fuzzy rules R, an output U is determined by the output of the fuzzy rules; the programmable device takes digital voice signals and online voice data as input, and finally finds out the most matched online voice data through calculation of a fuzzy matching algorithm, outputs the most matched online voice data to a processor for combination and text translation, and sends the most matched online voice data to a display and play module for playing and displaying;
s5, outputting the most matched online voice data.
CN202110223114.2A 2021-03-01 2021-03-01 Fuzzy matching training method for English speech translation with high accuracy Active CN112967717B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110223114.2A CN112967717B (en) 2021-03-01 2021-03-01 Fuzzy matching training method for English speech translation with high accuracy

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110223114.2A CN112967717B (en) 2021-03-01 2021-03-01 Fuzzy matching training method for English speech translation with high accuracy

Publications (2)

Publication Number Publication Date
CN112967717A CN112967717A (en) 2021-06-15
CN112967717B true CN112967717B (en) 2023-08-22

Family

ID=76277532

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110223114.2A Active CN112967717B (en) 2021-03-01 2021-03-01 Fuzzy matching training method for English speech translation with high accuracy

Country Status (1)

Country Link
CN (1) CN112967717B (en)

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101464896A (en) * 2009-01-23 2009-06-24 安徽科大讯飞信息科技股份有限公司 Voice fuzzy retrieval method and apparatus
CN103811008A (en) * 2012-11-08 2014-05-21 中国移动通信集团上海有限公司 Audio frequency content identification method and device
CN105118501A (en) * 2015-09-07 2015-12-02 徐洋 Speech recognition method and system
CN106528599A (en) * 2016-09-23 2017-03-22 深圳凡豆信息科技有限公司 A rapid fuzzy matching algorithm for strings in mass audio data
CN107329961A (en) * 2017-07-03 2017-11-07 西安市邦尼翻译有限公司 A kind of method of cloud translation memory library Fast incremental formula fuzzy matching
CN109074804A (en) * 2018-07-18 2018-12-21 深圳魔耳智能声学科技有限公司 Voice recognition processing method, electronic equipment and storage medium based on accent
CN109471953A (en) * 2018-10-11 2019-03-15 平安科技(深圳)有限公司 A kind of speech data retrieval method and terminal device
CN109616120A (en) * 2019-02-20 2019-04-12 上海昊沧系统控制技术有限责任公司 The interior exchange method of the voice-based application of one kind and system
CN112101111A (en) * 2020-08-13 2020-12-18 吕梁学院 English recognition and translation method

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101537370B1 (en) * 2013-11-06 2015-07-16 주식회사 시스트란인터내셔널 System for grasping speech meaning of recording audio data based on keyword spotting, and indexing method and method thereof using the system

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101464896A (en) * 2009-01-23 2009-06-24 安徽科大讯飞信息科技股份有限公司 Voice fuzzy retrieval method and apparatus
CN103811008A (en) * 2012-11-08 2014-05-21 中国移动通信集团上海有限公司 Audio frequency content identification method and device
CN105118501A (en) * 2015-09-07 2015-12-02 徐洋 Speech recognition method and system
CN106528599A (en) * 2016-09-23 2017-03-22 深圳凡豆信息科技有限公司 A rapid fuzzy matching algorithm for strings in mass audio data
CN107329961A (en) * 2017-07-03 2017-11-07 西安市邦尼翻译有限公司 A kind of method of cloud translation memory library Fast incremental formula fuzzy matching
CN109074804A (en) * 2018-07-18 2018-12-21 深圳魔耳智能声学科技有限公司 Voice recognition processing method, electronic equipment and storage medium based on accent
CN109471953A (en) * 2018-10-11 2019-03-15 平安科技(深圳)有限公司 A kind of speech data retrieval method and terminal device
CN109616120A (en) * 2019-02-20 2019-04-12 上海昊沧系统控制技术有限责任公司 The interior exchange method of the voice-based application of one kind and system
CN112101111A (en) * 2020-08-13 2020-12-18 吕梁学院 English recognition and translation method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
英文语义机器翻译中的模糊语义选取技术;皮谨煜;;现代电子技术(第22期);全文 *

Also Published As

Publication number Publication date
CN112967717A (en) 2021-06-15

Similar Documents

Publication Publication Date Title
CN110491382B (en) Speech recognition method and device based on artificial intelligence and speech interaction equipment
CN107369439B (en) Voice awakening method and device
CN101937431A (en) Emotional voice translation device and processing method
JP2007249212A (en) Method, computer program and processor for text speech synthesis
CN106935239A (en) The construction method and device of a kind of pronunciation dictionary
CN110459202B (en) Rhythm labeling method, device, equipment and medium
CN110782880B (en) Training method and device for prosody generation model
CN112349289B (en) Voice recognition method, device, equipment and storage medium
CN110853616A (en) Speech synthesis method, system and storage medium based on neural network
CN110010136A (en) The training and text analyzing method, apparatus, medium and equipment of prosody prediction model
KR101424193B1 (en) System And Method of Pronunciation Variation Modeling Based on Indirect data-driven method for Foreign Speech Recognition
Narendra et al. Optimal weight tuning method for unit selection cost functions in syllable based text-to-speech synthesis
WO2023245389A1 (en) Song generation method, apparatus, electronic device, and storage medium
CN112530402B (en) Speech synthesis method, speech synthesis device and intelligent equipment
CN112967717B (en) Fuzzy matching training method for English speech translation with high accuracy
CN112966528B (en) English speech translation fuzzy matching system
US20040181407A1 (en) Method and system for creating speech vocabularies in an automated manner
Dumpala et al. Analysis of constraints on segmental DTW for the task of query-by-example spoken term detection
KR100259777B1 (en) Optimal synthesis unit selection method in text-to-speech system
CN114783424A (en) Text corpus screening method, device, equipment and storage medium
Sefara et al. Web-based automatic pronunciation assistant
JP3981640B2 (en) Sentence list generation device for phoneme model learning and generation program
CN113870826A (en) Pronunciation duration prediction method based on duration prediction model and related equipment
Lu et al. The USTC system for blizzard challenge 2009
JP2001092482A (en) Speech synthesis system and speech synthesis method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant