CN107248409A - A kind of multi-language translation method of dialect linguistic context - Google Patents

A kind of multi-language translation method of dialect linguistic context Download PDF

Info

Publication number
CN107248409A
CN107248409A CN201710368832.2A CN201710368832A CN107248409A CN 107248409 A CN107248409 A CN 107248409A CN 201710368832 A CN201710368832 A CN 201710368832A CN 107248409 A CN107248409 A CN 107248409A
Authority
CN
China
Prior art keywords
dialect
code
feature
phonetic
voice
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710368832.2A
Other languages
Chinese (zh)
Inventor
李伊甸
戴沛景
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sichuan Xinyimai Technology Co Ltd
Original Assignee
Sichuan Xinyimai Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sichuan Xinyimai Technology Co Ltd filed Critical Sichuan Xinyimai Technology Co Ltd
Priority to CN201710368832.2A priority Critical patent/CN107248409A/en
Publication of CN107248409A publication Critical patent/CN107248409A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/02Feature extraction for speech recognition; Selection of recognition unit
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/40Processing or translation of natural language
    • G06F40/58Use of machine translation, e.g. for multi-lingual retrieval, for server-side translation for client devices or for real-time translation
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/04Segmentation; Word boundary detection
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/06Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice

Abstract

The invention discloses a kind of multi-language translation method of dialect linguistic context, belong to computer language translation technical field, solving language translation instrument of the prior art can not accurately dialect translating, the problem of increasing the mistranslation rate of translation.Condition code storehouse of the present invention including step 1, the condition code composition for setting up various dialect phonetics under mother tongue system;Step 2, the feature compensation code for producing according to the difference of condition code and standard base phonetic feature code dialect, composition characteristic compensation codes storehouse;Step 3, with the compound standard base voice produced under correspondence mother tongue system of the corresponding source voice of feature compensation code;Step 4, standard base voice are converted into the voice or text of object language.Translation for dialect and personalized speech.

Description

A kind of multi-language translation method of dialect linguistic context
Technical field
A kind of multilingual translation system and method for dialect linguistic context, for the translation of dialect and personalized speech, belongs to meter Calculation machine language translation technical field.
Background technology
In the contacts in each fields such as countries in the world culture, economy, military affairs, the communication of language is particularly important, in order to Accurately express macaronic meaning main in various foreign affairs contacts for a long time or based on translating with people.In recent years with The fast development of computer and digital technology, translation is done with computer and digital technology and has made significant headway, Various outstanding machine translation systems are continued to bring out, more perfect especially in terms of character translation.But active computer is translated System do in terms of simultaneous interpretation due to the more linguistic context of dialect is different under each mother tongue system or preference of personal pronunciation cause mistranslation rate compared with It is high, it is impossible to the meaning of accurate expression source voice.
All can only be by the base voice of the relative standard in the mother tongue system of source in numerous language translation systems and interpreting equipment It is translated as the standard base voice of the target family of languages.But because the source voice that need to be translated is under many circumstances in practical application environment Non-standard voice, there are some the local dialects in every kind of mother tongue system in major mother tongue systems in the world, and this causes the language of computerization Speech translation system has very high mistranslation rate because of the difference of linguistic context.
The patent of Patent No. 200820234990.5, is to be used as feature and mark by extracting the word of dialect phonetic, words and phrases Quasi- base voice carries out directly contrast to find the corresponding standard base voice of dialect phonetic, once the people's pronunciation spoken a dialect is inaccurate, It is easy for causing translation inaccurate, so that corresponding standard base voice is can not find, and also every dialect being translated is required for Storage, the problems such as causing carrying cost, high operation hardware cost and the slow operational speed of a computer.
The content of the invention
It is an object of the invention to:Solving language translation instrument of the prior art can not accurately dialect translating, increase The problem of mistranslation rate of translation, there is provided a kind of multi-language translation method of dialect linguistic context.
The technical solution adopted by the present invention is as follows:
A kind of multi-language translation method of dialect linguistic context, it is characterised in that comprise the following steps:
Step 1, set up various dialect phonetics under mother tongue system condition code composition condition code storehouse;
Step 2, the feature compensation code for producing according to the difference of condition code and standard base phonetic feature code dialect, composition characteristic Compensation codes storehouse;
Step 3, with the compound standard base voice produced under correspondence mother tongue system of the corresponding source voice of feature compensation code;
Step 4, standard base voice are converted into the voice or text of object language.
Further, the method for building up in condition code storehouse comprises the following steps:
Various dialect phonetic samples under step 1.1, acquisition mother tongue system;
Step 1.2, to dialect phonetic sample carry out pretreatment remove redundancy section, pre-filtering is carried out by bandpass filter Preemphasis is carried out by a high-pass filter again after processing, to being multiplied by Hamming window after the dialect phonetic sample noise abatement after preemphasis After carry out end-point detection;
Step 1.3, the pretreated dialect phonetic sample of Hamming window will be multiplied by carry out spectrum analysis, and then carry out feature and carry Take, the formant of the dialect phonetic sample after feature extraction spectrum analysis, pitch period feature, MFCC and LPCC parameter attributes Code;
Step 1.4 is by the formant of acquisition, pitch period feature, and it is right after de-redundancy that MFCC and LPCC parameter attributes code is carried out Multiple dialect phonetic files carry out the probability distribution statistical of condition code, find out condition code of its denominator as the dialect phonetic Key element;
Step 1.5 by the code with condition code key element use compress mode to recompile for 64 bytes dialect phonetic Condition code, assigns searching number by this feature code and is incorporated into condition code storehouse.
Further, the extraction step of the formant of the dialect phonetic after spectrum analysis is included in step 1.3:
Smooth spectrum is obtained after homomorphic filtering to the dialect phonetic after spectrum analysis discrete fourier is asked to the spectrum again Conversion, then extracts the formant parameter of voice signal with DFT spectrums.
Further, the extraction step of the pitch period feature of the dialect phonetic after spectrum analysis is included in step 1.3:
Pitch period feature is extracted using average magnitude difference function method to the dialect phonetic after spectrum analysis.
Further, the extraction step of the MFCC parameters of the dialect phonetic after spectrum analysis is included in step 1.3:
Dialect phonetic after spectrum analysis is subjected to Short Time Fourier Transform and obtains its frequency spectrum, then seeks square of spectrum amplitude Energy spectrum is obtained, bandpass filtering is carried out with triangle filter equalizer, the number of wave filter is close with critical band number, if wave filter number is Obtain being output as after M, filtering:X (k), k=l, 2 ..., M, the output to wave filter group are taken the logarithm, and then make 2M points against in Fu Leaf transformation is that can obtain MFCC parameters.
Further, the extraction step of the LPCC parameters of the dialect phonetic after spectrum analysis is included in step 1.3:
Dialect phonetic after spectrum analysis is carried out to the inverse Z-transform of logarithm modular function after transform, passes through the Fourier of signal Conversion, the logarithm of modulus, then Fourier transformation of negating obtain LPCC parameters.
Further, the foundation in feature compensation code storehouse uses following steps in step 2:
Step 2.1, the standard base speech samples obtained under mother tongue system, the condition code of extraction standard base voice;
Step 2.2, com-parison and analysis is carried out to dialect phonetic condition code under mother tongue system and standard base phonetic feature code, drawn The respective probability distribution variances frequency spectrum of general meaning;
Step 2.3, by the difference frequency spectrum carry out radix-minus-one complement superposition draw dialect phonetic feature compensation code;
Step 2.3, feature compensation code and the condition code of dialect phonetic are combined, the condition code after being combined again with standard base voice Verify and correct by n times, be combined after being corrected to the n times verification of dialect phonetic feature compensation code and draw standard base phonetic feature Code is in the range of allowable error, and this feature compensation codes are the feature compensation code of the dialect phonetic, and this feature compensation codes are assigned Searching number is incorporated into feature compensation code storehouse.
Further, the feature compensation code obtaining step in step 3:Pass through the feature compensation code and feature compensation code of source voice Storehouse carries out similarity retrieval and obtains corresponding feature compensation code.
Further, the feature compensation code in step 3, which is obtained, is set manually by user.
In summary, by adopting the above-described technical solution, the beneficial effects of the invention are as follows:
1st, the present invention can by match source voice and poster base voice feature compensation code, then with source voice and feature benefit Repay code to be combined, dialect or individualized language can accurately be translated, greatly reduce the mistranslation rate of existing interpreting equipment, it is accurate True rate may be up to more than 95%;
2nd, the present invention is applicable the dialect phonetic translation under different occasions, significantly improves the language simultaneous interpretation of computerization Application;
3rd, the present invention can most be represented by extraction each dialect or the feature of personalized speech so that the feature compensation of generation Code is applied to corresponding dialect or personalized speech, it is to avoid when deviation occurs in pronunciation, causes compound standard base voice to be forbidden Really the problem of;
4th, requirement of the present invention to storage hardware, operation hardware is low, so as to save hardware cost so that arithmetic speed It hurry up.
Brief description of the drawings
Fig. 1 sets up the schematic diagram in condition code storehouse and feature compensation code storehouse for system in the present invention;
Fig. 2 is carries out the block schematic illustration of the acquisition of feature compensation code during the translation of unknown languages in the present invention;
The block schematic illustration of the acquisition of feature compensation code Fig. 3 is carries out specifying languages translation in the present invention when;
Fig. 4 is the schematic diagram of the embodiment 1 in condition code storehouse in the present invention, time-domain diagram, sound spectrograph is followed successively by from left to right common Shake peak, pitch period figure, loudness of a sound figure;
Fig. 5 is the schematic diagram of the embodiment 2 in condition code storehouse in the present invention, time-domain diagram, sound spectrograph is followed successively by from left to right common Shake peak, pitch period figure, loudness of a sound figure;
Fig. 6 is the schematic diagram of the embodiment 3 in condition code storehouse in the present invention, time-domain diagram, sound spectrograph is followed successively by from left to right common Shake peak, pitch period figure, loudness of a sound figure;
Fig. 7 is the schematic diagram of the embodiment 4 in condition code storehouse in the present invention, time-domain diagram, sound spectrograph is followed successively by from left to right common Shake peak, pitch period figure, loudness of a sound figure;
Fig. 8 is the schematic diagram of the embodiment 5 in condition code storehouse in the present invention, time-domain diagram, sound spectrograph is followed successively by from left to right common Shake peak, pitch period figure, loudness of a sound figure;
Fig. 9 is the schematic diagram of the embodiment 6 in condition code storehouse in the present invention, time-domain diagram, sound spectrograph is followed successively by from left to right common Shake peak, pitch period figure, loudness of a sound figure.
Embodiment
In order to make the purpose , technical scheme and advantage of the present invention be clearer, it is right below in conjunction with drawings and Examples The present invention is further elaborated.It should be appreciated that specific embodiment described herein is only to explain the present invention, not For limiting the present invention.
A kind of multi-language translation method of dialect linguistic context, concrete implementation method is as follows:
It (can be each dialect or personalized language under each mother tongue system of the world that each dialect phonetic under mother tongue system should be set up first Sound or each dialect or personalized speech under the mother tongue system specified) condition code storehouse and feature compensation code storehouse.Condition code The foundation in storehouse uses following steps:
Massage voice reading is carried out with same section of representational word, declaimer is various dialect men, female voice under certain mother tongue system Voice, dialect phonetic declaimer, which tackles the dialect, has certain representative and general meaning.
Typing preserves for dialect phonetic file after the text sampling that dialect phonetic declaimer is read aloud, the parameter of voice document For:Sample rate 11025Hz, monophonic, sampling precision 16,1~30s of file size, every kind of dialect phonetic file is recorded respectively Each representational text-to-speech file of men and women's sound is at least each 50, and the dialect phonetic file of recording is more, last other side's speech The probability distribution statistical of sound condition code is more accurate.
Pretreatment to dialect phonetic:Remove redundancy section, pass through again after carrying out pre-filtering processing by bandpass filter One high-pass filter carries out preemphasis, and end-point detection is carried out to being multiplied by after the signal noise abatement after preemphasis after Hamming window.
The pretreated dialect phonetic of Hamming window will be multiplied by and carry out spectrum analysis, feature extraction, feature emphasis is then carried out It is to use most represent the formant of dialect phonetic feature, pitch period feature, MFCC and LPCC parameters are analyzed and feature Extract.
The extraction of formant:Smooth spectrum is obtained after homomorphic filtering to the dialect phonetic after spectrum analysis again to the spectrum Discrete Fourier transform (DFT) is asked, the formant parameter of voice signal is then extracted with DFT spectrums.
The extraction of pitch period feature:To the dialect phonetic after spectrum analysis using average magnitude difference function (AMDF) method come Extract pitch period feature.
MFCC extraction:Dialect phonetic after spectrum analysis is subjected to Short Time Fourier Transform and obtains its frequency spectrum, then seeks frequency Spectral amplitude square energy spectrum, bandpass filtering is carried out with triangle filter equalizer, the number of wave filter is close with critical band number, If wave filter number is M, obtain being output as after filtering:X (k), k=l, 2 ..., M, the output to wave filter group are taken the logarithm, then Make 2M point inverse Fourier transforms and can obtain MFCC parameters.
LPCC extraction:Dialect phonetic after spectrum analysis is carried out to the inverse Z-transform of logarithm modular function after transform, passed through The Fourier transformation of signal, the logarithm of modulus, then Fourier transformation of negating obtain LPCC parameters.
By the formant of acquisition, pitch period feature, MFCC and LPCC parameter attributes code is carried out after de-redundancy to multiple sides Speech voice document carries out having the probability distribution statistical of dialect phonetic condition code, and emphasis is the pronunciation to vowel, the first formant and Second formant, pitch curve, LPCC and MFCC probability distribution statistical, find out feature of the denominator as the dialect phonetic Code key element.
By the code with condition code key element use compress mode to recompile for 64 bytes dialect phonetic condition code, Searching number, which is assigned, by this feature code is incorporated into condition code storehouse, i.e. condition code storehouse.
The specific foundation in condition code storehouse is as follows:
Embodiment 1
The present embodiment says English model voice " This is a test sample for myself " using American.Adopt Sample frequency 11025Hz, sampling depth 16bit, monophonic, duration 2.157 seconds.The time-domain diagram of actual analysis, sound spectrograph formant, Pitch period figure, loudness of a sound figure, as shown in Figure 4;Formant is extracted, pitch period feature, MFCC and LPCC parameter attributes code is carried out The probability distribution statistical for having dialect phonetic condition code is carried out after de-redundancy to dialect phonetic file, by the generation with condition code key element Code uses compress mode to recompile for the condition code of the dialect phonetic of 64 bytes, and assigning searching number by this feature code is incorporated into feature Code storehouse, that is, be incorporated into spectrum signature code storehouse.
Embodiment 2
The present embodiment says English model voice " This is a test sample for myself " using Englishman.Adopt Sample frequency 11025Hz, sampling depth 16bit, monophonic, duration 2.267 seconds.The time-domain diagram of actual analysis, sound spectrograph formant, Pitch period figure, loudness of a sound figure, as shown in Figure 5;Formant is extracted, pitch period feature, MFCC and LPCC parameter attributes code is carried out The probability distribution statistical for having dialect phonetic condition code is carried out after de-redundancy to dialect phonetic file, by the generation with condition code key element Code uses compress mode to recompile for the condition code of the dialect of 64 bytes, and assigning searching number by this feature code is incorporated into condition code Storehouse, that is, be incorporated into spectrum signature code storehouse.
Embodiment 3
The present embodiment says English model voice " This is a test sample for myself " using Indian.Adopt Sample frequency 11025Hz, sampling depth 16bit, monophonic, duration 1.956 seconds.The time-domain diagram of actual analysis, sound spectrograph formant, Pitch period figure, loudness of a sound figure, as shown in Figure 6;Formant is extracted, pitch period feature, MFCC and LPCC parameter attributes code is carried out The probability distribution statistical for having dialect phonetic condition code is carried out after de-redundancy to dialect phonetic file, by the generation with condition code key element Code uses compress mode to recompile for the condition code of the dialect phonetic of 64 bytes, and assigning searching number by this feature code is incorporated into feature Code storehouse, that is, be incorporated into spectrum signature code storehouse.
Embodiment 4
The present embodiment says that Henan model voice " please input the sample that received pronunciation text makees speech sample using Chinese Plate ".Sample frequency 11025Hz, sampling depth 16bit, monophonic, duration 4.27 seconds.The time-domain diagram of actual analysis, sound spectrograph is common Shake peak, pitch period figure, loudness of a sound figure, as shown in Figure 7;Extract formant, pitch period feature, MFCC and LPCC parameter attributes code The probability distribution statistical for having dialect phonetic condition code after de-redundancy to the progress of dialect phonetic file is carried out, there will be condition code key element Code use compress mode to recompile for 64 bytes dialect phonetic condition code, by this feature code assign searching number be incorporated into Condition code storehouse, that is, be incorporated into spectrum signature code storehouse.
Embodiment 5
The present embodiment says that Sichuan model voice " please input the sample that received pronunciation text makees speech sample using Chinese Plate ".Sample frequency 11025Hz, sampling depth 16bit, monophonic, duration 4.928 seconds.The time-domain diagram of actual analysis, sound spectrograph Formant, pitch period figure, loudness of a sound figure, as shown in Figure 8;Extract formant, pitch period feature, MFCC and LPCC parameter attributes Code carries out the probability distribution statistical for having dialect phonetic condition code after de-redundancy to the progress of dialect phonetic file, will be wanted with condition code The code of element uses compress mode to recompile for the condition code of the dialect phonetic of 64 bytes, and assigning searching number by this feature code compiles Enter condition code storehouse, that is, be incorporated into spectrum signature code storehouse.
Embodiment 6
The present embodiment " please input the sample that received pronunciation text makees speech sample using Chinese's model voice that speaks standard Chinese pronunciation Plate ".Sample frequency 11025Hz, sampling depth 16bit, monophonic, duration 3.96 seconds.The time-domain diagram of actual analysis, sound spectrograph is common Shake peak, pitch period figure, loudness of a sound figure, as shown in Figure 9;Extract formant, pitch period feature, MFCC and LPCC parameter attributes code The probability distribution statistical for having dialect phonetic condition code after de-redundancy to the progress of dialect phonetic file is carried out, there will be condition code key element Code use compress mode to recompile for 64 bytes dialect phonetic condition code, by this feature code assign searching number be incorporated into Condition code storehouse, that is, be incorporated into spectrum signature code storehouse.
It is that source voice can be complex as to the standard base voice of the mother tongue system, Ying Jian after the condition code storehouse of dialect phonetic is set up Dialect phonetic under vertical mother tongue system is combined required feature compensation code storehouse.Corresponding with condition code storehouse feature compensation code storehouse is built It is vertical so that translation system can reduce hardware cost in practical application.The foundation in spectrum signature compensation codes storehouse uses following steps:
Massage voice reading is carried out with same section of representational word, declaimer is various standard men, female voice under certain mother tongue system Voice, massage voice reading person should be the mother tongue system standard base voice.
To the condition code of standard base voice extraction standard base voice, wherein, the extraction of the condition code of standard base voice is with building The extracting method of condition code is identical during vertical dialect phonetic condition code storehouse.
Feature compensation code and the condition code of dialect phonetic are combined, and the condition code after being combined again with standard base voice passes through n times Verification is corrected, and is combined after being corrected to the n times verification of dialect phonetic feature compensation code and is shown that standard base phonetic feature code is being permitted Perhaps in error range, this feature compensation codes are the feature compensation code of the dialect phonetic, and searching number is assigned by this feature compensation codes It is incorporated into feature compensation code storehouse.Available for later stage real-time dialect or the language translation system of personalized linguistic context.
Establish behind condition code storehouse and feature compensation code storehouse, produced with the corresponding source voice of feature compensation code is compound Standard base voice under correspondence mother tongue system;The acquisition of feature compensation code has two ways:
Feature compensation code obtains step:Similarity retrieval is carried out by the feature compensation code and feature compensation code storehouse of source voice Obtain corresponding feature compensation code.I.e. in the case of unknown languages, to source voice carry out condition code extraction, then with standard base language The feature compensation code of sound generation source voice, the feature compensation code of source voice carries out maximum comparability with feature compensation code storehouse and retrieved To corresponding feature compensation code, obtained feature compensation code carries out the compound standard produced under corresponding mother tongue system with source voice again Base voice.
Feature compensation code, which is obtained, to be set manually by user.In the case of being applicable known languages, source voice directly with The compound standard base voice produced under correspondence mother tongue system of given feature compensation code.
Standard base voice is converted into the voice or text of object language, and exports voice or text.
The present invention can translate the dialect for specifying languages according to user's request, can also translate the dialect of unknown languages.
The foregoing is merely illustrative of the preferred embodiments of the present invention, is not intended to limit the invention, all essences in the present invention Any modifications, equivalent substitutions and improvements made within refreshing and principle etc., should be included in the scope of the protection.

Claims (9)

1. a kind of multi-language translation method of dialect linguistic context, it is characterised in that comprise the following steps:
Step 1, set up various dialect phonetics under mother tongue system condition code composition condition code storehouse;
Step 2, the feature compensation code for producing according to the difference of condition code and standard base phonetic feature code dialect, composition characteristic compensation Code storehouse;
Step 3, with the compound standard base voice produced under correspondence mother tongue system of the corresponding source voice of feature compensation code;
Step 4, standard base voice are converted into the voice or text of object language.
2. a kind of multi-language translation method of dialect linguistic context according to claim 1, it is characterised in that condition code storehouse is built Cube method comprises the following steps:
Various dialect phonetic samples under step 1.1, acquisition mother tongue system;
Step 1.2, to dialect phonetic sample carry out pretreatment remove redundancy section, pass through bandpass filter carry out pre-filtering processing Preemphasis is carried out by a high-pass filter again afterwards, it is laggard to being multiplied by Hamming window after the dialect phonetic sample noise abatement after preemphasis Row end-point detection;
Step 1.3, the pretreated dialect phonetic sample of Hamming window will be multiplied by carry out spectrum analysis, and then carry out feature extraction, The formant of dialect phonetic sample after feature extraction spectrum analysis, pitch period feature, MFCC and LPCC parameter attributes code;
Step 1.4 carries out the formant of acquisition, pitch period feature, MFCC and LPCC parameter attributes code after de-redundancy to multiple Dialect phonetic file carries out the probability distribution statistical of condition code, and finding out its denominator will as the condition code of the dialect phonetic Element;
Step 1.5 by the code with condition code key element use compress mode to recompile for 64 bytes dialect phonetic feature Code, assigns searching number by this feature code and is incorporated into condition code storehouse.
3. the multi-language translation method of a kind of dialect linguistic context according to claim 2, it is characterised in that right in step 1.3 The extraction step of the formant of dialect phonetic after spectrum analysis includes:
Smooth spectrum is obtained after homomorphic filtering to the dialect phonetic after spectrum analysis discrete Fourier transform is asked to the spectrum again, Then the formant parameter of voice signal is extracted with DFT spectrums.
4. the multi-language translation method of a kind of dialect linguistic context according to claim 2, it is characterised in that right in step 1.3 The extraction step of the pitch period feature of dialect phonetic after spectrum analysis includes:
Pitch period feature is extracted using average magnitude difference function method to the dialect phonetic after spectrum analysis.
5. the multi-language translation method of a kind of dialect linguistic context according to claim 2, it is characterised in that right in step 1.3 The extraction step of the MFCC parameters of dialect phonetic after spectrum analysis includes:
Dialect phonetic after spectrum analysis is subjected to Short Time Fourier Transform and obtains its frequency spectrum, then ask spectrum amplitude square energy Amount spectrum, bandpass filtering is carried out with triangle filter equalizer, the number of wave filter is close with critical band number, if wave filter number is M, filter Obtain being output as after ripple:X (k), k=l, 2 ..., M, the output to wave filter group are taken the logarithm, and are then made 2M points and are become against Fourier Change and can obtain MFCC parameters.
6. the multi-language translation method of a kind of dialect linguistic context according to claim 2, it is characterised in that right in step 1.3 The extraction step of the LPCC parameters of dialect phonetic after spectrum analysis includes:
Dialect phonetic after spectrum analysis is carried out to the inverse Z-transform of logarithm modular function after transform, become by the Fourier of signal Change, the logarithm of modulus, then Fourier transformation of negating obtain LPCC parameters.
7. a kind of multi-language translation method of dialect linguistic context according to claim 1, it is characterised in that feature in step 2 The foundation in compensation codes storehouse uses following steps:
Step 2.1, the standard base speech samples obtained under mother tongue system, the condition code of extraction standard base voice;
Step 2.2, com-parison and analysis is carried out to dialect phonetic condition code under mother tongue system and standard base phonetic feature code, drawn general meaning Respective probability distribution variances frequency spectrum;
Step 2.3, by the difference frequency spectrum carry out radix-minus-one complement superposition draw dialect phonetic feature compensation code;
Step 2.3, feature compensation code and the condition code of dialect phonetic are combined, and the condition code after being combined again with standard base voice is passed through N times verification is corrected, and is combined after being corrected to the n times verification of dialect phonetic feature compensation code and is shown that standard base phonetic feature code exists In the range of allowable error, this feature compensation codes are the feature compensation code of the dialect phonetic, and this feature compensation codes are assigned and retrieved Number it is incorporated into feature compensation code storehouse.
8. according to a kind of multi-language translation method of any described dialect linguistic context of claim 1-7, it is characterised in that:Step 3 In feature compensation code obtaining step:Similarity retrieval is carried out by the feature compensation code of source voice with feature compensation code storehouse to obtain Corresponding feature compensation code.
9. according to a kind of multi-language translation method of any described dialect linguistic context of claim 1-7, it is characterised in that:Step 3 In feature compensation code obtain be to be set manually by user.
CN201710368832.2A 2017-05-23 2017-05-23 A kind of multi-language translation method of dialect linguistic context Pending CN107248409A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710368832.2A CN107248409A (en) 2017-05-23 2017-05-23 A kind of multi-language translation method of dialect linguistic context

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710368832.2A CN107248409A (en) 2017-05-23 2017-05-23 A kind of multi-language translation method of dialect linguistic context

Publications (1)

Publication Number Publication Date
CN107248409A true CN107248409A (en) 2017-10-13

Family

ID=60016797

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710368832.2A Pending CN107248409A (en) 2017-05-23 2017-05-23 A kind of multi-language translation method of dialect linguistic context

Country Status (1)

Country Link
CN (1) CN107248409A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109446535A (en) * 2018-10-22 2019-03-08 内蒙古工业大学 A kind of illiteracy Chinese nerve machine translation method based on triangle framework
CN113836945A (en) * 2021-09-23 2021-12-24 平安科技(深圳)有限公司 Intention recognition method and device, electronic equipment and storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1225567A2 (en) * 2001-01-23 2002-07-24 Matsushita Electric Industrial Co., Ltd. Method and apparatus for speech recognition
CN201403118Y (en) * 2008-12-12 2010-02-10 康佳集团股份有限公司 Device with dialect translating function and mobile terminal
CN102074234A (en) * 2009-11-19 2011-05-25 财团法人资讯工业策进会 Voice variation model building device and method as well as voice recognition system and method
CN102117614A (en) * 2010-01-05 2011-07-06 索尼爱立信移动通讯有限公司 Personalized text-to-speech synthesis and personalized speech feature extraction
CN102543073A (en) * 2010-12-10 2012-07-04 上海上大海润信息系统有限公司 Shanghai dialect phonetic recognition information processing method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1225567A2 (en) * 2001-01-23 2002-07-24 Matsushita Electric Industrial Co., Ltd. Method and apparatus for speech recognition
CN201403118Y (en) * 2008-12-12 2010-02-10 康佳集团股份有限公司 Device with dialect translating function and mobile terminal
CN102074234A (en) * 2009-11-19 2011-05-25 财团法人资讯工业策进会 Voice variation model building device and method as well as voice recognition system and method
CN102117614A (en) * 2010-01-05 2011-07-06 索尼爱立信移动通讯有限公司 Personalized text-to-speech synthesis and personalized speech feature extraction
CN102543073A (en) * 2010-12-10 2012-07-04 上海上大海润信息系统有限公司 Shanghai dialect phonetic recognition information processing method

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109446535A (en) * 2018-10-22 2019-03-08 内蒙古工业大学 A kind of illiteracy Chinese nerve machine translation method based on triangle framework
CN113836945A (en) * 2021-09-23 2021-12-24 平安科技(深圳)有限公司 Intention recognition method and device, electronic equipment and storage medium
WO2023045186A1 (en) * 2021-09-23 2023-03-30 平安科技(深圳)有限公司 Intention recognition method and apparatus, and electronic device and storage medium
CN113836945B (en) * 2021-09-23 2024-04-16 平安科技(深圳)有限公司 Intention recognition method, device, electronic equipment and storage medium

Similar Documents

Publication Publication Date Title
CN107945805B (en) A kind of across language voice identification method for transformation of intelligence
Gawali et al. Marathi isolated word recognition system using MFCC and DTW features
Schlippe et al. Grapheme-to-phoneme model generation for Indo-European languages
Ali et al. Advances in dialectal arabic speech recognition: A study using twitter to improve egyptian asr
CN110853629A (en) Speech recognition digital method based on deep learning
Bharali et al. A comparative study of different features for isolated spoken word recognition using HMM with reference to Assamese language
CN113744722A (en) Off-line speech recognition matching device and method for limited sentence library
Tantawi et al. A deep learning approach for automatic speech recognition of The Holy Qur’ān recitations
Cardenas et al. Siminchik: A speech corpus for preservation of southern quechua
CN107248409A (en) A kind of multi-language translation method of dialect linguistic context
Mukherjee et al. REARC-a Bangla Phoneme recognizer
Abbas et al. Pashto Spoken Digits database for the automatic speech recognition research
Gedam et al. Development of automatic speech recognition of Marathi numerals-a review
Ahmed et al. Technique for automatic sentence level alignment of long speech and transcripts.
Mahmood Arabic speaker recognition system based on phoneme fusion
Sharan et al. Speaker-independent recognition system for continuous Hindi speech using probabilistic model
Chit et al. Myanmar continuous speech recognition system using fuzzy logic classification in speech segmentation
Liu et al. Supra-Segmental Feature Based Speaker Trait Detection.
Jin et al. On continuous speech recognition of Indian English
Coats A pipeline for the large-scale acoustic analysis of streamed content
Zubi et al. Arabic Dialects System using Hidden Markov Models (HMMs)
Laleye et al. Automatic text-independent syllable segmentation using singularity exponents and rényi entropy
Sreeraj et al. Automatic dialect recognition using feature fusion
Shinde et al. Isolated Word Recognition System based on LPC and DTW Technique
Gadekar et al. Analysis of speech recognition techniques

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20171013

RJ01 Rejection of invention patent application after publication