CN102982801B - Phonetic feature extracting method for robust voice recognition - Google Patents
Phonetic feature extracting method for robust voice recognition Download PDFInfo
- Publication number
- CN102982801B CN102982801B CN201210449436.XA CN201210449436A CN102982801B CN 102982801 B CN102982801 B CN 102982801B CN 201210449436 A CN201210449436 A CN 201210449436A CN 102982801 B CN102982801 B CN 102982801B
- Authority
- CN
- China
- Prior art keywords
- speech
- power spectrum
- spectrum
- short
- formula
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 50
- 238000001228 spectrum Methods 0.000 claims abstract description 144
- 238000001914 filtration Methods 0.000 claims abstract description 21
- 230000008569 process Effects 0.000 claims abstract description 15
- 238000000605 extraction Methods 0.000 claims description 29
- 238000012545 processing Methods 0.000 claims description 16
- 230000000873 masking effect Effects 0.000 claims description 12
- 230000008901 benefit Effects 0.000 claims description 8
- 230000008859 change Effects 0.000 claims description 4
- NAWXUBYGYWOOIX-SFHVURJKSA-N (2s)-2-[[4-[2-(2,4-diaminoquinazolin-6-yl)ethyl]benzoyl]amino]-4-methylidenepentanedioic acid Chemical compound C1=CC2=NC(N)=NC(N)=C2C=C1CCC1=CC=C(C(=O)N[C@@H](CC(=C)C(O)=O)C(O)=O)C=C1 NAWXUBYGYWOOIX-SFHVURJKSA-N 0.000 claims 1
- 230000010354 integration Effects 0.000 claims 1
- 230000000694 effects Effects 0.000 abstract description 9
- 238000010606 normalization Methods 0.000 abstract description 9
- 238000012935 Averaging Methods 0.000 abstract description 3
- 238000004422 calculation algorithm Methods 0.000 description 11
- NGVDGCNFYWLIFO-UHFFFAOYSA-N pyridoxal 5'-phosphate Chemical compound CC1=NC=C(COP(O)(O)=O)C(C=O)=C1O NGVDGCNFYWLIFO-UHFFFAOYSA-N 0.000 description 7
- 238000012360 testing method Methods 0.000 description 4
- 238000006243 chemical reaction Methods 0.000 description 3
- 230000008447 perception Effects 0.000 description 3
- 230000003595 spectral effect Effects 0.000 description 3
- 238000012549 training Methods 0.000 description 3
- 239000000654 additive Substances 0.000 description 2
- 230000000996 additive effect Effects 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 230000007613 environmental effect Effects 0.000 description 2
- 230000006872 improvement Effects 0.000 description 2
- 230000002401 inhibitory effect Effects 0.000 description 2
- 230000001766 physiological effect Effects 0.000 description 2
- 230000009466 transformation Effects 0.000 description 2
- 241000201297 Chelone Species 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 238000012152 algorithmic method Methods 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000005764 inhibitory process Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000001629 suppression Effects 0.000 description 1
Landscapes
- Filters That Use Time-Delay Elements (AREA)
Abstract
The invention discloses a phonetic feature extracting method for robust voice recognition. The method comprises the steps of obtaining a power spectrum, handling the power spectrum by a filter set, obtaining a medium length power spectrum by means of frame averaging, carrying out unsymmetrical filtration process to the power spectrum, meanwhile carrying out a sheltering process to the power spectrum, obtaining a pure voice power spectrum, carrying out a channel averaging process to a ratio of the pure voice power spectrum and a voice power spectrum with noise so as to be smoothed, multiplying the smoothed ratio of the pure voice power spectrum and the voice power spectrum with noise by a power spectrum output by a filter bank so as to obtain a short power spectrum of pure voice, carrying out an energy normalization process to the short power spectrum so as to eliminate multiplicative noise, carrying out equal loudness aggravation to the power spectrum, carrying out an index operation to the power spectrum, carrying out an inverse Fourier transform to the power spectrum, obtaining cepstrum coefficients of signals, and carrying out a mean value normalization process to the cepstrum coefficients. The voice signals extracted by the method are high in speed and capable of achieving an online process. An acoustic model trained by characteristics extracted by the method has a good noise proof effect. The phonetic feature extracting method for robust voice recognition is of great importance to use.
Description
Technical field
The present invention relates to field of speech recognition, relate in particular to one in speech recognition, can have obvious inhibiting speech feature extraction algorithmic method to steady and nonstationary noise.
Background technology
It is one of sixty-four dollar question in speech recognition that speech recognition system recognition performance under complex environment sharply reduces.For example mobile phone speech polling geographic position on road, the residing acoustic enviroment of user is very complicated and variation is rapid, and this performance on speech recognition system has produced great impact.Original speech recognition system can obtain good processing and recognition effect not having under noisy environment, but in real world applications the performance of recognition system can due to time become uncertain environmental noise and the impact of channel, speaker's difference, the factors such as the variation of conversation content affect degradation.Become the key of speech recognition technology so how to improve the robustness of speech recognition system under the mismatch condition of training and testing environment.
In recent years, propose a lot of improvement technology and calculating methods this research field of speech recognition technology environmental robustness people, and obtained certain effect.According to the flow process of speech recognition, robust speech identification can be divided into four classes: the anti-noise of time-frequency domain; The noise compensation of property field; The noise self-adaptation in model territory and the self-adaptation of decoded domain.Technology is the earliest the anti-noise of time-frequency domain, and for example spectrum subtracts and Wiener filtering, also has two stage Wiener filterings of classical ETSI.Feature aspect squelch normally compensates noise in the process of extracting feature.Because PLP and MFCC feature occupy huge legendary turtle head always, thus the squelch of feature aspect mostly in these two kinds of features, carry out, such as vectorial Taylor series etc.Three phases is to carry out self-adaptation in model side's face of noise, comprises the HMM of multi-mode speech model, shared variable parameter etc.Four levels is the noise self-adaptation in decoding aspect, comprises uncertain decoding and replaces uncertain decode etc. with subband revaluation.
These all methods are all to seek under certain criterion unmatched a kind of the optimal compensation between training environment and test environment essentially.In a series of supposed premise conditions, as the independence between the independence of the Gaussian distribution of additivity noise, noise and voice signal, different noise, roll-off characteristic of channel etc., these methods have all been made useful exploration and contribution for the robustness of speech recognition technology, especially under stationary noise environment, have good noise suppression effect.But this also has very large gap with the application requirements of speech recognition system under true noisy environment, therefore for complex environment more, such as the environment such as burst noise are helpless.
Summary of the invention
(1) technical matters that will solve
Low in order to solve the above-mentioned phonetic recognization rate under complex environment, and the common feature extracting method shortcoming strong not to the inhibition ability of nonstationary noise, the present invention proposes a kind of feature extracting method that can improve its discrimination, object is to improve the discrimination with the voice of the additive noise such as burst noise and music noise impact, and the phonetic recognization rate under pure environment is not declined.
(2) technical scheme
The present invention based on a kind of Speech Feature Extraction for robust speech identification, comprise the following steps to realize:
Step 1, obtain the power spectrum of voice signal;
Step 2, by obtained power spectrum by bank of filters processing, obtain the short-time rating spectrum of noisy speech;
Step 3, according to the short-time rating spectrum of obtained noisy speech, adopt the average mode of frame to ask for the medium duration power spectrum of noisy speech;
Step 4, the medium duration power spectrum of obtained noisy speech is carried out asymmetric filtering and shelters anti-noise, to obtain the medium duration power spectrum of clean speech;
Step 5, obtain the short-time rating spectrum of clean speech according to the short-time rating spectrum of the medium duration power spectrum of the medium duration power spectrum of described clean speech, noisy speech and noisy speech;
Step 6, the short-time rating spectrum of clean speech is carried out to energy normalized processing, to eliminate the property taken advantage of noise;
Step 7, offset and the loudness such as carry out except the short-time rating spectrum of the clean speech of the property taken advantage of noise and increase the weight of;
The short-time rating spectrum of the clean speech after step 8, reciprocity loudness increase the weight of is carried out index nonlinear operation;
Step 9, the short-time rating spectrum of having carried out the clean speech after index nonlinear operation is carried out to inverse fourier transform, to ask for cepstrum coefficient, cepstrum coefficient is carried out to average normalized, finally obtain phonetic feature.
The present invention starts with from traditional Speech Feature Extraction, for the shortcoming a little less than traditional voice feature anti-noise ability, has proposed some means and has improved phonetic feature, finally forms a set of new Speech Feature Extraction.The present invention is directed to noise and change the feature slower than voice, utilize the average mode of frame short-time rating to be composed to the power spectrum that is converted to medium duration, for estimating noise; Utilize the mode of asymmetric filtering, estimate respectively the spectrum envelope of noise and voice in noisy speech; On the basis of asymmetric filtering, adopt the mode estimated snr of sheltering, and it is processed, the signal to noise ratio (S/N ratio) that is converted into short-time rating spectrum is carried out anti-noise; Also by energy normalized with index is non-linear that power spectrum is processed.The Speech Feature Extraction for robust speech identification that the present invention proposes not only can be estimated more accurately to noise, also can make phonetic feature more meet the auditory properties of people's ear.Therefore the feature that this feature extracting method is asked for has good inhibiting effect to noise.
(3) beneficial effect
The present invention starts with from traditional Speech Feature Extraction, in traditional Speech Feature Extraction, add anti-noise to process and meet the conversion process of human auditory system, make this feature extracting method various additive noises of not only can drawing up, and discrimination under pure environment is also higher than traditional Speech Feature Extraction.
Brief description of the drawings
Fig. 1 is the main-process stream block diagram of the present invention for the Speech Feature Extraction of robust speech identification;
Fig. 2 is the structure flow chart that comprises the asymmetric low-pass filtering anti-noise module of sheltering;
Fig. 3 is the structure flow chart of masking block in Fig. 2.
Embodiment
For making the object, technical solutions and advantages of the present invention clearer, below in conjunction with specific embodiment, and with reference to accompanying drawing, the present invention is described in more detail.
Fig. 1 is the main-process stream block diagram of the present invention for the Speech Feature Extraction of robust speech identification.As shown in Figure 1, a kind of Speech Feature Extraction for robust speech identification that the present invention proposes is mainly made up of following flow process: voice signal is carried out to pre-emphasis; To voice windowing and adopt short time discrete Fourier transform to ask for voice spectrum; To voice spectrum square, ask for power spectrum; Adopt bank of filters to process power spectrum, to obtain the short-time rating spectrum of noisy speech; Adopt the average mode of frame to ask for the medium duration power spectrum of noisy speech; Medium duration power spectrum to asked for noisy speech carries out asymmetric low-pass filtering treatment, follows the tracks of the noise in voice, asked for medium duration power spectrum is sheltered to processing simultaneously, obtains the short-time rating spectrum of clean speech; The ratio of the power spectrum to clean speech and noisy speech carries out passage average treatment, level and smooth to carry out; The power spectrum of the clean speech after level and smooth and noisy speech is multiplied each other with the short-time rating spectrum of the noisy speech of bank of filters output, obtain the short-time rating spectrum of clean speech; Short-time rating spectrum to clean speech is carried out energy normalized processing, to eliminate multiplicative noise; Short-time rating spectrum after normalization is carried out etc. to loudness and increase the weight of, make it meet human auditory system effect; Then by etc. the power spectrum utilization index operation of loudness after increasing the weight of carry out the conversion of intensity loudness, make it meet people's physiological characteristic; Afterwards the power spectrum after the conversion of intensity loudness is carried out to inverse fourier transform; The result obtaining according to inverse fourier transform is again asked for cepstrum coefficient; Finally the cepstrum coefficient of asking for is carried out to average normalized, finally obtain the phonetic feature of the inventive method.Below each step of invention is specifically set forth.
One, voice signal is carried out to pre-emphasis
The object of pre-emphasis is the impact that weakens low-frequency disturbance, the main composition of outstanding high-frequency signal.Conventionally use following formula to carry out pre-emphasis to speech sample point:
Wherein α is called as pre emphasis factor, and x is speech sample point, and y is the speech sample point value after pre-emphasis, the index that t is sampled point.
Two, to the voice signal windowing after described pre-emphasis and to adopt short time discrete Fourier transform to ask for voice spectrum voice signal be a continuous time varying signal, for voice are analyzed and researched, conventionally intercept one section of voice, think that voice are steady-state signals in this section, and this section of voice are called to a frame.In order to reduce truncation effect, take advantage of a window conventionally to this section of voice, common are Hanning window and hamming window.A frame voice signal after windowing is carried out to short time discrete Fourier transform and just can obtain the frequency spectrum of these frame voice.Specifically comprise: divide frame to voice, wherein frame length scope is 20ms~30ms, and it is 10ms~15ms that frame moves scope value; Each frame to voice carries out windowing, adopts Hanning window or hamming window; Voice after windowing are carried out to short time discrete Fourier transform, adopt original Fourier to change formula, or utilizing fast fourier transform to solve to the integer power of the voice zero padding to 2 after windowing obtains voice spectrum.
Three,, to voice spectrum square, ask for power spectrum
In order to obtain the power spectrum P (w) of voice signal, the real part after short time discrete Fourier transform and imaginary part are asked for respectively a square also summation by we.Formula is as follows:
P(w)=Re[S(w)]
2+Im[S(w)]
2 (2)
Wherein S (w) represents fourier spectra in short-term, Re[S (w)] and Im[S (w)] real part and the imaginary part of fourier spectra in short-term represented respectively.
Four, adopt bank of filters to process power spectrum
People's ear has different perceptions to the voice of different frequency, experiment discovery, and below 1000Hz, perception and frequency are linear, and more than 1000Hz, perception becomes logarithmic relationship with frequency.In order to simulate the apperceive characteristic of people's ear to different frequency voice, conventionally adopt bank of filters to change linear spectral.The bank of filters adopting can be Mel bank of filters (Mel-filterbank) or Gamma-tone bank of filters, and port number can be chosen different numbers according to different wave filters;
In a preferred embodiment of the invention, adopt Gamma-tone bank of filters.It has some passages, being distributed in equivalent rectangular bandwidth of the centre frequency linearity of these passages.
The short-time rating that utilizes so Gamma-tone bank of filters summation to obtain noisy speech compose as shown in the formula:
Wherein m and l represent respectively the index of frame and passage, and what K was Fourier transform counts, w
k=2 π/F
s, F
srepresent the sample frequency of voice signal.
represent voice m frame the
the amplitude of frequency,
represent l passage
the Gamma-tone filter value of frequency.
Five, adopt the average mode of frame to ask for medium duration power spectrum
Because the variation of noise is often slow than the variation of voice, therefore need to ask for a window longer than general window when estimating noise, in feature extracting method of the present invention, the average that adopts the average mode of frame to obtain several windows is described a longer window.But can not all use so long window to all voice, be because the long too large words of window can make phonetic recognization rate reduce.The formula that frame is on average asked for the medium duration power spectrum of noisy speech is expressed as follows:
Wherein m and l represent respectively the index of frame and passage, when M represents to ask for medium duration, and the frame number sum got forward and backward respectively.
Six, the medium duration power spectrum of noisy speech is carried out asymmetric filtering and shelters anti-noise
Because noise is very fast in the variation of some frequency, so for tracking noise more accurately, need to carry out not identical processing to the noise of different passages, therefore introduce herein and comprised the asymmetric low-pass filtering anti-noise module of sheltering.Idiographic flow as shown in Figure 2.
Fig. 2 is the structure flow chart that comprises the asymmetric low low-pass filtering anti-noise module of sheltering in the present invention.
In Fig. 2, first asymmetric low-pass filter can be described with following formula:
Wherein λ
aand λ
bfor adjustable parameter, span is (0~1).The Q obtaining by above formula
le[m, l] same to Q[m, l] subtract each other after, then obtain Q through a half-wave rectification block
o[m, l].Subtract each other with the concrete operations of rectification as shown in formula (6).
Q
o[m,l]=max(Q[m,l]-Q
le[m,l],0) (6)
Q
o[m, l] is sent to respectively masking block and second asymmetric low-pass filter.Second asymmetric low-pass filter is identical with first above-mentioned asymmetric low-pass filter, and inner structure still can obtain with formula (5), and just input is by Q[m, l] become Q
o[m, l], output is by Q
le[m, l] become Q
f[m, l].Second value Q that asymmetric low-pass filter produces
f[m, l] will serve as spectrum end power, the i.e. minimum value of power spectrum.Second asymmetric low-pass filter is too little and cause unnecessary music noise in order to prevent asymmetric filtering and to shelter the value of output valve of mass action.On the other hand, Q
o[m, l] obtains Q through masking block
tm[m, l], this step describes in detail in the back.Q
tm[m, l] and Q
f[m, l] is input to maximal value module jointly, utilizes following formula to obtain R
sp[m, l]:
R
sp[m,l]=max(Q
tm[m,l],Q
f[m,l]) (7)
Determine that finally by crossing selective switch result exports R[m, l] value.This optionally switch determined by following formula:
Wherein c is adjustable parameter, as selected c=2.The meaning of this formula is, if the medium duration power of a sound bite can not be greater than c=2 himself spectrum end power doubly, just thinks that this section of voice are mute state, and therefore output valve should be spectrum end power.
Processing procedure recited above is all to be realized by the asymmetric low-pass filtering anti-noise module of sheltering that comprises of describing in Fig. 2.Specifically describe the masking block in Fig. 2 below.Its structure as shown in Figure 3.First input Q
o[m, l] obtains Q through MAX module
p[m, l], formula is as follows:
Q
p[m,l]=max(λ
tQ
p[m-1,l],Q
o[m,l]) (9)
Wherein λ
tbe Forgetting coefficient, span is (0~1).The output valve Q that masking block is last
tm[m, l] determined by selective switch, and its formula is described below:
Wherein μ
tfor corresponding parameter, span is (0~1).The output valve Q of masking block
tm[m, l] is with the output Q of second dissymetrical filter
fthe maximal value module that [m, l] describes through publicity (7) has obtained R
sp[m, l], R
sp[m, l] is last with spectrum end Q
fafter the selective switch that [m, l] describes through formula (8), finally obtain the medium duration power spectrum of noisy speech to carry out asymmetric filtering and shelter the result R[m after anti-noise, l].
By description above, calculate asymmetric filtering and sheltered the output R[m after anti-noise, l], this value represents the medium duration power spectrum of clean speech, the medium duration power spectrum of noisy speech Q[m after it and frame are average, l] ratio clean speech power proportion in noisy speech power spectrum can be described, we are with H[m, l] represent.Be formulated as follows:
Seven, the average and anti-noise of passage is integrated
Because the threshold value between passage and passage is different, and to process be usually also based on a sound bite, between passage, is therefore smoothly necessary.It is average that we carry out passage with following formula, obtains passage average weight H
s[m, l]:
Wherein l
2=min (l+N, L), l
1=max (l-N, 1), L represents the number of filter channel, N represents to ask for passage average time, the total number of channels of looking around forward and backward.Described through the average weight H of passage
s[m, l] is used for modulating the short-time rating spectrum of noisy speech, and to obtain the short-time rating spectrum of clean speech, formula is as follows:
T[m,l]=P[m,l]H
s[m,l] (13)
Eight, the short-time rating spectrum of clean speech is carried out to energy normalized processing, to eliminate multiplicative noise
In traditional feature extraction algorithm such as MFCC, for matching people's physiological property, adopt logarithm operation, so just the property taken advantage of in feature extraction algorithm is operated to the noise bringing and has become the information of additivity, finally can remove by average normalization.But in feature extracting method of the present invention, adopt the operation of index to carry out matching people's physiological property, the property taken advantage of operates the noise bringing and can not eliminate by average normalization like this, therefore adds this step, in order that can eliminate this multiplicative noise.
Because feature extracting method of the present invention is online feature, therefore can not obtain the average of all frames.In the present invention, adopt the average dynamically updating to replace the average of whole piece voice, formula is as follows:
Wherein L represents the number of filter channel, λ
μrepresent Forgetting coefficient, span is (0~1).Utilize this average to be normalized the impact that just can eliminate multiplicative noise to the clean speech short-time rating spectrum of each passage.The formula of this step is as follows:
Wherein k is arbitrary constant, utilizes so online processing, can make online feature reach the effect of off-line.
Nine, the short-time rating spectrum of energy normalized clean speech after treatment being carried out etc. to loudness increases the weight of
When the sound such as different frequency, their sound pressure is different.For this deviation of compensator's ear to frequency, need to be to loudness pre-emphasis processing such as power spectrum carry out.Conventionally this passage is compensated as the frequency of this passage by the centre frequency of each passage, and that the formula compensating has is varied, in the present invention, adopt etc. loudness weight formula as follows:
Wherein w represents frequency, and l represents the index of passage, w
lthe frequency of l passage, i.e. the centre frequency of l passage.
The short-time rating spectrum of energy normalized clean speech after treatment is carried out etc. to loudness and increases the weight of to adopt formula below:
O[m,l]=U[m,l]·E(w
l) (17)
Wherein m, l is respectively the index of frame and passage.
Ten, the clean speech short-time rating spectrum of reciprocity loudness after increasing the weight of carried out index operation
For better matching people's auditory model, intensity is converted into loudness, need to carry out nonlinear compression to power spectrum, in traditional PLP feature, adopt the non-linear of cubic root; And in traditional MFCC, adopted the nonlinear mode of logarithm.In feature extracting method of the present invention, adopt the nonlinear mode of index.Formula is as follows:
L[m,l]=O[m,l]
θ (18)
Wherein θ is the nonlinear parameter of index.
11, the clean speech short-time rating spectrum after index nonlinear transformation is carried out to inverse fourier transform
It is the cepstrum coefficient in order to ask for voice signal that clean speech short-time rating spectrum after index nonlinear transformation is carried out to inverse fourier transform, and then obtains phonetic feature.The Fourier inversion has here adopted basic inverse fourier transform method.
12, ask for the cepstrum coefficient of signal
In order to obtain cepstrum coefficient, in method of the present invention, first adopt Durbin recursive algorithm to ask for linear predictor coefficient, then utilize the linear predictor coefficient of asking for, obtain corresponding cepstrum coefficient according to recursion formula.Recursion formula is as follows:
Wherein a is linear predictor coefficient, and k is reflection coefficient, and they are all to be tried to achieve according to the autocoorrelation of inverse fourier transform described in step 11 by Durbin recursive algorithm, and n is the index to spectral coefficient in addition, and p is the exponent number of model.
13, cepstrum coefficient is carried out to average normalization
Although carried out energy normalized in step 8, average normalization is still necessary, at least average normalization can not bring negative impact.Average normalization is the mean value of all dimensions of cepstrum coefficient being asked for respectively to all frames, then every one dimension of each frame cepstrum coefficient is all deducted to the average to spectral coefficient of respective dimension.Because feature extracting method of the present invention is online, therefore average is also that all frames before present frame are averaging.
Below in conjunction with accompanying drawing, the example of a kind of Speech Feature Extraction for robust speech identification of the present invention is described, for the voice of 16KHz sample frequency, be specifically described as follows.
1. pair voice signal carries out pre-emphasis, increases the weight of factor alpha and adopts 0.97.System function is as shown in formula (1).
2. voice frame length adopts 25ms, and frame moves and adopts 10ms, adds hamming window, and by 400 afterbody zero padding to 512 points of frame voice, then adopts fast fourier transform to ask for voice spectrum.
3. utilize the voice spectrum of asking for, ask for phonetic speech power spectrum according to formula (2).
4. adopt Gamma-tone bank of filters to process power spectrum, passage number adopts 40, and the formula of employing is as shown in (3).
5. adopt the average mode of frame to ask for the power spectrum of medium duration, computing formula formula described above (4), wherein M=2, utilize present frame and before it two frames and below the average power of two frames replace the medium duration power of former single frame as the medium duration power of noisy speech, this duration is [(2M+1)-1] * 10ms+25ms=65ms.
6. pair power spectrum carries out asymmetric filtering processing, follows the tracks of the noise in voice, power spectrum is sheltered to processing simultaneously, obtains clean speech power spectrum.In this step, calculate according to the formula in implementation method comprising the asymmetric low-pass filtering anti-noise module of sheltering.Formula used has (5,6,7,8,9,10,11), and wherein the value of design parameter is described below:
λ
a=0.999,λ
b=0.5
c=2
λ
t=0.85,μ
t=0.2
7. it is average that the ratio of pair clean speech and noisy speech power spectrum carries out passage, level and smooth to carry out, the formula adopting is (12), wherein N=4 (port number of looking around respectively forward and is backward 4), and the value that is about to 9 passages is carried out smoothly.The power spectrum of the clean speech after level and smooth and noisy speech is multiplied each other with the short-time rating spectrum of the noisy speech of bank of filters output, i.e. formula (13), the short-time rating that obtains clean speech is composed.
8. the short-time rating of pair clean speech spectrum is carried out energy normalized processing, to eliminate multiplicative noise, as shown in formula (15).Wherein average adopts dynamic estimation, and as shown in formula (14), initial value obtains from data centralization statistics.
9. power spectrum the loudness such as carries out and increases the weight of, and makes it meet human auditory system effect.
10. pair power spectrum carries out index operation, makes its physiological characteristic that meets people, and index nonlinear parameter θ chooses 1/15 here.
11. pairs of power spectrum carry out inverse fourier transform, can utilize the fundamental formular of inverse fourier transform to calculate, because itself count just seldom, calculated amount is little.
12. while asking for the cepstrum coefficient of signal, and the linear predictor coefficient of choosing is 12, and cepstrum coefficient is also 12, and formula is as shown in (17).
13. pairs of cepstrum coefficients carry out average normalized, finally obtain the phonetic feature of the inventive method.
The described feature extracting method that the present invention proposes and the Contrast on effect of common feature extracting method:
Utilize feature extracting method of the present invention on 863 desktop voice collection, to extract the feature of voice, use the phonetic feature of senior anti-noise leading portion (AFE) the feature extraction 863 desktop voice collection of PLP feature extracting method and ETSI (ETSI) simultaneously.Utilize this three feature sets, under identical condition, adopt respectively HTK instrument training acoustic model.Then, choose 1000 and purely read aloud recording, add simulating the white noise, then utilize three kinds of above-mentioned feature extracting methods to extract respectively feature.In addition, a random recording collection of talking is marked, obtain 7072 pure recording and 360 band Recording Noises, still use three kinds of above-mentioned feature extracting methods to extract respectively phonetic feature.
Utilize above-mentioned acoustic model and its characteristic of correspondence to carry out speech recognition, language model all adopts same 3 gram language model, and recognizer adopts the demoder in HTK instrument.Here adopt Word Error Rate (WER) to assess speech recognition performance, wherein, PNPLP is the title of feature extraction algorithm of the present invention.The computing formula of WER is as follows:
Under the test condition of simulating the white noise, the performance of various features is as shown in table 1.In table 1, can find out not have in noisy clean speech situation, PLP mark sheet reveals good performance, but along with the increasing of noise, PLP performance variation gradually.The anti-noise feature (AFE) of ETSI can show certain effect on noise, but feature extracting method noiseproof feature of the present invention will be far superior to the anti-noise algorithm of ETSI.
Table 2 is various feature extraction algorithms experimental results on real test set.From subordinate list, can find out, feature extracting method noiseproof feature of the present invention is outstanding, will get well much than the noiseproof feature of ETSI.In addition, anti-noise feature extraction algorithm of the present invention has slightly and declines with respect to classical PLP algorithm on clean speech collection, but contrasts with the anti-noise algorithm of ETSI, and anti-noise feature extraction algorithm of the present invention is still well a lot.
Table 1
Table 2
WER | PLP | AFE | PNPLP |
clear | 11.64% | 13.68% | 12.07% |
noise | 35.89% | 34.21% | 33.36% |
Above-described specific embodiment; object of the present invention, technical scheme and beneficial effect are further described; institute is understood that; the foregoing is only specific embodiments of the invention; be not limited to the present invention; within the spirit and principles in the present invention all, any amendment of making, be equal to replacement, improvement etc., within all should being included in protection scope of the present invention.
Claims (9)
1. for a Speech Feature Extraction for robust speech identification, it is characterized in that, the method comprises:
Step 1, obtain the power spectrum of voice signal;
Step 2, by obtained power spectrum by bank of filters processing, obtain the short-time rating spectrum of noisy speech;
Step 3, according to the short-time rating spectrum of obtained noisy speech, adopt the average mode of frame to ask for the medium duration power spectrum of noisy speech;
Step 4, the medium duration power spectrum of obtained noisy speech is carried out asymmetric filtering and shelters anti-noise, to obtain the medium duration power spectrum of clean speech;
Step 5, obtain the short-time rating spectrum of clean speech according to the short-time rating spectrum of the medium duration power spectrum of the medium duration power spectrum of described clean speech, noisy speech and noisy speech;
Step 6, the short-time rating spectrum of clean speech is carried out to energy normalized processing, to eliminate the property taken advantage of noise;
Step 7, offset and the loudness such as carry out except the short-time rating spectrum of the clean speech of the property taken advantage of noise and increase the weight of;
The short-time rating spectrum of the clean speech after step 8, reciprocity loudness increase the weight of is carried out index nonlinear operation;
Step 9, the short-time rating spectrum of having carried out the clean speech after index nonlinear operation is carried out to inverse fourier transform, to ask for cepstrum coefficient, cepstrum coefficient is carried out to average normalized, finally obtain phonetic feature;
Wherein, the frequency spectrum that obtains voice signal described in step 1 further comprises following content:
Step 11, to voice signal adopt formula (1) carry out pre-emphasis:
Wherein α is called as pre emphasis factor, and x is speech sample point, and y is the speech sample point value after pre-emphasis, the index that t is sampled point;
Step 12, each frame of the voice after pre-emphasis is carried out to windowing, adopt Hanning window or hamming window, voice after windowing are carried out to short time discrete Fourier transform, comprise and adopt original Fourier to change formula, or utilize fast fourier transform to solve to the integer power of the voice zero padding to 2 after windowing;
Step 13, the real part after Short Time Fourier Transform and imaginary part are asked for respectively square and summation, to obtain the power spectrum of voice signal, as shown in formula (2):
P(w)=Re[S(w)]
2+Im[S(w)]
2 (2)
The power spectrum that wherein P (w) is voice signal, S (w) represents fourier spectra in short-term, Re[S (w)] and Im[S (w)] real part and the imaginary part of fourier spectra in short-term represented respectively.
2. the Speech Feature Extraction for robust speech identification according to claim 1, it is characterized in that, the bank of filters adopting in described step 2 is Mel bank of filters Mel-filter bank or Gamma-tone bank of filters, port number is chosen different numbers according to different wave filters, wherein, the short-time rating that utilizes the summation of Gamma-tone bank of filters to obtain noisy speech is composed, as shown in formula (3):
Wherein p[m, l] be the short-time rating spectrum of noisy speech, m and l represent respectively the index of frame and bank of filters passage, what K was Fourier transform counts, w
k=2 π/F
s, F
srepresent the sample frequency of voice signal,
represent voice signal m frame the
the amplitude of frequency,
represent l passage
the Gamma-tone filter value of frequency.
3. the Speech Feature Extraction for robust speech identification according to claim 1, is characterized in that, adopts the average mode of frame to ask for the medium duration power spectrum of noisy speech in step 3, as shown in formula (4):
Wherein Q[m, l] be the medium duration power spectrum of noisy speech, m and l represent respectively the index of frame and bank of filters passage, when M represents to ask for medium duration, the frame number sum got forward and backward respectively, P[m ', l] be the noisy speech short-time rating spectrum of the m ' frame.
4. the Speech Feature Extraction for robust speech identification according to claim 1, is characterized in that, described in step 4, the medium duration power spectrum of obtained noisy speech is carried out asymmetric filtering and shelter anti-noise specifically comprising following step:
Step 41, medium obtained noisy speech duration power spectrum is carried out to filtering through first asymmetric low-pass filter, and the Output rusults that medium described noisy speech duration power spectrum is deducted to described first asymmetric low-pass filter is to integrate; Wherein said first asymmetric low-pass filter as formula (5) represent:
Wherein m and l represent respectively the index of frame and bank of filters passage, Q
le[m, l] is the output of described first asymmetric low-pass filter; Q[m, l] be the medium duration power spectrum of described noisy speech, λ
aand λ
bfor adjustable parameter, span is (0~1);
Step 42, the result after described integration is obtained to Q through a half-wave rectification block
o[m, l], by Q
o[m, l] sends into respectively masking block and second asymmetric low-pass filter processed, and described second asymmetric low-pass filter is identical with first asymmetric low-pass filter, and the output of second dissymetrical filter is as spectrum end power; The Q that wherein half-wave rectification block obtains
othe formula of [m, l] is expressed as follows:
Q
o[m,l]=max(Q[m,l]-Q
le[m,l],0) (6)
Step 43, the described Q obtaining through half-wave rectification block
o[m, l] obtains result Q after masking block is processed
tm[m, l], and Q
o[m, l] obtains result Q after second asymmetric low-pass filter processed
f[m, l], afterwards, described Q
tm[m, l] and described Q
f[m, l] is input to maximal value module and obtains result R
sp[m, l]; Wherein maximal value module is as shown in formula (7):
R
sp[m,l]=max(Q
tm[m,l],Q
f[m,l]) (7)
Step 44, determined the medium duration power spectrum R[m of clean speech by the first selective switch, l], described the first selective switch is as shown in formula (8):
Wherein c is adjustable parameter.
5. the Speech Feature Extraction for robust speech identification according to claim 4, is characterized in that, the operating process of described masking block comprises following content:
The described Q obtaining through half-wave rectification block
o[m, l] obtains Q through the MAX module of masking block
p[m, l], as shown in formula (9):
Q
p[m,l]=max(λ
tQ
p[m-1,l],Q
o[m,l]) (9)
Wherein λ
tbe Forgetting coefficient, span is (0~1), the output valve Q that masking block is last
tm[m, l] determined by the second selective switch, and described the second selective switch is as shown in formula (10):
Wherein μ
tfor corresponding parameter, span is (0~1), the output valve Q of masking block
tmafter the selective switch that [m, l] describes through formula (8), final, the medium duration power spectrum of noisy speech is carried out to asymmetric filtering and the result of sheltering after anti-noise is R[m, l].
6. the Speech Feature Extraction for robust speech identification according to claim 1, is characterized in that, the short-time rating spectrum of obtaining clean speech in described step 5 comprises following content:
Step 51, calculates the ratio H[m of the medium duration power spectrum of obtained clean speech and the medium duration power spectrum of noisy speech, l], as shown in formula (11):
Wherein said R[m, l] be the medium duration power spectrum of clean speech, Q[m, l] be the medium duration power spectrum of noisy speech;
Step 52, carries out passage average, to obtain passage average weight H
s[m, l], as shown in formula (12):
Wherein l
2=min (l+N, L), l
1=max (l-N, 1), L represents the number of filter channel, N represents to ask for passage average time, the total number of channels of looking around forward and backward;
Step 53, utilizes described passage average weight H
sthe short-time rating spectrum P[m of [m, l] modulation noisy speech, l], and obtain the short-time rating spectrum T[m of clean speech, l], as shown in formula (13):
T[m,l]=P[m,l]H
s[m,l] (13)。
7. the Speech Feature Extraction for robust speech identification according to claim 1, is characterized in that, the short-time rating spectrum to clean speech in described step 6 is carried out energy normalized processing, as shown in formula (15):
Wherein k is arbitrary constant, T[m, l] be the short-time rating spectrum of clean speech, μ [m] is as shown in formula (14):
Wherein L represents the number of filter channel, λ
μrepresent Forgetting coefficient, span is (0~1).
8. the Speech Feature Extraction for robust speech identification according to claim 1, is characterized in that, in described step 7, the described spectrum of the short-time rating to clean speech the loudness such as is carried out and increased the weight of as shown in formula (17):
O[m,l]=U[m,l]·E(w
l) (17)
Wherein m, l is respectively the index of frame and passage, U[m, l] for voice are through anti-noise short-time rating spectrum after treatment, E (w
l) as shown in formula (16):
Wherein w represents frequency, w
lit is the frequency of l passage.
9. the Speech Feature Extraction for robust speech identification according to claim 1, it is characterized in that, in described step 8, the short-time rating spectrum of the clean speech after described reciprocity loudness increases the weight of is carried out index nonlinear operation as shown in formula (18):
L[m,l]=O[m,l]
θ (18)
Wherein θ is the nonlinear parameter of index, O[m, l] be the short-time rating spectrum of the clean speech of the loudness such as described after increasing the weight of.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201210449436.XA CN102982801B (en) | 2012-11-12 | 2012-11-12 | Phonetic feature extracting method for robust voice recognition |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201210449436.XA CN102982801B (en) | 2012-11-12 | 2012-11-12 | Phonetic feature extracting method for robust voice recognition |
Publications (2)
Publication Number | Publication Date |
---|---|
CN102982801A CN102982801A (en) | 2013-03-20 |
CN102982801B true CN102982801B (en) | 2014-12-10 |
Family
ID=47856712
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201210449436.XA Active CN102982801B (en) | 2012-11-12 | 2012-11-12 | Phonetic feature extracting method for robust voice recognition |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN102982801B (en) |
Families Citing this family (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104091593B (en) * | 2014-04-29 | 2017-02-15 | 苏州大学 | Speech endpoint detection algorithm adopting perceptual speech spectrum structure boundary parameters |
CN105989835B (en) * | 2015-02-05 | 2019-08-13 | 宏碁股份有限公司 | Voice identification apparatus and speech identifying method |
CN104835495B (en) * | 2015-05-30 | 2018-05-08 | 宁波摩米创新工场电子科技有限公司 | A kind of high definition speech recognition system based on low-pass filtering |
CN105390134B (en) * | 2015-10-20 | 2019-01-11 | 河海大学 | A kind of model self-adapting method based on subband VTS |
CN105578115B (en) * | 2015-12-22 | 2016-10-26 | 深圳市鹰硕音频科技有限公司 | A kind of Network teaching method with Speech Assessment function and system |
CN105632501B (en) * | 2015-12-30 | 2019-09-03 | 中国科学院自动化研究所 | A kind of automatic accent classification method and device based on depth learning technology |
CN106653048B (en) * | 2016-12-28 | 2019-10-15 | 云知声(上海)智能科技有限公司 | Single channel sound separation method based on voice model |
CN107346659B (en) * | 2017-06-05 | 2020-06-23 | 百度在线网络技术(北京)有限公司 | Speech recognition method, device and terminal based on artificial intelligence |
CN107845389B (en) * | 2017-12-21 | 2020-07-17 | 北京工业大学 | Speech enhancement method based on multi-resolution auditory cepstrum coefficient and deep convolutional neural network |
CN108694951B (en) * | 2018-05-22 | 2020-05-22 | 华南理工大学 | Speaker identification method based on multi-stream hierarchical fusion transformation characteristics and long-and-short time memory network |
CN108922514B (en) * | 2018-09-19 | 2023-03-21 | 河海大学 | Robust feature extraction method based on low-frequency log spectrum |
CN109256127B (en) * | 2018-11-15 | 2021-02-19 | 江南大学 | Robust voice feature extraction method based on nonlinear power transformation Gamma chirp filter |
IT201900016328A1 (en) * | 2019-09-13 | 2021-03-13 | Elenos S R L | METHOD FOR MEASURING AND DISPLAYING THE SIGNAL / AUDIO NOISE RATIO |
CN110556125B (en) * | 2019-10-15 | 2022-06-10 | 出门问问信息科技有限公司 | Feature extraction method and device based on voice signal and computer storage medium |
CN111667838B (en) * | 2020-06-22 | 2022-10-14 | 清华大学 | Low-power-consumption analog domain feature vector extraction method for voiceprint recognition |
CN113077787A (en) * | 2020-12-22 | 2021-07-06 | 珠海市杰理科技股份有限公司 | Voice data identification method, device, chip and readable storage medium |
CN112738338B (en) * | 2020-12-25 | 2022-10-14 | 平安科技(深圳)有限公司 | Telephone recognition method, device, equipment and medium based on deep learning |
CN113270097B (en) * | 2021-05-18 | 2022-05-17 | 成都傅立叶电子科技有限公司 | Unmanned mechanical control method, radio station voice instruction conversion method and device |
CN117690421B (en) * | 2024-02-02 | 2024-06-04 | 深圳市友杰智新科技有限公司 | Speech recognition method, device, equipment and medium of noise reduction recognition combined network |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE102008039329A1 (en) * | 2008-01-25 | 2009-07-30 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | An apparatus and method for calculating control information for an echo suppression filter and apparatus and method for calculating a delay value |
-
2012
- 2012-11-12 CN CN201210449436.XA patent/CN102982801B/en active Active
Also Published As
Publication number | Publication date |
---|---|
CN102982801A (en) | 2013-03-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN102982801B (en) | Phonetic feature extracting method for robust voice recognition | |
Valin et al. | A perceptually-motivated approach for low-complexity, real-time enhancement of fullband speech | |
CN108831499B (en) | Speech enhancement method using speech existence probability | |
CN105513605B (en) | The speech-enhancement system and sound enhancement method of mobile microphone | |
CN108447495B (en) | Deep learning voice enhancement method based on comprehensive feature set | |
Kim et al. | Nonlinear enhancement of onset for robust speech recognition. | |
CN109256127B (en) | Robust voice feature extraction method based on nonlinear power transformation Gamma chirp filter | |
CN103544961B (en) | Audio signal processing method and device | |
WO2001033550A1 (en) | Speech parameter compression | |
CN103021405A (en) | Voice signal dynamic feature extraction method based on MUSIC and modulation spectrum filter | |
Shi et al. | Robust speaker recognition based on improved GFCC | |
Wang et al. | Joint noise and mask aware training for DNN-based speech enhancement with sub-band features | |
CN103400578B (en) | Anti-noise voiceprint recognition device with joint treatment of spectral subtraction and dynamic time warping algorithm | |
Maganti et al. | Auditory processing-based features for improving speech recognition in adverse acoustic conditions | |
CN112863517B (en) | Speech recognition method based on perceptual spectrum convergence rate | |
Alam et al. | Robust feature extraction for speech recognition by enhancing auditory spectrum | |
CN112233657A (en) | Speech enhancement method based on low-frequency syllable recognition | |
Bahadur et al. | Performance measurement of a hybrid speech enhancement technique | |
Rao et al. | Speech enhancement using sub-band cross-correlation compensated Wiener filter combined with harmonic regeneration | |
Flynn et al. | Combined speech enhancement and auditory modelling for robust distributed speech recognition | |
Laaksonen et al. | Artificial bandwidth expansion method to improve intelligibility and quality of AMR-coded narrowband speech | |
TWI749547B (en) | Speech enhancement system based on deep learning | |
Upadhyay et al. | Robust recognition of English speech in noisy environments using frequency warped signal processing | |
Maganti et al. | A perceptual masking approach for noise robust speech recognition | |
Mehta et al. | Robust front-end and back-end processing for feature extraction for Hindi speech recognition |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant |