WO2004040555A1 - 音声強調装置 - Google Patents
音声強調装置 Download PDFInfo
- Publication number
- WO2004040555A1 WO2004040555A1 PCT/JP2002/011332 JP0211332W WO2004040555A1 WO 2004040555 A1 WO2004040555 A1 WO 2004040555A1 JP 0211332 W JP0211332 W JP 0211332W WO 2004040555 A1 WO2004040555 A1 WO 2004040555A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- vocal tract
- unit
- spectrum
- formant
- characteristic
- Prior art date
Links
- 230000001755 vocal effect Effects 0.000 claims abstract description 119
- 230000003321 amplification Effects 0.000 claims abstract description 89
- 238000003199 nucleic acid amplification method Methods 0.000 claims abstract description 89
- 238000004364 calculation method Methods 0.000 claims abstract description 40
- 238000012937 correction Methods 0.000 claims abstract description 30
- 238000000926 separation method Methods 0.000 claims abstract description 14
- 230000002194 synthesizing effect Effects 0.000 claims abstract description 14
- 230000005236 sound signal Effects 0.000 claims abstract description 12
- 238000000605 extraction Methods 0.000 claims abstract description 10
- 238000001228 spectrum Methods 0.000 claims description 168
- 238000005311 autocorrelation function Methods 0.000 claims description 36
- 230000015572 biosynthetic process Effects 0.000 claims description 24
- 238000003786 synthesis reaction Methods 0.000 claims description 24
- 239000002131 composite material Substances 0.000 claims description 12
- 238000004458 analytical method Methods 0.000 claims description 7
- 239000000284 extract Substances 0.000 claims description 6
- 238000010586 diagram Methods 0.000 description 39
- 238000000034 method Methods 0.000 description 39
- 238000012545 processing Methods 0.000 description 15
- 230000001965 increasing effect Effects 0.000 description 13
- 238000007796 conventional method Methods 0.000 description 8
- 230000003595 spectral effect Effects 0.000 description 8
- 230000002708 enhancing effect Effects 0.000 description 7
- 230000006870 function Effects 0.000 description 7
- 238000005516 engineering process Methods 0.000 description 4
- 238000005070 sampling Methods 0.000 description 4
- 238000004891 communication Methods 0.000 description 3
- 102000001690 Factor VIII Human genes 0.000 description 2
- 108010054218 Factor VIII Proteins 0.000 description 2
- 238000005314 correlation function Methods 0.000 description 2
- 230000002349 favourable effect Effects 0.000 description 2
- 240000004050 Pentaglottis sempervirens Species 0.000 description 1
- 235000004522 Pentaglottis sempervirens Nutrition 0.000 description 1
- 230000002238 attenuated effect Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000012886 linear function Methods 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 230000001737 promoting effect Effects 0.000 description 1
- 238000012887 quadratic function Methods 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 210000001260 vocal cord Anatomy 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/04—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
- G10L19/06—Determination or coding of the spectral characteristics, e.g. of the short-term prediction coefficients
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0316—Speech enhancement, e.g. noise reduction or echo cancellation by changing the amplitude
- G10L21/0364—Speech enhancement, e.g. noise reduction or echo cancellation by changing the amplitude for improving intelligibility
Definitions
- the present invention relates to a voice emphasizing device that makes it easy to hear a received voice of a mobile phone or the like in an environment where there is ambient background noise.
- the easiest way to make the received voice easier to hear in a noisy environment is to increase the receiving volume according to the noise level.
- the receiving volume is set too high, the input to the speaker of the mobile phone becomes excessive and the sound is distorted, which may degrade the sound quality.
- increasing the volume of the received sound increases the burden on the hearing of the listener (user) and is not favorable for health.
- Figure 1 shows an example of a voice frequency spectrum.
- Figure 1 shows the case where there are three peaks (formants) in the spectrum.
- the first formant, second formant, and third formant are called in order from the lowest frequency, and the peak frequency of each formant (1), (2), or (3) is called the formant frequency.
- amplitude (power) of a speech spectrum tends to decrease as the frequency increases.
- speech intelligibility is closely related to formants, and emphasis on higher (second and third) formants can improve speech intelligibility. It is known.
- Fig. 2 shows an example of spectrum enhancement.
- the solid line in FIG. 2 (a) and the dotted line in FIG. 2 (b) represent the speech spectrum before emphasis.
- the solid line in FIG. 2 (b) shows the speech spectrum after emphasis.
- the slope of the entire spectrum is flattened by increasing the amplitude of the higher-order formants, and this can improve the intelligibility of the entire speech.
- FIG. 3 shows a block diagram of this conventional technique.
- the spectrum of the input voice is obtained by a spectrum estimating unit 100, and the convex band (peak) and the concave band (valley) determining unit 101 are used to determine the convex band and the concave portion from the obtained spectrum. Find the band, and calculate the amplification factor (or attenuation factor) for the convex band and the concave band.
- a coefficient for realizing the amplification rate (or attenuation rate) is provided to the filter section 103 by the filter section 102, and the input voice is input to the filter section 103 to obtain a spectrum. Achieve emphasis. .
- voice enhancement is realized by individually amplifying and attenuating peaks and valleys of the voice spectrum.
- Figure 4 shows a model for speech generation.
- the sound source signal generated by the sound source (vocal cord) 110 is input to the articulatory system (vocal tract) 111, and after the vocal tract characteristics are added to the vocal tract 11
- the speech is output as a speech waveform from the lips 112 (see “Highly efficient speech coding”, pp. 69-71, written by Kazuo Nakata, Morikita Publishing).
- Figures 5 and 6 show examples.
- Figure 5 shows the input speech spectrum before emphasis processing.
- Fig. 6 shows the spectrum when the input speech in Fig. 5 is emphasized by the method using the band division filter.
- the amplitude of high-frequency components above 2 kHz is amplified while maintaining the general shape of the spectrum.
- the portion between 500 Hz and 2 kHz is significantly different from the vector in Fig. 5 before emphasis, indicating that the sound source characteristics are broken.
- the distortion of the sound source characteristic is increased and the sound quality may be deteriorated.
- Figures 7 and 8 show that higher-order formants are emphasized. However, at around 0.95 seconds and around 1.03 seconds in Fig. 8, there is a discontinuity in the spectrum after enhancement. That is, in the spectrum before emphasis in FIG. 7, the formant frequency changes smoothly, but in FIG. 8, the formant changes discontinuously. This formant discontinuity is perceived as noise when the processed speech is actually heard.
- An object of the present invention has been devised in view of the above-described problems in the conventional technology.
- a speech enhancement device that achieves the object of the present invention includes, as a first aspect, a signal separation unit that separates an input speech signal into sound source characteristics and vocal tract characteristics, and a feature that extracts characteristic information from the vocal tract characteristics.
- An extraction unit a corrected vocal tract characteristic calculation unit that obtains vocal tract characteristic correction information from the vocal tract characteristics and the characteristic information, and a vocal tract characteristic correction unit that corrects the vocal tract characteristics using the vocal tract characteristic correction information.
- a signal synthesizing unit for synthesizing the modified vocal tract characteristics from the vocal tract characteristic correction unit and the sound source characteristics; It is characterized by force.
- a speech enhancement device that achieves the above object of the present invention includes, as a second aspect, an autocorrelation calculation unit that obtains an autocorrelation function from input speech of a current frame; and an autocorrelation unit that stores the autocorrelation of the current frame.
- a buffer unit that outputs a correlation function, an average autocorrelation calculating unit that calculates a weighted average of the autocorrelation of the current frame and the autocorrelation function of the past frame, and calculates an inverse filter coefficient from the weighted average of the autocorrelation function
- a first filter coefficient calculator an inverse filter configured by the inverse filter coefficient, a spectrum calculator that calculates a frequency spectrum from the inverse filter coefficient, and the calculated frequency spectrum
- a formant estimator for estimating a formant frequency and a formant amplitude from a vector; the calculated frequency spectrum;
- An amplification factor calculating unit that obtains an amplification factor from the gain frequency and the formant amplitude; a spectrum emphasis unit that changes the calculated frequency spectrum based on the amplification factor, and obtains a changed frequency spectrum;
- a second filter coefficient calculating unit that calculates a synthesis filter coefficient from the changed frequency spectrum; and a synthesis filter including the synthesis filter coefficient, and inputs the input voice to the inverse filter
- a speech enhancement apparatus that achieves the above object of the present invention includes, as a third aspect, a linear prediction coefficient analysis unit that analyzes a linear prediction coefficient of an input speech signal of a current frame to obtain an autocorrelation function and a linear prediction coefficient, An inverse filter composed of coefficients, a first spectrum calculating unit for obtaining a frequency spectrum from the linear prediction coefficients, and storing an autocorrelation of the current frame and outputting an autocorrelation function of a past frame A buffer unit, an average autocorrelation calculating unit for obtaining a weighted average of the autocorrelation function of the current frame and the autocorrelation function of the past frame, and a first filter for calculating an average filter coefficient from the weighted average of the autocorrelation function.
- Frequency spectrum A spectrum emphasizing unit that obtains a changed frequency spectrum, a second filter coefficient calculating unit that calculates a synthetic filter coefficient from the changed frequency spectrum, and The input signal is input to the inverse filter to obtain a residual signal, and the residual signal is input to the synthetic filter to obtain an output sound.
- a voice emphasizing apparatus for achieving the above object, comprising: an autocorrelation calculating unit for obtaining an autocorrelation function from an input voice of a current frame; storing an autocorrelation of the current frame; A buffer for outputting a correlation function; an average for calculating a weighted average of the autocorrelation of the current frame and the autocorrelation function of the past frame; and an autocorrelation calculator for calculating an inverse filter coefficient from the weighted average of the autocorrelation function.
- a first filter coefficient calculating unit to calculate, an inverse filter coefficient configured by the inverse filter coefficient, a spectrum calculating unit to calculate a frequency spectrum from the inverse filter coefficient, and the frequency spectrum
- a formant estimating unit for estimating a formant frequency and a formant amplitude from a vector, the frequency spectrum, the formant frequency and the formant amplitude
- a temporary gain calculating unit that calculates a temporary gain of the current frame from the current frame, a difference calculating unit that calculates a differential gain between the temporary gain and the gain of the previous frame, and a difference calculated from a predetermined threshold value.
- a spectrum emphasizing unit that has an amplification factor determining unit that changes the frequency spectrum based on the amplification factor of the current frame, and obtains a changed frequency spectrum.
- a second filter coefficient calculator for calculating a composite filter coefficient from the frequency spectrum; a composite filter composed of the composite filter coefficient; and a pitch enhancement calculating a pitch enhancement coefficient from the residual signal. Coefficient calculation And a pitch emphasis filter configured by the pitch emphasis coefficient. The input voice is input to the inverse filter, a residual signal is obtained, and the residual signal is input to the pitch emphasis filter. To obtain a residual signal in which the pitch periodicity is emphasized, and input the pre-emphasized residual signal to the synthesized filter to obtain an output voice.
- a voice emphasizing apparatus for achieving the above object.
- An enhancement filter that enhances a part of the frequency band of the signal, a signal separation unit that separates the input audio signal enhanced by the enhancement filter into a sound source characteristic and a vocal tract characteristic, and extracts feature information from the vocal tract characteristic
- a vocal tract characteristic correction unit for obtaining vocal tract characteristic correction information from the vocal tract characteristic and the characteristic information; and a vocal tract characteristic correction for correcting the vocal tract characteristic using the vocal tract characteristic correction information.
- a signal synthesizing unit for synthesizing the modified vocal tract characteristics from the vocal tract characteristic correcting unit and the sound source characteristics, and outputs a voice synthesized by the signal synthesizing unit.
- a voice emphasizing device that achieves the above object is a signal separating unit that separates an input voice signal into a sound source characteristic and a vocal tract characteristic, and a feature that extracts characteristic information from the vocal tract characteristic.
- a signal synthesizer for combining the modified vocal tract characteristics from the vocal tract characteristic corrector with the sound source characteristics; and a filter for enhancing a partial frequency band of the signal synthesized by the signal synthesizer. It is characterized by the following.
- FIG. 1 is a diagram illustrating an example of a frequency spectrum of speech.
- FIG. 2 is a diagram illustrating an example of a frequency spectrum of a voice before and after emphasis.
- FIG. 3 shows a block diagram of a prior art described in Japanese Patent Application Publication No. 2000-1-177573.
- FIG. 4 is a diagram showing a speech generation model.
- FIG. 5 is a diagram illustrating an example of the spectrum of the input voice.
- FIG. 6 is a diagram illustrating an example of a spectrum when emphasized in a frame unit.
- FIG. 7 is a diagram showing the spectrum of the input voice (before emphasis).
- FIG. 8 is a diagram showing an audio spectrum when the spectrum is emphasized in frame units.
- FIG. 9 is a diagram showing a principle diagram of the present invention.
- FIG. 10 is a diagram showing a configuration block of the first embodiment of the present invention.
- FIG. 11 is a flowchart showing the processing of the amplification factor calculating section 6 in the embodiment of FIG.
- FIG. 12 is a diagram showing a state when the amplitude of the formant E () is adjusted to the reference power Pmv-re / in the embodiment of FIG.
- FIG. 13 is a diagram for explaining that the amplification factor at the frequency between formants is obtained by an interpolation curve R (,).
- FIG. 14 is a diagram showing a configuration block of a second embodiment of the present invention.
- FIG. 15 is a diagram showing a configuration block of a third embodiment of the present invention.
- FIG. 16 is a diagram showing a configuration block of a fourth embodiment of the present invention.
- FIG. 17 is a diagram showing a configuration walk of the fifth embodiment of the present invention.
- FIG. 18 is a diagram showing a configuration block of a sixth embodiment of the present invention.
- FIG. 19 is a diagram showing a spectrum emphasized according to the present invention.
- FIG. 20 is a diagram illustrating the principle of the present invention for solving the problem that the sense of noise increases when the amplification factor greatly changes between frames.
- FIG. 21 is a diagram showing another principle of the present invention, which solves the problem that the sense of noise increases when the transmission rate greatly changes between frames.
- FIG. 22 is a diagram showing a configuration block of an embodiment of the present invention according to the principle diagram of FIG. BEST MODE FOR CARRYING OUT THE INVENTION
- FIG. 9 is a diagram showing the principle of the present invention.
- the separating unit 20 separates an input voice into a sound source characteristic and a vocal tract characteristic, emphasizes the sound source characteristic and the vocal tract characteristic individually, and thereafter,
- the synthesizing unit 21 synthesizes and outputs. The processing in FIG. 9 is described below.
- the average spectrum calculation unit 1 of the unit 20 calculates the average spectrum ((0 ⁇ / ⁇ N f )).
- the average spectrum calculation unit 1 which is a linear prediction circuit first calculates the autocorrelation function of the current frame.
- an average autocorrelation is obtained by a weighted average of the autocorrelation function of the current frame and the autocorrelation function of the past frame. From the average autocorrelation, the average spectrum (/ ⁇ ) is obtained.
- N f is the number of data points in the spectrum, and N ⁇ N F.
- the input speech is input to the inverse filter 3 in the separation unit 20 composed of the obtained inverse filter coefficients ⁇ ), and the residual signal? ( ⁇ ), ( ⁇ w ⁇ N) is determined.
- the input speech can be separated into a residual signal r (which is a sound source characteristic) and a spectrum ⁇ (/) which is a vocal tract characteristic.
- the residual signal r (is input to the pitch emphasizing unit 4, where a residual signal whose pitch periodicity is emphasized is obtained.
- the spectrum s Pl (l) which is a vocal tract characteristic, is input to the formant estimating unit 5 as a feature extracting unit, and the formant frequency or (), (l ⁇ A: ⁇ A: max ) and the formant amplitude flm; Estimate? (:), (l ⁇ : ⁇ A max ).
- max is the number of formants to be estimated.
- A: The value of max is arbitrary, but max 4 or 5 for audio with a sampling frequency of 8 kHz.
- spectrum iota (Ri and formant or frequency () and enter the formant amplitude amplification factor calculating unit 6 calculates the amplification factor (Ri with respect to the scan Bae spectrum iota (Zeta).
- the spectrum ⁇ (R) and the amplification factor (R are input to the spectrum emphasizing unit 7 to obtain the emphasized vector 2 (R.
- the enhanced spectrum 2 (R Input to the second filter coefficient calculator 8 for calculating the coefficient of 9 to obtain the combined filter coefficient ⁇ 2 ( ⁇ ⁇ ), (1 ⁇ ⁇ ⁇ ;? 2 ) where 2 is the combined filter 9 This is the order of the filter evening.
- the residual signal after pitch emphasis by the pitch emphasizing unit 4 is input to the synthesis filter coefficient ⁇ 2 (synthesis filter 9 composed of a filter), and output speech y ("), (0 ⁇ w ⁇ N) is obtained.
- the emphasized sound source characteristics and the vocal tract characteristics are synthesized.
- the input speech is separated into the sound source characteristics (residual signal) and the vocal tract characteristics (spectral envelope)
- enhancement processing suitable for each characteristic can be performed.
- the voice periodicity can be improved by enhancing the pitch periodicity for the sound source characteristics and the formant for the vocal tract characteristics.
- FIG. 10 is a block diagram of the configuration of the first embodiment according to the present invention.
- the average spectrum calculation unit 1 in the separation unit 20 is divided before and after the filter coefficient calculation unit 2, and the input sound of the current frame is input before the filter coefficient calculation unit 2.
- the signal (, (0 ⁇ « ⁇ N) is input to the autocorrelation calculator 10, and the autocorrelation function c () (), (0 ⁇ ) of the current frame is calculated by equation (1).
- N is the frame length, is the frame number of the current frame, and is the order of the inverse filter 3 described later.
- the separation unit 20 outputs the auto-correlation function flc (m-)), (1 ⁇ , 0 ⁇ /) in the L frame immediately before from the buffer unit 11.
- the autocorrelation calculation unit The autocorrelation function of the current frame obtained in step 10 "c (m)" From the previous autocorrelation, the average autocorrelation calculating section 12 obtains the average autocorrelation c AE (r).
- the method of obtaining the average autocorrelation i? C AE (i ') is arbitrary.
- the weighted average of the equation (2) can be used.
- w is a weight coefficient.
- the state of the buffer unit 11 is updated as follows. First, the oldest)) in the past autocorrelation function stored in the buffer section 11 is discarded. Next, the calculated in the current frame is stored in the buffer unit 11. Further, in the separating unit 20, the first filter coefficient calculating unit 2 calculates the average autocorrelation fl ⁇ (calculated by the average autocorrelation calculating unit 12 by a known method such as the Levinson algorithm from the inverse filter coefficient ( ), (1 ⁇ ⁇ ).
- the input voice; c is input to the inverse filter 3 composed of the filter coefficients, and the residual signal r (), ⁇ « ⁇ ) is obtained as a sound source characteristic by the equation (3).
- R (n) ⁇ ⁇ ) + ⁇ ⁇ (x (n - ⁇ ), (0 ⁇ n ⁇ N) (3)
- the spectrum calculation section 112 placed downstream of the filter coefficient calculation section 2 converts the coefficient obtained by the filter coefficient calculation section 2) into the following equation (4).
- Ru asked the LP C scan Bae spectrum ⁇ (Ri as vocal tract characteristics.
- N f is the number of data points in the spectrum.
- the sampling frequency is E s
- the frequency resolution of LPC spectrum is The variable ⁇ is an index of the spectrum and represents the discrete frequency. / Is converted to frequency [Hz] It becomes. Int [x] means that the variable X is converted to an integer (the same applies to the following description).
- the input voice can be separated into the sound source signal (residual signal r (, (0 ⁇ n ⁇ N)) and the vocal tract characteristic (LPC spectrum ⁇ ( ⁇ )) by the separation unit 20.
- the spectrum ⁇ ( ⁇ ) is input to the formant estimating unit 5 as an example of the feature extracting unit, and the formant frequency is (), (1 ⁇ ⁇ : max ) and ⁇ formant amplitude " (), (1 ⁇ ⁇ max ) where; max is the number of formants to be estimated.
- a known method such as a method of obtaining from a root of a higher-order equation having a coefficient of inverse fill coefficient or a peak picking method of estimating formant from a frequency spectrum peak is used. it can. From the lowest formant frequency, or (1), (2), ⁇ , or (: max ).
- a threshold may be set for the formant bandwidth, and only the frequency at which the bandwidth is equal to or less than the threshold may be set as the formant frequency.
- the spectrum ( ⁇ ), the discrete formant frequency or / (), and the formant amplitude amp () are input to the amplification factor calculation unit 6, and the amplification factor for the spectrum ( ⁇ ) is calculated.
- the processing of the amplification factor calculating section 6 includes calculation of the reference power (processing step P1), calculation of the formant amplification factor (processing step P2), and interpolation of the amplification factor (processing step P2). Processing is performed in the order of the process P 3). Hereinafter, each process will be described in order.
- Processing step P 1 Calculate reference power Pmv—re / from spectrum ⁇ ( ⁇ ).
- the calculation method is arbitrary, for example, the average power of the entire frequency band or the average power of the low frequency band can be used as the reference power.
- Pow-re / is expressed by the following equation (5).
- Processing step P 2 For adjusting the amplitude of the formant () to the reference power re /
- the amplification factor is calculated by the following equation (6).
- FIG. 12 shows how the amplitude of the formant E (jt) is adjusted to the reference power-re /. Further, in FIG. 12, the amplification factor) at the frequency between formants is obtained by the interpolation curve R (, /).
- the shape of the interpolation curve R (, /) is arbitrary, but for example, a linear function or a quadratic function can be used.
- Fig. 13 shows an example in which a quadratic curve is used as the interpolation curve? (, G. Define the interpolation curve R (Z) as in equation (7). It is a paramete that determines the shape of the curve.
- the amplification factor between the adjacent formants F () and E (+1) in this interpolation curve Set the minimum point.
- the setting method of the minimum point is arbitrary.
- the frequency (or Z (R + // (+ 1)) / 2 can be set to the minimum point and the amplification factor at that time can be set to G ().
- ⁇ is a constant, and 0 ⁇ y ⁇ 1.
- the input voice can be separated into the sound source characteristics and the vocal tract characteristics, and only the vocal tract characteristics can be emphasized.
- the pitch emphasis unit 4 is omitted in the embodiment of FIG. 10, the pitch emphasis unit 4 is arranged on the output side of the inverse filter 3 according to the principle diagram shown in FIG. It is also possible to perform pitch enhancement processing on r (.
- FIG. 14 is a block diagram showing the configuration of the second embodiment of the present invention.
- FIG. 10 shows that the LPC coefficient obtained from the input speech of the current frame is used as the inverse filter coefficient. Different from the first embodiment shown, the rest is the same as the first embodiment.
- the input signal of the current frame is compared with the case of using the LPC coefficient having the average frequency characteristic as in the first embodiment.
- the LPC coefficient obtained from the signal as the coefficient of the inverse filter 3 has a higher prediction gain, and the vocal tract characteristics and the sound source characteristics can be accurately separated.
- the input speech of the current frame is subjected to LPC analysis by the LPC analysis unit 13 and the obtained LPC coefficient (R, (lsi ⁇ A)) is used as a coefficient of the inverse filter 3.
- the LPC coefficient (the spectrum is calculated from the spectrum by the second spectrum calculator 1-2B).
- the method of calculating the spectrum ⁇ (R) is the same as that of the equation (4) in the first embodiment.
- the first spectrum calculating section 1-2 calculates the average spectrum, and calculates the average spectrum.
- the formant estimator 5 obtains the formant frequency () and the formant amplitude mp (A :).
- the gain is calculated by the gain calculating unit 6 from the spectrum 3 ⁇ 4 (the formant frequency or () and the formant amplitude ap ()), and the gain is calculated based on the gain.
- the spectrum emphasis unit 7 performs spectrum emphasis and obtains the emphasized spectrum 2). From the emphasized spectrum 2 ( ⁇ ), the composite filter coefficient 2 (set in the composite filter 9) to be set in the composite filter 9 is obtained, and the residual signal r (is input to the composite filter 9 and the output voice _y (Is obtained.
- the vocal tract characteristics and the sound source characteristics of the current frame are accurately separated, and based on the average spectrum.
- the clarity can be improved by smoothly enhancing the vocal tract characteristics.
- the third embodiment differs from the first embodiment in that an automatic gain control unit (AGC unit) 14 is provided to control the amplitude of the synthesis filter 9); This is the same as the first embodiment.
- A is set so that the power ratio between the input audio signal (and the final output audio signal z (becomes 1
- the gain is adjusted by the 0 ⁇ unit 14.
- the AGC unit 14 any method is possible for the AGC unit 14, for example, the following method can be used.
- the amplitude ratio is calculated from the input audio signal x (n) and the synthesized output y (n) according to equation (14). Ask for.
- N is the frame length.
- the input voice x (n) is determined by the sound source characteristics and the vocal tract. It is possible to separate the vocal tract characteristics from the vocal tract characteristics. As a result, it is possible to suppress spectral distortion that occurs when simultaneously emphasizing the vocal tract characteristics and the sound source characteristics, which are problems in the related art, and to improve clarity.
- FIG. 16 shows a block diagram of the fourth embodiment of the present invention.
- the present embodiment differs from the first embodiment in that pitch enhancement processing is performed on the residual signal ro) output from the inverse filter 3 according to the principle diagram of FIG. This is the same as the embodiment of FIG.
- the method of pitch enhancement by the pitch enhancement filter 4 is arbitrary.
- a pitch coefficient calculator 4-1 is provided, and the following method can be used.
- the autocorrelation rscor (i) of the residual signal of the current frame is obtained by equation (17), and the pitch lag ⁇ ⁇ at which the autocorrelation racor (i) is maximized is obtained.
- the method of calculating the pitch prediction coefficient can be obtained by a known method such as the Levinson algorithm.
- the inverse fill output r (w) is input to the pitch emphasis fill 4 to obtain a voice: whose pitch periodicity is emphasized.
- the pitch enhancement filter 4 a filter represented by the transfer function of equation (18) can be used. Here, is a weighting factor.
- the pitch period component included in the residual signal can be emphasized by adding the pitch emphasis filter 4, and the voice clarity compared with the first embodiment is improved. Can be further improved.
- FIG. 17 shows a configuration block diagram of a fifth embodiment of the present invention.
- the second embodiment is different from the first embodiment in that the second buffer unit 15 for holding the amplification factor of the previous frame is provided, and the other configuration is the same as that of the first embodiment.
- a temporary gain sggi(/) is obtained in the gain calculator 6 from the spectrum ⁇ ( ⁇ ), the formant frequency (), and the amplitude ampij from the spectrum calculator 1-2.
- the method of calculating the provisional amplification factor is the same as the method of calculating the amplification factor 8 (Ri in the first embodiment.
- the provisional amplification factor s solicit(Ri and the pre-frame output from the buffer 15 are output.
- Frame amplification factor S— (Amplification factor of the current frame from the frame) S (Z) is calculated, where: Previous frame width ratio — 0 W (R is the final gain calculated in the previous frame.
- Previous frame width ratio — 0 W (R is the final gain calculated in the previous frame.
- the procedure for obtaining the amplification factor is as follows.
- Provisional amplification factor 8 ⁇ (Ri and previous frame amplification factor jS—oW (Ri difference
- the gain of the previous frame is the same as that of the first embodiment except for the part for obtaining the amplification factor 3 Description is omitted.
- the amplification factor in the previous frame is selectively used so that the amplification factor does not change rapidly between frames. This makes it possible to improve clarity while suppressing an increase in noise due to spectral enhancement.
- FIG. 18 shows a configuration block diagram of a sixth embodiment of the present invention.
- This embodiment shows a configuration in which the first and third to fifth embodiments are combined. The overlapping part is the same as in the other embodiments, and the description is omitted.
- FIG. 19 is a diagram showing a speech spectrum emphasized by the above embodiment. The effect of the present invention is clear when the spectrum of FIG. 19 is compared with the input speech spectrum before emphasis shown in FIG. 7 and the spectrum emphasized in frame units shown in FIG.
- the input voice is separated into the sound source characteristics and the vocal tract characteristics, and the vocal tract characteristics and the sound source characteristics are individually It is possible to emphasize. This makes it possible to suppress the distortion of the spectrum, which has been a problem in the prior art for emphasizing the voice itself, and to improve clarity.
- FIGS. 20 and 21 the principle configuration of FIGS. 20 and 21 is applied by the present invention.
- the principle configuration shown in FIGS. 20 and 21 is characterized in that it has a two-stage configuration of dynamic fill I and fixed fill.
- Fig. 20 shows a principle diagram in which the fixed filter II is placed after the dynamic filter I.
- the fixed filter II is replaced with the dynamic filter I. It may be placed in the first stage.
- the parameters used in the dynamic filter I are calculated by analyzing the input speech.
- the dynamic filter I uses a configuration according to the principle shown in FIG. 9 described above.
- FIGS. 20 and 21 schematically show the principle configuration shown in FIG.
- the dynamic filter I is composed of a separation function unit 20 that separates input speech into sound source characteristics and vocal tract characteristics, a feature extraction function unit 5 that extracts formant features from vocal tract characteristics, and a feature extraction function unit 5.
- Amplification ratio calculation function unit 6 that calculates the amplification factor based on the obtained formant characteristics, a spectrum function unit 7 that emphasizes the spectrum of the vocal tract characteristics according to the calculated amplification factor, and sound source characteristics and spectrum It has a synthesis function unit 21 for synthesizing the emphasized vocal tract characteristics.
- the fixed filter II has a filter characteristic having a constant pass band in a predetermined range of frequency width.
- the frequency band to be emphasized by the fixed filter II is arbitrary.
- a high-frequency band of 2 kHz or more or a band-enhanced filter that emphasizes an intermediate band component of 1 kHz to 3 kHz can be used.
- the fixed filter II amplifies some frequency bands, and the dynamic filter I enhances the formants. Since the gain of Fixed Fill II is fixed, there is no change in gain between frames. With such a configuration, excessive emphasis by the dynamic filter I can be prevented, and clarity can be improved.
- FIG. 22 is a block diagram of the configuration of a further embodiment of the present invention based on the principle diagram of FIG. This embodiment uses the configuration of the third embodiment described above as the dynamic filter I. Therefore, the detailed description is omitted.
- an input voice is separated into a sound source characteristic and a vocal tract characteristic by a dynamic filter I, and only the vocal tract characteristic is emphasized.
- the vocal tract features It is possible to suppress the spectral distortion that occurs when simultaneously enhancing the characteristics and the sound source characteristics, and improve the clarity. Further, by adjusting the gain by the AGC unit 14 so that the amplitude of the output sound does not become excessively large compared to the input signal due to the spectrum enhancement, it is possible to obtain a smooth and highly natural output sound.
- the vocal tract characteristics and the sound source characteristics can be individually enhanced by the present invention.
- the vocal tract characteristics are emphasized, by emphasizing based on the average spectrum, a rapid change in the amplification factor between frames is reduced, so that good sound quality with little noise can be realized.
- the present invention enables favorable voice communication in a mobile phone, and can contribute to further spread of the mobile phone.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Signal Processing (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Quality & Reliability (AREA)
- Spectroscopy & Molecular Physics (AREA)
- Telephone Function (AREA)
- Electrophonic Musical Instruments (AREA)
- Circuit For Audible Band Transducer (AREA)
Abstract
Description
Claims
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP02779956.8A EP1557827B8 (en) | 2002-10-31 | 2002-10-31 | Voice intensifier |
CNB028295854A CN100369111C (zh) | 2002-10-31 | 2002-10-31 | 话音增强装置 |
PCT/JP2002/011332 WO2004040555A1 (ja) | 2002-10-31 | 2002-10-31 | 音声強調装置 |
JP2004547997A JP4219898B2 (ja) | 2002-10-31 | 2002-10-31 | 音声強調装置 |
US11/060,188 US7152032B2 (en) | 2002-10-31 | 2005-02-17 | Voice enhancement device by separate vocal tract emphasis and source emphasis |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/JP2002/011332 WO2004040555A1 (ja) | 2002-10-31 | 2002-10-31 | 音声強調装置 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/060,188 Continuation US7152032B2 (en) | 2002-10-31 | 2005-02-17 | Voice enhancement device by separate vocal tract emphasis and source emphasis |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2004040555A1 true WO2004040555A1 (ja) | 2004-05-13 |
Family
ID=32260023
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2002/011332 WO2004040555A1 (ja) | 2002-10-31 | 2002-10-31 | 音声強調装置 |
Country Status (5)
Country | Link |
---|---|
US (1) | US7152032B2 (ja) |
EP (1) | EP1557827B8 (ja) |
JP (1) | JP4219898B2 (ja) |
CN (1) | CN100369111C (ja) |
WO (1) | WO2004040555A1 (ja) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8255222B2 (en) | 2007-08-10 | 2012-08-28 | Panasonic Corporation | Speech separating apparatus, speech synthesizing apparatus, and voice quality conversion apparatus |
JP2013101255A (ja) * | 2011-11-09 | 2013-05-23 | Nippon Telegr & Teleph Corp <Ntt> | 音声強調装置とその方法とプログラム |
JP2013218147A (ja) * | 2012-04-10 | 2013-10-24 | Nippon Telegr & Teleph Corp <Ntt> | 音声明瞭度変換装置、音声明瞭度変換方法及びそのプログラム |
JP2021157082A (ja) * | 2020-03-27 | 2021-10-07 | 株式会社トランストロン | 基本周波数推定装置、アクティブノイズコントロール装置、基本周波数の推定方法及び基本周波数の推定プログラム |
Families Citing this family (37)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4076887B2 (ja) * | 2003-03-24 | 2008-04-16 | ローランド株式会社 | ボコーダ装置 |
DE60330715D1 (de) | 2003-05-01 | 2010-02-04 | Fujitsu Ltd | Sprachdecodierer, sprachdecodierungsverfahren, programm,aufzeichnungsmedium |
US20070011009A1 (en) * | 2005-07-08 | 2007-01-11 | Nokia Corporation | Supporting a concatenative text-to-speech synthesis |
EP1850328A1 (en) * | 2006-04-26 | 2007-10-31 | Honda Research Institute Europe GmbH | Enhancement and extraction of formants of voice signals |
JP4827661B2 (ja) * | 2006-08-30 | 2011-11-30 | 富士通株式会社 | 信号処理方法及び装置 |
US8050434B1 (en) | 2006-12-21 | 2011-11-01 | Srs Labs, Inc. | Multi-channel audio enhancement system |
PL2232700T3 (pl) | 2007-12-21 | 2015-01-30 | Dts Llc | System regulacji odczuwanej głośności sygnałów audio |
US8831936B2 (en) * | 2008-05-29 | 2014-09-09 | Qualcomm Incorporated | Systems, methods, apparatus, and computer program products for speech signal processing using spectral contrast enhancement |
KR101475724B1 (ko) * | 2008-06-09 | 2014-12-30 | 삼성전자주식회사 | 오디오 신호 품질 향상 장치 및 방법 |
US8538749B2 (en) * | 2008-07-18 | 2013-09-17 | Qualcomm Incorporated | Systems, methods, apparatus, and computer program products for enhanced intelligibility |
WO2010035438A1 (ja) * | 2008-09-26 | 2010-04-01 | パナソニック株式会社 | 音声分析装置および音声分析方法 |
US9202456B2 (en) | 2009-04-23 | 2015-12-01 | Qualcomm Incorporated | Systems, methods, apparatus, and computer-readable media for automatic control of active noise cancellation |
WO2011004579A1 (ja) * | 2009-07-06 | 2011-01-13 | パナソニック株式会社 | 声質変換装置、音高変換装置および声質変換方法 |
US8538042B2 (en) | 2009-08-11 | 2013-09-17 | Dts Llc | System for increasing perceived loudness of speakers |
KR20120054081A (ko) * | 2009-08-25 | 2012-05-29 | 난양 테크놀러지컬 유니버시티 | 속삭임을 포함하는 입력 신호로부터 음성을 재구성하는 방법 및 시스템 |
WO2011026247A1 (en) * | 2009-09-04 | 2011-03-10 | Svox Ag | Speech enhancement techniques on the power spectrum |
US8204742B2 (en) | 2009-09-14 | 2012-06-19 | Srs Labs, Inc. | System for processing an audio signal to enhance speech intelligibility |
TWI459828B (zh) * | 2010-03-08 | 2014-11-01 | Dolby Lab Licensing Corp | 在多頻道音訊中決定語音相關頻道的音量降低比例的方法及系統 |
US9053697B2 (en) | 2010-06-01 | 2015-06-09 | Qualcomm Incorporated | Systems, methods, devices, apparatus, and computer program products for audio equalization |
EP2610865B1 (en) * | 2010-08-23 | 2014-07-23 | Panasonic Corporation | Audio signal processing device and audio signal processing method |
PL2737479T3 (pl) * | 2011-07-29 | 2017-07-31 | Dts Llc | Adaptacyjna poprawa zrozumiałości głosu |
JP2013073230A (ja) * | 2011-09-29 | 2013-04-22 | Renesas Electronics Corp | オーディオ符号化装置 |
CN102595297B (zh) * | 2012-02-15 | 2014-07-16 | 嘉兴益尔电子科技有限公司 | 数字式助听器增益控制优化方法 |
US9312829B2 (en) | 2012-04-12 | 2016-04-12 | Dts Llc | System for adjusting loudness of audio signals in real time |
CN102779527B (zh) * | 2012-08-07 | 2014-05-28 | 无锡成电科大科技发展有限公司 | 基于窗函数共振峰增强的语音增强方法 |
DE112012006876B4 (de) * | 2012-09-04 | 2021-06-10 | Cerence Operating Company | Verfahren und Sprachsignal-Verarbeitungssystem zur formantabhängigen Sprachsignalverstärkung |
CN104464746A (zh) * | 2013-09-12 | 2015-03-25 | 索尼公司 | 语音滤波方法、装置以及电子设备 |
CN104143337B (zh) * | 2014-01-08 | 2015-12-09 | 腾讯科技(深圳)有限公司 | 一种提高音频信号音质的方法和装置 |
CN107851433B (zh) * | 2015-12-10 | 2021-06-29 | 华侃如 | 基于谐波模型和声源-声道特征分解的语音分析合成方法 |
CN106970771B (zh) * | 2016-01-14 | 2020-01-14 | 腾讯科技(深圳)有限公司 | 音频数据处理方法和装置 |
WO2018084305A1 (ja) * | 2016-11-07 | 2018-05-11 | ヤマハ株式会社 | 音声合成方法 |
EP3688754A1 (en) * | 2017-09-26 | 2020-08-05 | Sony Europe B.V. | Method and electronic device for formant attenuation/amplification |
JP6991041B2 (ja) * | 2017-11-21 | 2022-01-12 | ヤフー株式会社 | 生成装置、生成方法、および生成プログラム |
JP6962269B2 (ja) * | 2018-05-10 | 2021-11-05 | 日本電信電話株式会社 | ピッチ強調装置、その方法、およびプログラム |
CN109346058B (zh) * | 2018-11-29 | 2024-06-28 | 西安交通大学 | 一种语音声学特征扩大系统 |
CN113571079A (zh) * | 2021-02-08 | 2021-10-29 | 腾讯科技(深圳)有限公司 | 语音增强方法、装置、设备及存储介质 |
CN115206142B (zh) * | 2022-06-10 | 2023-12-26 | 深圳大学 | 一种基于共振峰的语音训练方法及系统 |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH08160992A (ja) * | 1994-12-02 | 1996-06-21 | Atr Ningen Joho Tsushin Kenkyusho:Kk | 音質改善装置 |
JPH08248996A (ja) * | 1995-03-10 | 1996-09-27 | Nippon Telegr & Teleph Corp <Ntt> | ディジタルフィルタのフィルタ係数決定方法 |
JPH08305397A (ja) * | 1995-05-12 | 1996-11-22 | Mitsubishi Electric Corp | 音声加工フィルタ及び音声合成装置 |
Family Cites Families (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4969192A (en) * | 1987-04-06 | 1990-11-06 | Voicecraft, Inc. | Vector adaptive predictive coder for speech and audio |
JP2588004B2 (ja) | 1988-09-19 | 1997-03-05 | 日本電信電話株式会社 | 後処理フィルタ |
JP2626223B2 (ja) * | 1990-09-26 | 1997-07-02 | 日本電気株式会社 | 音声符号化装置 |
US5233660A (en) * | 1991-09-10 | 1993-08-03 | At&T Bell Laboratories | Method and apparatus for low-delay celp speech coding and decoding |
WO1993018505A1 (en) * | 1992-03-02 | 1993-09-16 | The Walt Disney Company | Voice transformation system |
FR2734389B1 (fr) * | 1995-05-17 | 1997-07-18 | Proust Stephane | Procede d'adaptation du niveau de masquage du bruit dans un codeur de parole a analyse par synthese utilisant un filtre de ponderation perceptuelle a court terme |
US5774837A (en) * | 1995-09-13 | 1998-06-30 | Voxware, Inc. | Speech coding system and method using voicing probability determination |
JPH09160595A (ja) | 1995-12-04 | 1997-06-20 | Toshiba Corp | 音声合成方法 |
US6240384B1 (en) * | 1995-12-04 | 2001-05-29 | Kabushiki Kaisha Toshiba | Speech synthesis method |
KR100269255B1 (ko) * | 1997-11-28 | 2000-10-16 | 정선종 | 유성음 신호에서 성문 닫힘 구간 신호의 가변에의한 피치 수정방법 |
US6003000A (en) * | 1997-04-29 | 1999-12-14 | Meta-C Corporation | Method and system for speech processing with greatly reduced harmonic and intermodulation distortion |
US6073092A (en) * | 1997-06-26 | 2000-06-06 | Telogy Networks, Inc. | Method for speech coding based on a code excited linear prediction (CELP) model |
US6098036A (en) * | 1998-07-13 | 2000-08-01 | Lockheed Martin Corp. | Speech coding system and method including spectral formant enhancer |
GB2342829B (en) * | 1998-10-13 | 2003-03-26 | Nokia Mobile Phones Ltd | Postfilter |
US6950799B2 (en) * | 2002-02-19 | 2005-09-27 | Qualcomm Inc. | Speech converter utilizing preprogrammed voice profiles |
-
2002
- 2002-10-31 CN CNB028295854A patent/CN100369111C/zh not_active Expired - Fee Related
- 2002-10-31 WO PCT/JP2002/011332 patent/WO2004040555A1/ja active Application Filing
- 2002-10-31 JP JP2004547997A patent/JP4219898B2/ja not_active Expired - Fee Related
- 2002-10-31 EP EP02779956.8A patent/EP1557827B8/en not_active Expired - Lifetime
-
2005
- 2005-02-17 US US11/060,188 patent/US7152032B2/en not_active Expired - Lifetime
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH08160992A (ja) * | 1994-12-02 | 1996-06-21 | Atr Ningen Joho Tsushin Kenkyusho:Kk | 音質改善装置 |
JPH08248996A (ja) * | 1995-03-10 | 1996-09-27 | Nippon Telegr & Teleph Corp <Ntt> | ディジタルフィルタのフィルタ係数決定方法 |
JPH08305397A (ja) * | 1995-05-12 | 1996-11-22 | Mitsubishi Electric Corp | 音声加工フィルタ及び音声合成装置 |
Non-Patent Citations (1)
Title |
---|
See also references of EP1557827A4 * |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8255222B2 (en) | 2007-08-10 | 2012-08-28 | Panasonic Corporation | Speech separating apparatus, speech synthesizing apparatus, and voice quality conversion apparatus |
JP2013101255A (ja) * | 2011-11-09 | 2013-05-23 | Nippon Telegr & Teleph Corp <Ntt> | 音声強調装置とその方法とプログラム |
JP2013218147A (ja) * | 2012-04-10 | 2013-10-24 | Nippon Telegr & Teleph Corp <Ntt> | 音声明瞭度変換装置、音声明瞭度変換方法及びそのプログラム |
JP2021157082A (ja) * | 2020-03-27 | 2021-10-07 | 株式会社トランストロン | 基本周波数推定装置、アクティブノイズコントロール装置、基本周波数の推定方法及び基本周波数の推定プログラム |
JP7461192B2 (ja) | 2020-03-27 | 2024-04-03 | 株式会社トランストロン | 基本周波数推定装置、アクティブノイズコントロール装置、基本周波数の推定方法及び基本周波数の推定プログラム |
Also Published As
Publication number | Publication date |
---|---|
JPWO2004040555A1 (ja) | 2006-03-02 |
JP4219898B2 (ja) | 2009-02-04 |
CN1669074A (zh) | 2005-09-14 |
EP1557827A1 (en) | 2005-07-27 |
EP1557827A4 (en) | 2008-05-14 |
EP1557827B1 (en) | 2014-10-01 |
EP1557827B8 (en) | 2015-01-07 |
US7152032B2 (en) | 2006-12-19 |
CN100369111C (zh) | 2008-02-13 |
US20050165608A1 (en) | 2005-07-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2004040555A1 (ja) | 音声強調装置 | |
US7158932B1 (en) | Noise suppression apparatus | |
JP3457293B2 (ja) | 雑音抑圧装置及び雑音抑圧方法 | |
US8560308B2 (en) | Speech sound enhancement device utilizing ratio of the ambient to background noise | |
JP4423300B2 (ja) | 雑音抑圧装置 | |
JP4018571B2 (ja) | 音声強調装置 | |
US8126176B2 (en) | Hearing aid | |
JPH10509256A (ja) | ピッチ操作器を使用する音声信号の変換方法 | |
JP2000347688A (ja) | 雑音抑圧装置 | |
JPH09204196A (ja) | 雑音抑圧器及び移動局並びに雑音抑圧方法 | |
JP6073456B2 (ja) | 音声強調装置 | |
JP2004061617A (ja) | 受話音声処理装置 | |
JP4413480B2 (ja) | 音声処理装置及び移動通信端末装置 | |
US7606702B2 (en) | Speech decoder, speech decoding method, program and storage media to improve voice clarity by emphasizing voice tract characteristics using estimated formants | |
JP2007251354A (ja) | マイクロホン、音声生成方法 | |
JP2008309955A (ja) | ノイズサプレス装置 | |
Chanda et al. | Speech intelligibility enhancement using tunable equalization filter | |
JP4135242B2 (ja) | 受信装置及び方法、通信装置及び方法 | |
JP4227421B2 (ja) | 音声強調装置および携帯端末 | |
RU2589298C1 (ru) | Способ повышения разборчивости и информативности звуковых сигналов в шумовой обстановке | |
JPH09311696A (ja) | 自動利得調整装置 | |
JP2006201622A (ja) | 帯域分割型雑音抑圧装置及び帯域分割型雑音抑圧方法 | |
KR100746680B1 (ko) | 음성 강조 장치 | |
JP4135240B2 (ja) | 受信装置及び方法、通信装置及び方法 | |
JP2011141540A (ja) | 音声信号処理装置、テレビジョン受像機、音声信号処理方法、プログラム、および、記録媒体 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): CN JP KR US |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR IE IT LU MC NL PT SE SK TR |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2004547997 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 11060188 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1020057002760 Country of ref document: KR |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2002779956 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 20028295854 Country of ref document: CN |
|
WWP | Wipo information: published in national office |
Ref document number: 1020057002760 Country of ref document: KR |
|
WWP | Wipo information: published in national office |
Ref document number: 2002779956 Country of ref document: EP |