WO2008072671A1 - 音声復号化装置およびパワ調整方法 - Google Patents
音声復号化装置およびパワ調整方法 Download PDFInfo
- Publication number
- WO2008072671A1 WO2008072671A1 PCT/JP2007/073968 JP2007073968W WO2008072671A1 WO 2008072671 A1 WO2008072671 A1 WO 2008072671A1 JP 2007073968 W JP2007073968 W JP 2007073968W WO 2008072671 A1 WO2008072671 A1 WO 2008072671A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- coefficient
- signal
- post filter
- amplitude ratio
- output signal
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 21
- 238000009499 grossing Methods 0.000 claims abstract description 39
- 238000004364 calculation method Methods 0.000 claims abstract description 10
- 230000001133 acceleration Effects 0.000 claims description 5
- 230000015572 biosynthetic process Effects 0.000 abstract description 11
- 238000003786 synthesis reaction Methods 0.000 abstract description 11
- 230000005284 excitation Effects 0.000 description 50
- 239000013598 vector Substances 0.000 description 42
- 230000003044 adaptive effect Effects 0.000 description 26
- 238000013139 quantization Methods 0.000 description 10
- 239000002131 composite material Substances 0.000 description 9
- 238000005516 engineering process Methods 0.000 description 8
- 238000005070 sampling Methods 0.000 description 7
- 238000004458 analytical method Methods 0.000 description 6
- 238000012545 processing Methods 0.000 description 6
- 238000010586 diagram Methods 0.000 description 4
- 230000000694 effects Effects 0.000 description 4
- 230000005236 sound signal Effects 0.000 description 4
- 238000009792 diffusion process Methods 0.000 description 3
- 230000010354 integration Effects 0.000 description 3
- 238000010295 mobile communication Methods 0.000 description 3
- 230000006870 function Effects 0.000 description 2
- 238000012805 post-processing Methods 0.000 description 2
- 241001123248 Arma Species 0.000 description 1
- 208000018583 New-onset refractory status epilepticus Diseases 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000010365 information processing Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 230000000630 rising effect Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 238000007493 shaping process Methods 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/04—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
- G10L19/26—Pre-filtering or post-filtering
-
- H—ELECTRICITY
- H03—ELECTRONIC CIRCUITRY
- H03G—CONTROL OF AMPLIFICATION
- H03G3/00—Gain control in amplifiers or frequency changers
- H03G3/20—Automatic control
- H03G3/30—Automatic control in amplifiers having semiconductor devices
- H03G3/3005—Automatic control in amplifiers having semiconductor devices in amplifiers suitable for low-frequencies, e.g. audio amplifiers
Definitions
- the present invention relates to an audio decoding apparatus and a power adjustment method for decoding an encoded audio signal.
- a post-filter is generally applied to the synthesized sound before output. Most of the standard codecs for mobile phones use this post filter.
- CELP post-filters use pole-zero type (ARMA type) pole enhancement filters, high frequency band enhancement filters, and pitch filters using LPC parameters.
- Patent Documents 1 and 2 disclose a technique that uses a smoothing coefficient so that the power is gradually adjusted for each sample after obtaining the adjustment coefficient.
- the smoothing coefficient is ⁇
- (l- ⁇ ) is the acceleration coefficient.
- Patent Document 1 JP-A-9 190195
- Patent Document 2 JP-A-9 127996
- the filter gain of the post filter is large at a place where the power rises at the rising edge of the sound, the output signal of the post filter tends to be abruptly larger than the input signal. In this case, The adjustment factor for power must be adapted quickly.
- the input / output ratio of the post filter fluctuates significantly over time, it is necessary to make a quick adjustment.
- sound quality distortion due to a rapid change in the adjustment coefficient becomes a problem in sections where the input / output fluctuations of the input / output of the post filter are small and in steady speech sections such as vowels. .
- An object of the present invention is to provide a speech decoding apparatus and a power adjustment method capable of obtaining a good synthesized sound with a stable volume.
- the speech decoding apparatus includes a post filter that filters a signal having a subframe length at a predetermined sample timing interval, and an amplitude ratio or a path ratio between an input signal and an output signal of the post filter.
- the first calculated value and the second calculated value that is the fluctuation amount of the first calculated value for each subframe, and smoothing based on the first calculated value and the second calculated value
- Smoothing coefficient setting means for setting a coefficient for each subframe, adjustment coefficient setting means for setting an adjustment coefficient for each sample based on the first calculated value and the smoothing coefficient, and the output signal of the post filter in the output signal
- a power adjustment means for obtaining a decoded speech signal by multiplying by the adjustment coefficient.
- the power adjustment method of the present invention is a method of adjusting the output signal of a post filter that filters a subframe length signal at a predetermined sample timing interval, and the input signal and output of the post filter
- the first calculated value which is the amplitude ratio or the signal ratio with the signal
- a step of calculating the second calculated value which is a variation amount of the first calculated value, for each subframe, and a smoothing coefficient is set for each subframe based on the first calculated value and the second calculated value.
- the power when the post filter greatly changes the power, the power can be quickly adjusted when the temporal fluctuation of the power ratio becomes large. For steady intervals such as vowels, smooth power adjustment without discontinuity can be realized. Therefore, according to the present invention, a good synthesized sound can be obtained with a stable volume.
- FIG. 1 is a block diagram showing a configuration of a voice encoding apparatus that transmits encoded data to a voice decoding apparatus according to an embodiment of the present invention.
- FIG. 2 is a block diagram showing a configuration of a speech decoding apparatus according to an embodiment of the present invention.
- FIG. 3 is a flowchart for explaining an algorithm for adjusting the power of the speech decoding apparatus according to the embodiment of the present invention.
- FIG. 4 is a flowchart for explaining a power adjustment algorithm of a speech decoding apparatus according to an embodiment of the present invention.
- FIG. 1 is a block diagram showing a configuration of a speech encoding apparatus that transmits encoded data to the speech decoding apparatus according to the present embodiment.
- the pre-processing unit 101 performs a waveform shaping process and a pre-facility process on the input audio signal to improve the performance of the high-pass filter process that removes the DC component and the subsequent encoding process.
- the signal (Xin) is output to the LPC analysis unit 102 and the addition unit 105.
- the LPC analysis unit 102 performs linear prediction analysis using Xin, and the analysis result (linear prediction coefficient) Is output to the LPC quantization unit 103.
- the LPC quantization unit 103 performs a quantization process on the linear prediction coefficient (LPC) output from the LPC analysis unit 102, outputs the quantized LPC to the synthesis filter 104, and multiplexes a code (U is multiplexed) representing the quantized LPC. Output to part 114.
- LPC linear prediction coefficient
- the synthesis filter 104 generates a synthesized signal by performing filter synthesis on a driving sound source output from an adder 111 described later using a filter coefficient based on the quantized LPC, and adds the synthesized signal to the adder 105. Output to.
- Adder 105 calculates the error signal by inverting the polarity of the combined signal and adding it to Xin, and outputs the error signal to auditory weighting unit 112.
- Adaptive excitation codebook 106 stores in the buffer the driving excitation that was output in the past by addition section 111, and one frame from the past driving excitation specified by the signal output from parameter determination section 113. Are extracted as adaptive sound source vectors and output to the multiplier 109.
- Gain codebook 107 outputs the gain of the adaptive excitation vector and the gain of the fixed excitation vector specified by the signal output from parameter determining section 113 to multiplication section 109 and multiplication section 110, respectively.
- Fixed excitation codebook 108 stores a plurality of predetermined excitation source vectors in a buffer, and multiplies a pulse source vector having a shape specified by the signal output from parameter determining unit 113 by a diffusion vector.
- the fixed sound source vector obtained in this way is output to multiplication section 110.
- Multiplying section 109 multiplies the gain output from gain codebook 107 by the adaptive excitation vector output from adaptive excitation codebook 106 and outputs the result to adding section 111.
- Multiplication section 110 multiplies the gain output from gain codebook 107 by the fixed excitation vector output from fixed excitation codebook 108 and outputs the result to addition section 111.
- Adder 111 receives the adaptive excitation vector and fixed excitation vector after gain multiplication from multiplication unit 109 and multiplication unit 110, respectively, adds these vectors, and adds the drive sound source that is the addition result to the synthesis filter.
- 104 and adaptive excitation codebook 106 receives the adaptive excitation vector and fixed excitation vector after gain multiplication from multiplication unit 109 and multiplication unit 110, respectively, adds these vectors, and adds the drive sound source that is the addition result to the synthesis filter.
- 104 and adaptive excitation codebook 106 The driving sound source input to the adaptive sound source code book 106 is stored in the buffer.
- the auditory weighting unit 112 performs auditory weighting on the error signal output from the adding unit 105. Then, it is output to the parameter determination unit 113 as coding distortion.
- the parameter determination unit 113 searches for the adaptive excitation vector, the fixed excitation vector and the quantization gain code that minimizes the coding distortion output from the perceptual weighting unit 112, and searches for the searched adaptive excitation source.
- a code (A) representing a vector, a code (F) representing a fixed excitation vector, and a code (G) representing a quantization gain are output to the multiplexing unit 114.
- Multiplexer 114 receives code (L) representing quantized LPC from LPC quantizer 103, code (A) representing adaptive excitation vector, parameter representing fixed excitation vector from parameter determining unit 113 (F) and a code (G) representing the quantization gain are input, and the information is multiplexed and output as encoded information.
- FIG. 2 is a block diagram showing a configuration of the speech decoding apparatus according to the present embodiment.
- the encoded information is separated into individual codes (L, A, G, F) by the multiplexing / separating unit 201.
- Code representing quantization LPC U is output to LPC decoding section 202
- code (A) representing adaptive sound source vector is output to adaptive excitation codebook 203
- code (G) representing quantization gain is gain code
- the code (F) that is output to the book 204 and represents the fixed excitation vector is output to the fixed excitation codebook 205.
- the LPC decoding unit 202 decodes the quantized LSP parameter from the code (L) representing the quantized LPC, re-converts the obtained quantized LSP parameter into the quantized LPC parameter, and combines the synthesized finalizer parameters. 209 ⁇ output.
- Adaptive excitation codebook 203 stores the past driving excitation used in synthesis filter 209, and from the past driving excitation specified by the adaptive codebook lag corresponding to code (A) representing the adaptive excitation vector. A sample for one frame is taken out as an adaptive excitation vector and output to the multiplication unit 206. In addition, adaptive excitation codebook 203 updates the stored driving excitation with the driving excitation output from addition section 208.
- Gain codebook 204 decodes the adaptive excitation vector gain and fixed excitation vector gain specified by the code (G) representing the quantization gain, and outputs the adaptive excitation vector gain to multiplication section 206.
- the gain of the fixed sound source vector is output to the multiplication unit 207.
- Fixed excitation codebook 205 stores a plurality of predetermined excitation source vectors in a buffer, and generates a pulse excitation vector having a shape specified by code (F) representing the fixed excitation vector. Generate a fixed sound source vector obtained by multiplying the diffusion vector, and output to the multiplication unit 207
- Multiplying section 206 multiplies the adaptive excitation vector by a gain and outputs the result to adding section 208.
- Multiplier 207 multiplies the fixed sound source vector by the gain and outputs the result to adder 208.
- Adder 208 adds the adaptive excitation vector after gain multiplication output from multipliers 206 and 207 and the fixed excitation vector to generate a drive excitation, and generates this as synthesis filter 209 and adaptive excitation codebook Output to 203.
- the synthesis filter 209 performs filter synthesis of the driving sound source output from the addition unit 208 using the filter coefficients decoded by the LPC decoding unit 202, and obtains an obtained signal (hereinafter referred to as “first synthesis”).
- Signal is output to the post filter 210 and the amplitude ratio / variation calculation unit 211.
- the post-filter 210 performs processing for improving the subjective quality of the voice such as formant enhancement and pitch enhancement for the signal output from the synthesis filter 209. Processing for improving the quality is performed, and the obtained signal (hereinafter referred to as “second synthesized signal”) is output to the amplitude ratio / variation calculation unit 211 and the power adjustment unit 214.
- second synthesized signal the obtained signal
- pitch analysis may be omitted, and a filter created using the adaptive codebook lag of adaptive excitation codebook 203 and the gain of the adaptive excitation vector may be applied.
- Amplitude ratio and change amount calculation section 211 has an amplitude ratio between the first synthesized signal that is the input signal of post filter 210 and the second synthesized signal that is the output signal of post filter 210, and the fluctuation amount of this amplitude ratio. Is calculated for each subframe, the calculated amplitude ratio is output to the smoothing coefficient setting unit 212 and the adjustment coefficient setting unit 213, and the fluctuation amount of the calculated amplitude ratio is output to the smoothing coefficient setting unit 212.
- Smoothing coefficient setting section 212 sets the smoothing coefficient for each subframe using the amplitude ratio between the first combined signal and the second combined signal and the fluctuation amount of the amplitude ratio, and sets the smoothing The coefficient is output to the adjustment coefficient setting unit 213.
- the adjustment coefficient setting unit 213 sets the adjustment coefficient for each sample using the amplitude ratio between the first combined signal and the second combined signal and the smoothing coefficient, and the set adjustment coefficient is used as the power adjustment unit 214. Output to.
- the noise adjustment unit 214 adjusts the power of the second synthesized signal by multiplying the second synthesized signal by the adjustment coefficient to obtain a final decoded speech signal.
- the first synthesized signal and the second synthesized signal at all sampling timings are input in subframe units (ST302), the first synthesized signal part p0, the second synthesized signal part pl, and the sample value n. Is initialized to 0 (ST303), and then the first synthesized signal ⁇ and the second synthesized signal pl of the current subframe are obtained (ST304, ST305, ST306).
- the smoothing coefficient ⁇ is set for each case according to the amplitude ratio g of the current subframe.
- the smoothing coefficient ⁇ is set closer to 1.0 as the amplitude ratio g of the current subframe is closer to 1.0.
- the acceleration coefficient (1 ⁇ ) becomes closer to 0.0 as the smoothing coefficient ⁇ becomes closer to 1.0.
- the smoothing coefficient ⁇ is multiplied by the stationarity scale 0 to obtain a new smoothing coefficient ⁇ (ST3 20).
- the smoothing factor ⁇ is reduced (the acceleration factor (1 ⁇ ) is increased) when the fluctuations in time are large, and the performance is quickly increased. Can be adjusted!
- an adjustment coefficient g is calculated based on the obtained amplitude ratio g of the current subframe and the smoothing coefficient ⁇ . Specifically, a new adjustment coefficient is obtained by adding the adjustment coefficient g of the previous sample multiplied by the smoothing coefficient ⁇ and the amplitude ratio g of the current subframe multiplied by the acceleration coefficient (1 ⁇ ). Calculate g. Then, the final synthesized speech signal q [n] is obtained by multiplying the second composite signal pfti] by the adjustment coefficient g (ST321, ST322, ST323, ST324).
- the above processing is repeated for the next subframe (ST326).
- the adjustment factor g used last is used as it is in the next subframe.
- the amplitude ratio g of the current subframe obtained in ST308 or ST309 is used as the amplitude ratio g of the previous subframe in the processing of the next subframe.
- the power can be quickly adjusted when the post filter greatly changes the power or the amplitude ratio fluctuates with time. Therefore, smooth adjustment without discontinuity can be achieved in sections where there are no fluctuations in the frequency or in sections that are steady in time. Therefore, according to the present embodiment, a good synthesized sound can be obtained with a stable volume.
- the present invention is in other cases where the sampling frequency and the subframe length are not limited. Even if it exists, it is effective.
- the sampling is double 16kHz sampling
- the subframe unit is 80 samples, and good performance can be obtained by setting the smoothing coefficient to a larger value.
- the smoothing coefficient constant ⁇ 0.9, 0.96, 0.99, 0.999 ⁇ in this embodiment is ⁇ 0.95, 0.98, 0.993, 0.999 ⁇ .
- the power described for the case where the amplitude ratio is referred to in the determination of the smoothing coefficient or the stationarity scale is not limited to this, and the power ratio is used instead of the amplitude ratio. However, the same effect can be obtained.
- the power ratio has a strong correlation with the square of the amplitude ratio.
- the power for calculating the square root of the ratio of the sum of squares of two signals to obtain the adjustment coefficient of the current subframe is not limited to this, and the sum of the absolute values of the signals is not limited to this. The same effect can be obtained even if the ratio of is used.
- the power shown for the power adjustment unit for adjusting the fluctuation of the input / output power of the post filter is not limited to the post filter, and there is a power fluctuation in the input / output. It is effective in the case. For example, even in phonological enhancement processing used in hearing aids and the like, it is necessary to adjust the sound so that the sound does not fluctuate suddenly. can do.
- the present invention is effective even with the force S used for CELP and other codecs. This is because the power adjustment unit of the present invention is used for the post-processing of the decoder process and does not depend on the type of codec.
- the power for generating the fixed excitation vector by multiplying the pulse excitation vector by the diffusion vector in the fixed excitation codebook is not limited to this, and the Norse excitation vector itself is used as the fixed excitation source. It can be a vector.
- the speech decoding apparatus can be mounted on a communication terminal apparatus and a base station apparatus in a mobile communication system, and thereby has a similar effect to the above.
- a base station apparatus, and a mobile communication system can be provided.
- the present invention can also be realized by software.
- the ability to realize the same function as the speech decoding apparatus according to the present invention by describing the algorithm according to the present invention in a programming language, storing the program in a memory, and causing the information processing means to execute the algorithm. it can.
- Each functional block used in the description of the above embodiment is typically realized as an LSI which is an integrated circuit. These may be individually made into one chip, or may be made into one chip so as to include a part or all of them. [0064] Although LSI is used here, depending on the degree of integration, IC, system LSI, super L
- the method of circuit integration is not limited to LSI's, and implementation using dedicated circuitry or general purpose processors is also possible.
- FPGA Field Programmable Gate Array
- the present invention is suitable for use in a speech decoding apparatus that decodes an encoded speech signal.
Landscapes
- Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Signal Processing (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Compression, Expansion, Code Conversion, And Decoders (AREA)
Abstract
Description
Claims
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
BRPI0720266-0A BRPI0720266A2 (pt) | 2006-12-13 | 2007-12-12 | Dispositivo de decodificação de aúdio e método de ajuste de potência |
EP07859788A EP2096631A4 (en) | 2006-12-13 | 2007-12-12 | TONE DECODING DEVICE AND POWER ADJUSTMENT METHOD |
JP2008549343A JPWO2008072671A1 (ja) | 2006-12-13 | 2007-12-12 | 音声復号化装置およびパワ調整方法 |
US12/517,603 US20100332223A1 (en) | 2006-12-13 | 2007-12-12 | Audio decoding device and power adjusting method |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2006336272 | 2006-12-13 | ||
JP2006-336272 | 2006-12-13 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2008072671A1 true WO2008072671A1 (ja) | 2008-06-19 |
Family
ID=39511688
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2007/073968 WO2008072671A1 (ja) | 2006-12-13 | 2007-12-12 | 音声復号化装置およびパワ調整方法 |
Country Status (5)
Country | Link |
---|---|
US (1) | US20100332223A1 (ja) |
EP (1) | EP2096631A4 (ja) |
JP (1) | JPWO2008072671A1 (ja) |
BR (1) | BRPI0720266A2 (ja) |
WO (1) | WO2008072671A1 (ja) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110689488A (zh) * | 2019-08-22 | 2020-01-14 | 稿定(厦门)科技有限公司 | 图像调色方法、介质、设备及装置 |
Families Citing this family (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5245622B2 (ja) * | 2008-07-31 | 2013-07-24 | 富士通株式会社 | ノイズ検出装置及びノイズ検出方法 |
CN103928031B (zh) | 2013-01-15 | 2016-03-30 | 华为技术有限公司 | 编码方法、解码方法、编码装置和解码装置 |
US9728200B2 (en) | 2013-01-29 | 2017-08-08 | Qualcomm Incorporated | Systems, methods, apparatus, and computer-readable media for adaptive formant sharpening in linear prediction coding |
US9042462B2 (en) | 2013-04-24 | 2015-05-26 | Commscope Technologies Llc | Differential signal transmission |
US9620134B2 (en) | 2013-10-10 | 2017-04-11 | Qualcomm Incorporated | Gain shape estimation for improved tracking of high-band temporal characteristics |
US10083708B2 (en) | 2013-10-11 | 2018-09-25 | Qualcomm Incorporated | Estimation of mixing factors to generate high-band excitation signal |
US10614816B2 (en) | 2013-10-11 | 2020-04-07 | Qualcomm Incorporated | Systems and methods of communicating redundant frame information |
US9384746B2 (en) | 2013-10-14 | 2016-07-05 | Qualcomm Incorporated | Systems and methods of energy-scaled signal processing |
US10163447B2 (en) | 2013-12-16 | 2018-12-25 | Qualcomm Incorporated | High-band signal modeling |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH07160296A (ja) * | 1993-12-10 | 1995-06-23 | Nec Corp | 音声復号装置 |
JPH09127996A (ja) | 1995-10-26 | 1997-05-16 | Sony Corp | 音声復号化方法及び装置 |
JPH09138697A (ja) * | 1995-09-14 | 1997-05-27 | Toshiba Corp | ホルマント強調方法 |
JPH09190195A (ja) | 1995-09-18 | 1997-07-22 | Toshiba Corp | 音声信号のスペクトル形状調整方法および装置 |
JPH10214100A (ja) * | 1997-01-31 | 1998-08-11 | Sony Corp | 音声合成方法 |
JP2000305598A (ja) * | 1990-02-23 | 2000-11-02 | Toshiba Corp | 適応ポストフィルタ |
Family Cites Families (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3076086B2 (ja) * | 1991-06-28 | 2000-08-14 | シャープ株式会社 | 音声合成装置用ポストフィルタ |
GB9512284D0 (en) * | 1995-06-16 | 1995-08-16 | Nokia Mobile Phones Ltd | Speech Synthesiser |
EP0763818B1 (en) * | 1995-09-14 | 2003-05-14 | Kabushiki Kaisha Toshiba | Formant emphasis method and formant emphasis filter device |
US5864798A (en) * | 1995-09-18 | 1999-01-26 | Kabushiki Kaisha Toshiba | Method and apparatus for adjusting a spectrum shape of a speech signal |
US5890138A (en) * | 1996-08-26 | 1999-03-30 | Bid.Com International Inc. | Computer auction system |
US6092041A (en) * | 1996-08-22 | 2000-07-18 | Motorola, Inc. | System and method of encoding and decoding a layered bitstream by re-applying psychoacoustic analysis in the decoder |
JPH1084284A (ja) * | 1996-09-06 | 1998-03-31 | Sony Corp | 信号再生方法および装置 |
US6691082B1 (en) * | 1999-08-03 | 2004-02-10 | Lucent Technologies Inc | Method and system for sub-band hybrid coding |
US6819275B2 (en) * | 2000-09-08 | 2004-11-16 | Koninklijke Philips Electronics N.V. | Audio signal compression |
DE60139144D1 (de) * | 2000-11-30 | 2009-08-13 | Nippon Telegraph & Telephone | Audio-dekodierer und audio-dekodierungsverfahren |
CA2388439A1 (en) * | 2002-05-31 | 2003-11-30 | Voiceage Corporation | A method and device for efficient frame erasure concealment in linear predictive based speech codecs |
CA2457988A1 (en) * | 2004-02-18 | 2005-08-18 | Voiceage Corporation | Methods and devices for audio compression based on acelp/tcx coding and multi-rate lattice vector quantization |
DE602004025517D1 (de) * | 2004-05-17 | 2010-03-25 | Nokia Corp | Audiocodierung mit verschiedenen codierungsrahmenlängen |
US7596486B2 (en) * | 2004-05-19 | 2009-09-29 | Nokia Corporation | Encoding an audio signal using different audio coder modes |
CN101006495A (zh) * | 2004-08-31 | 2007-07-25 | 松下电器产业株式会社 | 语音编码装置、语音解码装置、通信装置以及语音编码方法 |
WO2006030864A1 (ja) * | 2004-09-17 | 2006-03-23 | Matsushita Electric Industrial Co., Ltd. | 音声符号化装置、音声復号装置、通信装置及び音声符号化方法 |
JP4732730B2 (ja) * | 2004-09-30 | 2011-07-27 | パナソニック株式会社 | 音声復号装置 |
US7676362B2 (en) * | 2004-12-31 | 2010-03-09 | Motorola, Inc. | Method and apparatus for enhancing loudness of a speech signal |
JPWO2007088853A1 (ja) * | 2006-01-31 | 2009-06-25 | パナソニック株式会社 | 音声符号化装置、音声復号装置、音声符号化システム、音声符号化方法及び音声復号方法 |
-
2007
- 2007-12-12 WO PCT/JP2007/073968 patent/WO2008072671A1/ja active Application Filing
- 2007-12-12 EP EP07859788A patent/EP2096631A4/en not_active Withdrawn
- 2007-12-12 US US12/517,603 patent/US20100332223A1/en not_active Abandoned
- 2007-12-12 BR BRPI0720266-0A patent/BRPI0720266A2/pt not_active IP Right Cessation
- 2007-12-12 JP JP2008549343A patent/JPWO2008072671A1/ja not_active Withdrawn
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2000305598A (ja) * | 1990-02-23 | 2000-11-02 | Toshiba Corp | 適応ポストフィルタ |
JPH07160296A (ja) * | 1993-12-10 | 1995-06-23 | Nec Corp | 音声復号装置 |
JPH09138697A (ja) * | 1995-09-14 | 1997-05-27 | Toshiba Corp | ホルマント強調方法 |
JPH09190195A (ja) | 1995-09-18 | 1997-07-22 | Toshiba Corp | 音声信号のスペクトル形状調整方法および装置 |
JPH09127996A (ja) | 1995-10-26 | 1997-05-16 | Sony Corp | 音声復号化方法及び装置 |
JPH10214100A (ja) * | 1997-01-31 | 1998-08-11 | Sony Corp | 音声合成方法 |
Non-Patent Citations (1)
Title |
---|
See also references of EP2096631A4 * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110689488A (zh) * | 2019-08-22 | 2020-01-14 | 稿定(厦门)科技有限公司 | 图像调色方法、介质、设备及装置 |
Also Published As
Publication number | Publication date |
---|---|
BRPI0720266A2 (pt) | 2014-01-28 |
EP2096631A4 (en) | 2012-07-25 |
JPWO2008072671A1 (ja) | 2010-04-02 |
EP2096631A1 (en) | 2009-09-02 |
US20100332223A1 (en) | 2010-12-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2008072671A1 (ja) | 音声復号化装置およびパワ調整方法 | |
US8069040B2 (en) | Systems, methods, and apparatus for quantization of spectral envelope representation | |
CN101180676B (zh) | 用于谱包络表示的向量量化的方法和设备 | |
US20100010810A1 (en) | Post filter and filtering method | |
EP2282309A2 (en) | Sub-band voice with multi-stage codebooks and redundant coding | |
CN1947173B (zh) | 分层编码装置及分层编码方法 | |
RU2636685C2 (ru) | Решение относительно наличия/отсутствия вокализации для обработки речи | |
JPH09127996A (ja) | 音声復号化方法及び装置 | |
CA2659197A1 (en) | Time-warping frames of wideband vocoder | |
JP2003501675A (ja) | 時間同期波形補間によるピッチプロトタイプ波形からの音声を合成するための音声合成方法および音声合成装置 | |
JP6644848B2 (ja) | ベクトル量子化装置、音声符号化装置、ベクトル量子化方法、及び音声符号化方法 | |
JPWO2007037359A1 (ja) | 音声符号化装置および音声符号化方法 | |
EP3079151A1 (en) | Audio encoder and method for encoding an audio signal | |
JPWO2008001866A1 (ja) | 音声符号化装置及び音声符号化方法 | |
WO2007129726A1 (ja) | 音声符号化装置及び音声符号化方法 | |
WO2012053146A1 (ja) | 符号化装置及び符号化方法 | |
WO2011048810A1 (ja) | ベクトル量子化装置及びベクトル量子化方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 07859788 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2008549343 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2007859788 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 12517603 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: PI0720266 Country of ref document: BR Kind code of ref document: A2 Effective date: 20090612 |