WO2007129726A1 - 音声符号化装置及び音声符号化方法 - Google Patents
音声符号化装置及び音声符号化方法 Download PDFInfo
- Publication number
- WO2007129726A1 WO2007129726A1 PCT/JP2007/059580 JP2007059580W WO2007129726A1 WO 2007129726 A1 WO2007129726 A1 WO 2007129726A1 JP 2007059580 W JP2007059580 W JP 2007059580W WO 2007129726 A1 WO2007129726 A1 WO 2007129726A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- codebook
- sound source
- encoding
- value
- vector
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims description 34
- 239000013598 vector Substances 0.000 claims abstract description 67
- 230000003044 adaptive effect Effects 0.000 claims description 33
- 230000005284 excitation Effects 0.000 claims description 23
- 230000001755 vocal effect Effects 0.000 claims description 6
- 230000005236 sound signal Effects 0.000 claims description 5
- 230000003595 spectral effect Effects 0.000 claims description 4
- 238000001228 spectrum Methods 0.000 abstract description 3
- 230000006870 function Effects 0.000 description 32
- 238000012545 processing Methods 0.000 description 14
- 238000004364 calculation method Methods 0.000 description 11
- 230000015572 biosynthetic process Effects 0.000 description 9
- 238000003786 synthesis reaction Methods 0.000 description 9
- 238000005516 engineering process Methods 0.000 description 7
- 238000010586 diagram Methods 0.000 description 5
- 238000007781 pre-processing Methods 0.000 description 5
- 238000004458 analytical method Methods 0.000 description 4
- 238000013139 quantization Methods 0.000 description 4
- YGLMVCVJLXREAK-MTVMDMGHSA-N 1,1-dimethyl-3-[(1S,2R,6R,7S,8R)-8-tricyclo[5.2.1.02,6]decanyl]urea Chemical compound C([C@H]12)CC[C@@H]1[C@@H]1C[C@@H](NC(=O)N(C)C)[C@H]2C1 YGLMVCVJLXREAK-MTVMDMGHSA-N 0.000 description 2
- 230000006835 compression Effects 0.000 description 2
- 238000007906 compression Methods 0.000 description 2
- 238000009795 derivation Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 230000010354 integration Effects 0.000 description 2
- 239000011159 matrix material Substances 0.000 description 2
- 238000010295 mobile communication Methods 0.000 description 2
- 235000018185 Betula X alpestris Nutrition 0.000 description 1
- 235000018212 Betula X uliginosa Nutrition 0.000 description 1
- 241001237745 Salamis Species 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 235000015175 salami Nutrition 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/04—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
- G10L19/08—Determination or coding of the excitation function; Determination or coding of the long-term prediction parameters
- G10L19/12—Determination or coding of the excitation function; Determination or coding of the long-term prediction parameters the excitation function being a code excitation, e.g. in code excited linear prediction [CELP] vocoders
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/04—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
- G10L19/08—Determination or coding of the excitation function; Determination or coding of the long-term prediction parameters
- G10L19/10—Determination or coding of the excitation function; Determination or coding of the long-term prediction parameters the excitation function being a multipulse excitation
- G10L19/107—Sparse pulse excitation, e.g. by using algebraic codebook
Definitions
- the present invention relates to a speech coding apparatus and speech coding method for performing fixed codebook search.
- Non-patent document 1 describes a speech coding technique whose performance has been greatly improved by “CEL PJ (Code Excited Linear Prediction)”, a basic method that skillfully applies vector quantization by modeling speech utterance mechanisms.
- CEL PJ Code Excited Linear Prediction
- the target technology has also come out.
- Patent Document 1 discloses that a coding distortion of a noisy code band is calculated when encoding with a CELP fixed excitation codebook, and the calculation result is noisy. While multiplying a fixed weighting value according to the degree of noise, the coding distortion of the non-noisy driving excitation vector is calculated, and the calculation result is multiplied by a fixed weighting value according to the degree of noise. However, it is described that the driving excitation code related to the multiplication result with the smaller value is selected.
- a non-noisy (pulse-like) code vector tends to be selected because the distance from the input signal to be encoded tends to be smaller than a noisy code vector.
- the technology described in Patent Document 1 divides the codebook into two parts, noise and non-noise, and calculates the distance for each. According to the result of the above, a weight (multiplication to the distance) is applied so that a non-noisy code envelope is easily selected. This encodes the noisy input speech and restores it. The sound quality of the synthesized speech can be improved.
- Patent Document 1 Japanese Patent No. 3404016
- Non-Patent Document 1 Salami, Laflamme, Adoul, "8kbit / s ACELP Coding of Speech with 10ms Speech-Frame: a Candidate for CCITT Standardization, ⁇ Proc. ICASSP94, pp. II-97n
- the distance calculation result is weighted by multiplication, and the weight by multiplication does not affect the absolute value of the distance. This means that the same weight is applied both when the distance is long and when it is close, and it cannot be said that the tendency according to the noise and non-noise characteristics of the input signal to be encoded is fully utilized.
- An object of the present invention is to provide a speech encoding apparatus and speech encoding method that can sufficiently utilize the tendency according to noise characteristics and non-noise characteristics of an input signal to be encoded to obtain good sound quality. It is.
- the speech coding apparatus includes first coding means for coding vocal tract information in the input speech signal into spectral envelope information, and excitation information in the input speech signal as the adaptive codebook and fixed code.
- a configuration having weighting means for performing weighting according to the number of pulses forming the sound source vector to the calculated value serving as a search reference is adopted.
- the speech coding method converts the vocal tract information of the input speech signal into spectral envelope information.
- FIG. 1 is a block diagram showing a configuration of a CELP coding apparatus according to an embodiment of the present invention.
- FIG. 2 is a block diagram showing the internal configuration of the distortion minimizing section shown in FIG.
- FIG. 1 is a block diagram showing a configuration of CELP encoding apparatus 100 according to the embodiment of the present invention.
- This CELP code encoder 100 encodes an audio signal S 11 composed of vocal tract information and sound source information by obtaining an LPC parameter (linear prediction coefficient) for the vocal tract information, and the sound source information is pre- By obtaining an index for identifying which of the stored speech models is used, that is, an index for identifying what excitation vector (code vector) is generated in the adaptive codebook 103 and the fixed codebook 104.
- the sound source information is encoded.
- each unit of CELP encoding apparatus 100 performs the following operation.
- the LPC analysis unit 101 performs linear prediction analysis on the speech signal S11, obtains an LPC parameter that is spectrum envelope information, and outputs the LPC parameter to the LPC quantization unit 102 and the perceptual weighting unit 111.
- the LPC quantization unit 102 quantizes the LPC parameters obtained by the LPC analysis unit 101, and obtains them.
- the quantized LPC parameter is output to the LPC synthesis filter 109, and the quantized LPC parameter index is output to the outside of the CELP encoder 100.
- adaptive codebook 103 stores past driving sound sources used in LPC synthesis filter 109, and stores them according to the adaptive codebook lag corresponding to the index instructed from distortion minimizing section 112.
- a sound source vector for one subframe is generated from the driving sound source.
- This excitation vector is output to multiplier 106 as an adaptive codebook vector.
- Fixed codebook 104 stores a plurality of excitation vectors having a predetermined shape in advance, and uses the excitation vector corresponding to the index designated by distortion minimizing section 112 as a fixed codebook vector. Output to 107.
- fixed codebook 104 is an algebraic codebook, and the configuration in the case where an algebraic codebook using two kinds of pulses is used will be described in the case where weighting is performed by addition.
- An algebraic sound source is a sound source used in many standard codecs, and its position and polarity (
- the adaptive codebook 103 is used to express a component with strong periodicity such as voiced sound, while the fixed codebook 104 is a component with weak periodicity such as white noise. Used to express
- Gain codebook 105 is output from adaptive codebook vector gain (adaptive codebook gain) output from adaptive codebook 103 and from fixed codebook 104 in accordance with instructions from distortion minimizing section 112.
- a fixed codebook vector gain (fixed codebook gain) is generated and output to multipliers 106 and 107, respectively.
- Multiplier 106 multiplies the adaptive codebook gain output from gain codebook 105 by the adaptive codebook vector output from adaptive codebook 103 and outputs the result to adder 108.
- Multiplier 107 multiplies the fixed codebook gain output from gain codebook 105 by the fixed codebook vector output from fixed codebook 104 and outputs the result to adder 108.
- Adder 108 adds the adaptive codebook vector output from multiplier 106 and the fixed codebook vector output from multiplier 107, and uses the added excitation vector as the driving excitation. Output to LPC synthesis filter 109.
- LPC synthesis filter 109 uses the quantized LPC parameters output from LPC quantization section 102 as filter coefficients, and uses the excitation code generated in adaptive codebook 103 and fixed codebook 104 as the driving excitation.
- the synthesized signal is generated using the filtered filter function, that is, the LPC synthesis filter. This synthesized signal is output to adder 110.
- the adder 110 calculates an error signal by subtracting the synthesized signal generated by the LPC synthesis filter 109 from the audio signal S 11, and outputs the error signal to the audibility weighting unit 111.
- This error signal corresponds to coding distortion.
- the perceptual weighting unit 111 performs perceptual weighting on the sign distortion output from the adder 110 and outputs the result to the distortion minimizing unit 112.
- Distortion minimizing section 112 subframes each index of adaptive codebook 103, fixed codebook 104, and gain codebook 105 so that the code distortion output from perceptual weighting section 111 is minimized. Each index is obtained and output as an index key information to the outside of the CELP encoding apparatus 100. More specifically, a series of processes for generating a composite signal based on the adaptive codebook 103 and the fixed codebook 104 described above and obtaining the encoding distortion of this signal is closed-loop control (feedback control). The distortion minimizing section 112 searches for each codebook by changing the index to instruct each codebook in one subframe, and finally obtains each codebook that minimizes the coding distortion. Print the book index.
- the driving sound source when the code distortion is minimized is the adaptive codebook 1 for each subframe.
- Adaptive codebook 103 updates the stored driving sound source by this feedback.
- the search for the excitation vector and the derivation of the code are performed by searching for the excitation vector that minimizes the coding distortion in Eq. (1) below.
- E coding distortion
- X coding target
- p adaptive codebook vector gain
- H auditory weight Attached synthesis filter
- a adaptive codebook vector
- q fixed codebook vector gain
- s fixed codebook vector
- E coding distortion
- X coding target (perceptual weighted speech signal)
- p optimal gain of adaptive codebook extraneous
- H perceptual weighting synthesis filter
- a adaptive codebook vector
- q fixed codebook Vector gain
- s Fixed codebook vector
- y Target code for fixed codebook search
- the function C can be calculated with a small amount of calculation if y H and HH are calculated in advance.
- FIG. 2 is a block diagram showing an internal configuration of distortion minimizing section 112 shown in FIG.
- adaptive codebook search section 201 uses perceptual weighting in perceptual weighting section 111.
- the adaptive codebook 103 is searched using the distorted coding distortion.
- the code of the adaptive codebook vector is output to the preprocessing unit 203 and the adaptive codebook 103 of the fixed codebook search unit 202.
- Preprocessing section 203 of fixed codebook search section 202 calculates vector yH and matrix HH using coefficient H of the synthesis filter in perceptual weighting section 111.
- yH is obtained by convolving the matrix H with the target vector y reversed, and then reversing the result.
- HH is obtained by multiplying the matrices.
- the additional value g is obtained from the y value and the fixed value G to be added as shown in the following equation (5).
- the preprocessing unit 203 determines the polarity of the pulse in advance from the polarity (+ _) of the element of the vector yH. Specifically, the polarity of the pulse at each position is matched to the polarity of the value of yH at that position, and the polarity of the value of yH is stored in another array. After storing the polarity of each position in a separate array, all yH values are absolute values and converted to positive values. Also, the HH value is converted by multiplying the polarity according to the polarity of each stored position. The obtained yH and HH are output to the correlation value / sound source value adding sections 205 and 209 in the search loops 204 and 208, and the additional value g is output to the weight adding section 206.
- the search loop 204 includes a correlation value / sound source value adding unit 205, a weight addition unit 206 and a magnitude determination unit 207, and the search loop 208 includes a correlation value / sound source value addition unit 209 and a magnitude determination unit 210. ing.
- Correlation value / sound source power adding section 205 obtains function C by adding the values of yH and HH output from preprocessing section 203 for the case of two pulses, and weights the obtained function C. Outputs to appendix 206.
- the weight addition unit 206 performs addition processing on the function C using the additional value g shown in the above equation (5).
- the added function C is output to the magnitude determination unit 207.
- the magnitude determination unit 207 compares the value of the function C value added in the weight addition unit 206 and overwrites and stores the numerator denominator of the function C when a larger function value is indicated. And thus, the largest function C in the entire search loop 204 is output to the magnitude determination unit 210 of the search loop 208.
- the correlation value.sound source value adding unit 209 for the case of 3 pulses, is similar to the correlation value / sound source value adding unit 205 in the search loop 204, and the yH and HH output from the preprocessing unit 203.
- the function C is obtained by adding the values, and the obtained function C is output to the magnitude determination unit 210.
- the magnitude determination unit 210 compares the magnitudes of the values of the function C output from the correlation value / sound source value addition unit 209, including the function C output from the magnitude determination unit 207 of the search loop 204, Overwrites the numerator denominator of function C for large function values. Then, a search is made for a combination of pulse positions that is the largest in the entire search loop 208.
- the size determination unit 210 combines the code of the position of each pulse and the code of the polarity into the code of the fixed codebook vector, and outputs this code to the fixed codebook 104 and the gain codebook search unit 211.
- Gain codebook search section 211 searches for a gain codebook based on the code of a fixed codebook vector obtained by combining the code of the position and the polarity of each pulse output from magnitude determination section 210, The search result is output to gain codebook 105.
- FIG. 3 a series of processing procedures using the above two search loops 204 and 208 are shown in FIG. 3 and FIG.
- a candidate position of codebook 0 (number of pulses 2) is set, initialization is performed in ST302, and in ST303, it is confirmed that i0 is less than 20. If iO is less than 20, the position of the first pulse from codebook 0 is output, the values are extracted from yH and HH, and are set as correlation value syO and path shO, respectively (ST304). Repeat this calculation until iO reaches 20 (the number of pulse positions) (ST303 to ST306). In ST302 to ST309, f or no. The codebook search process is performed when there are two nores.
- the size of the function C is compared (ST308), and the numerator and denominator of the function C showing the larger function value are stored (3 Ding 309) This calculation is repeated until 1 is 20 (the number of pulse position candidates) (ST305 to ST310).
- ST312 confirms that iO is less than 10, and if iO is less than 10, it outputs the codebook 1 power, the position of the first pulse, and extracts the value from yH and HH.
- the correlation value is syO and the value is shO respectively (ST313). This calculation is repeated until iO reaches 10 (number of pulse position candidates) (ST312 to ST315).
- the position of the third pulse from the codebook 1 is output, the values are extracted from yH and HH, and added to the correlation values syl and ⁇ shl, The correlation value is sy2 and the path is sh2 (ST319).
- the numerator and denominator of function C showing the maximum function value is compared with the function value consisting of correlation values sy2 and sh2 (ST 320), and the numerator and denominator of function C showing the larger function value are Store (ST321) This calculation is repeated until i2 reaches 8 (number of pulse position candidates) (ST317 to ST322).
- the number of pulses of 3 is more easily selected than the number of pulses of 2 due to the effect of the additional value g .
- weighting based on a clear criterion “number of pulses” can be realized.
- additional processing as a weighting method, if the error from the target vector that is the target of the sign is large (unvoiced (noise) with distributed energy), weighting is performed. If has a relatively large meaning and the error is small (concentrated voiced), the weighting has a relatively small meaning. Therefore, a higher quality synthesized sound can be obtained. The reason is shown qualitatively as follows.
- the target scale is voiced (non-noise)
- a function value serving as a selection criterion has a high part and a low part.
- the sound source vector is selected based on the magnitude of the function value alone, and there is no significant change in the addition processing of the fixed value of the present invention. Therefore, the sound source beta is selected based on the magnitude of the function value alone.
- the effectiveness of the addition processing has been described as the weighting method, but the present invention is also effective when using multiplication. This is because if the corresponding part in Fig. 3 is replaced as shown in the following equation (6), the weighting process based on a clear criterion "number of pulses" can be realized.
- two and three fixed codebook vector pulses are used, but this may be a combination of any number of pulses. This is because the present invention depends on the number of pulses.
- the force S using two kinds of variations of the number of pulses which can be any number. This can be easily achieved by using a smaller number and a smaller value, and the search process can be the concatenation process shown in FIG.
- the inventor used it to search for five types of fixed codebook vectors with 1 to 5 pulses it was confirmed by encoding and decoding experiments that good performance was obtained with the following numerical values. Yes.
- the present invention is applied to a codebook in which the number of sets is divided.
- fixed codebook vectors having different numbers of pulses may be mixed in the codebook. That This is because the set of fixed codebook vectors having a predetermined number of pulses does not need to be gathered because the additional processing of the present invention is used in the function value determination part.
- an algebraic codebook is used as an example of the fixed codebook in this embodiment, but this is because a fixed codebook vector is directly stored in a conventional multipulse codebook or ROM. It is obvious that the present invention can be applied to a learning codebook in a written format. The number of multipulses itself is used in the same way as in the present invention, and even when all fixed codebook vectors have values, it is easy to extract information such as the number of amplitudes, etc. Yes, if you use it.
- the present invention can be applied to any encoding / decoding method in which there is a codebook in which excitation vectors for which the number of powers used for CELP is known are stored. It is clear. This is because the present invention is only in the search of the fixed codebook vector, and does not depend on the presence / absence of the adaptive codebook and the analysis method of the spectrum envelope, whether it is LPC power, FFT or finole bank.
- Each functional block used in the description of the present embodiment is typically realized as an LSI which is an integrated circuit. These may be individually made into one chip, or may be made into one chip so as to include a part or all of them. Here, it is sometimes called IC, system LSI, super LSI, or ultra LSI, depending on the difference in power integration.
- circuit integration is not limited to LSI's, and implementation using dedicated circuitry or general purpose processors is also possible.
- An FPGA Field Programmable Gate Array
- reconfigurable 'processor that can reconfigure the connection and settings of circuit cells inside the LSI may be used.
- the adaptive codebook used in the description of the present embodiment is sometimes called an adaptive excitation codebook.
- the fixed codebook is sometimes called a fixed excitation codebook.
- the speech coding apparatus and speech coding method according to the present invention can fully utilize the tendency according to the noise characteristics and non-noise characteristics of the input signal to be encoded, and can obtain good sound quality. For example, it can be applied to a mobile phone or the like in a mobile communication system.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- Multimedia (AREA)
- Acoustics & Sound (AREA)
- Human Computer Interaction (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Health & Medical Sciences (AREA)
- Signal Processing (AREA)
- Mathematical Analysis (AREA)
- Theoretical Computer Science (AREA)
- Pure & Applied Mathematics (AREA)
- Mathematical Physics (AREA)
- Mathematical Optimization (AREA)
- General Physics & Mathematics (AREA)
- Algebra (AREA)
- Compression, Expansion, Code Conversion, And Decoders (AREA)
Abstract
Description
Claims
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2008514506A JPWO2007129726A1 (ja) | 2006-05-10 | 2007-05-09 | 音声符号化装置及び音声符号化方法 |
US12/299,986 US20090164211A1 (en) | 2006-05-10 | 2007-05-09 | Speech encoding apparatus and speech encoding method |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2006-131851 | 2006-05-10 | ||
JP2006131851 | 2006-05-10 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2007129726A1 true WO2007129726A1 (ja) | 2007-11-15 |
Family
ID=38667834
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2007/059580 WO2007129726A1 (ja) | 2006-05-10 | 2007-05-09 | 音声符号化装置及び音声符号化方法 |
Country Status (3)
Country | Link |
---|---|
US (1) | US20090164211A1 (ja) |
JP (1) | JPWO2007129726A1 (ja) |
WO (1) | WO2007129726A1 (ja) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2099025A4 (en) * | 2006-12-14 | 2010-12-22 | Panasonic Corp | AUDIO CODING DEVICE AND AUDIO CODING METHOD |
CN100578620C (zh) * | 2007-11-12 | 2010-01-06 | 华为技术有限公司 | 固定码书搜索方法及搜索器 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1999034354A1 (en) * | 1997-12-24 | 1999-07-08 | Mitsubishi Denki Kabushiki Kaisha | Sound encoding method and sound decoding method, and sound encoding device and sound decoding device |
JP2002169595A (ja) * | 2000-11-30 | 2002-06-14 | Matsushita Electric Ind Co Ltd | 固定音源符号帳及び音声符号化/復号化装置 |
JP2002518694A (ja) * | 1998-06-09 | 2002-06-25 | 松下電器産業株式会社 | 音声符号化装置及び音声復号化装置 |
JP2002196799A (ja) * | 2000-12-26 | 2002-07-12 | Mitsubishi Electric Corp | 音声符号化装置及び音声符号化方法 |
Family Cites Families (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
FI98104C (fi) * | 1991-05-20 | 1997-04-10 | Nokia Mobile Phones Ltd | Menetelmä herätevektorin generoimiseksi ja digitaalinen puhekooderi |
US5396576A (en) * | 1991-05-22 | 1995-03-07 | Nippon Telegraph And Telephone Corporation | Speech coding and decoding methods using adaptive and random code books |
ATE192259T1 (de) * | 1995-11-09 | 2000-05-15 | Nokia Mobile Phones Ltd | Verfahren zur synthetisierung eines sprachsignalblocks in einem celp-kodierer |
EP1071081B1 (en) * | 1996-11-07 | 2002-05-08 | Matsushita Electric Industrial Co., Ltd. | Vector quantization codebook generation method |
FI113571B (fi) * | 1998-03-09 | 2004-05-14 | Nokia Corp | Puheenkoodaus |
US6173257B1 (en) * | 1998-08-24 | 2001-01-09 | Conexant Systems, Inc | Completed fixed codebook for speech encoder |
EP1959435B1 (en) * | 1999-08-23 | 2009-12-23 | Panasonic Corporation | Speech encoder |
CN1202514C (zh) * | 2000-11-27 | 2005-05-18 | 日本电信电话株式会社 | 编码和解码语音及其参数的方法、编码器、解码器 |
JP4245288B2 (ja) * | 2001-11-13 | 2009-03-25 | パナソニック株式会社 | 音声符号化装置および音声復号化装置 |
CN101615396B (zh) * | 2003-04-30 | 2012-05-09 | 松下电器产业株式会社 | 语音编码设备、以及语音解码设备 |
SG123639A1 (en) * | 2004-12-31 | 2006-07-26 | St Microelectronics Asia | A system and method for supporting dual speech codecs |
JP3981399B1 (ja) * | 2006-03-10 | 2007-09-26 | 松下電器産業株式会社 | 固定符号帳探索装置および固定符号帳探索方法 |
-
2007
- 2007-05-09 JP JP2008514506A patent/JPWO2007129726A1/ja not_active Withdrawn
- 2007-05-09 US US12/299,986 patent/US20090164211A1/en not_active Abandoned
- 2007-05-09 WO PCT/JP2007/059580 patent/WO2007129726A1/ja active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1999034354A1 (en) * | 1997-12-24 | 1999-07-08 | Mitsubishi Denki Kabushiki Kaisha | Sound encoding method and sound decoding method, and sound encoding device and sound decoding device |
JP2002518694A (ja) * | 1998-06-09 | 2002-06-25 | 松下電器産業株式会社 | 音声符号化装置及び音声復号化装置 |
JP2002169595A (ja) * | 2000-11-30 | 2002-06-14 | Matsushita Electric Ind Co Ltd | 固定音源符号帳及び音声符号化/復号化装置 |
JP2002196799A (ja) * | 2000-12-26 | 2002-07-12 | Mitsubishi Electric Corp | 音声符号化装置及び音声符号化方法 |
Non-Patent Citations (1)
Title |
---|
YAMAURA T. ET AL.: "Onsei Yotai ni Motozuku Omomi Tsuki Yugami ni yoru CELP Ongen Tansaku", PROCEEDINGS OF THE 2001 IEICE GENERAL CONFERENCE JOHO.SYSTEM 1, vol. D-14-25, 7 March 2001 (2001-03-07), pages 195, XP003016947 * |
Also Published As
Publication number | Publication date |
---|---|
US20090164211A1 (en) | 2009-06-25 |
JPWO2007129726A1 (ja) | 2009-09-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6480822B2 (en) | Low complexity random codebook structure | |
TW448417B (en) | Speech encoder adaptively applying pitch preprocessing with continuous warping | |
CN101180676B (zh) | 用于谱包络表示的向量量化的方法和设备 | |
CN100369112C (zh) | 可变速率语音编码 | |
US6393390B1 (en) | LPAS speech coder using vector quantized, multi-codebook, multi-tap pitch predictor and optimized ternary source excitation codebook derivation | |
EP1224662A1 (en) | Variable bit-rate celp coding of speech with phonetic classification | |
WO2007047037A2 (en) | An adaptive equalizer for a coded speech signal | |
WO2008072701A1 (ja) | ポストフィルタおよびフィルタリング方法 | |
AU2008222241A1 (en) | Encoding device and encoding method | |
US7596491B1 (en) | Layered CELP system and method | |
WO2008072671A1 (ja) | 音声復号化装置およびパワ調整方法 | |
WO2007105587A1 (ja) | 固定符号帳探索装置および固定符号帳探索方法 | |
CN100593195C (zh) | 在语音编码系统中对增益信息进行编码的方法和装置 | |
JP5388849B2 (ja) | 音声符号化装置および音声符号化方法 | |
JP6644848B2 (ja) | ベクトル量子化装置、音声符号化装置、ベクトル量子化方法、及び音声符号化方法 | |
WO2007129726A1 (ja) | 音声符号化装置及び音声符号化方法 | |
EP2099025A1 (en) | Audio encoding device and audio encoding method | |
JPWO2008001866A1 (ja) | 音声符号化装置及び音声符号化方法 | |
EP2116996A1 (en) | Encoding device and encoding method | |
JP3510168B2 (ja) | 音声符号化方法及び音声復号化方法 | |
JP2013068847A (ja) | 符号化方法及び符号化装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 07743015 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2008514506 Country of ref document: JP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 12299986 Country of ref document: US |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 07743015 Country of ref document: EP Kind code of ref document: A1 |