WO2008072737A1 - 符号化装置、復号装置およびこれらの方法 - Google Patents
符号化装置、復号装置およびこれらの方法 Download PDFInfo
- Publication number
- WO2008072737A1 WO2008072737A1 PCT/JP2007/074141 JP2007074141W WO2008072737A1 WO 2008072737 A1 WO2008072737 A1 WO 2008072737A1 JP 2007074141 W JP2007074141 W JP 2007074141W WO 2008072737 A1 WO2008072737 A1 WO 2008072737A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- spectrum
- decoding
- band
- decoded
- encoding
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/04—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
- G10L19/16—Vocoder architecture
- G10L19/18—Vocoders using multiple modes
- G10L19/24—Variable rate codecs, e.g. for generating different qualities using a scalable representation such as hierarchical encoding or layered encoding
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/038—Speech enhancement, e.g. noise reduction or echo cancellation using band spreading techniques
Definitions
- the present invention relates to an encoding device, a decoding device, and a method thereof used in a communication system that encodes and transmits a signal.
- Non-Patent Document 1 parameters that generate high-frequency spectrum data from low-frequency spectrum data by converting input signals into frequency-domain components and using the correlation between low-frequency spectrum data and high-frequency spectrum data.
- Non-Patent Document 1 Masahiro Oshikiri, Hiroyuki Ehara, Koji Yoshida, ⁇ Improvement of Ultra Wideband Scalable Speech Coding Using Spectral Coding Based on Pitch Filtering, '' Sound Lecture 2-4-13, p. 297-298, Sep. 2004.
- the upper layer layer on the decoding side uses the spectrum data as it is in the high frequency band obtained by extending the bandwidth in the lower layer. It cannot be said that the high-frequency spectrum data with sufficient accuracy is reproduced.
- An object of the present invention is to calculate high-frequency spectrum data with high accuracy using low-frequency spectrum data on the decoding side, and obtain a decoded signal with higher quality. Coding apparatus, decoding apparatus, and methods thereof.
- the encoding device of the present invention includes a first encoding means for generating a first encoded data by encoding a low-frequency part of the input signal that is a band lower than a predetermined frequency, and the first code First decoding means for decoding the encoded data to generate a first decoded signal, and a second encoded data by encoding a predetermined band portion of the residual signal of the input signal and the first decoded signal
- the low-frequency part of one of the input signal, the first decoded signal, and the calculated signal calculated using the first decoded signal is obtained as a finalizer.
- a filtering unit that obtains a pitch coefficient and a filtering coefficient for obtaining a high frequency part that is a band higher than the predetermined frequency of the input signal.
- the decoding device of the present invention is a decoding device using a scalable codec having a layer configuration of r layers (r is an integer of 2 or more), and is an m-th layer (m is an integer of r or less). ) And receiving means for receiving the band extension parameter calculated using the decoded signal, and using the band extension parameter for the low frequency component of the decoded signal of the nth layer (n is an integer equal to or less than r). And a decoding means for generating a high frequency component.
- the decoding device of the present invention includes first encoded data that is transmitted from the encoding device and that encodes a low-frequency portion that is a band lower than a predetermined frequency in the input signal in the encoding device; Second encoded data obtained by encoding a predetermined band portion of the residual between the first decoded spectrum obtained by decoding the first encoded data and the spectrum of the input signal; the input signal; Of the decoded spectrum and the first added spectrum obtained by adding the first decoded spectrum and the second decoded spectrum obtained by decoding the second encoded data, any one of the low-frequency parts And receiving means for receiving a pitch coefficient and a filtering coefficient for obtaining a high-frequency portion that is higher than the predetermined frequency of the input signal and decoding the first encoded data.
- a first decoding means for generating a third decoded spectrum in a low frequency Before A first decoding means for generating a third decoded spectrum in a low frequency; a second decoding means for decoding the second encoded data to generate a fourth decoded spectrum in the predetermined band portion; and the pitch coefficient and Using a filtering coefficient, the third decoded spectrum, the fourth decoded spectrum, And a fifth decoding spectrum generated using both of them, by band-extending any one of the fifth decoded spectrum, thereby decoding a band portion that has not been decoded by the first decoding means and the second decoding means. And a decoding means.
- the encoding method of the present invention includes a first encoding step of generating a first encoded data by encoding a low-frequency part of a band lower than a predetermined frequency in the input signal, and the first code A decoding step of decoding the encoded data to generate a first decoded signal, and a second encoded data by encoding a predetermined band portion of the residual signal of the input signal and the first decoded signal Filtering the low frequency part of one of the two encoding steps and the input signal, the first decoded signal, and a calculated signal calculated using the first decoded signal. And a filtering step for obtaining a pitch coefficient and a filtering coefficient for obtaining a high-frequency portion that is a band higher than the predetermined frequency of the input signal.
- a decoding method of the present invention is a decoding method using a scalable codec having a layer configuration of r layers (r is an integer of 2 or more), and is an m-th layer (m is an integer of r or less)
- the band extension parameter is used for the reception step of receiving the band extension parameter calculated using the decoded signal of (1) and the low frequency component of the decoded signal of the nth layer (n is an integer equal to or less than r).
- a decoding step for generating a high frequency component.
- the decoding method includes first encoded data obtained by encoding a low-frequency portion, which is a band lower than a predetermined frequency, of the input signal in the encoding device transmitted from the encoding device, and the first Second encoded data obtained by encoding a predetermined band portion of a residual between the first decoded spectrum obtained by decoding the encoded data and the spectrum of the input signal, the input signal, the first decoded spectrum, Of the first addition spectrum obtained by adding the first decoded spectrum and the second decoded spectrum obtained by decoding the second encoded data, one of the low-frequency portions is filtered.
- a first decoding step of generating a third decoded scan Bae spectrum in a second decoding step of generating a fourth decoded spectrum in said predetermined band portion by decoding the second encoded data, the first condensate And the second decoding step are generated using the third decoded spectrum, the fourth decoding spectrum, and both using the pitch coefficient and the filtering coefficient.
- the coding band is selected in the higher layer on the coding side, the band is expanded on the decoding side, and the band components that could not be decoded in the lower layer and the higher layer are reduced.
- decoding high-frequency spectrum data with high accuracy can be calculated flexibly according to the coding band selected in the higher layer on the coding side, and a V-quality decoded signal with better quality can be obtained. .
- FIG. 1 is a block diagram showing the main configuration of an encoding apparatus according to Embodiment 1 of the present invention.
- FIG. 2 is a block diagram showing the main configuration inside the second layer coding section according to Embodiment 1 of the present invention.
- FIG. 3 is a block diagram showing the main configuration inside the spectrum encoding section according to Embodiment 1 of the present invention.
- FIG. 4 is a diagram for explaining an outline of filtering processing of the filtering unit according to Embodiment 1 of the present invention.
- FIG. 5 is a diagram for explaining how the spectrum of the estimated value of the input spectrum changes as the pitch coefficient T according to the first embodiment of the present invention changes.
- FIG. 6 is a diagram for explaining how the spectrum of the estimated value of the input spectrum changes as the pitch coefficient T according to the first embodiment of the present invention changes.
- FIG. 7 is a flowchart showing a procedure of processes performed in a pitch coefficient setting unit, a filtering unit, and a search unit according to Embodiment 1 of the present invention.
- FIG. 8 is a block diagram showing the main configuration of the decoding apparatus according to Embodiment 1 of the present invention.
- FIG. 9 is a block diagram showing the main configuration inside the second layer decoding section according to Embodiment 1 of the present invention.
- FIG. 10 is a block diagram showing the main components inside the spectrum decoding section according to Embodiment 1 of the present invention. Lock figure
- FIG. 11 is a diagram showing a decoding spectrum generated in the filtering unit according to Embodiment 1 of the present invention.
- FIG. 12 is a diagram showing a case where the band of the second spectrum S2 (k) completely overlaps the band of the first spectrum S1 (k) according to Embodiment 1 of the present invention.
- FIG. 13 is a diagram showing a case where the band of the first spectrum S 1 (k) and the band of the second spectrum S2 (k) according to Embodiment 1 of the present invention are not adjacent but separated from each other
- FIG. 14 is a block diagram showing the main configuration of an encoding apparatus according to Embodiment 2 of the present invention.
- FIG. 15 is a block diagram showing the main configuration inside the spectrum encoding section according to Embodiment 2 of the present invention.
- FIG. 16 is a block diagram showing the main configuration of an encoding apparatus according to Embodiment 3 of the present invention.
- FIG. 17 is a block diagram showing the main components inside the spectrum encoding unit according to Embodiment 3 of the present invention.
- FIG. 1 is a block diagram showing the main configuration of coding apparatus 100 according to Embodiment 1 of the present invention.
- an encoding apparatus 100 includes a downsampling unit 101, a first layer encoding unit 102, a first layer decoding unit 103, an upsampling unit 104, a delay unit 105, a second layer encoding unit 106, A spectrum encoding unit 107 and a multiplexing unit 108 are provided, and a scalable configuration consisting of two layers is adopted.
- the first layer of the encoding device 100 encodes the speech 'audio signal input using the CELP (Code Exited Linear Prediction) method encoding method
- the second layer encoding inputs the first layer decoded signal. Encode the residual signal with the signal.
- the encoding apparatus 100 divides the input signal into N (N is a natural number) samples, and encodes each frame with N samples as one frame.
- the down-sampling unit 101 performs down-sampling processing on an input audio signal and / or audio signal (hereinafter referred to as audio “audio signal”), and The sampling frequency of one audio signal is converted into Rate 1 force and Rate 2 (Rate 1> Rate 2), and output to the first layer encoding unit 102.
- audio signal an input audio signal and / or audio signal
- the sampling frequency of one audio signal is converted into Rate 1 force and Rate 2 (Rate 1> Rate 2), and output to the first layer encoding unit 102.
- First layer encoding section 102 performs CELP speech encoding on the down-sampled speech 'audio signal input from down sampling section 101, and obtains the obtained first layer encoded information. Output to first layer decoding section 103 and multiplexing section 108. Specifically, the first layer encoding unit 102 obtains an LPC (Linear Prediction Coefficient) parameter by using the vocal tract information for an audio signal composed of vocal tract information and sound source information. For sound source information, an index that specifies which of the previously stored speech models is used, that is, an index that specifies which excitation vector of the adaptive codebook or fixed codebook is to be generated is obtained.
- LPC Linear Prediction Coefficient
- First layer decoding section 103 performs CELP speech decoding on the first layer encoded information input from first layer encoding section 102, and up-sampling the obtained first layer decoded signal Output to 104.
- Upsampling section 104 performs upsampling processing on the first layer decoded signal input from first layer decoding section 103, converts the sampling frequency of the first layer decoded signal from Rate2 to Ratel Output to second layer encoding section 106.
- the delay unit 105 outputs the delayed audio / audio signal to the second layer encoding unit 106 by storing the input audio signal in a built-in buffer and outputting it after a predetermined time.
- the predetermined time that is delayed is a time that takes into account the algorithm delay that occurs in downsampling section 101, first layer encoding section 102, first layer decoding section 103, and upsampling section 104.
- Second layer encoding section 106 converts the audio signal input from delay section 105 into a residual signal between the up-sampled first layer decoded signal input from up sampling section 104.
- second layer encoding is performed by performing gain 'shape quantization, and the obtained second layer encoded information is output to multiplexing section 108.
- the internal configuration and specific operation of second layer encoding section 106 will be described later.
- the spectrum encoding unit 107 converts the input speech 'audio signal into the frequency domain. Analyze the correlation between the low-frequency component and high-frequency component of the input spectrum obtained, calculate the parameters for estimating the high-frequency component by expanding the bandwidth on the decoding side, and calculate the vector coding The information is output to the multiplexing unit 108 as information.
- the internal configuration and specific operation of spectrum encoding section 107 will be described later.
- Multiplexer 108 receives first layer encoded information input from first layer encoder 102, second layer encoded information input from second layer encoder 106, and spectral encoding.
- the spectrum coding information input from unit 107 is multiplexed, and the obtained bit stream is transmitted to the decoding device.
- FIG. 2 is a block diagram showing the main configuration inside second layer encoding section 106.
- second layer encoding section 106 includes frequency domain transform sections 161 and 162, residual MDCT coefficient calculation section 163, band selection section 164, shape quantization section 165, and predictive coding presence / absence determination section. 166, gain quantization section 167, and multiplexing section 168.
- Frequency domain transform section 161 performs a modified discrete cosine transform (MDCT) using the delayed speech 'audio signal input from delay section 105, and obtains the resulting input MDCT coefficient as a residual. Output to MDCT coefficient calculation section 163.
- MDCT modified discrete cosine transform
- Frequency domain transform section 162 performs MDCT using the up-sampled first layer decoded signal input from up-sampling section 104, and obtains the obtained first layer MDCT coefficient as a residual MDCT coefficient calculation section. Output to 163.
- Residual MDCT coefficient calculation section 163 calculates and obtains a residual between the input MD CT coefficient input from frequency domain transform section 161 and the first layer MDCT coefficient input from frequency domain transform section 162. The obtained residual MDCT coefficients are output to band selection section 164 and shape quantization section 165.
- Band selection section 164 divides residual MDCT coefficient input from residual MDCT coefficient calculation section 163 into a plurality of subbands, and a band to be quantized (quantization target band) from the plurality of subbands , And outputs band information indicating the selected band to the shape quantizing unit 165, the predictive coding presence / absence determining unit 166, and the multiplexing unit 168.
- a method of selecting the quantization target band a method of selecting a band having the highest energy, or a method of selecting in consideration of the correlation with the quantization target band selected in the past and energy at the same time. There are laws.
- the shape quantizing unit 165 includes the MDCT corresponding to the quantization target band indicated by the band information input from the band selecting unit 164 among the residual MDC T coefficients input from the residual MDCT coefficient calculating unit 163. Shape quantization is performed using the coefficients, that is, the second layer MDCT coefficients! /, And the obtained shape coding information is output to the multiplexing unit 168.
- the shape quantizing unit 165 obtains an ideal gain value for shape quantization, and outputs the obtained ideal gain value to the gain quantizing unit 167.
- Predictive coding presence / absence determination section 166 uses the band information input from band selection section 164 to determine the sub-subbands common between the quantization target band of the current frame and the quantization target band of the past frame. Find a number. Then, when the number of common sub-subbands is equal to or greater than a predetermined value, the predictive coding presence / absence determining unit 166 applies the residual MDCT coefficient of the quantization target band indicated by the band information, that is, the second layer MDCT coefficient. If it is determined that predictive encoding is to be performed and the number of common sub-subbands is smaller than the predetermined value, it is determined that predictive encoding is not performed on the second layer MDCT coefficient! /. Predictive coding presence / absence determination section 166 outputs the determination result to gain quantization section 167.
- the gain quantizing unit 167 stores the past data stored in the built-in buffer. Using the quantization gain value of the frame and the built-in gain codebook, the gain encoding information is obtained by performing predictive coding of the gain of the quantization target band of the current frame. On the other hand, when the determination result input from the predictive coding presence / absence determining unit 166 indicates a determination result indicating that the predictive encoding is not performed, the gain quantizing unit 167 determines the ideal gain value input from the shape quantizing unit 165. Gain coding information is obtained by performing direct quantization as a quantization target. Gain quantization section 167 outputs gain coding information obtained to multiplexing section 168.
- Multiplexer 168 receives band information input from band selector 164, shape encoded information input from shape quantizer 165, and gain encoded information input from gain quantizer 167. And the obtained bit stream is transmitted to the multiplexing unit 108 as second layer encoded information.
- the band information generated by second layer encoding section 106, the shape encoding information, the gain The encoded information may be input directly to the multiplexing unit 108 without passing through the multiplexing unit 168 and multiplexed with the first layer encoded information and the spectrum encoded information.
- FIG. 3 is a block diagram showing the main configuration inside spectrum coding section 107.
- spectrum coding section 107 has frequency domain conversion section 171, internal state setting section 172, pitch coefficient setting section 173, filtering section 174, search section 175, and filter coefficient calculation section 176.
- the frequency domain transform unit 171 performs frequency transform on an input audio-audio signal whose effective frequency band is 0 ⁇ k ⁇ FH, and calculates an input spectrum S (k).
- DFT discrete Fourier transform
- DCT discrete cosine transform
- MDCT modified discrete cosine transform
- Internal state setting section 172 sets the internal state of the filter used in filtering section 174 using input spectrum S (k) whose effective frequency band is 0 ⁇ k ⁇ FH. The setting of the internal state of this filter will be described later.
- the pitch coefficient setting unit 173 sequentially outputs the pitch coefficient T to the filtering unit 174 while changing the pitch coefficient T little by little within a predetermined search range Tmin to Tma X.
- Filtering section 174 filters the input spectrum using the internal state of the filter set by internal state setting section 172 and pitch coefficient T output from pitch coefficient setting section 173, and estimates the input spectrum. S ′ (k) is calculated. Details of this filtering process will be described later.
- Search section 175 is a parameter indicating the similarity between input spectrum S (k) input from frequency domain transform section 171 and estimated value S '(k) of the input spectrum output from filtering section 174. Similarity is calculated. The similarity calculation process will be described in detail later. This similarity calculation process is performed every time the pitch coefficient T is given from the pitch coefficient setting unit 173 to the filtering unit 174, and the pitch coefficient that maximizes the calculated similarity, that is, the optimum pitch coefficient T ′ (Tmin ⁇ The range of Tmax) is given to the filter coefficient calculation unit 176.
- the filter coefficient calculation unit 176 uses the optimal pitch coefficient T ′ given from the search unit 175 and the input spectrum S (k) inputted from the frequency domain conversion unit 171 to use the filter coefficient / 3
- the filter coefficient / 3 i and the optimum pitch coefficient T are output to the multiplexing unit 108 as spectral coding information. Details of the filter coefficient / 3 calculation process in the filter coefficient calculation unit 176 will be described later.
- FIG. 4 is a diagram for explaining an outline of the filtering process of the filtering unit 174.
- T represents the pitch coefficient input from pitch coefficient setting unit 173
- the input spectrum S (k) is stored as the internal state of the filter in the band of 0 ⁇ k ⁇ FL of S (k).
- the estimated value S ′ (k) of the input spectrum obtained using the following equation (2) is stored in the FL ⁇ k ⁇ FH band of S (k).
- the above filtering process is performed by clearing S (k) to zero each time in the range of FL ⁇ k ⁇ FH every time the pitch coefficient T is given from the pitch coefficient setting unit 173.
- S (k) is calculated and output to search section 175 every time pitch coefficient T changes.
- ⁇ ⁇ represents the square error between S (k) and S '(k).
- the input term on the right side is a fixed value regardless of the pitch coefficient T, so the pitch coefficient that generates S, (k) that maximizes the second term on the right side is searched.
- the second term on the right side of the above equation (3) is defined as the similarity. That is, a pitch coefficient T ′ that maximizes the similarity A expressed by the following equation (4) is searched.
- FIG. 5 is a diagram for explaining how the spectrum of the estimated value S ′ (k) of the input spectrum changes as the pitch coefficient T changes.
- FIG. 5A is a diagram showing an input spectrum S (k) having a harmonic structure stored as an internal state.
- Figures 5B to 5D show the spectra of the input spectrum estimate S '(k) calculated by filtering using the three types of pitch coefficients TO, Tl, and ⁇ 2, respectively. It is.
- FIG. 6 is a diagram for explaining how the spectrum of the estimated value S ′ (k) of the input spectrum changes as the pitch coefficient T changes, as in FIG. However, the phase of the input spectrum stored as the internal state is different from that shown in Fig. 5. In the example shown in Fig. 6 as well, the pitch coefficient T at which the harmonic structure is maintained is T1.
- changing the pitch coefficient T to find T having the maximum similarity means that the pitch of the harmonic structure of the spectrum is an integer multiple thereof) by trie-and-error. This is equivalent to finding it.
- the filter coefficient calculation unit 176 uses the optimum pitch coefficient T ′ given from the search unit 175 to obtain a filter coefficient / 3 that minimizes the square distortion E expressed by the following equation (5).
- FIG. 7 is a flowchart showing a procedure of processes performed in pitch coefficient setting section 173, filtering section 174, and search section 175.
- pitch coefficient setting section 173 sets pitch coefficient T and optimum pitch coefficient T ′ to lower limit value Tmin of the search range, and sets maximum similarity Amax to 0.
- the fine lettering section 174 performs input spectral filtering. And calculate the estimated value S ′ (k) of the input spectrum.
- search section 175 calculates similarity A between input spectrum S (k) and estimated value S ′ (k) of the input spectrum.
- search section 175 compares calculated similarity A with maximum similarity Am ax.
- search section 175 updates maximum similarity Amax using similarity A, and updates optimum pitch coefficient T ′ using pitch coefficient T.
- search section 175 compares pitch coefficient T with search range upper limit value Tmax.
- search section 175 outputs optimum pitch coefficient T 'in ST1080.
- the encoding apparatus 100 includes the input signal divided into the low-frequency part (0 ⁇ k ⁇ FL) and the high-frequency part (FL ⁇ k ⁇ FH) in the spectrum encoding unit 107.
- the shape of the high-frequency spectrum is estimated using the filtering unit 174 having the low-frequency spectrum as the internal state.
- the parameter T ′ and 0 itself representing the filter characteristics of the filtering unit 174 indicating the correlation between the low-frequency spectrum and the high-frequency spectrum are transmitted to the decoding device instead of the high-frequency spectrum.
- the optimum pitch coefficient T which indicates the correlation between the low-frequency spectrum and the high-frequency spectrum
- the filter coefficient / 3 are also estimation parameters for estimating the high-frequency spectrum from the low-frequency spectrum.
- the filtering unit 174 of the spectrum encoding unit 107 uses the low-frequency spectrum to increase the frequency.
- the pitch coefficient setting unit 173 outputs the frequency difference between the low-frequency spectrum and the high-frequency spectrum used as the estimation reference, that is, the pitch coefficient T in various ways, and outputs it. 175 searches for a pitch coefficient T that maximizes the similarity between the low-frequency spectrum and the high-frequency spectrum. Therefore, the shape of the high-frequency spectrum can be estimated based on the pitch of the harmonic structure of the entire spectrum, and encoding can be performed while maintaining the harmonic structure of the entire spectrum, improving the quality of the decoded speech signal. can do.
- the bandwidth of the low-frequency spectrum can be set arbitrarily without having to align the bandwidth of the low-frequency spectrum with the pitch of the harmonic structure (an integral multiple of that). Therefore, the spectrum can be smoothly connected at the connection portion between the low-frequency spectrum and the high-frequency spectrum with a simple operation, and the quality of the decoded speech signal can be improved.
- FIG. 8 is a block diagram showing the main configuration of decoding apparatus 200 according to the present embodiment.
- decoding apparatus 200 includes control section 201, first layer decoding section 202, upsampling section 203, second layer decoding section 204, spectrum decoding section 205, and switch 206.
- Control section 201 separates the first layer encoded information, the second layer encoded information, and the spectrum encoded information that constitute the bit stream transmitted from encoding apparatus 100, and obtains the first obtained
- the encoded information is output to first layer decoding section 202
- the second layer encoded information is output to second layer decoding section 204
- the spectral encoding information is output to spectrum decoding section 205.
- the control unit 201 adaptively generates control information for controlling the switch 206 according to the constituent elements of the bit stream transmitted from the encoding apparatus 100, and outputs the control information to the switch 206.
- First layer decoding section 202 performs CELP decoding on the first layer encoded information input from control section 201, and obtains the obtained first layer decoded signal by upsampling section 203 and switch 206. Output to.
- Upsampling section 203 performs upsampling processing on the first layer decoded signal input from first layer decoding section 202, converts the sampling frequency of the first layer decoded signal from Rate2 to Ratel, Output to spectrum decoding section 205.
- Second layer decoding section 204 performs gain 'shape inverse quantization using the second layer encoded information input from control section 201, and obtains the obtained second layer MDCT coefficients, that is, quantization targets Band residual MDCT coefficients are output to spectrum decoding section 205. The internal configuration and specific operation of second layer decoding section 204 will be described later.
- Spectral decoding section 205 receives second layer MDCT coefficients input from second layer decoding section 204, spectral coding information input from control section 201, upsampling section 203, and post-upsampling input.
- the first layer decoded signal is used to perform band extension processing, and the obtained second layer decoded signal is output to the switch 206.
- the internal configuration and specific operation of spectrum decoding section 205 will be described later.
- switch 206 is a bit stream transmitted from encoding apparatus 100 to decoding apparatus 200, for the first layer encoded information, the second layer encoded information, And the spectrum coding information, the bit stream is composed of the first layer coding information, the extra coding information power, or the bit stream is the first layer coding information,
- the second layer encoded information is V
- the second layer decoded signal input from spectrum decoding section 205 is output as a decoded signal.
- switch 206 outputs the first layer decoded signal input from first layer decoding section 202 as a decoded signal when the bit stream is composed only of the first layer encoded information.
- FIG. 9 is a block diagram showing a main configuration inside second layer decoding section 204.
- second layer decoding section 204 includes separation section 241, shape inverse quantization section 242, predictive decoding presence / absence determination section 243, and gain inverse quantization section 244.
- Separating section 241 separates the band information, shape encoded information, and gain encoded information from the second layer encoded information input from control section 201, and changes the obtained band information to the shape inverse quantization section 242 and predictive decoding presence / absence determination unit 243, shape encoding information is output to shape inverse quantization unit 242 and gain encoding information is output to gain inverse quantization unit 244.
- the shape inverse quantization unit 242 decodes the shape encoded information input from the separation unit 241, and the MDC corresponding to the quantization target band indicated by the band information input from the separation unit 241.
- the shape value of the T coefficient is obtained and output to the gain inverse quantization unit 244.
- Predictive decoding presence / absence determining section 243 uses the band information input from demultiplexing section 241 to determine the common subband between the quantization target band of the current frame and the quantization target band of the past frame. Find a number. Then, when the number of common subbands is equal to or greater than a predetermined value, the predictive decoding presence / absence determining unit 243 determines that predictive decoding is performed on the MDCT coefficient of the quantization target band indicated by the band information, and When the number of subbands is smaller than the predetermined value, it is determined that predictive decoding is not performed on the MDCT coefficient of the quantization target band indicated by the band information. Predictive decoding presence / absence determination section 243 outputs the determination result to gain inverse quantization section 244.
- the gain dequantization unit 244 determines the past frame stored in the built-in buffer. Using the gain value and the built-in gain codebook, the gain coding information input from the separation unit 241 is subjected to predictive decoding to obtain a gain value. On the other hand, when the determination result input from the predictive decoding presence / absence determining unit 243 indicates that the predictive decoding is not performed! /, And! /, The gain dequantizing unit 244 uses the built-in gain codebook. Thus, the gain encoded information input from the separation unit 241 is directly inversely quantized to obtain a gain value. Gain dequantization section 244 obtains the second layer MDCT coefficient, that is, the residual MDCT coefficient of the quantization target band, using the obtained gain value and the shape value input from shape inverse quantization section 242. Output.
- second layer decoding section 204 having the above configuration is the same as that of second layer encoding section 1
- FIG. 10 is a block diagram showing the main configuration inside spectrum decoding section 205.
- spectrum decoding section 205 has frequency domain conversion section 251, addition spectrum calculation section 252, internal state setting section 253, filtering section 254, and time domain conversion section 255.
- Frequency domain transform section 251 performs frequency transform on the first layer decoded signal after up-sampling input from up-sampling section 203, calculates first spectrum Sl (k), and adds spectrum calculation section Output to 252.
- the effective frequency band of the first layer decoded signal after upsampling is 0 ⁇ k ⁇ FL
- the frequency conversion method is discrete Fourier transform. Conversion (DFT), discrete cosine transform (DCT), modified discrete cosine transform (MDCT), etc.
- Addition spectrum calculation section 252 receives first spectrum Sl (k) from frequency domain transform section 251 and second layer MDCT coefficients (hereinafter referred to as second spectrum S2) from second layer decoding section 204. (denoted as (k)) is input, the first spectrum Sl (k) and the second spectrum S2 (k) are added, and the addition result is output to the internal state setting unit 253 as the addition spectrum S3 (k) To do.
- the addition spectrum calculation unit 252 receives only the first spectrum Sl (k) from the frequency domain conversion unit 251 and does not receive the second spectrum S2 (k) from the second layer decoding unit 204. 1Spectrum S l (k) is output to internal state setting unit 253 as addition spectrum S3 (k)
- the internal state setting unit 253 sets the internal state of the filter used in the filtering unit 254 using the addition spectrum S3 (k).
- Filtering section 254 calculates the internal state of the filter set by internal state setting section 253, and the optimum pitch coefficient T 'and filter coefficient / 3 included in the spectral coding information input from control section 201. Then, the sum spectrum S3 (k) is filtered to generate the sum spectrum estimate S3 ′ (k). Then, filtering section 254 outputs decoded spectrum S ′ (k) composed of added spectrum S3 (k) and estimated value S3 ′ (k) of the added spectrum to time domain converting section 255. In such a case, the filtering unit 254 uses the filter function represented by the above equation (1).
- FIG. 11 is a diagram showing the decoded spectrum S ′ (k) generated by the filtering unit 254.
- the fine-lettering unit 254 does not use the first layer MDCT coefficient (0 ⁇ k ⁇ FU and the second layer MDCT coefficient (FL, ⁇ Filtering is performed using the sum spectrum S3 (k) where the band is 0 ⁇ k ⁇ FL ", which is the sum of k ⁇ FL", and an estimated value S3 '(k) of the sum spectrum is obtained. As shown in FIG.
- the decoded spectrum S ′ (k) in the band to be quantized indicated by the band information is composed of the added spectrum S3 (k)
- the decoded spectrum S ′ (k) in FH is composed of the estimated value S3 ′ (k) of the added spectrum.
- the decoded spectrum S '(k) in the frequency band FL' ⁇ k ⁇ FL is the estimated value S3' (k) of the added spectrum obtained by the filtering process of the filtering unit 254 using the added spectrum S3 (k). Then it takes the value of the additive spectrum S3 (k) itself.
- FIG. 11 shows an example in which the band of the first spectrum S 1 (k) and the band of the second spectrum S2 (k) partially overlap.
- the band of the first spectrum Sl (k) completely overlaps the band of the second spectrum S2 (k), or the first spectrum Sl (k)
- the band of the second spectrum S2 (k) is not adjacent and may be separated.
- FIG. 12 is a diagram showing a case where the band of the second spectrum S2 (k) completely overlaps the band of the first spectrum Sl (k).
- the decoding spectrum S ′ (k) in the frequency band FL ⁇ k ⁇ FH takes the value of the estimated value S3 ′ (k) itself of the added spectrum.
- the value of the addition spectrum S3 (k) is obtained by caloring the value of the first spectrum SI (k) and the value of the second spectrum S2 (k). The accuracy of the estimated value S3 ′ (k) is improved, and hence the quality of the decoded speech signal is improved.
- FIG. 13 is a diagram showing a case where the band of the first spectrum SI (k) and the band of the second spectrum S2 (k) are not adjacent but separated from each other.
- the filtering unit 254 obtains an estimated value S3 ′ (k) of the added spectrum using the first spectrum tunnel S l (k), and performs a band expansion process to the frequency band FL ⁇ k ⁇ FH.
- the portion of the estimated value S3 ′ (k) corresponding to the band of the second spectrum S 2 (k) is replaced using the second spectrum S 2 (k).
- the reason is that the accuracy of the second spectrum S2 (k) is higher than the estimated value S3 ′ (k) of the added spectrum, which improves the quality of the decoded speech signal.
- Time domain conversion section 255 converts decoded spectrum S '(k) input from filtering section 254 into a time domain signal and outputs it as a second layer decoded signal.
- the time domain conversion unit 255 performs processing such as appropriate windowing and overlay addition as necessary to avoid discontinuities between frames.
- the encoding band is selected in the upper layer on the encoding side, and the decoded spectrum of the lower layer and the upper layer is added on the decoding side to obtain
- the band is extended using the added spectrum, and the band components that cannot be decoded by the lower and upper layers are decoded. Therefore, high-frequency spectrum data with high accuracy can be calculated flexibly according to the coding band selected in the higher layer on the coding side, and a decoded signal with higher quality can be obtained.
- second layer encoding section 106 selects force band S to be quantized and performs second layer encoding as an example. Not limited to this, second layer encoding section 106 may encode components in the same band as the bands encoded in first layer encoding section 102, which may encode fixed band components. .
- decoding apparatus 200 uses the optimal pitch coefficient T and the filter coefficient / 3 included in the spectral encoding information to add spectrum S3 (k).
- the present invention is not limited to this, and the decoding apparatus 200 is not limited to this, but has been described as an example in which the high-band part vector is estimated by filtering and generating the estimated value S3 ′ (k) of the added spectrum. By filtering the first spectrum SI (k), the spectrum of the high band may be estimated.
- 1SM is not limited to this. It is possible to use an integer (natural number) of 0 or more.
- the CELP type encoding / decoding scheme is applied in the first layer, but other encoding / decoding schemes may be used.
- encoding apparatus 10 that performs hierarchical encoding (scalable encoding).
- the present invention is not limited to this, and may be applied to an encoding apparatus that performs encoding in a method other than hierarchical encoding.
- encoding apparatus 100 includes frequency domain transform sections 161 and 162, these are necessary when time domain signals are used as input signals.
- the present invention is not limited to this, and when a vector is directly input to the spectrum encoding unit 107, the frequency domain transform units 161 and 162 may not be provided.
- the filter coefficient is calculated by the filter coefficient calculation unit 176 after the pitch coefficient is calculated by the filtering unit 174. 1.
- the present invention is not limited to this, and the filter coefficient calculation unit 176 may not be provided and the filter coefficient may not be calculated. Alternatively, the filter coefficient calculation unit 176 may not be provided, and the filtering unit 174 may perform filtering using the pitch coefficient and the filter coefficient to search for the optimum pitch coefficient and filter coefficient at the same time. In such a case, the following equations (6) and (7) are used instead of the above equations (1) and (2).
- the case where the high frequency spectrum is encoded using the low frequency spectrum that is, using the low frequency spectrum as a reference for encoding has been described as an example.
- the reference spectrum may be set in other ways. For example, it is not desirable from the viewpoint of effective use of energy, but it is acceptable to encode a low-frequency spectrum using a high-frequency spectrum, or to encode a spectrum in an intermediate frequency band. You may encode the spectrum of another band as a reference
- FIG. 14 is a block diagram showing the main configuration of coding apparatus 300 according to Embodiment 2 of the present invention.
- coding apparatus 300 has the same basic configuration as coding apparatus 100 (see FIGS. 1 to 3) shown in Embodiment 1, and the same components are denoted by the same reference numerals. The description is omitted.
- the spectrum encoding unit 307 of the encoding device 300 and the spectrum encoding unit 107 of the encoding device 100 are different in part of the processing, and different codes are attached to indicate this.
- Spectral coding section 307 converts the speech 'audio signal that is an input signal of coding apparatus 300 and the first layer decoded signal after upsampling input from upsampling section 104 into the frequency domain. Obtain the input spectrum and the first layer decoding spectrum. Then, the spectrum encoding unit 307 includes a low frequency component of the first layer decoded spectrum, Analyzing the correlation with the high-frequency component of the input spectrum, calculating the parameters for estimating the high-frequency component from the low-frequency component on the decoding side, and outputting it to the multiplexing unit 108 as spectral coding information .
- FIG. 15 is a block diagram showing the main components inside spectrum coding section 307.
- the spectrum encoding unit 307 has the same basic configuration as the spectrum encoding unit 107 (see FIG. 3) shown in Embodiment 1, and the same components are assigned the same reference numerals. The explanation is omitted.
- Spectrum coding section 307 is different from spectrum coding section 107 in that frequency coding section 377 is further provided.
- the setting unit 172, the filtering unit 174, the search unit 175, and the filter coefficient calculation unit 176 are different in part of the processing, and different reference numerals are given to indicate this.
- Frequency domain transform section 377 performs frequency transform on the input audio-audio signal whose effective frequency band is 0 ⁇ k ⁇ FH, and calculates input spectrum S (k).
- DFT discrete Fourier transform
- DCT discrete cosine transform
- MDCT modified discrete cosine transform
- the frequency domain transform unit 371 performs upsampling in which the effective frequency band input from the upsampling unit 104 is 0 ⁇ k ⁇ FH instead of the audio 'audio signal in which the effective frequency band is 0 ⁇ k ⁇ FH.
- a frequency conversion is performed on the subsequent first layer decoded signal to calculate a first layer decoded spectrum S (k).
- the frequency conversion method is discrete
- DFT discrete cosine transform
- DCT discrete cosine transform
- MDCT modified discrete cosine transform
- the internal state setting unit 372 performs the first layer decoded spectrum S (k )
- the internal state of the filter used in the filtering unit 374 is set using DEC1.
- the internal state of this filter is set by adding spectrum S instead of input spectrum S (k).
- Filtering section 374 performs filtering of the first layer decoded vector using the internal state of the filter set by internal state setting section 372 and pitch coefficient T output from pitch coefficient setting section 173. Calculate estimated value S '(k) of 1-layer decoded spectrum
- Search section 375 is similar to input spectrum S (k) input from frequency domain transform section 377 and estimated value S ′ (k) of the first layer decoded spectrum output from filtering section 374.
- Similarity which is a parameter indicating sex, is calculated.
- the similarity calculation process is the same as the similarity calculation process performed by the search unit 175 except that the following expression (9) is used instead of the expression (4), and thus detailed description thereof is omitted.
- This similarity calculation process is performed every time the pitch coefficient T is given from the pitch coefficient setting unit 173 to the filtering unit 374, and the pitch coefficient that maximizes the calculated similarity, that is, the optimum pitch coefficient T '(Range from Tmin to Tmax) is given to the filter coefficient calculation unit 376.
- Filter coefficient calculation section 376 has an optimum pitch coefficient T ′ given from search section 375, input spectrum S (k) input from frequency domain conversion section 377, and first input from frequency domain conversion section 371. Using the layer decoded spectrum S (k), find the filter coefficient / 3,
- Multiplexer 10 uses DEC1 i filter coefficient 13 and optimum pitch coefficient T 'as spectral coding information. Output to 8.
- the filter coefficient calculation process in the filter coefficient calculation unit 376 is the same as the calculation process of the filter coefficient / 3 in the filter coefficient calculation unit 176 except that the following expression (10) is used instead of the expression (5). Since it is the same, detailed explanation is omitted.
- encoding apparatus 300 uses spectral encoding section 307 to perform a filter that uses first layer decoded spectrum S (k) whose effective frequency band is 0 ⁇ k ⁇ FH as an internal state.
- the shape of the high-frequency part (FL ⁇ k ⁇ FH) of the first layer decoded spectrum S (k) where the effective frequency band is 0 ⁇ k ⁇ FH is estimated using the scaling part 374.
- the optimum pitch coefficient T ′ and the filter coefficient / 3 representing the filter characteristics of the filtering unit 374 are obtained, and these are transmitted to the decoding device instead of the encoded information of the high frequency part of the input spectrum.
- decoding apparatus Since the decoding apparatus according to the present embodiment has the same configuration as that of decoding apparatus 100 according to Embodiment 1 and performs the same operation, description thereof is omitted.
- the decoding spectrum of the lower layer and the upper layer is added on the decoding side, and the obtained addition spectrum is band-extended to be used for obtaining the estimated value of the adding spectrum.
- the optimal pitch coefficient and filter coefficient are not the correlation between the estimated value S '(k) of the input spectrum and the high band (FL ⁇ k ⁇ FH) of the input spectrum S (k).
- FIG. 16 is a block diagram showing the main configuration of encoding apparatus 400 according to Embodiment 3 of the present invention.
- coding apparatus 400 has the same basic configuration as coding apparatus 100 (see FIGS. 1 to 3) shown in the first embodiment, and the same reference numerals are given to the same components. A description thereof will be omitted.
- Encoding apparatus 400 is different from encoding apparatus 100 in that it further includes second layer decoding section 409. Note that the spectrum encoding unit 407 of the encoding device 400 and the spectrum encoding unit 107 of the encoding device 100 are different in part of the processing, and different codes are attached to indicate this.
- Second layer decoding section 409 has the same configuration as second layer decoding section 204 (FIGS. 8 to 10) in decoding apparatus 200 according to Embodiment 1, and performs the same operation. The explanation is omitted. However, while the output of second layer decoding section 204 is referred to as the second layer MDCT coefficient, here the output of second layer decoding section 409 is referred to as the second layer decoded spectrum, and S (k)
- Spectral coding section 407 converts the speech 'audio signal that is an input signal of coding apparatus 400 and the first layer decoded signal after up-sampling input from up-sampling section 104 into the frequency domain. Obtain the input spectrum and the first layer decoding spectrum. Then, spectrum encoding section 407 adds the low-frequency component of the first layer decoded spectrum and the second layer decoded spectrum input from second layer decoding section 409, adds the addition spectrum that is the addition result, and the input spectrum The band is expanded on the decoding side to calculate a parameter for estimating the low frequency component power and the high frequency component, and outputs to the multiplexing unit 108 as spectrum coding information.
- FIG. 17 is a block diagram showing the main components inside spectrum coding section 407.
- the spectrum encoding unit 407 has the same basic configuration as the spectrum encoding unit 107 (see FIG. 3) shown in Embodiment 1, and the same components are assigned the same reference numerals. The explanation is omitted.
- Spectrum encoding section 407 is replaced with frequency domain transform section 171 instead of frequency domain transform section 171.
- the filter coefficient calculation unit 176 has a difference in part of the processing, and a different reference numeral is attached to indicate this.
- the frequency domain transform unit 471 performs upsampling in which the effective frequency band input from the upsampling unit 104 is 0 ⁇ k ⁇ FH instead of the voice / audio signal in which the effective frequency band is 0 ⁇ k ⁇ FH.
- the frequency conversion is performed on the subsequent first layer decoded signal to calculate the first layer decoded spectrum S (k) and output it to the added spectrum calculating section 478
- DFT discrete Fourier transform
- DCT discrete cosine transform
- MDCT modified discrete cosine transform
- Addition spectrum calculation section 478 receives the low frequency (0 ⁇ k ⁇ FU component of first layer decoded spectrum S (k) input from frequency domain transform section 471 and the second layer decoding section 409.
- the band of the added spectrum S (k) is the low band (0 ⁇ k ⁇ FL) and the second layer encoding.
- the frequency domain transform unit 477 performs frequency transform on the input audio-audio signal whose effective frequency band is 0 ⁇ k ⁇ FH, and calculates the input spectrum S (k).
- DFT discrete Fourier transform
- DCT discrete cosine transform
- MDCT modified discrete cosine transform
- the internal state setting unit 472 uses the sum spectrum S (k) whose effective frequency band is 0 ⁇ k ⁇ FH instead of the input spectrum S (k) whose effective frequency band is 0 ⁇ k ⁇ FH.
- the internal state of the filter used in the filtering unit 474 is set.
- the internal state of this filter is set using the summed spectrum S (k) instead of the input spectrum S (k).
- the internal state setting unit 172 is the same as the internal state setting, and thus detailed description thereof is omitted.
- the filtering unit 474 uses the internal state of the filter set by the internal state setting unit 472 and the pitch coefficient T output from the pitch coefficient setting unit 473 to add the spectrum S
- Search section 475 is a parameter indicating the similarity between input spectrum S (k) input from frequency domain transform section 477 and estimated value S '(k) of the added spectrum output from filtering section 474.
- the similarity that is a parameter is calculated.
- the similarity calculation process is the same as the similarity calculation process performed by the search unit 175 except that the following expression (12) is used instead of the expression (4), and thus detailed description thereof is omitted.
- This similarity calculation process is performed each time the pitch coefficient T is given from the pitch coefficient setting unit 173 to the filtering unit 474, and the pitch coefficient that maximizes the calculated similarity, ie, the optimum
- the pitch coefficient T ′ (range from Tmin to Tmax) is given to the filter coefficient calculation unit 476.
- Filter coefficient calculation section 476 includes optimum pitch coefficient T ′ provided from search section 475, input spectrum S (k) input from frequency domain conversion section 477, and addition input from addition spectrum calculation section 478. Using the spectrum S (k), find the filter coefficient / 3 and filter
- Coefficient / 3 and optimum pitch coefficient T ′ are output to multiplexing section 108 as spectral coding information.
- the filter coefficient calculation unit 476 calculates the filter coefficient / 3 in the filter coefficient calculation unit 176 except that the following expression (13) is used instead of the expression (5). Detailed description will be omitted.
- the encoding apparatus 400 includes a filtering unit 47 that uses the spectrum encoding unit 407 to set the added spectrum S (k) in which the effective frequency band is 0 ⁇ k ⁇ FH to an internal state.
- the parameters indicating the correlation with the high-frequency part (FL ⁇ k ⁇ FH), that is, the optimum pitch coefficient T 'and the filter coefficient / 3 that represent the filter characteristics of the filtering part 474 are obtained, and the high-frequency part of the input spectrum is encoded. It is transmitted to the decoding device instead of the information.
- decoding apparatus Since the decoding apparatus according to the present embodiment has the same configuration as that of decoding apparatus 100 according to Embodiment 1 and performs the same operation, the description thereof is omitted.
- the first layer decoding spectrum and the second layer decoded spectrum are added to calculate the added spectrum, and the added spectrum and the input are calculated. Based on the correlation with the spectrum, the optimum pitch coefficient and filter coefficient are obtained.
- the addition spectrum is calculated by adding the decoded spectra of the lower layer and the upper layer, and the estimated value of the addition spectrum is obtained using the optimum pitch coefficient and filter coefficient transmitted from the encoding side. Perform bandwidth expansion. Therefore, it is possible to further suppress the influence of coding distortion of the first layer coding and the second layer coding on the band expansion on the decoding side, and further improve the quality of the decoded signal.
- the first layer decoded spectrum and the second layer decoded spectrum are added to calculate an added spectrum, and the correlation between the added spectrum and the input spectrum is calculated.
- the present invention is not limited to this, and the spectrum for which the correlation with the input spectrum is obtained is described.
- either the addition spectrum or the first decoding spectrum may be selected.
- the optimal pitch coefficient and filter coefficient for band expansion are calculated based on the correlation between the first layer decoded spectrum and the input spectrum.
- the optimal pitch coefficient and filter coefficient for band expansion are calculated based on the correlation between the added spectrum and the input spectrum.
- auxiliary information input to the encoding device or the state of the transmission line can be used.
- the first layer coding where the use efficiency of the transmission line is very high. If you are unable to transmit information,
- the correlation between the low frequency component and the high frequency component of the input spectrum is compared with the case of calculating the optimum pitch coefficient and the filter coefficient. You may add when you ask. For example, if the distortion between the first layer decoded spectrum and the input spectrum is very small, calculating the optimum pitch coefficient and filter coefficient from the low-frequency component and high-frequency component of the input spectrum, the higher the layer, the higher the layer. Therefore, it is possible to provide a higher quality output signal.
- the present invention provides a first layer decoded signal or a first layer decoded signal used in a scalable codec to calculate a band extension parameter in an encoding device. And a band extension parameter for band extension by the decoding device, and a low band component of a calculated signal (for example, an added signal obtained by adding the first layer decoded signal and the second layer decoded signal). Apply the first layer decoded signal or the calculated signal calculated by using the first layer decoded signal (for example, the sum signal obtained by adding the first layer decoded signal and the second layer decoded signal).
- An advantageous effect can be obtained by configuring the band components to be different. It is possible to configure the low frequency components to be the same as each other, or to use the low frequency components of the input signal in the encoding device.
- the force shown as an example of using a pitch coefficient and a filter coefficient as parameters used for band expansion is not limited to this.
- one coefficient may be fixed on the encoding side and the decoding side, and only the other coefficient may be transmitted as a parameter from the encoding side.
- separate parameters to be used for transmission can be obtained and used as band extension parameters. They may be used in combination.
- the encoding apparatus adjusts the energy for each high-frequency subband (a band obtained by dividing the entire band into a plurality of frequency component areas) after filtering.
- the decoding apparatus may receive the gain information and use it for band expansion.
- gain information used for energy adjustment for each subband obtained by the encoding device is transmitted to the decoding device as a parameter used for band expansion, and this gain information is applied to the band expansion by the decoding device.
- the pitch coefficient for estimating the high band spectrum from the low band spectrum and the filtering coefficient are fixed between the encoding device and the decoding device, so that the energy for each subband is fixed. It is possible to use only gain information for adjusting the bandwidth as a parameter for bandwidth expansion. Therefore, band expansion can be performed by using at least one of the three types of information of pitch coefficient, filtering coefficient, and gain information.
- the encoding apparatus, decoding apparatus, and these methods according to the present invention are not limited to the above embodiments, and can be implemented with various modifications. For example, each embodiment can be implemented in combination as appropriate.
- the encoding device and the decoding device according to the present invention can be mounted on a communication terminal device and a base station device in a mobile communication system, and thereby have a similar operation effect as described above.
- a base station apparatus and a mobile communication system can be mounted on a communication terminal device and a base station device in a mobile communication system, and thereby have a similar operation effect as described above.
- the power described with reference to the case where the present invention is configured by hardware can be realized by software.
- an encoding device and a decoding device according to the present invention are described by describing an algorithm of the encoding method and the decoding method according to the present invention in a programming language, storing the program in a memory, and causing the information processing means to execute the program. The same function can be realized.
- each functional block used in the description of each of the above embodiments is typically realized as an LSI which is an integrated circuit. These may be individually made into one chip, or may be made into one chip so as to include some or all of them. [0149] Although LSI is used here, depending on the degree of integration, IC, system LSI, super L
- the method of circuit integration is not limited to LSI, and may be realized by a dedicated circuit or a general-purpose processor. You can use FPGA (Field Programmable Gate Array) that can be programmed after LSI manufacturing, or a reconfigurable processor that can reconfigure the connection or setting of circuit cells inside the LSI! / .
- FPGA Field Programmable Gate Array
- the first invention of the present invention is the first encoding means for generating a first encoded data by encoding a low-frequency part of the input signal, which is a band lower than a predetermined frequency, and the first A first decoding means for decoding the encoded data to generate a first decoded signal; and a second encoding by encoding a predetermined band portion of the residual signal between the input signal and the first decoded signal. Filtering the low-frequency part of the second encoding means for generating data and the first decoded signal or a calculated signal calculated using the first decoded signal, and And a filtering unit that obtains a band extension parameter for obtaining a high-frequency part that is a band higher than the frequency.
- the second encoded data is decoded and the second encoded data is decoded.
- the filtering means comprises the addition Applying a signal as the calculated signal and filtering the low-frequency portion of the sum signal, the band extension parameter for obtaining a high-frequency portion that is higher than the predetermined frequency of the input signal.
- the third invention of the present invention further comprises gain information generating means for calculating gain information for adjusting energy for each subband after the filtering in the first or second invention. It is an encoding device.
- a fourth invention of the present invention is a decoding device using a scalable codec having a layer configuration of r layers (r is an integer of 2 or more), wherein the encoding device uses the m-th layer (m is equal to or less than r).
- the decoding device includes decoding means for generating a high frequency component.
- the decoding means uses the band extension parameter to decode a decoded signal of an nth layer (m ⁇ n) different from the mth layer.
- This is a decoding device that generates high-frequency components.
- the receiving means further receives gain information transmitted from the encoding device, and the decoding means is configured to expand the band.
- the decoding apparatus generates the high-frequency component of the decoded signal of the nth layer using the gain information instead of the extension parameter, or using the band extension parameter and the gain information.
- the seventh invention of the present invention is a first encoded data obtained by encoding a low-frequency portion, which is a band lower than a predetermined frequency, of the input signal in the encoding device transmitted from the encoding device.
- Data second encoded data obtained by encoding a predetermined band portion of the residual between the first decoded spectrum obtained by decoding the first encoded data and the spectrum of the input signal, and the first Filtering and filtering the low-frequency part of the first decoded spectrum or the first added spectrum obtained by adding the first decoded spectrum and the second decoded spectrum obtained by decoding the second encoded data
- Receiving means for receiving a band extension parameter for obtaining a high-frequency portion that is a band higher than the predetermined frequency of the signal, and generating the third decoded spectrum in the low-frequency band by decoding the first encoded data First decryption Means, second decoding means for decoding the second encoded data to generate a fourth decoded spectrum in the predetermined band portion, and using the band extension parameter
- the receiving means includes the first encoded data, the second encoded data, and the low-frequency portion of the first addition spectrum. And a band extension parameter for obtaining a high-frequency part that is a band higher than the predetermined frequency of the input signal.
- the third decoding means adds the third decoded spectrum and the fourth decoded spectrum to generate a second added spectrum.
- Filtering means for filtering the third decoded spectrum, the fourth decoded spectrum, or the second added spectrum as the fifth decoded spectrum by using the band extension parameter to perform the band extension.
- a decoding device comprising:
- the receiving means further receives gain information transmitted from the encoding device, and the third decoding means is the band extension parameter. 5th decoding generated using the third decoded spectrum, the fourth decoded spectrum, or both using the gain information instead of the above or using the band extension parameter and the gain information.
- This is a decoding device that decodes a band portion that has not been decoded by the first decoding means and the second decoding means by band-extending one of the spectra.
- An eleventh aspect of the present invention is the encoding apparatus' decoding apparatus according to any of the first to tenth aspects, comprising at least one of a band extension parameter, a pitch coefficient, and a filtering coefficient.
- the encoding apparatus and the like according to the present invention can be applied to applications such as a communication terminal apparatus and a base station apparatus in a mobile communication system.
Abstract
Description
Claims
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP07850645.8A EP2101322B1 (en) | 2006-12-15 | 2007-12-14 | Encoding device, decoding device, and method thereof |
CN2007800444142A CN101548318B (zh) | 2006-12-15 | 2007-12-14 | 编码装置、解码装置以及其方法 |
US12/518,371 US8560328B2 (en) | 2006-12-15 | 2007-12-14 | Encoding device, decoding device, and method thereof |
JP2008549379A JP5339919B2 (ja) | 2006-12-15 | 2007-12-14 | 符号化装置、復号装置およびこれらの方法 |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2006338341 | 2006-12-15 | ||
JP2006-338341 | 2006-12-15 | ||
JP2007-053496 | 2007-03-02 | ||
JP2007053496 | 2007-03-02 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2008072737A1 true WO2008072737A1 (ja) | 2008-06-19 |
Family
ID=39511750
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2007/074141 WO2008072737A1 (ja) | 2006-12-15 | 2007-12-14 | 符号化装置、復号装置およびこれらの方法 |
Country Status (5)
Country | Link |
---|---|
US (1) | US8560328B2 (ja) |
EP (1) | EP2101322B1 (ja) |
JP (1) | JP5339919B2 (ja) |
CN (1) | CN101548318B (ja) |
WO (1) | WO2008072737A1 (ja) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2012046447A1 (ja) * | 2010-10-06 | 2012-04-12 | パナソニック株式会社 | 符号化装置、復号装置、符号化方法及び復号方法 |
WO2013035257A1 (ja) * | 2011-09-09 | 2013-03-14 | パナソニック株式会社 | 符号化装置、復号装置、符号化方法および復号方法 |
JP5565914B2 (ja) * | 2009-10-23 | 2014-08-06 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 符号化装置、復号装置およびこれらの方法 |
Families Citing this family (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2198426A4 (en) * | 2007-10-15 | 2012-01-18 | Lg Electronics Inc | METHOD AND DEVICE FOR PROCESSING A SIGNAL |
JP5098569B2 (ja) * | 2007-10-25 | 2012-12-12 | ヤマハ株式会社 | 帯域拡張再生装置 |
CN103366755B (zh) * | 2009-02-16 | 2016-05-18 | 韩国电子通信研究院 | 对音频信号进行编码和解码的方法和设备 |
WO2010137300A1 (ja) | 2009-05-26 | 2010-12-02 | パナソニック株式会社 | 復号装置及び復号方法 |
JP5754899B2 (ja) | 2009-10-07 | 2015-07-29 | ソニー株式会社 | 復号装置および方法、並びにプログラム |
JP5295380B2 (ja) * | 2009-10-20 | 2013-09-18 | パナソニック株式会社 | 符号化装置、復号化装置およびこれらの方法 |
JP5850216B2 (ja) | 2010-04-13 | 2016-02-03 | ソニー株式会社 | 信号処理装置および方法、符号化装置および方法、復号装置および方法、並びにプログラム |
JP5609737B2 (ja) | 2010-04-13 | 2014-10-22 | ソニー株式会社 | 信号処理装置および方法、符号化装置および方法、復号装置および方法、並びにプログラム |
JP6073215B2 (ja) * | 2010-04-14 | 2017-02-01 | ヴォイスエイジ・コーポレーション | Celp符号器および復号器で使用するための柔軟で拡張性のある複合革新コードブック |
JPWO2011158657A1 (ja) * | 2010-06-17 | 2013-08-19 | シャープ株式会社 | 画像フィルタ装置、復号装置、符号化装置、および、データ構造 |
US20120029926A1 (en) | 2010-07-30 | 2012-02-02 | Qualcomm Incorporated | Systems, methods, apparatus, and computer-readable media for dependent-mode coding of audio signals |
US9208792B2 (en) | 2010-08-17 | 2015-12-08 | Qualcomm Incorporated | Systems, methods, apparatus, and computer-readable media for noise injection |
JP5707842B2 (ja) | 2010-10-15 | 2015-04-30 | ソニー株式会社 | 符号化装置および方法、復号装置および方法、並びにプログラム |
WO2012053150A1 (ja) * | 2010-10-18 | 2012-04-26 | パナソニック株式会社 | 音声符号化装置および音声復号化装置 |
JP5992427B2 (ja) * | 2010-11-10 | 2016-09-14 | コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. | 信号におけるピッチおよび/または基本周波数に関するパターンを推定する方法および装置 |
CN103443856B (zh) * | 2011-03-04 | 2015-09-09 | 瑞典爱立信有限公司 | 音频编码中的后量化增益校正 |
JP5704397B2 (ja) * | 2011-03-31 | 2015-04-22 | ソニー株式会社 | 符号化装置および方法、並びにプログラム |
JP5817499B2 (ja) * | 2011-12-15 | 2015-11-18 | 富士通株式会社 | 復号装置、符号化装置、符号化復号システム、復号方法、符号化方法、復号プログラム、及び符号化プログラム |
WO2013162450A1 (en) * | 2012-04-24 | 2013-10-31 | Telefonaktiebolaget L M Ericsson (Publ) | Encoding and deriving parameters for coded multi-layer video sequences |
CN103971691B (zh) * | 2013-01-29 | 2017-09-29 | 鸿富锦精密工业(深圳)有限公司 | 语音信号处理系统及方法 |
MX353240B (es) * | 2013-06-11 | 2018-01-05 | Fraunhofer Ges Forschung | Dispositivo y método para extensión de ancho de banda para señales acústicas. |
CN105531762B (zh) | 2013-09-19 | 2019-10-01 | 索尼公司 | 编码装置和方法、解码装置和方法以及程序 |
KR102356012B1 (ko) | 2013-12-27 | 2022-01-27 | 소니그룹주식회사 | 복호화 장치 및 방법, 및 프로그램 |
US10468035B2 (en) * | 2014-03-24 | 2019-11-05 | Samsung Electronics Co., Ltd. | High-band encoding method and device, and high-band decoding method and device |
JP6798312B2 (ja) * | 2014-09-08 | 2020-12-09 | ソニー株式会社 | 符号化装置および方法、復号装置および方法、並びにプログラム |
CN105513601A (zh) * | 2016-01-27 | 2016-04-20 | 武汉大学 | 一种音频编码带宽扩展中频带复制的方法及装置 |
PL3443557T3 (pl) * | 2016-04-12 | 2020-11-16 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Koder audio do kodowania sygnału audio, sposób kodowania sygnału audio i program komputerowy, z uwzględnieniem wykrytego regionu widmowego pełnego w wyższym pasmie częstotliwości |
US10825467B2 (en) * | 2017-04-21 | 2020-11-03 | Qualcomm Incorporated | Non-harmonic speech detection and bandwidth extension in a multi-source environment |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2005112001A1 (ja) * | 2004-05-19 | 2005-11-24 | Matsushita Electric Industrial Co., Ltd. | 符号化装置、復号化装置、およびこれらの方法 |
WO2006049205A1 (ja) * | 2004-11-05 | 2006-05-11 | Matsushita Electric Industrial Co., Ltd. | スケーラブル復号化装置およびスケーラブル符号化装置 |
WO2006049204A1 (ja) * | 2004-11-05 | 2006-05-11 | Matsushita Electric Industrial Co., Ltd. | 符号化装置、復号化装置、符号化方法及び復号化方法 |
JP2006338341A (ja) | 2005-06-02 | 2006-12-14 | Hitachi Ltd | ストレージシステム及び構成変更方法 |
JP2007053496A (ja) | 2005-08-16 | 2007-03-01 | Sony Corp | 映像表示方法、映像表示方法のプログラム、映像表示方法のプログラムを記録した記録媒体及び映像表示装置 |
Family Cites Families (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5752225A (en) * | 1989-01-27 | 1998-05-12 | Dolby Laboratories Licensing Corporation | Method and apparatus for split-band encoding and split-band decoding of audio information using adaptive bit allocation to adjacent subbands |
JP2779886B2 (ja) * | 1992-10-05 | 1998-07-23 | 日本電信電話株式会社 | 広帯域音声信号復元方法 |
JP2964879B2 (ja) * | 1994-08-22 | 1999-10-18 | 日本電気株式会社 | ポストフィルタ |
JP3653826B2 (ja) * | 1995-10-26 | 2005-06-02 | ソニー株式会社 | 音声復号化方法及び装置 |
JP3255047B2 (ja) * | 1996-11-19 | 2002-02-12 | ソニー株式会社 | 符号化装置および方法 |
JP3541680B2 (ja) * | 1998-06-15 | 2004-07-14 | 日本電気株式会社 | 音声音楽信号の符号化装置および復号装置 |
EP1132892B1 (en) * | 1999-08-23 | 2011-07-27 | Panasonic Corporation | Speech encoding and decoding system |
FI109393B (fi) * | 2000-07-14 | 2002-07-15 | Nokia Corp | Menetelmä mediavirran enkoodaamiseksi skaalautuvasti, skaalautuva enkooderi ja päätelaite |
DE60233032D1 (de) * | 2001-03-02 | 2009-09-03 | Panasonic Corp | Audio-kodierer und audio-dekodierer |
JP3888097B2 (ja) * | 2001-08-02 | 2007-02-28 | 松下電器産業株式会社 | ピッチ周期探索範囲設定装置、ピッチ周期探索装置、復号化適応音源ベクトル生成装置、音声符号化装置、音声復号化装置、音声信号送信装置、音声信号受信装置、移動局装置、及び基地局装置 |
EP1701340B1 (en) * | 2001-11-14 | 2012-08-29 | Panasonic Corporation | Decoding device, method and program |
US7752052B2 (en) * | 2002-04-26 | 2010-07-06 | Panasonic Corporation | Scalable coder and decoder performing amplitude flattening for error spectrum estimation |
JP3881943B2 (ja) * | 2002-09-06 | 2007-02-14 | 松下電器産業株式会社 | 音響符号化装置及び音響符号化方法 |
US7844451B2 (en) * | 2003-09-16 | 2010-11-30 | Panasonic Corporation | Spectrum coding/decoding apparatus and method for reducing distortion of two band spectrums |
JP4789430B2 (ja) * | 2004-06-25 | 2011-10-12 | パナソニック株式会社 | 音声符号化装置、音声復号化装置、およびこれらの方法 |
WO2006041055A1 (ja) * | 2004-10-13 | 2006-04-20 | Matsushita Electric Industrial Co., Ltd. | スケーラブル符号化装置、スケーラブル復号装置及びスケーラブル符号化方法 |
BRPI0517513A (pt) * | 2004-10-26 | 2008-10-14 | Matsushita Electric Ind Co Ltd | aparelho de codificação de som e processo de sua realização |
JP4859670B2 (ja) * | 2004-10-27 | 2012-01-25 | パナソニック株式会社 | 音声符号化装置および音声符号化方法 |
KR100818268B1 (ko) * | 2005-04-14 | 2008-04-02 | 삼성전자주식회사 | 오디오 데이터 부호화 및 복호화 장치와 방법 |
EP1876586B1 (en) * | 2005-04-28 | 2010-01-06 | Panasonic Corporation | Audio encoding device and audio encoding method |
EP1990800B1 (en) * | 2006-03-17 | 2016-11-16 | Panasonic Intellectual Property Management Co., Ltd. | Scalable encoding device and scalable encoding method |
US20080059154A1 (en) * | 2006-09-01 | 2008-03-06 | Nokia Corporation | Encoding an audio signal |
JP4871894B2 (ja) * | 2007-03-02 | 2012-02-08 | パナソニック株式会社 | 符号化装置、復号装置、符号化方法および復号方法 |
CN101771417B (zh) * | 2008-12-30 | 2012-04-18 | 华为技术有限公司 | 信号编码、解码方法及装置、系统 |
CA2770287C (en) * | 2010-06-09 | 2017-12-12 | Panasonic Corporation | Bandwidth extension method, bandwidth extension apparatus, program, integrated circuit, and audio decoding apparatus |
-
2007
- 2007-12-14 US US12/518,371 patent/US8560328B2/en active Active
- 2007-12-14 WO PCT/JP2007/074141 patent/WO2008072737A1/ja active Application Filing
- 2007-12-14 CN CN2007800444142A patent/CN101548318B/zh not_active Expired - Fee Related
- 2007-12-14 EP EP07850645.8A patent/EP2101322B1/en not_active Not-in-force
- 2007-12-14 JP JP2008549379A patent/JP5339919B2/ja not_active Expired - Fee Related
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2005112001A1 (ja) * | 2004-05-19 | 2005-11-24 | Matsushita Electric Industrial Co., Ltd. | 符号化装置、復号化装置、およびこれらの方法 |
WO2006049205A1 (ja) * | 2004-11-05 | 2006-05-11 | Matsushita Electric Industrial Co., Ltd. | スケーラブル復号化装置およびスケーラブル符号化装置 |
WO2006049204A1 (ja) * | 2004-11-05 | 2006-05-11 | Matsushita Electric Industrial Co., Ltd. | 符号化装置、復号化装置、符号化方法及び復号化方法 |
JP2006338341A (ja) | 2005-06-02 | 2006-12-14 | Hitachi Ltd | ストレージシステム及び構成変更方法 |
JP2007053496A (ja) | 2005-08-16 | 2007-03-01 | Sony Corp | 映像表示方法、映像表示方法のプログラム、映像表示方法のプログラムを記録した記録媒体及び映像表示装置 |
Non-Patent Citations (4)
Title |
---|
KATAOKA A. ET AL.: "G.729 o Kosei Yoso to shite Mochiiru Scalable Kotaiiki Onsei Fugoka", THE TRANSACTIONS OF THE INSTITUTE OF ELECTRONICS, INFORMATION AND COMMUNICATION ENGINEERS D-II, vol. J86-D-II, no. 3, 1 March 2003 (2003-03-01), pages 379 - 387, XP002996342 * |
MASAHIRO OSHIKIRI; HIROYUKI EHARA; KOJI YOSHIDA: "Improvement of the super-wideband scalable coder using pitch filtering based spectrum coding", ANNUAL MEETING OF ACOUSTIC SOCIETY OF JAPAN 2-4-13, September 2004 (2004-09-01), pages 297 - 298 |
OSHIKIRI M. ET AL.: "AMR o Core ni shita Kyotaiiki/Kotaiiki Scalable Onsei Fugoka Hoshiki", THE ACOUSTICAL SOCIETY OF JAPAN (ASJ) KOEN RONBUNSHU CD-ROM, no. 1-Q-28, 7 March 2006 (2006-03-07), pages 389 - 390 * |
See also references of EP2101322A4 |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5565914B2 (ja) * | 2009-10-23 | 2014-08-06 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 符号化装置、復号装置およびこれらの方法 |
US8898057B2 (en) | 2009-10-23 | 2014-11-25 | Panasonic Intellectual Property Corporation Of America | Encoding apparatus, decoding apparatus and methods thereof |
WO2012046447A1 (ja) * | 2010-10-06 | 2012-04-12 | パナソニック株式会社 | 符号化装置、復号装置、符号化方法及び復号方法 |
JPWO2012046447A1 (ja) * | 2010-10-06 | 2014-02-24 | パナソニック株式会社 | 符号化装置、復号装置、符号化方法及び復号方法 |
US9117461B2 (en) | 2010-10-06 | 2015-08-25 | Panasonic Corporation | Coding device, decoding device, coding method, and decoding method for audio signals |
WO2013035257A1 (ja) * | 2011-09-09 | 2013-03-14 | パナソニック株式会社 | 符号化装置、復号装置、符号化方法および復号方法 |
US9384749B2 (en) | 2011-09-09 | 2016-07-05 | Panasonic Intellectual Property Corporation Of America | Encoding device, decoding device, encoding method and decoding method |
JP2017027069A (ja) * | 2011-09-09 | 2017-02-02 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 符号化装置および符号化方法 |
US9741356B2 (en) | 2011-09-09 | 2017-08-22 | Panasonic Intellectual Property Corporation Of America | Coding apparatus, decoding apparatus, and methods |
US9886964B2 (en) | 2011-09-09 | 2018-02-06 | Panasonic Intellectual Property Corporation Of America | Encoding apparatus, decoding apparatus, and methods |
US10269367B2 (en) | 2011-09-09 | 2019-04-23 | Panasonic Intellectual Property Corporation Of America | Encoding apparatus, decoding apparatus, and methods |
US10629218B2 (en) | 2011-09-09 | 2020-04-21 | Panasonic Intellectual Property Corporation Of America | Encoding apparatus, decoding apparatus, and methods |
Also Published As
Publication number | Publication date |
---|---|
EP2101322A1 (en) | 2009-09-16 |
US8560328B2 (en) | 2013-10-15 |
JPWO2008072737A1 (ja) | 2010-04-02 |
US20100017198A1 (en) | 2010-01-21 |
EP2101322A4 (en) | 2011-08-31 |
CN101548318B (zh) | 2012-07-18 |
EP2101322B1 (en) | 2018-02-21 |
JP5339919B2 (ja) | 2013-11-13 |
CN101548318A (zh) | 2009-09-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5339919B2 (ja) | 符号化装置、復号装置およびこれらの方法 | |
JP5328368B2 (ja) | 符号化装置、復号装置、およびこれらの方法 | |
RU2488897C1 (ru) | Кодирующее устройство, декодирующее устройство и способ | |
JP5230782B2 (ja) | 符号化装置、階層符号化装置、復号化装置、通信端末装置、基地局装置、符号化方法、および復号化方法 | |
US7983904B2 (en) | Scalable decoding apparatus and scalable encoding apparatus | |
JP5404418B2 (ja) | 符号化装置、復号装置および符号化方法 | |
US8554549B2 (en) | Encoding device and method including encoding of error transform coefficients | |
JP5030789B2 (ja) | サブバンド符号化装置およびサブバンド符号化方法 | |
EP2012305B1 (en) | Audio encoding device, audio decoding device, and their method | |
JP5036317B2 (ja) | スケーラブル符号化装置、スケーラブル復号化装置、およびこれらの方法 | |
JP5511785B2 (ja) | 符号化装置、復号装置およびこれらの方法 | |
WO2009113316A1 (ja) | 符号化装置、復号装置およびこれらの方法 | |
WO2009093466A1 (ja) | 符号化装置、復号装置およびこれらの方法 | |
KR20080049085A (ko) | 음성 부호화 장치 및 음성 부호화 방법 | |
US20100017199A1 (en) | Encoding device, decoding device, and method thereof | |
WO2006041055A1 (ja) | スケーラブル符号化装置、スケーラブル復号装置及びスケーラブル符号化方法 | |
WO2008053970A1 (fr) | Dispositif de codage de la voix, dispositif de décodage de la voix et leurs procédés | |
WO2013057895A1 (ja) | 符号化装置及び符号化方法 | |
RU2459283C2 (ru) | Кодирующее устройство, декодирующее устройство и способ | |
JP5774490B2 (ja) | 符号化装置、復号装置およびこれらの方法 | |
WO2011045927A1 (ja) | 符号化装置、復号装置およびこれらの方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 200780044414.2 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 07850645 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2008549379 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 12518371 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2007850645 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |