JP2012078866A - Audio coding system using characteristics of decoded signal to adapt synthesized spectral components - Google Patents

Audio coding system using characteristics of decoded signal to adapt synthesized spectral components Download PDF

Info

Publication number
JP2012078866A
JP2012078866A JP2011287052A JP2011287052A JP2012078866A JP 2012078866 A JP2012078866 A JP 2012078866A JP 2011287052 A JP2011287052 A JP 2011287052A JP 2011287052 A JP2011287052 A JP 2011287052A JP 2012078866 A JP2012078866 A JP 2012078866A
Authority
JP
Japan
Prior art keywords
signal
components
subband signals
spectral
subband
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
JP2011287052A
Other languages
Japanese (ja)
Other versions
JP5253565B2 (en
Inventor
Allen Davidson Grant
Stuart Vinton Mark
Conrad Fellers Matthew
Mead Truman Michael
スチュアート ヴィントン,マーク
アレン デイビッドソン,グラント
ミード トゥルーマン,マイケル
コンラッド フェラーズ,マシュー
Original Assignee
Dolby Lab Licensing Corp
ドルビー ラボラトリーズ ライセンシング コーポレイション
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to US10/174,493 priority Critical patent/US7447631B2/en
Priority to US10/174,493 priority
Priority to US10/238,047 priority
Priority to US10/238,047 priority patent/US7337118B2/en
Application filed by Dolby Lab Licensing Corp, ドルビー ラボラトリーズ ライセンシング コーポレイション filed Critical Dolby Lab Licensing Corp
Publication of JP2012078866A publication Critical patent/JP2012078866A/en
Application granted granted Critical
Publication of JP5253565B2 publication Critical patent/JP5253565B2/en
Application status is Active legal-status Critical
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/02Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders
    • G10L19/032Quantisation or dequantisation of spectral components
    • G10L19/035Scalar quantisation
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/038Speech enhancement, e.g. noise reduction or echo cancellation using band spreading techniques

Abstract

PROBLEM TO BE SOLVED: To provide techniques that can be used in low bit rate audio coding systems to improve the perceived quality of audio signals generated by such systems.SOLUTION: A receiver in an audio coding system receives a signal conveying frequency subband signals representing an audio signal. The subband signals are examined to assess one or more characteristics of the audio signal. Spectral components are synthesized having the assessed characteristics. The synthesized spectral components are integrated with the subband signals and passed through a synthesis filterbank to generate an output signal. In one implementation, the assessed characteristic is temporal shape and noise-like spectral components are synthesized having the temporal shape of the audio signal.

Description

  The present invention relates generally to audio coding systems, and more specifically to improving the perceived quality of audio signals obtained from audio coding systems.

  The audio coding system codes the audio signal into a coded signal suitable for transmission or storage, and then receives or recovers the coded signal, decodes it, and obtains the original audio signal for playback. A perceptual audio coding system codes an audio signal into a coded signal having a lower information capacity than the original audio signal, and then decodes the coded signal to produce a perceptually indistinguishable output from the original audio signal. Try to provide. One example of a perceptual audio coding system is described in the Advanced Television Systems Committee (ATSC) A / 52A document published on August 20, 2001, entitled “Digital Audio Compression (AC-3) Standard Revision A”. And it is called Dolby Digital. Another example is J. AES, vol. 45, no. 10, October 1997, pages 789-814, described by Bosi et al. In “ISO / TEC MPEG-2 Advanced Audio Coding”, which is called Advanced Audio Coding (AAC).

  In these two coding systems, like many other perceptual coding systems, a split band transmitter applies an analysis filter bank to the audio signal to obtain spectral components that are arranged in groups or in frequency bands, The spectral component is coded according to psychoacoustic principles to generate a coded signal. The bandwidth is usually different and is usually equal to the critical bandwidth of the so-called human hearing instrument. A supplemental split band receiver receives and decodes the coded signal to recover the spectral components and applies a synthesis filter bank to the decoded spectral components to obtain a copy of the original audio signal.

  Using a perceptual coding system can preserve a subjective or perceptual standard of sound quality and reduce the information capacity of the audio signal, thus using less bandwidth of the encoded audio signal Without being carried over the communication channel, or it can be stored on the recording medium using that less space. Information capacity is reduced by quantizing the spectral components. Quantization adds noise to the quantized signal, but since perceptual audio coding systems generally use psychoacoustic models in an attempt to control the amplitude of the quantized noise, it is masked or inaudible by the spectral components of the signal To be done.

  Traditional perceptual coding techniques work fairly well in audio coding systems that can transport or record coded signals with medium to high bit rates, but the coded signals are constrained to low bit rates. When done, it does not provide very good sound quality. Other techniques have been used in connection with perceptual coding techniques in an attempt to provide high quality signals at very low bit rates.

  One technique called “HF Regeneration” (HFR) was filed on March 28, 2002 by Truman et al. In US Patent Application No. 10 / 113,858, entitled “High Frequency Regeneration”. Wideband frequency shift for ". The entire contents of this patent are incorporated herein for reference. In an audio coding system that uses HFR, the transmitter removes high frequency components from the coded signal and the receiver regenerates or synthesizes alternative components such as noise for the missing high frequency components. The resulting signal provided at the output of the receiver is generally not perceptually identical to the original signal provided at the transmitter input, but sophisticated regeneration techniques are possible at low bit rates. Can provide an output signal that is a fairly good approximation of the original input signal with much higher quality. In such a relationship, high quality (high quality) usually means a wide bandwidth and a low level of perceived noise.

  Another synthetic technique called “Spectral Hole Filling” (SHF) is US Patent Application No. 10 / 174,493, filed June 17, 2002 by Truman et al., Entitled “Spectral Hole Filling. Improved audio coding system to use ". The entire contents of this patent are incorporated herein for reference. According to this technique, the transmitter quantizes and codes the spectral components of the input signal such that the spectral component bands are omitted from the coded signal. The missing spectral component bands are called spectral holes. The receiver synthesizes the spectral components to close the spectral holes. Although SHF technology generally does not provide an output signal that is perceptually identical to the original input signal, it does not provide a perceived quality of the output signal in a system that is constrained to work with low bit rate coded signals. It can be improved.

  Technologies such as HFR and SHF can provide advantages in many situations, but they do not work well in all those situations. One particularly troublesome situation arises when audio signals with rapidly changing amplitudes are coded by analysis filter banks and systems that use block transforms to perform analysis filter banks. In this situation, components such as audible noise will be smeared over the time corresponding to the transform block.

  One technique that can be used to reduce the audible effect of time-blurring noise is to reduce the block length of the analysis and synthesis transforms during intervals where the input signal is not significantly invariant. This technique works well with audio coding systems that can transmit or record medium to high bit rate coded signals, but does not work very well with lower bit rate systems. This is because using shorter blocks reduces the coding gain achieved by the transform.

  In another technique, the transmitter modifies the input signal so that sudden changes in amplitude are removed or reduced before applying the composite transform. The receiver applies the reverse operation after applying the composite transform. Unfortunately, this technique obscures the true spectral characteristics of the input signal, thus distorting the information required for effective perceptual coding, and the transmitter uses a portion of the transmitted signal to allow the receiver to make the correction This is because parameters necessary for applying the reverse operation must be transmitted.

  In a third technique, known as time domain noise waveform shaping, the transmitter applies a prediction filter to the spectral components obtained from the analysis filter bank, communicates the prediction error and prediction filter coefficients of the transmitted signal, and The receiver applies an inverse prediction filter to the prediction error to recover the spectral components. This technique is undesirable in low bit rate systems. This is because the signal overhead needs to convey the prediction filter coefficients.

  It is an object of the present invention to provide a technique that can be used in low bit rate audio coding systems and that improves the perceived quality of audio signals generated by such systems.

  According to the present invention, the coded acoustic information is received to obtain a subband signal that represents some spectral content rather than all the spectral content of the audio signal, the subband signal is examined to obtain the characteristics of the audio signal, Generating a composite spectral component having characteristics of the audio signal, integrating the composite spectral component with the subband signal to generate a set of modified subband signals, and applying a synthetic filter bank to the set of modified subband signals; Apply. In this way, the coded acoustic information is processed.

  The various features and preferred embodiments of the present invention will be better understood with reference to the following discussion and the accompanying drawings. The following discussion and the contents of the drawings are described in detail by way of example and should not be understood as limiting the scope of the invention.

2 is a schematic block diagram of a transmitter of an audio coding system. 2 is a schematic block diagram of a receiver of an audio coding system. 2 is a schematic block diagram of an apparatus that can be used to carry out various aspects of the present invention.

A. Overview Various aspects of the present invention can be incorporated into various signal processing methods and apparatus, including the apparatus shown in FIGS. Some aspects may be performed by processing done only at the receiver. Other aspects require cooperative processing performed at both the receiver and transmitter. A description of the processes that may be used to perform these various aspects of the invention follows below in accordance with an overview of exemplary apparatus that may be used to perform these processes.

  FIG. 1 illustrates one embodiment of a split-band audio transmitter, in which an analysis filter bank 12 receives acoustic information representing an audio signal from a path 11 and, in response, a frequency representing the spectral content of the audio signal. Provides subband signals. Each subband signal is passed to encoder 14, which generates a coded representation of the subband signal and passes the coded representation to formatter 16. The formatter 16 assembles the coded display into an output signal suitable for transmission or storage, and passes this output signal to the path 17.

  FIG. 2 illustrates an example of a split-band audio receiver, in which a formatter 22 receives an input signal from path 21 that conveys a coded representation of a frequency subband signal that represents the spectral content of the audio signal. The formatter 22 obtains a coded representation from the input signal and passes it to the decoder 24. Decoder 24 decodes the coded representation into a frequency subband signal. The analyzer 25 examines the subband signal to obtain one or more characteristics of the audio signal represented by the subband signal. The characteristic indication is passed to the component synthesizer 26, which generates a composite spectral component by using a process corresponding to the characteristic. The integrator 27 generates a set of modified subband signals by integrating the subband signals provided by the decoder 24 with the combined spectral components generated by the component synthesizer 26. In response to the modified subband signal set, the synthesis filter bank 28 generates acoustic information representing the audio signal in the path 29. In the particular embodiment shown in the figure, neither analyzer 25 nor component synthesizer 26 adapts the processing in response to any control information obtained from the input signal by deformator 22. In other embodiments, analyzer 25 and / or component synthesizer 26 may be responsive to control information obtained from an input signal.

  In the apparatus shown in FIGS. 1 and 2, a filter bank for three frequency subbands is shown. Although many more subbands can be used in a typical implementation, only three are shown for clarity. In the present invention, any particular number is not important.

  The analysis filter bank and the synthesis filter bank may be implemented with essentially any block transform including a discrete Fourier transform or a discrete cosine transform (DCT). In one audio coding system having a transmitter and receiver, such as the transmitter and receiver discussed above, the analysis filter bank 12 and the synthesis filter bank 28 are connected to the ICASSP1987 Conf. Proc. 1987, pp. Time-Domain Aliasing described in 2161-64, Princen et al., "Subband / Transform Coding Using Filter Bank Designs Based on Time Domain Aliasing Cancellation" Performed by a modified DCT known as Cancellation (TDAC) transformation.

  An analysis filter bank implemented by block transformation transforms a block or interval of an input signal into a set of transform coefficients that represent the spectral content of that interval of the signal. One or more groups of adjacent transform coefficients represent spectral content in a particular frequency subband having a bandwidth equal to the number of coefficients in that group. The term “subband signal” means a group of one or more adjacent transform coefficients, and the term “spectral component” means a transform coefficient.

The terms “encoder” and “coding” as used in this disclosure refer to an information processing apparatus and method that can be used to represent an audio signal with coded information having a smaller information capacity than the audio signal itself. . The terms “decoder” and “decoding” refer to an information processing apparatus and method that can be used to recover an audio signal from a coded display. Two examples of reduced information capacity are the Dolby Digital described above and the coding required to process a bitstream compatible with the AAC coding standard. Any particular type of coding or decoding is not critical to the present invention.
B. Receiver Various aspects of the invention can be performed in a receiver that does not require any special processing or information from the transmitter. These aspects will be described first.
1. Analysis of Signal Characteristics The present invention may be used in a coding system that represents an audio signal with a very low bit rate coded signal. The coded information of an ultra low bit rate system usually carries a subband signal that represents only a portion of the spectral component of the audio signal. The analyzer 25 examines these subband signals to obtain one or more characteristics of the portion of the audio signal represented by the subband signals. The representation of one or more characteristics is passed to the component synthesizer 26 and used to generate a composite spectral component. Some examples of properties that can be used are described below.
a) Amplitude Coded information generated by many coding systems represents a spectral component quantized to some required bit length, or quantization resolution. Small spectral components having a magnitude smaller than the level represented by the least significant bit (LSB) of the quantized component can be omitted from the coded information, or alternatively the quantized value is zero or zero. It can be expressed in any display form that is considered. The level corresponding to the LSB of the quantized spectral component carried by the coded information can be considered as an upper limit on the size of the small spectral component that is omitted from the coded information.

Component synthesizer 26 can use this level to limit the amplitude of any component that is synthesized to replace missing spectral components.
b) Spectral shape The spectral shape of the subband signal carried by the coded information is immediately available from the subband signal itself. However, other information regarding the spectral shape can be derived by applying a filter to the subband signal in the frequency domain. The filter may be a prediction filter, a low pass filter, or essentially any other type of filter desired.

The spectral shape or filter output instruction is appropriately passed to the component synthesizer 26. If necessary, an indication of which filter should be used should also be passed.
c) Masking A perceptual model may be applied to estimate the psychoacoustic masking effect of the spectral components of the subband signal. Since these masking effects vary with frequency, the masking provided by the first spectral component at one frequency is caused by the second spectral component even when the first spectral component has the same amplitude as the second spectral component at another frequency. It is not necessary to provide the same level of masking as provided.

An indication of the estimated masking effect is passed to the component synthesizer 26, where the estimated masking effect of the synthesized component is preferably the estimated masking effect of the spectral components of the subband signal. Control the synthesis of spectral components to have a relationship.
d) Color tone The color tone of the subband signal can be evaluated by various methods. One method is the calculation of Spectral Flatness Measure, which is normalized by dividing the arithmetic mean of the subband signal samples by the geometric mean of the subband signal samples. It is a quotient. The tone can also be evaluated by analyzing the arrangement or distribution of spectral components in the subband signal. For example, if several large spectral components are separated by a long interval of much smaller components, the subband signal is considered to be a tone rather than a noise. Another method is to apply a prediction filter to the subband signal to determine the prediction gain. A large prediction gain tends to indicate that the signal is more tonal.

Since the tone indication is passed to the component synthesizer 26, the component synthesizer 26 controls the synthesis so that the synthesized spectral components have the appropriate level of tone. This is done by performing a weighted combination of synthetic components such as tones and synthetic components such as noise to achieve the required level of tone.
e) Temporal shape The temporal shape of the signal represented by the subband signal can be estimated directly from the subband signal. The technical basis for one implementation of the temporal shape estimator is described with respect to the linear system represented by Equation 1.
y (t) = h (t) ・ x (t) (1)
Where y (t) is a signal having an estimated temporal shape,
h (t) is the temporal shape of the signal y (t),
The dot symbol (•) represents multiplication,
x (t) is a temporally flat signal of the signal y (t).

Equation 1 can be written as Equation 2 below.
Y [k] = H [k] * X [k] (2)
Where Y [k] is the frequency domain representation of the signal y (t)
H [k] is the frequency domain display of h (t),
An asterisk (*) represents a convolution,
X [k] is a frequency domain representation of the signal x (t).

  The frequency domain display Y [k] corresponds to one or more of the subband signals obtained by the decoder 24. The analyzer 25 solves a set of equations obtained from an automatic regression moving average (ARMA) model of Y [k] and X [k] to obtain a frequency domain representation H [k] of the temporal shape h (t). You can get an estimate of Additional information on the use of the ARMA model can be found in “Digital Signal Processing: Principles, Algorithms and Applications” by Proakis and Manolakis published in 1988, New York Macmillan Publishers (see especially pages 818-821). ) Can be obtained from.

  The frequency domain display Y [k] is configured as a block of transform coefficients. Each block of transform coefficients represents a short time spectrum of the signal y (t). The frequency domain display X [k] is also configured as a block. Each block of coefficients in the frequency domain representation X [k] represents a block of samples of a temporarily flat signal x (t) that is assumed to be stationary (invariant) in a broad sense. It is assumed that the coefficients within each block of the X [k] representation are also distributed separately. Under these assumptions, the signal is represented below as Equation 3 by the ARMA model.

Where L is the length of the autoregressive part of the ARMA model,
Q is the length of the moving average part of the ARMA model,
Equation 3 is solved as Equation 4 below for a1 and bq by solving the autocorrelation of Y [k].

Here, E {} is an expected value function.

  Equation 4 can be rewritten as Equation 5 below.

Where R YY [n] represents the autocorrelation of Y [n]
R YY [k] represents the cross-correlation between Y [k] and X [k].

  If we further assume that the linear system represented by H [k] is just autoregressive, we can ignore the second term on the right-hand side of Equation 5 and write Equation 5 as .

This represents a set of L order linear equations to be solved to obtain L coefficients ai.

Using this equation, one embodiment of a temporal shape estimator using the frequency domain method can now be described. In this implementation, the temporal shape estimator receives the frequency domain representation Y [k] of one or more subband signals y (t) and auto-correlation sequence R YY [n] when −L <m <L. Calculate These values are used to establish a set of linear equations to solve to obtain the coefficients ai (these coefficients represent the poles of the linear all-pole filter FR shown below in Equation 7).

This filter is applied to the frequency domain representation of any temporally flat signal, such as a noise-like signal, to have this temporal shape having a temporal shape substantially equal to the temporal shape of the signal y (t). A flat signal frequency domain display can be obtained.

  A description of the poles of the filter FR is passed to the component synthesizer 26, which can use the filter to generate a composite spectral component that represents a signal having the required temporal shape.

2. Synthetic component generation component synthesizer 26 can generate synthetic spectral components in a variety of ways. Two methods are described below. Multiplexing methods may be used. For example, different methods may be selected corresponding to the characteristics obtained from the subband signal or as a function related to frequency.

  The first method generates a noise-like signal. For example, essentially any of a variety of time domain and frequency domain methods can be used to generate a noise-like signal.

  The second method uses a frequency domain method called spectral shift or spectral replication that replicates spectral components from one or more frequency subbands. Good low frequency spectral components are usually replicated to higher frequencies. This is because higher frequency components are often associated in some way with lower frequency components. However, in principle, the spectral components can be copied to higher or lower frequencies. If desired, noise may be added or mixed to the transferred component, and the amplitude may be modified as desired. Desirably, phase discontinuities in the composite component can be eliminated or at least reduced by adjusting as necessary.

  The synthesis of the spectral components is controlled by information received from the analyzer 25 so that the synthesized components have one or more characteristics derived from the subband signal.

3. Integration of Signal Components Various methods can integrate the synthesized spectral components with the spectral components of the subband signals. One method uses the synthesized component as a dither form by combining each synthesized component representing a corresponding frequency and a subband component. Another method uses one or more composite components in place of selected spectral components present in the subband signal. Yet another method combines spectral components with subband signal components to represent spectral components that are not present in the subband signal. These methods and other methods may be used in various combinations.

C. Even if the transmitter does not provide more information than the control information required by the receiver to receive and decode the subband signal without the functions of the present invention, the features of the present invention described above can be obtained. It can be executed at the receiver. If additional control information is provided, these features of the present invention can be improved. One example is discussed below.

  The degree to which time domain waveform shaping is applied to the composite component is adapted by the control information provided in the coded information. One way in which this can be done is to use the parameters shown in Equation 8 below.

When β = 0, the filter does not provide any time domain waveform shaping. When β = 1, the filter provides time domain waveform shaping so that the correlation between the temporal shape of the synthesized component and the temporal shape of the subband signal is maximal. Other values of β provide an intermediate level of time domain waveform shaping.

  In one implementation, the transmitter provides control information so that the receiver can set β to one of eight values.

  The transmitter can provide other control information that can be used by the receiver to adapt the component synthesis process in any way desired.

D. Implementation Various aspects of the present invention include software in other devices, including more specialized components such as general purpose computer devices or digital signal processor (DSP) circuitry coupled to components similar to components of general purpose computer devices. It can be implemented in various ways. FIG. 3 is a block diagram of an apparatus 70 that can be used to implement various aspects of the present invention at a transmitter or receiver. The DSP 72 provides computer resources. The RAM 73 is a system random access memory (RAM) used by the DSP 72 for signal processing. ROM 74 represents some form of storage, such as read only memory (ROM), for storing the programs necessary to operate device 70 and perform various aspects of the present invention. The input / output controller 75 represents an interface circuit for sending and receiving signals through the communication channels 76 and 77. If it is desired to receive and / or transmit an analog audio signal, the input / output controller 75 may include an analog / digital converter and a digital / analog converter. In the illustrated embodiment, all major system components are connected to bus 71 (which may be one or more physical buses), but the bus architecture is necessary to implement the present invention. Not.

  In an embodiment implemented on a general-purpose computer device, additional components such as a keyboard, mouse, and display that interface the device, and an additional component that controls a storage device having a storage medium such as magnetic tape, disk, or optical media It may be included. Storage media can be used to record instruction programs, utility software, and application software for an operating system, and the storage media can include embodiments of programs that perform various aspects of the invention.

  The functions necessary to carry out various aspects of the present invention may be performed by components that include discrete logic components, one or more ASICs and / or program controlled processors, and that are executed in various ways. The manner in which these components are implemented is not critical to the present invention.

  Media that can be read by many machines via spectrums including supersonic to ultraviolet frequencies, such as baseband and modulated communication paths, or essentially any magnetic or optical recording, including magnetic tape, magnetic disks, and optical disks The implementation software of the present invention can be carried by storage media including media that carries information using technology. Various aspects may also be performed on various components of computer system 70 by processing circuitry such as ASICs, general purpose integrated circuits, microprocessors controlled by programs embodied in various aspects of ROM or RAM, and other technologies. Can be done.

  The following items are further disclosed with respect to the above embodiments.

(1) A method of processing coded acoustic information,
Receiving coded acoustic information and obtaining from the acoustic information a subband signal representing some spectral content instead of the full spectral content of the audio signal;
Examine the subband signal to obtain the characteristics of the audio signal,
Generating a synthesized spectral component having the characteristics of the audio signal;
Integrating the combined spectral component with the subband signal to generate a set of modified subband signals;
Generating the acoustic information by applying a synthesis filter bank to the set of modified subband signals;
A method comprising that.

  (2) The method of (1), wherein the characteristic is a temporal shape, the method comprising generating a spectral component and convolving the generated spectral component into a frequency domain representation of the temporal shape. A method of obtaining a temporal shape by generating a synthetic spectral component.

  (3) The method according to (1), wherein the temporal shape is obtained by calculating an autocorrelation function of at least some components of the subband signal.

  (4) The method of (1), wherein the characteristic is a temporal shape, the method generating the spectral component and applying a filter to at least some of the generated spectral components. A method of generating a component to obtain the temporal shape.

  (5) The method of (4), wherein control information is obtained from the coded information and a filter is adapted in response to the control information.

  (6) The method according to (1), wherein the set of modified subband signals is generated by merging the combined spectral component with the component of the subband signal.

  (7) The method of (1), wherein the set of modified subband signals is generated by combining the combined spectral component with each component of the subband signal.

  (8) The method according to (1), wherein the one set of modified subband signals is generated by using the combined spectral component instead of each component of the subband signal.

(9) The method of (1),
Obtaining the characteristic of the audio signal by examining components of one or more subband signals in the first part of the spectrum;
One or more components of the subband signal of the first part of the spectrum are copied to a second part of the spectrum to form the composite subband signal, and the composite subband signal has the characteristics of the audio signal Generating the composite spectral component by modifying the copied component as follows:
Combining the synthesized spectral component with the subband signal by combining the synthesized subband signal with the subband signal;
Method.

  (10) The method according to (1), wherein the characteristic is any one of a set of amplitude, spectral shape, psychoacoustic masking effect, color tone, and temporal shape.

(11) A medium that is readable by a device and conveys an instruction program executable by the device to perform a method of processing coded acoustic information, the method comprising:
Receiving the coded acoustic information and obtaining from the coded acoustic information subband signals representing some spectral content rather than all spectral content of the audio signal;
Examine the subband signal to obtain the characteristics of the audio signal,
Generating a synthesized spectral component having the characteristics of the audio signal;
Integrating the combined spectral component with the subband signal to generate a set of modified subband signals;
Generating the acoustic information by applying a synthesis filter bank to the set of modified subband signals;
A media comprising steps for performing an action.

  (12) The medium according to (11), wherein the characteristic is a temporal shape, and the method generates a composite spectral component and convolves the generated spectral component with a frequency domain display of the temporal shape. To generate the synthesized spectral component to obtain the temporal shape.

  (13) The medium according to (11), wherein the method obtains the temporal shape by calculating an autocorrelation function of at least some components of the subband signal.

  (14) The medium of (11), wherein the characteristic is a temporal shape, and the method generates the spectral component and applies the filter to at least some of the generated spectral component. A medium that generates spectral components to obtain the temporal shape.

  (15) The medium according to (14), wherein the method obtains control information from the coded information and adapts the filter in response to the control information.

  (16) The medium according to (11), wherein the signal method generates a modified set of subband signals by merging the combined spectral component with the subband component.

  (17) The medium of (11), wherein the method generates the set of modified subband signals by combining the combined spectral components with respective components of the subband signals.

  (18) The medium according to (11), wherein the method generates the set of modified subband signals by using the synthesized spectral components instead of the respective components of the subband signals.

(19) The medium according to (11), wherein the method includes:
Obtaining the characteristic of the audio signal by examining the components of one or more subband signals of the first part of the spectrum;
One or more components of the subband signal of the first part of the spectrum are copied to a second part of the spectrum to form a composite subband signal, and the composite subband signal has the characteristics of the audio signal Generating the composite spectral component by modifying the copied component to
Media that combines the combined spectral component with the subband signal by combining the combined subband signal with the subband signal.

  (20) The medium according to (11), wherein the characteristic is any one of a set of amplitude, spectral shape, psychoacoustic masking effect, color tone, and temporal shape.

(21) A device for processing coded acoustic information,
An input terminal for receiving the coded acoustic information, a memory, and a processing circuit coupled to the input terminal and the memory;
The processing circuitry is
Receiving the coded acoustic information and obtaining from the coded acoustic information subband signals representing some spectral content rather than all spectral content of the audio signal;
Examine the subband signal to obtain the characteristics of the audio signal,
Generating a synthesized spectral component having the characteristics of the audio signal;
Integrating the combined spectral component with a subband signal to generate a set of modified subband signals;
Generating acoustic information by applying a synthesis filter bank to the set of modified subband signals;
apparatus.

  (22) The apparatus of (21), wherein the characteristic is a temporal shape, and the processing circuitry generates a synthesized spectral component and convolves the generated spectral component with the temporal domain frequency domain display. Thereby generating the synthesized spectral component to obtain the temporal shape.

  (23) The apparatus according to (21), wherein the processing circuitry obtains the temporal shape by calculating an autocorrelation function of at least some components of the subband signal.

  (24) The apparatus of (21), wherein the characteristic is a temporal shape, and the processing circuitry generates a composite spectral component and applies a filter to at least some of the generated spectral components. An apparatus for generating the synthesized spectral component to obtain the temporal shape.

  (25) The apparatus according to (24), wherein the processing circuitry obtains control information from the coded information and adapts the filter in response to the control information.

  (26) The apparatus according to (21), wherein the processing circuitry generates the set of modified subband signals by merging the synthesized spectral components with the components of the subband signals.

  (27) The apparatus of (21), wherein the processing circuitry generates the set of modified subband signals by combining the combined spectral components with respective components of the subband signals.

  (28) The apparatus of (21), wherein the processing circuitry generates the set of modified subband signals by using the synthesized spectral components instead of the respective components of the subband signals. The device provided in.

(29) The apparatus according to (21), wherein the processing circuitry is
Obtaining the characteristic of the audio signal by examining components of one or more subband signals in the first part of the spectrum;
One or more components of the subband signal of the first part of the spectrum are copied to a second part of the spectrum to form the composite subband signal, and the composite subband signal has the characteristics of the audio signal Generating the composite spectral component by modifying the copied component as follows:
Combining the synthesized spectral component with the subband signal by combining the synthesized subband signal with the subband signal;
apparatus.

  (30) The device according to (21), wherein the characteristic is one of amplitude, spectral shape, psychoacoustic masking effect, color tone, and temporal shape.

Claims (9)

  1. A method for processing coded acoustic information, comprising:
    Receiving the coded acoustic information and obtaining from the acoustic information subband signals representing some spectral components instead of all spectral components of the audio signal;
    Examine the subband signal to obtain an estimated temporal shape,
    Generating a composite spectral component using a process adapted in response to the estimated temporal shape;
    Integrating the combined spectral component with a subband signal representing the spectral component of the audio signal to generate a set of modified subband signals;
    Generating the acoustic information by applying a synthesis filter bank to the set of modified subband signals;
    A method involving that.
  2.   The method of claim 1, wherein the method generates the composite spectral component in response to the estimated temporal shape by applying a filter to at least some of the generated composite spectral components. A method characterized by:
  3.   The method of claim 2, wherein control information is obtained from the coded information and the filter is adapted in response to the control information.
  4.   4. A method according to any one of claims 1 to 3, wherein the set of modified subband signals is generated by merging the combined spectral components with the components of the subband signals. And how to.
  5.   4. A method as claimed in any preceding claim, wherein the set of modified subband signals is generated by combining the combined spectral components with respective components of the subband signals. A method characterized by.
  6.   4. The method according to any one of claims 1 to 3, wherein the set of modified subband signals is generated by using the combined spectral components instead of the respective components of the subband signals. A method characterized by that.
  7. The method of claim 1, comprising:
    Obtaining the estimated temporal shape of the audio signal by examining the components of one or more subband signals in the first part of the spectrum;
    One or more components of the subband signal of the first part of the spectrum are copied to a second part of the spectrum to form a composite subband signal and the copied in response to the estimated temporal shape Generating the composite spectral component by modifying
    A method characterized by that.
  8. A recording medium readable by the apparatus,
    A recording medium on which a program of instructions capable of executing the procedure of the method according to any one of claims 1 to 7 is recorded by the apparatus.
  9. An apparatus for processing coded acoustic information,
    An apparatus comprising means for executing the procedure of the method according to any one of claims 1 to 7.
JP2011287052A 2002-06-17 2011-12-28 Audio coding system that uses the characteristics of the decoded signal to fit the synthesized spectral components Active JP5253565B2 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
US10/174,493 US7447631B2 (en) 2002-06-17 2002-06-17 Audio coding system using spectral hole filling
US10/174,493 2002-06-17
US10/238,047 2002-09-06
US10/238,047 US7337118B2 (en) 2002-06-17 2002-09-06 Audio coding system using characteristics of a decoded signal to adapt synthesized spectral components

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
JP2004514061 Division 2003-06-09

Publications (2)

Publication Number Publication Date
JP2012078866A true JP2012078866A (en) 2012-04-19
JP5253565B2 JP5253565B2 (en) 2013-07-31

Family

ID=29733607

Family Applications (6)

Application Number Title Priority Date Filing Date
JP2004514060A Active JP4486496B2 (en) 2002-06-17 2003-05-30 Audio coding system using spectral hole filling
JP2010030139A Active JP5063717B2 (en) 2002-06-17 2010-02-15 Audio information generation method
JP2011287051A Active JP5253564B2 (en) 2002-06-17 2011-12-28 Audio coding system that uses the characteristics of the decoded signal to fit the synthesized spectral components
JP2011287052A Active JP5253565B2 (en) 2002-06-17 2011-12-28 Audio coding system that uses the characteristics of the decoded signal to fit the synthesized spectral components
JP2012149087A Active JP5345722B2 (en) 2002-06-17 2012-07-03 Audio information generation method
JP2013146451A Active JP5705273B2 (en) 2002-06-17 2013-07-12 Audio information generation method

Family Applications Before (3)

Application Number Title Priority Date Filing Date
JP2004514060A Active JP4486496B2 (en) 2002-06-17 2003-05-30 Audio coding system using spectral hole filling
JP2010030139A Active JP5063717B2 (en) 2002-06-17 2010-02-15 Audio information generation method
JP2011287051A Active JP5253564B2 (en) 2002-06-17 2011-12-28 Audio coding system that uses the characteristics of the decoded signal to fit the synthesized spectral components

Family Applications After (2)

Application Number Title Priority Date Filing Date
JP2012149087A Active JP5345722B2 (en) 2002-06-17 2012-07-03 Audio information generation method
JP2013146451A Active JP5705273B2 (en) 2002-06-17 2013-07-12 Audio information generation method

Country Status (20)

Country Link
US (4) US7447631B2 (en)
EP (6) EP1514261B1 (en)
JP (6) JP4486496B2 (en)
KR (5) KR100991450B1 (en)
CN (1) CN100369109C (en)
AT (7) AT526661T (en)
CA (6) CA2736046A1 (en)
DE (3) DE60310716T8 (en)
DK (3) DK1514261T3 (en)
ES (1) ES2275098T3 (en)
HK (6) HK1070728A1 (en)
IL (2) IL165650A (en)
MX (1) MXPA04012539A (en)
MY (2) MY159022A (en)
PL (1) PL208344B1 (en)
PT (1) PT2216777E (en)
SG (3) SG10201702049SA (en)
SI (2) SI2209115T1 (en)
TW (1) TWI352969B (en)
WO (1) WO2003107328A1 (en)

Families Citing this family (120)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7742927B2 (en) * 2000-04-18 2010-06-22 France Telecom Spectral enhancing method and device
DE10134471C2 (en) * 2001-02-28 2003-05-22 Fraunhofer Ges Forschung Method and device for characterizing a signal and method and device for generating an indexed signal
US7240001B2 (en) 2001-12-14 2007-07-03 Microsoft Corporation Quality improvement techniques in an audio encoder
US7447631B2 (en) 2002-06-17 2008-11-04 Dolby Laboratories Licensing Corporation Audio coding system using spectral hole filling
JP2005532586A (en) * 2002-07-08 2005-10-27 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィKoninklijke Philips Electronics N.V. Audio processing
US7889783B2 (en) * 2002-12-06 2011-02-15 Broadcom Corporation Multiple data rate communication system
PL1629463T3 (en) 2003-05-28 2008-01-31 Dolby Laboratories Licensing Corp Method, apparatus and computer program for calculating and adjusting the perceived loudness of an audio signal
US7461003B1 (en) * 2003-10-22 2008-12-02 Tellabs Operations, Inc. Methods and apparatus for improving the quality of speech signals
US7460990B2 (en) 2004-01-23 2008-12-02 Microsoft Corporation Efficient coding of digital media spectral data using wide-sense perceptual similarity
AU2004317678C1 (en) * 2004-03-12 2009-09-24 Nokia Technologies Oy Synthesizing a mono audio signal based on an encoded multichannel audio signal
US8417515B2 (en) * 2004-05-14 2013-04-09 Panasonic Corporation Encoding device, decoding device, and method thereof
EP1742202B1 (en) * 2004-05-19 2008-05-07 Matsushita Electric Industrial Co., Ltd. Encoding device, decoding device, and method thereof
CN101006496B (en) * 2004-08-17 2012-03-21 皇家飞利浦电子股份有限公司 Scalable audio coding
KR20070065401A (en) * 2004-09-23 2007-06-22 코닌클리케 필립스 일렉트로닉스 엔.브이. A system and a method of processing audio data, a program element and a computer-readable medium
MX2007005027A (en) 2004-10-26 2007-06-19 Dolby Lab Licensing Corp Calculating and adjusting the perceived loudness and/or the perceived spectral balance of an audio signal.
US8199933B2 (en) 2004-10-26 2012-06-12 Dolby Laboratories Licensing Corporation Calculating and adjusting the perceived loudness and/or the perceived spectral balance of an audio signal
KR100657916B1 (en) * 2004-12-01 2006-12-14 삼성전자주식회사 Apparatus and method for processing audio signal using correlation between bands
KR100707173B1 (en) * 2004-12-21 2007-04-13 삼성전자주식회사 Low bitrate encoding/decoding method and apparatus
KR100851970B1 (en) * 2005-07-15 2008-08-12 삼성전자주식회사 Method and apparatus for extracting ISCImportant Spectral Component of audio signal, and method and appartus for encoding/decoding audio signal with low bitrate using it
US7562021B2 (en) * 2005-07-15 2009-07-14 Microsoft Corporation Modification of codewords in dictionary used for efficient coding of digital media spectral data
US7630882B2 (en) * 2005-07-15 2009-12-08 Microsoft Corporation Frequency segmentation to obtain bands for efficient coding of digital media
US7546240B2 (en) 2005-07-15 2009-06-09 Microsoft Corporation Coding with improved time resolution for selected segments via adaptive block transformation of a group of samples from a subband decomposition
US8121848B2 (en) * 2005-09-08 2012-02-21 Pan Pacific Plasma Llc Bases dictionary for low complexity matching pursuits data coding and decoding
US20070053603A1 (en) * 2005-09-08 2007-03-08 Monro Donald M Low complexity bases matching pursuits data coding and decoding
US7848584B2 (en) * 2005-09-08 2010-12-07 Monro Donald M Reduced dimension wavelet matching pursuits coding and decoding
US7813573B2 (en) * 2005-09-08 2010-10-12 Monro Donald M Data coding and decoding with replicated matching pursuits
US8126706B2 (en) * 2005-12-09 2012-02-28 Acoustic Technologies, Inc. Music detector for echo cancellation and noise reduction
TWI517562B (en) 2006-04-04 2016-01-11 杜比實驗室特許公司 Method, apparatus, and computer program for scaling the overall perceived loudness of a multichannel audio signal by a desired amount
JP5185254B2 (en) 2006-04-04 2013-04-17 ドルビー ラボラトリーズ ライセンシング コーポレイション Audio signal volume measurement and improvement in MDCT region
DK1869669T3 (en) * 2006-04-24 2008-12-01 Nero Ag Advanced audio coding device
CN102684628B (en) 2006-04-27 2014-11-26 杜比实验室特许公司 Method for modifying parameters of audio dynamic processor and device executing the method
US20070270987A1 (en) * 2006-05-18 2007-11-22 Sharp Kabushiki Kaisha Signal processing method, signal processing apparatus and recording medium
EP2082480B1 (en) 2006-10-20 2019-07-24 Dolby Laboratories Licensing Corporation Audio dynamics processing using a reset
US8521314B2 (en) 2006-11-01 2013-08-27 Dolby Laboratories Licensing Corporation Hierarchical control path with constraints for audio dynamics processing
US8639500B2 (en) * 2006-11-17 2014-01-28 Samsung Electronics Co., Ltd. Method, medium, and apparatus with bandwidth extension encoding and/or decoding
KR101379263B1 (en) * 2007-01-12 2014-03-28 삼성전자주식회사 Method and apparatus for decoding bandwidth extension
AU2012261547B2 (en) * 2007-03-09 2014-04-17 Skype Speech coding system and method
GB0704622D0 (en) * 2007-03-09 2007-04-18 Skype Ltd Speech coding system and method
KR101411900B1 (en) * 2007-05-08 2014-06-26 삼성전자주식회사 Method and apparatus for encoding and decoding audio signal
US7761290B2 (en) 2007-06-15 2010-07-20 Microsoft Corporation Flexible frequency and time partitioning in perceptual transform coding of audio
US7774205B2 (en) * 2007-06-15 2010-08-10 Microsoft Corporation Coding of sparse digital media spectral data
US8046214B2 (en) 2007-06-22 2011-10-25 Microsoft Corporation Low complexity decoder for complex transform coding of multi-channel sound
US7885819B2 (en) 2007-06-29 2011-02-08 Microsoft Corporation Bitstream syntax for multi-process audio decoding
JP5192544B2 (en) 2007-07-13 2013-05-08 ドルビー ラボラトリーズ ライセンシング コーポレイション Acoustic processing using auditory scene analysis and spectral distortion
ES2526333T3 (en) * 2007-08-27 2015-01-09 Telefonaktiebolaget L M Ericsson (Publ) Adaptive transition frequency between noise refilling and bandwidth extension
CA2698031C (en) 2007-08-27 2016-10-18 Telefonaktiebolaget Lm Ericsson (Publ) Method and device for noise filling
US8583426B2 (en) 2007-09-12 2013-11-12 Dolby Laboratories Licensing Corporation Speech enhancement with voice clarity
CN101802909B (en) * 2007-09-12 2013-07-10 杜比实验室特许公司 Speech enhancement with noise level estimation adjustment
US8249883B2 (en) 2007-10-26 2012-08-21 Microsoft Corporation Channel extension coding for multi-channel source
EP2229676B1 (en) * 2007-12-31 2013-11-06 LG Electronics Inc. A method and an apparatus for processing an audio signal
KR101400513B1 (en) 2008-07-11 2014-05-28 프라운호퍼 게젤샤프트 쭈르 푀르데룽 데어 안겐반텐 포르슝 에. 베. Providing a Time Warp Activation Signal and Encoding an Audio Signal Therewith
EP2304719B1 (en) * 2008-07-11 2017-07-26 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Audio encoder, methods for providing an audio stream and computer program
MY154452A (en) * 2008-07-11 2015-06-15 Fraunhofer Ges Forschung An apparatus and a method for decoding an encoded audio signal
EP2320416B1 (en) * 2008-08-08 2014-03-05 Panasonic Corporation Spectral smoothing device, encoding device, decoding device, communication terminal device, base station device, and spectral smoothing method
US8532983B2 (en) * 2008-09-06 2013-09-10 Huawei Technologies Co., Ltd. Adaptive frequency prediction for encoding or decoding an audio signal
US8532998B2 (en) 2008-09-06 2013-09-10 Huawei Technologies Co., Ltd. Selective bandwidth extension for encoding/decoding audio/speech signal
US8515747B2 (en) * 2008-09-06 2013-08-20 Huawei Technologies Co., Ltd. Spectrum harmonic/noise sharpness control
WO2010028299A1 (en) * 2008-09-06 2010-03-11 Huawei Technologies Co., Ltd. Noise-feedback for spectral envelope quantization
WO2010031003A1 (en) 2008-09-15 2010-03-18 Huawei Technologies Co., Ltd. Adding second enhancement layer to celp based core layer
WO2010031049A1 (en) * 2008-09-15 2010-03-18 GH Innovation, Inc. Improving celp post-processing for music signals
US8364471B2 (en) * 2008-11-04 2013-01-29 Lg Electronics Inc. Apparatus and method for processing a time domain audio signal with a noise filling flag
GB2466201B (en) * 2008-12-10 2012-07-11 Skype Ltd Regeneration of wideband speech
GB0822537D0 (en) 2008-12-10 2009-01-14 Skype Ltd Regeneration of wideband speech
US9947340B2 (en) * 2008-12-10 2018-04-17 Skype Regeneration of wideband speech
TWI597939B (en) 2009-02-18 2017-09-01 杜比國際公司 Complex-valued synthesis filter bank with phase shift
KR101078378B1 (en) * 2009-03-04 2011-10-31 주식회사 코아로직 Method and Apparatus for Quantization of Audio Encoder
WO2010111876A1 (en) * 2009-03-31 2010-10-07 华为技术有限公司 Method and device for signal denoising and system for audio frequency decoding
JP5754899B2 (en) 2009-10-07 2015-07-29 ソニー株式会社 Decoding apparatus and method, and program
KR101419148B1 (en) * 2009-10-20 2014-07-11 프라운호퍼 게젤샤프트 쭈르 푀르데룽 데어 안겐반텐 포르슝 에. 베. Audio encoder, audio decoder, method for encoding an audio information, method for decoding an audio information and computer program using an iterative interval size reduction
US9117458B2 (en) * 2009-11-12 2015-08-25 Lg Electronics Inc. Apparatus for processing an audio signal and method thereof
US9838784B2 (en) 2009-12-02 2017-12-05 Knowles Electronics, Llc Directional audio capture
WO2011086065A1 (en) 2010-01-12 2011-07-21 Fraunhofer Gesellschaft zur Förderung der angewandten Forschung e.V. Audio encoder, audio decoder, method for encoding and audio information, method for decoding an audio information and computer program using a hash table describing both significant state values and interval boundaries
RU2518682C2 (en) * 2010-01-19 2014-06-10 Долби Интернешнл Аб Improved subband block based harmonic transposition
TWI557723B (en) 2010-02-18 2016-11-11 杜比實驗室特許公司 Decoding method and system
US9047876B2 (en) 2010-03-30 2015-06-02 Panasonic Intellectual Property Managment Co., Ltd. Audio device
JP5609737B2 (en) 2010-04-13 2014-10-22 ソニー株式会社 Signal processing apparatus and method, encoding apparatus and method, decoding apparatus and method, and program
JP5850216B2 (en) 2010-04-13 2016-02-03 ソニー株式会社 Signal processing apparatus and method, encoding apparatus and method, decoding apparatus and method, and program
US8798290B1 (en) 2010-04-21 2014-08-05 Audience, Inc. Systems and methods for adaptive signal equalization
US9558755B1 (en) 2010-05-20 2017-01-31 Knowles Electronics, Llc Noise suppression assisted automatic speech recognition
WO2011156905A2 (en) * 2010-06-17 2011-12-22 Voiceage Corporation Multi-rate algebraic vector quantization with supplemental coding of missing spectrum sub-bands
US20120029926A1 (en) 2010-07-30 2012-02-02 Qualcomm Incorporated Systems, methods, apparatus, and computer-readable media for dependent-mode coding of audio signals
JP6075743B2 (en) * 2010-08-03 2017-02-08 ソニー株式会社 Signal processing apparatus and method, and program
US9208792B2 (en) * 2010-08-17 2015-12-08 Qualcomm Incorporated Systems, methods, apparatus, and computer-readable media for noise injection
US9008811B2 (en) 2010-09-17 2015-04-14 Xiph.org Foundation Methods and systems for adaptive time-frequency resolution in digital data coding
JP5707842B2 (en) 2010-10-15 2015-04-30 ソニー株式会社 Encoding apparatus and method, decoding apparatus and method, and program
WO2012053150A1 (en) * 2010-10-18 2012-04-26 パナソニック株式会社 Audio encoding device and audio decoding device
US10121481B2 (en) 2011-03-04 2018-11-06 Telefonaktiebolaget Lm Ericsson (Publ) Post-quantization gain correction in audio coding
US9009036B2 (en) 2011-03-07 2015-04-14 Xiph.org Foundation Methods and systems for bit allocation and partitioning in gain-shape vector quantization for audio coding
WO2012122303A1 (en) 2011-03-07 2012-09-13 Xiph. Org Method and system for two-step spreading for tonal artifact avoidance in audio coding
WO2012122297A1 (en) * 2011-03-07 2012-09-13 Xiph. Org. Methods and systems for avoiding partial collapse in multi-block audio coding
HUE026874T2 (en) * 2011-03-10 2016-07-28 ERICSSON TELEFON AB L M (publ) Filling of non-coded sub-vectors in transform coded audio signals
ES2637031T3 (en) 2011-04-15 2017-10-10 Telefonaktiebolaget Lm Ericsson (Publ) Decoder for attenuation of reconstructed signal regions with low accuracy
RU2705052C2 (en) 2011-05-13 2019-11-01 Самсунг Электроникс Ко., Лтд. Bit allocation, audio encoding and decoding
JP5986565B2 (en) * 2011-06-09 2016-09-06 パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America Speech coding apparatus, speech decoding apparatus, speech coding method, and speech decoding method
JP2013007944A (en) 2011-06-27 2013-01-10 Sony Corp Signal processing apparatus, signal processing method, and program
US20130006644A1 (en) * 2011-06-30 2013-01-03 Zte Corporation Method and device for spectral band replication, and method and system for audio decoding
JP5997592B2 (en) * 2012-04-27 2016-09-28 株式会社Nttドコモ Speech decoder
WO2013188562A2 (en) * 2012-06-12 2013-12-19 Audience, Inc. Bandwidth extension via constrained synthesis
EP2717263B1 (en) * 2012-10-05 2016-11-02 Nokia Technologies Oy Method, apparatus, and computer program product for categorical spatial analysis-synthesis on the spectrum of a multichannel audio signal
CN103854653B (en) 2012-12-06 2016-12-28 华为技术有限公司 The method and apparatus of signal decoding
KR101757341B1 (en) * 2013-01-29 2017-07-14 프라운호퍼 게젤샤프트 쭈르 푀르데룽 데어 안겐반텐 포르슝 에.베. Low-complexity tonality-adaptive audio signal quantization
KR101757347B1 (en) * 2013-01-29 2017-07-26 프라운호퍼 게젤샤프트 쭈르 푀르데룽 데어 안겐반텐 포르슝 에.베. Noise filling in perceptual transform audio coding
RU2640722C2 (en) 2013-04-05 2018-01-11 Долби Интернешнл Аб Improved quantizer
JP6157926B2 (en) * 2013-05-24 2017-07-05 株式会社東芝 Audio processing apparatus, method and program
EP2830060A1 (en) * 2013-07-22 2015-01-28 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Noise filling in multichannel audio coding
EP2830061A1 (en) * 2013-07-22 2015-01-28 Fraunhofer Gesellschaft zur Förderung der angewandten Forschung e.V. Apparatus and method for encoding and decoding an encoded audio signal using temporal noise/patch shaping
EP2830055A1 (en) * 2013-07-22 2015-01-28 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Context-based entropy coding of sample values of a spectral envelope
WO2015041070A1 (en) 2013-09-19 2015-03-26 ソニー株式会社 Encoding device and method, decoding device and method, and program
EP2919232A1 (en) * 2014-03-14 2015-09-16 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Encoder, decoder and method for encoding and decoding
JP6035270B2 (en) 2014-03-24 2016-11-30 株式会社Nttドコモ Speech decoding apparatus, speech encoding apparatus, speech decoding method, speech encoding method, speech decoding program, and speech encoding program
RU2572664C2 (en) * 2014-06-04 2016-01-20 Российская Федерация, От Имени Которой Выступает Министерство Промышленности И Торговли Российской Федерации Device for active vibration suppression
CN107112025A (en) 2014-09-12 2017-08-29 美商楼氏电子有限公司 System and method for recovering speech components
US10460736B2 (en) * 2014-11-07 2019-10-29 Samsung Electronics Co., Ltd. Method and apparatus for restoring audio signal
US20160173808A1 (en) 2014-12-16 2016-06-16 Psyx Research, Inc. System and method for level control at a receiver
DE112016000545B4 (en) 2015-01-30 2019-08-22 Knowles Electronics, Llc Context-related switching of microphones
TW201643864A (en) 2015-03-13 2016-12-16 杜比國際公司 Decoding audio bitstreams with enhanced spectral band replication metadata in at least one fill element
US20170024495A1 (en) * 2015-07-21 2017-01-26 Positive Grid LLC Method of modeling characteristics of a musical instrument
DE102016104665A1 (en) * 2016-03-14 2017-09-14 Ask Industries Gmbh Method and device for processing a lossy compressed audio signal
JP2019526691A (en) * 2016-12-09 2019-09-19 エルジー・ケム・リミテッド Sealant composition
EP3544005A1 (en) 2018-03-22 2019-09-25 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Audio encoder, audio decoder, audio encoding method and audio decoding method for dithered quantization for frequency-domain speech and audio coding

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2001080223A1 (en) * 2000-04-18 2001-10-25 France Telecom Sa Spectral enhancing method and device
JP2001521648A (en) * 1997-06-10 2001-11-06 コーディング テクノロジーズ スウェーデン アクチボラゲット Strengthening of primitive coding using a spectral band replication
JP2001324996A (en) * 2000-05-15 2001-11-22 Japan Music Agency Co Ltd Method and device for reproducing mp3 music data
JP2001525079A (en) * 1997-05-15 2001-12-04 ヒューレット・パッカード・カンパニー Speech coding system and method

Family Cites Families (64)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US36478A (en) * 1862-09-16 Improved can or tank for coal-oil
US3995115A (en) * 1967-08-25 1976-11-30 Bell Telephone Laboratories, Incorporated Speech privacy system
US3684838A (en) * 1968-06-26 1972-08-15 Kahn Res Lab Single channel audio signal transmission system
JPS6011360B2 (en) * 1981-12-15 1985-03-25 Kokusai Denshin Denwa Co Ltd
US4667340A (en) * 1983-04-13 1987-05-19 Texas Instruments Incorporated Voice messaging system with pitch-congruent baseband coding
WO1986003873A1 (en) * 1984-12-20 1986-07-03 Gte Laboratories Incorporated Method and apparatus for encoding speech
US4885790A (en) * 1985-03-18 1989-12-05 Massachusetts Institute Of Technology Processing of acoustic waveforms
US4790016A (en) * 1985-11-14 1988-12-06 Gte Laboratories Incorporated Adaptive method and apparatus for coding speech
US4935963A (en) * 1986-01-24 1990-06-19 Racal Data Communications Inc. Method and apparatus for processing speech signals
JPS62234435A (en) * 1986-04-04 1987-10-14 Kokusai Denshin Denwa Co Ltd <Kdd> Voice coding system
DE3683767D1 (en) * 1986-04-30 1992-03-12 Ibm Speech coding method and device for implementing this method.
US4776014A (en) * 1986-09-02 1988-10-04 General Electric Company Method for pitch-aligned high-frequency regeneration in RELP vocoders
US5054072A (en) * 1987-04-02 1991-10-01 Massachusetts Institute Of Technology Coding of acoustic waveforms
US5127054A (en) * 1988-04-29 1992-06-30 Motorola, Inc. Speech quality improvement for voice coders and synthesizers
JPH02183630A (en) * 1989-01-10 1990-07-18 Fujitsu Ltd Voice coding system
US5109417A (en) * 1989-01-27 1992-04-28 Dolby Laboratories Licensing Corporation Low bit rate transform coder, decoder, and encoder/decoder for high-quality audio
US5054075A (en) * 1989-09-05 1991-10-01 Motorola, Inc. Subband decoding method and apparatus
CN1062963C (en) * 1990-04-12 2001-03-07 多尔拜实验特许公司 Encoder/decoder for producing high-quality audio signals
ES2087522T3 (en) * 1991-01-08 1996-07-16 Dolby Lab Licensing Corp Decoding / encoding for multi-dimensional sound fields.
JP3134337B2 (en) * 1991-03-30 2001-02-13 ソニー株式会社 Digital signal encoding method
EP0551705A3 (en) * 1992-01-15 1993-08-18 Ericsson Ge Mobile Communications Inc. Method for subbandcoding using synthetic filler signals for non transmitted subbands
JP2563719B2 (en) 1992-03-11 1996-12-18 技術研究組合医療福祉機器研究所 Voice processing apparatus and the hearing aid
JP2693893B2 (en) * 1992-03-30 1997-12-24 松下電器産業株式会社 Stereo audio encoding method
JP3508146B2 (en) * 1992-09-11 2004-03-22 ソニー株式会社 Digital signal encoding / decoding device, digital signal encoding device, and digital signal decoding device
JP3127600B2 (en) * 1992-09-11 2001-01-29 ソニー株式会社 Digital signal decoding apparatus and method
US5402124A (en) * 1992-11-25 1995-03-28 Dolby Laboratories Licensing Corporation Encoder and decoder with improved quantizer using reserved quantizer level for small amplitude signals
US5394466A (en) * 1993-02-16 1995-02-28 Keptel, Inc. Combination telephone network interface and cable television apparatus and cable television module
US5623577A (en) * 1993-07-16 1997-04-22 Dolby Laboratories Licensing Corporation Computationally efficient adaptive bit allocation for encoding method and apparatus with allowance for decoder spectral distortions
JPH07225598A (en) 1993-09-22 1995-08-22 Massachusetts Inst Of Technol <Mit> Method and device for acoustic coding using dynamically determined critical band
JP3186489B2 (en) * 1994-02-09 2001-07-11 ソニー株式会社 Digital signal processing method and apparatus
JP3277682B2 (en) * 1994-04-22 2002-04-22 ソニー株式会社 Information encoding method and apparatus, information decoding method and apparatus, and information recording medium and the information transmission method
WO1995032499A1 (en) * 1994-05-25 1995-11-30 Sony Corporation Encoding method, decoding method, encoding-decoding method, encoder, decoder, and encoder-decoder
US5748786A (en) * 1994-09-21 1998-05-05 Ricoh Company, Ltd. Apparatus for compression using reversible embedded wavelets
JP3254953B2 (en) 1995-02-17 2002-02-12 日本ビクター株式会社 Voice and high efficiency coding device
DE19509149A1 (en) 1995-03-14 1996-09-19 Donald Dipl Ing Schulz Audio signal coding for data compression factor
JPH08328599A (en) 1995-06-01 1996-12-13 Mitsubishi Electric Corp Mpeg audio decoder
DE69620967T2 (en) * 1995-09-19 2002-11-07 At & T Corp Synthesis of speech signals in the absence of encoded parameters
US5692102A (en) * 1995-10-26 1997-11-25 Motorola, Inc. Method device and system for an efficient noise injection process for low bitrate audio compression
US6138051A (en) * 1996-01-23 2000-10-24 Sarnoff Corporation Method and apparatus for evaluating an audio decoder
JP3189660B2 (en) * 1996-01-30 2001-07-16 ソニー株式会社 Signal encoding method
JP3519859B2 (en) * 1996-03-26 2004-04-19 三菱電機株式会社 Encoder and decoder
DE19628293C1 (en) * 1996-07-12 1997-12-11 Fraunhofer Ges Forschung Encoding and decoding of audio signals using intensity stereo and prediction
US6092041A (en) * 1996-08-22 2000-07-18 Motorola, Inc. System and method of encoding and decoding a layered bitstream by re-applying psychoacoustic analysis in the decoder
JPH1091199A (en) * 1996-09-18 1998-04-10 Mitsubishi Electric Corp Recording and reproducing device
US5924064A (en) * 1996-10-07 1999-07-13 Picturetel Corporation Variable length coding using a plurality of region bit allocation patterns
JP3213582B2 (en) * 1997-05-29 2001-10-02 シャープ株式会社 Image encoding apparatus and image decoding apparatus
CN1144179C (en) * 1997-07-11 2004-03-31 索尼株式会社 Information decorder and decoding method, information encoder and encoding method and distribution medium
DE19730130C2 (en) * 1997-07-14 2002-02-28 Fraunhofer Ges Forschung Method for coding an audio signal
WO1999050828A1 (en) * 1998-03-30 1999-10-07 Voxware, Inc. Low-complexity, low-delay, scalable and embedded speech and audio coding with adaptive frame loss concealment
US6115689A (en) * 1998-05-27 2000-09-05 Microsoft Corporation Scalable audio coder and decoder
JP2000148191A (en) * 1998-11-06 2000-05-26 Matsushita Electric Ind Co Ltd Coding device for digital audio signal
US6300888B1 (en) * 1998-12-14 2001-10-09 Microsoft Corporation Entrophy code mode switching for frequency-domain audio coding
SE9903553D0 (en) 1999-01-27 1999-10-01 Lars Liljeryd Enhancing percepptual performance of SBR and related coding methods by adaptive noise addition (ANA) and noise substitution limiting (NSL)
US6363338B1 (en) * 1999-04-12 2002-03-26 Dolby Laboratories Licensing Corporation Quantization in perceptual audio coders with compensation for synthesis filter noise spreading
CA2368453C (en) * 1999-04-16 2009-12-08 Grant Allen Davidson Using gain-adaptive quantization and non-uniform symbol lengths for audio coding
JP3616307B2 (en) * 2000-05-22 2005-02-02 日本電信電話株式会社 Voice / musical sound signal encoding method and recording medium storing program for executing the method
SE0001926D0 (en) * 2000-05-23 2000-05-23 Lars Liljeryd Improved spectral translation / folding in the sub-band domain
JP2001343998A (en) * 2000-05-31 2001-12-14 Yamaha Corp Digital audio decoder
JP3538122B2 (en) 2000-06-14 2004-06-14 株式会社ケンウッド Frequency interpolation device, frequency interpolation method, and recording medium
SE0004187D0 (en) 2000-11-15 2000-11-15 Coding Technologies Sweden Ab Enhancing the performance of coding systems That use high frequency reconstruction methods
GB0103245D0 (en) * 2001-02-09 2001-03-28 Radioscape Ltd Method of inserting additional data into a compressed signal
US6963842B2 (en) * 2001-09-05 2005-11-08 Creative Technology Ltd. Efficient system and method for converting between different transform-domain signal representations
US20030187663A1 (en) 2002-03-28 2003-10-02 Truman Michael Mead Broadband frequency translation for high frequency regeneration
US7447631B2 (en) * 2002-06-17 2008-11-04 Dolby Laboratories Licensing Corporation Audio coding system using spectral hole filling

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001525079A (en) * 1997-05-15 2001-12-04 ヒューレット・パッカード・カンパニー Speech coding system and method
JP2001521648A (en) * 1997-06-10 2001-11-06 コーディング テクノロジーズ スウェーデン アクチボラゲット Strengthening of primitive coding using a spectral band replication
WO2001080223A1 (en) * 2000-04-18 2001-10-25 France Telecom Sa Spectral enhancing method and device
JP2001324996A (en) * 2000-05-15 2001-11-22 Japan Music Agency Co Ltd Method and device for reproducing mp3 music data

Also Published As

Publication number Publication date
KR20050010945A (en) 2005-01-28
JP5253564B2 (en) 2013-07-31
DE60332833D1 (en) 2010-07-15
JP2012103718A (en) 2012-05-31
KR100991450B1 (en) 2010-11-04
DE60310716T2 (en) 2007-10-11
JP2013214103A (en) 2013-10-17
SI2209115T1 (en) 2012-05-31
CA2736060C (en) 2015-02-17
CA2736046A1 (en) 2003-12-24
CA2736065A1 (en) 2003-12-24
CN100369109C (en) 2008-02-13
EP2207170A1 (en) 2010-07-14
JP2010156990A (en) 2010-07-15
SI2207169T1 (en) 2012-05-31
SG10201702049SA (en) 2017-04-27
CA2489441C (en) 2012-04-10
SG2014005300A (en) 2016-10-28
KR100986152B1 (en) 2010-10-07
JP4486496B2 (en) 2010-06-23
CA2736055C (en) 2015-02-24
JP5705273B2 (en) 2015-04-22
KR100986153B1 (en) 2010-10-07
PL208344B1 (en) 2011-04-29
JP2005530205A (en) 2005-10-06
IL216069A (en) 2015-11-30
TWI352969B (en) 2011-11-21
WO2003107328A1 (en) 2003-12-24
HK1141623A1 (en) 2012-02-24
MY159022A (en) 2016-11-30
AT470220T (en) 2010-06-15
IL216069D0 (en) 2011-12-29
KR100986150B1 (en) 2010-10-07
JP2012212167A (en) 2012-11-01
IL165650D0 (en) 2006-01-15
US20090138267A1 (en) 2009-05-28
KR20100063141A (en) 2010-06-10
AT349754T (en) 2007-01-15
US20030233236A1 (en) 2003-12-18
AT529858T (en) 2011-11-15
AT526661T (en) 2011-10-15
US8050933B2 (en) 2011-11-01
US7337118B2 (en) 2008-02-26
HK1146146A1 (en) 2012-08-17
DK1514261T3 (en) 2007-03-19
PL372104A1 (en) 2005-07-11
DE60310716T8 (en) 2008-01-31
US8032387B2 (en) 2011-10-04
AT536615T (en) 2011-12-15
HK1146145A1 (en) 2012-08-17
EP2207169A1 (en) 2010-07-14
JP5253565B2 (en) 2013-07-31
AU2003237295A1 (en) 2003-12-31
HK1070728A1 (en) 2011-02-18
DE60333316D1 (en) 2010-08-19
HK1070729A1 (en) 2007-04-13
EP2209115A1 (en) 2010-07-21
KR20050010950A (en) 2005-01-28
EP1736966A3 (en) 2007-11-07
AT473503T (en) 2010-07-15
EP2216777A1 (en) 2010-08-11
EP1736966B1 (en) 2010-07-07
KR100991448B1 (en) 2010-11-04
KR20100086068A (en) 2010-07-29
AT529859T (en) 2011-11-15
CA2736055A1 (en) 2003-12-24
TW200404273A (en) 2004-03-16
EP1736966A2 (en) 2006-12-27
PT2216777E (en) 2012-03-16
EP1514261B1 (en) 2006-12-27
IL165650A (en) 2010-11-30
CN1662958A (en) 2005-08-31
SG177013A1 (en) 2012-01-30
CA2735830A1 (en) 2003-12-24
MY136521A (en) 2008-10-31
EP2209115B1 (en) 2011-09-28
ES2275098T3 (en) 2007-06-01
CA2736065C (en) 2015-02-10
CA2489441A1 (en) 2003-12-24
US20030233234A1 (en) 2003-12-18
US7447631B2 (en) 2008-11-04
EP2207170B1 (en) 2011-10-19
JP5063717B2 (en) 2012-10-31
CA2736060A1 (en) 2003-12-24
EP1514261A1 (en) 2005-03-16
JP5345722B2 (en) 2013-11-20
EP2216777B1 (en) 2011-12-07
CA2735830C (en) 2014-04-08
EP2207169B1 (en) 2011-10-19
MXPA04012539A (en) 2005-04-28
US20090144055A1 (en) 2009-06-04
DE60310716D1 (en) 2007-02-08
KR20100086067A (en) 2010-07-29
DK1736966T3 (en) 2010-11-01
DK2207169T3 (en) 2012-02-06
HK1141624A1 (en) 2012-05-25

Similar Documents

Publication Publication Date Title
KR100209870B1 (en) Perceptual coding of audio signals
ES2664397T3 (en) Enhanced audio coding systems and methods that use a coupling of spectral components and regeneration of spectral components
AU746416B2 (en) Method and apparatus for encoding and decoding multiple audio channels at low bit rates
KR100949232B1 (en) Encoding device, decoding device and methods thereof
KR100978018B1 (en) Parametric representation of spatial audio
AU2007208482B2 (en) Complex-transform channel coding with extended-band frequency coding
DE60303689T2 (en) Audio decoding device and method
KR100898879B1 (en) Modulating One or More Parameter of An Audio or Video Perceptual Coding System in Response to Supplemental Information
JP4625084B2 (en) Shaped diffuse sound for binaural cue coding method etc.
CA2286068C (en) Method for coding an audio signal
CN1838239B (en) Apparatus for enhancing audio source decoder and method thereof
EP2250641B1 (en) Apparatus for mixing a plurality of input data streams
JP4574626B2 (en) Apparatus and method for constructing a multi-channel output signal or apparatus and method for generating a downmix signal
JP4664371B2 (en) Individual channel time envelope shaping for binaural cue coding method etc.
CA2197128C (en) Enhanced joint stereo coding method using temporal envelope shaping
EP2216777B1 (en) Audio coding system using spectral hole filling
CN101297356B (en) Audio compression
AU2002318813B2 (en) Audio signal decoding device and audio signal encoding device
ES2237706T3 (en) Reconstruction of high frequency components.
ES2686088T3 (en) Advanced processing based on a filter bank with complex exponential modulation and adaptive temporal signaling procedures
TWI307248B (en) Apparatus and method for generating multi-channel synthesizer control signal and apparatus and method for multi-channel synthesizing
US6680972B1 (en) Source coding enhancement using spectral-band replication
JP5307889B2 (en) Noise filling device, noise filling parameter computing device, method for providing noise filling parameter, method for providing noise filled spectral representation of audio signal, corresponding computer program and encoded audio signal
EP1798724A1 (en) Encoder, decoder, encoding method, and decoding method
JP4805540B2 (en) Stereo signal encoding

Legal Events

Date Code Title Description
TRDD Decision of grant or rejection written
A01 Written decision to grant a patent or to grant a registration (utility model)

Free format text: JAPANESE INTERMEDIATE CODE: A01

Effective date: 20130319

A61 First payment of annual fees (during grant procedure)

Free format text: JAPANESE INTERMEDIATE CODE: A61

Effective date: 20130416

R150 Certificate of patent or registration of utility model

Free format text: JAPANESE INTERMEDIATE CODE: R150

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20160426

Year of fee payment: 3

R250 Receipt of annual fees

Free format text: JAPANESE INTERMEDIATE CODE: R250

R250 Receipt of annual fees

Free format text: JAPANESE INTERMEDIATE CODE: R250

R250 Receipt of annual fees

Free format text: JAPANESE INTERMEDIATE CODE: R250

R250 Receipt of annual fees

Free format text: JAPANESE INTERMEDIATE CODE: R250