WO2011071610A1 - Decoding of multichannel aufio encoded bit streams using adaptive hybrid transformation - Google Patents

Decoding of multichannel aufio encoded bit streams using adaptive hybrid transformation Download PDF

Info

Publication number
WO2011071610A1
WO2011071610A1 PCT/US2010/054480 US2010054480W WO2011071610A1 WO 2011071610 A1 WO2011071610 A1 WO 2011071610A1 US 2010054480 W US2010054480 W US 2010054480W WO 2011071610 A1 WO2011071610 A1 WO 2011071610A1
Authority
WO
WIPO (PCT)
Prior art keywords
audio
channel
block
frame
encoded
Prior art date
Application number
PCT/US2010/054480
Other languages
English (en)
French (fr)
Inventor
Kamalanathan Ramamoorthy
Original Assignee
Dolby Laboratories Licensing Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to JP2012541085A priority Critical patent/JP5547297B2/ja
Priority to PL10776017T priority patent/PL2510515T3/pl
Application filed by Dolby Laboratories Licensing Corporation filed Critical Dolby Laboratories Licensing Corporation
Priority to SI201030604T priority patent/SI2510515T1/sl
Priority to AU2010328635A priority patent/AU2010328635B2/en
Priority to ES10776017.5T priority patent/ES2463840T3/es
Priority to NZ599981A priority patent/NZ599981A/en
Priority to UAA201208193A priority patent/UA100353C2/uk
Priority to RS20140226A priority patent/RS53288B/en
Priority to DK10776017.5T priority patent/DK2510515T3/da
Priority to KR1020137026329A priority patent/KR101629306B1/ko
Priority to CA2779453A priority patent/CA2779453C/en
Priority to CN201080051553.XA priority patent/CN102687198B/zh
Priority to AP2012006289A priority patent/AP3301A/xx
Priority to KR1020127012464A priority patent/KR101370522B1/ko
Priority to BR112012013745-0A priority patent/BR112012013745B1/pt
Priority to CN201410410643.3A priority patent/CN104217724B/zh
Priority to EA201270642A priority patent/EA024310B1/ru
Priority to MX2012005723A priority patent/MX2012005723A/es
Priority to US13/514,286 priority patent/US8891776B2/en
Priority to EP10776017.5A priority patent/EP2510515B1/en
Publication of WO2011071610A1 publication Critical patent/WO2011071610A1/en
Priority to IL219304A priority patent/IL219304A/en
Priority to ZA2012/03290A priority patent/ZA201203290B/en
Priority to TNP2012000211A priority patent/TN2012000211A1/en
Priority to MA34912A priority patent/MA33775B1/fr
Priority to HK12110658.8A priority patent/HK1170058A1/xx
Priority to HRP20140400AT priority patent/HRP20140400T1/hr
Priority to US14/512,755 priority patent/US9620132B2/en

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/02Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/008Multichannel audio signal coding or decoding using interchannel correlation to reduce redundancy, e.g. joint-stereo, intensity-coding or matrixing
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/02Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders
    • G10L19/0212Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders using orthogonal transformation

Definitions

  • the present invention pertains generally to audio coding systems and pertains more specifically to methods and devices that decode encoded digital audio signals.
  • ATSC Advanced Television Systems Committee
  • JCIC Joint Committee on InterSociety Coordination
  • bit stream syntax defining structural and syntactical features of the encoded information that a compliant decoder must be capable of decoding.
  • Many applications that comply with the ATSC Standards will transmit encoded digital audio information as binary data in a serial manner.
  • the encoded data is often referred to as a bit stream but other arrangements of the data are permissible.
  • bit stream is used herein to refer to an encoded digital audio signal regardless of the format or the recording or transmission technique that is used.
  • a bit stream that complies with the ATSC Standards is arranged in a series of "synchronization frames.”
  • Each frame is a unit of the bit stream that is capable of being fully decoded into one or more channels of pulse code modulated (PCM) digital audio data.
  • PCM pulse code modulated
  • Each frame includes "audio blocks" and frame metadata that is associated with the audio blocks.
  • Each of the audio blocks contain encoded audio data representing digital audio samples for one or more audio channels and block metadata associated with the encoded audio data.
  • a bit stream that complies with the ATSC Standards can be very complex because a large number of variations are possible.
  • a few examples mentioned here only briefly include channel coupling, channel rematrixing, dialog normalization, dynamic range compression, channel downmixing and block-length switching for standard AC-3 bit streams, and multiple independent streams, dependent substreams, spectral extension and adaptive hybrid transformation for enhanced AC-3 bit streams. Details for these features can be obtained from the A/52B document.
  • the preceding text and the following disclosure refer to encoded bit streams that comply with the ATSC Standards but the present invention is not limited to use with only these bit streams. Principles of the present invention may be applied to essentially any encoded bit stream that has structural features similar to the frames, blocks and channels used in AC-3 coding algorithms.
  • a method decodes a frame of an encoded digital audio signal by receiving the frame and examining the encoded digital audio signal in a single pass to decode the encoded audio data for each audio block in order by block.
  • Each frame comprises frame metadata and a plurality of audio blocks.
  • Each audio block comprises block metadata and encoded audio data for one or more audio channels.
  • the block metadata comprises control information describing coding tools used by an encoding process that produced the encoded audio data.
  • One of the coding tools is hybrid transform processing that applies an analysis filter bank implemented by a primary transform to one or more audio channels to generate spectral coefficients representing spectral content of the one or more audio channels, and applies a secondary transform to the spectral coefficients for at least some of the one or more audio channels to generate hybrid transform coefficients.
  • the decoding of each audio block determines whether the encoding process used adaptive hybrid transform processing to encode any of the encoded audio data. If the encoding process used adaptive hybrid transform processing, the method obtains all hybrid transform coefficients for the frame from the encoded audio data in the first audio block in the frame and applies an inverse secondary transform to the hybrid transform coefficients to obtain inverse secondary transform coefficients and obtains spectral coefficients from the inverse secondary transform coefficients.
  • spectral coefficients are obtained from the encoded audio data in the respective audio block.
  • An inverse primary transform is applied to the spectral coefficients to generate an output signal representing the one or more channels in the respective audio block.
  • Fig. 1 is a schematic block diagram of exemplary implementations of an encoder.
  • Fig. 2 is a schematic block diagram of exemplary implementations of a decoder.
  • Figs. 3A and 3B are schematic illustrations of frames in bit streams complying with standard and enhanced syntactical structures.
  • Figs. 4A and 4B are schematic illustrations of audio blocks that comply with standard and enhanced syntactical structures.
  • Figs. 5A to 5C are schematic illustrations of exemplary bit streams carrying data with program and channel extensions.
  • Fig. 6 is a schematic block diagram of an exemplary process implemented by a decoder that process encoded audio data in channel order.
  • Fig. 7 is a schematic block diagram of an exemplary process implemented by a decoder that process encoded audio data in block order.
  • Fig. 8 is a schematic block diagram of a device that may be used to implement various aspects of the present invention.
  • Figs. 1 and 2 are schematic block diagrams of exemplary implementations of an encoder and a decoder for an audio coding system in which the decoder may incorporate various aspects of the present invention. These implementations conform to what is disclosed in the A/52B document cited above.
  • the purpose of the coding system is to generate an encoded representation of input audio signals that can be recorded or transmitted and subsequently decoded to produce output audio signals that sound essentially identical to the input audio signals while using a minimum amount of digital information to represent the encoded signal.
  • Coding systems that comply with the basic ATSC Standards are capable of encoding and decoding information that can represent from one to so-called 5.1 channels of audio signals, where 5.1 is understood to mean five channels that can carry full-bandwidth signals and one channel of limited-bandwidth that is intended to carry signals for low-frequency effects (LFE).
  • the encoder receives a series of pulse code modulated (PCM) samples representing one or more input channels of audio signals from the input signal path 1, and applies an analysis filter bank 2 to the series of samples to generate digital values representing the spectral composition of the input audio signals.
  • PCM pulse code modulated
  • the analysis filter bank is implemented by a Modified Discrete Cosine Transform (MDCT) described in the A/52B document.
  • MDCT Modified Discrete Cosine Transform
  • the MDCT is applied to overlapping segments or blocks of samples for each input channel of audio signal to generate blocks of transform coefficients that represent the spectral composition of that input channel signal.
  • the MDCT is part of an analysis/synthesis system that uses specially designed window functions and overlap/add processes to cancel time- domain aliasing.
  • BFP block-floating point
  • This description refers to audio data expressed as floating-point exponents and mantissas because this form of representation is used in bit streams that comply with the ATSC Standards; however, this particular representation is merely one example of numerical representations that use scale factors and associated scaled values.
  • the BFP exponents for each block collectively provide an approximate spectral envelope for the input audio signal.
  • These exponents are encoded by delta modulation and other coding techniques to reduce information requirements, passed to the formatter 5, and input into a psychoacoustic model to estimate the psychoacoustic masking threshold of the signal being encoded.
  • the results from the model are used by the bit allocator 3 to allocate digital information in the form of bits for quantization of the mantissas in such a manner that the level of noise produced by quantization is kept below the psychoacoustic masking threshold of the signal being encoded.
  • the quantizer 4 quantizes the mantissas according to the bit allocations received from the bit allocator 3 and passed to the formatter 5.
  • the formatter 5 multiplexes or assembles the encoded exponents, the quantized mantissas and other control information, sometimes referred to as block metadata, into audio blocks.
  • the data for six successive audio blocks are assembled into units of digital information called frames.
  • the frames themselves also contain control information or frame metadata.
  • the encoded information for successive frames are output as a bit stream along the path 6 for recording on an information storage medium or for transmission along a communication channel.
  • the format of each frame in the bit stream complies with the syntax specified in the A/52B document.
  • the coding algorithm used by typical encoders that comply with the ATSC Standards are more complicated than what is illustrated in Fig. 1 and described above.
  • error detection codes are inserted into the frames to allow a receiving decoder to validate the bit stream.
  • a coding technique known as block-length switching may be used to adapt the temporal and spectral resolution of the analysis filter bank to optimize its performance with changing signal characteristics.
  • the floating-point exponents may be encoded with variable time and frequency resolution.
  • Two or more channels may be combined into a composite representation using a coding technique known as channel coupling.
  • Another coding technique known as channel rematrixing may be used adaptively for two-channel audio signals. Additional coding techniques may be used that are not mentioned here. A few of these other coding techniques are discussed below. Many other details of implementation are omitted because they are not needed to understand the present invention. These details may be obtained from the A/52B document as desired.
  • the decoder performs a decoding algorithm that is essentially the inverse of the coding algorithm that is performed in the encoder. Referring to the exemplary
  • the decoder receives an encoded bit stream representing a series of frames from the input signal path 11.
  • the encoded bit stream may be retrieved from an information storage medium or received from a communication channel.
  • the deformatter 12 demultiplexes or disassembles the encoded information for each frame into frame metadata and six audio blocks.
  • the audio blocks are disassembled into their respective block metadata, encoded exponents and quantized mantissas.
  • the encoded exponents are used by a psychoacoustic model in the bit allocator 13 to allocate digital information in the form of bits for dequantization of the quantized mantissas in the same manner as bits were allocated in the encoder.
  • the dequantizer 14 dequantizes the quantized mantissas according to the bit allocations received from the bit allocator 13 and passes the dequantized mantissas to the synthesis filter bank 15.
  • the encoded exponents are decoded and passed to the synthesis filter bank 15.
  • the decoded exponents and dequantized mantissas constitute a BFP representation of the spectral content of the input audio signal as encoded by the encoder.
  • the synthesis filter bank 15 is applied to the representation of spectral content to reconstruct an inexact replica of the original input audio signals, which is passed along the output signal path 16.
  • the synthesis filter bank is implemented by an Inverse Modified Discrete Cosine Transform (IMDCT) described in the A/52B document.
  • IMDCT Inverse Modified Discrete Cosine Transform
  • a few decoding techniques that are the inverse of the coding techniques described above include error detection for error correction or concealment, block-length switching to adapt the temporal and spectral resolution of the synthesis filter bank, channel decoupling to recover channel information from coupled composite representations, and matrix operations for recovery of rematrixed two-channel representations. Information about other techniques and additional detail may be obtained from the A/52B document as desired.
  • An encoded bit stream that complies with the ATSC Standards comprises a series of encoded information units called "synchronization frames" that are sometimes referred to more simply as frames.
  • each frame contains frame metadata and six audio blocks.
  • Each audio block contains block metadata and encoded BFP exponents and mantissas for a concurrent interval of one or more channels of audio signals.
  • the structure for the standard bit stream is illustrated schematically in Fig. 3A.
  • the structure for an enhanced AC-3 bit stream as described in Annex E of the A/52B document is illustrated in Fig. 3B.
  • the portion of each bit stream within the marked interval from SI to CRC is one frame.
  • a special bit pattern or synchronization word is included in synchronization information (SI) that is provided at the start of each frame so that a decoder may identify the start of a frame and maintain synchronization of its decoding processes with the encoded bit stream.
  • SI synchronization information
  • a bit stream information (BSI) section immediately following the SI carries parameters that are needed by the decoding algorithm to decode the frame.
  • the BSI specifies the number, type and order of channels that are represented by encoded information in the frame, and the dynamic range compression and dialogue normalization information to be used by the decoder.
  • Each frame contains six audio blocks (ABO to AB5), which may be followed by auxiliary (AUX) data if desired.
  • Error detection information in the form of a cyclical redundancy check (CRC) word is provided at the end of each frame.
  • CRC cyclical redundancy check
  • a frame in the enhanced AC-3 bit stream also contains audio frame (AFRM) data that contains flags and parameters that pertain to additional coding techniques that are not available for use in coding a standard bit stream.
  • AFRM audio frame
  • SPX spectral extension
  • AHT adaptive hybrid transform
  • Each audio block contains encoded representations of BFP exponents and quantized mantissas for 256 transform coefficients, and block metadata needed to decode the encoded exponents and quantized mantissas.
  • This structure is illustrated schematically in Fig. 4A.
  • the structure for the audio block in an enhanced AC-3 bit stream as described in Annex E of the A/52B document is illustrated in Fig. 4B.
  • An audio block structure in an alternate version of the bit stream as described in Annex D of the A/52B document is not discussed here because its unique features are not pertinent to the present invention.
  • block metadata include flags and parameters for block switching (BLKSW), dynamic range compression (DYNRNG), channel coupling (CPL), channel rematrixing (REMAT), exponent coding technique or strategy (EXPSTR) used to encode the BFP exponents, the encoded BFP exponents (EXP), bit allocation (BA) information for the mantissas, adjustments to bit allocation known as delta bit allocation (DBA) information, and the quantized mantissas (MANT).
  • BLKSW block switching
  • DYNRNG dynamic range compression
  • CPL channel coupling
  • REMAT channel rematrixing
  • EXPSTR exponent coding technique or strategy
  • BA bit allocation
  • DBA delta bit allocation
  • MANT quantized mantissas
  • Each audio block in an enhanced AC-3 bit stream may contain information for additional coding techniques including spectral extension (SPX).
  • the ATSC Standards impose some constraints on the contents of the bit stream that are pertinent to the present invention. Two constraints are mentioned here: (1) the first audio block in the frame, which is referred to as ABO, must contain all of the information needed by the decoding algorithm to begin decoding all of the audio blocks in the frame, and
  • the ATSC Standards describe a number of bit stream syntactical features in terms of encoding processes or "coding tools" that may be used to generate an encoded bit stream.
  • An encoder need not employ all of the coding tools but a decoder that complies with the standard must be able to respond to the coding tools that are deemed essential for compliance. This response is implemented by performing an appropriate decoding tool that is essentially the inverse of the corresponding coding tool.
  • decoding tools are particularly relevant to the present invention because their use or lack of use affects how aspects of the present invention should be implemented.
  • a few decoding processes and a few decoding tools are discussed briefly in the following paragraphs. The following descriptions are not intended to be a complete description. Various details and optional features are omitted. The descriptions are intended only to provide a high-level introduction to those who are not familiar with the techniques and to refresh memories of those who may have forgotten which techniques these terms describe.
  • All decoders must unpack or demultiplex the encoded bit stream to obtain parameters and encoded data.
  • This process is represented by the deformatter 12 discussed above. This process is essentially one that reads data in the incoming bit stream and copies portions of the bit stream to registers, copies portions to memory locations, or stores pointers or other references to data in the bit stream that are stored in a buffer. Memory is required to store the data and pointers and a tradeoff can be made between storing this information for later use or re-reading the bit stream to obtain the information whenever it is needed.
  • the values of all BFP exponents are needed to unpack the data in the audio blocks for each frame because these values indirectly indicate the numbers of bits that are allocated to the quantized mantissas.
  • the exponent values in the bit stream are encoded, however, by differential coding techniques that may be applied across both time and frequency. As a result, the data representing the encoded exponents must be unpacked from the bit stream and decoded before they can be used for other decoding processes.
  • Each of the quantized BFP mantissas in the bit stream are represented by a varying number of bits that are a function of the BFP exponents and possibly other metadata contained in the bit stream.
  • the BFP exponents are input to a specified model, which calculates a bit allocation for each mantissa. If an audio block also contains delta bit allocation (DBA) information, this additional information is used to adjust the bit allocation calculated by the model.
  • DBA delta bit allocation
  • the quantized BFP mantissas constitute most of the data in an encoded bit stream.
  • the bit allocation is used both to determine the location of each mantissa in the bit stream for unpacking as well as to select the appropriate dequantization function to obtain the dequantized mantissas.
  • Some data in the bit stream can represent multiple mantissas by a single value. In this situation, an appropriate number of mantissas are derived from the single value. Mantissas that have an allocation equal to zero may be reproduced either with a value equal to zero or as a pseudo-random number.
  • the channel coupling coding technique allows an encoder to represent multiple audio channels with less data.
  • the technique combines spectral components from two or more selected channels, referred to as the coupled channels, to form a single channel of composite spectral components, referred to as the coupling channel.
  • the spectral components of the coupling channel are represented in BFP format.
  • a set of scale factors describing the energy difference between the coupling channel and each coupled channel, known as coupling coordinates, is derived for each of the coupled channels and included in the encoded bit stream. Coupling is used for only a specified portion of the bandwidth of each channel.
  • a decoder uses a decoding technique known as channel decoupling to derive an inexact replica of the BFP exponents and mantissas for each coupled channel from the spectral components of the coupling channel and the coupling coordinates. This is done by multiplying each coupled channel spectral component by the appropriate coupling coordinate. Additional details may be obtained from the A/52B document.
  • the channel rematrixing coding technique allows an encoder to represent two-channel signals with less data by using a matrix to convert two independent audio channels into sum and difference channels.
  • a decoder When rematrixing is used, as indicated by a flag in the bit stream, a decoder obtains values representing the two audio channels by applying an appropriate matrix to the sum and difference values. Additional details may be obtained from the A/52B document.
  • Annex E of the A/52B describes features of the enhanced AC-3 bit stream syntax that permits the use of additional coding tools. A few of these tools and related processes are described briefly below.
  • the adaptive hybrid transform (AHT) coding technique provides another tool in addition to block switching for adapting the temporal and spectral resolution of the analysis and synthesis filter banks in response to changing signal characteristics by applying two transforms in cascade. Additional information for AHT processing may be obtained from the A/52B document and U.S. patent 7,516,064 entitled “Adaptive Hybrid Transform for Signal Analysis and Synthesis” by Vinton et al., which issued April 7, 2009 and is incorporated herein by reference in its entirety.
  • Encoders employ a primary transform implemented by the MDCT analysis transform mentioned above in front of and in cascade with a secondary transform implemented by a Type-II Discrete Cosine Transform (DCT-II).
  • the MDCT is applied to overlapping blocks of audio signal samples to generate spectral coefficients representing spectral content of the audio signal.
  • the DCT-II may be switched in and out of the signal processing path as desired and, when switched in, is applied to non-overlapping blocks of the MDCT spectral coefficients representing the same frequency to generate hybrid transform coefficients.
  • the DCT-II is switched on when the input audio signal is deemed to be sufficiently stationary because its use significantly increases the effective spectral resolution of the analysis filter bank by decreasing its effective temporal resolution from 256 samples to 1536 samples.
  • Decoders employ an inverse primary transform implemented by the IMDCT synthesis filter bank mentioned above that follows and is in cascade with an inverse secondary transform implemented by a Type-II Inverse Discrete Cosine Transform (IDCT-II).
  • IDCT-II is switched in and out of the signal processing path in response to metadata provided by the encoder. When switched in, the IDCT-II is applied to non-overlapping blocks of hybrid transform coefficients to obtain inverse secondary transform coefficients.
  • the inverse secondary transform coefficients may be spectral coefficients for direct input into the IMDCT if no other coding tool like channel coupling or SPX was used.
  • the MDCT spectral coefficients may be derived from the inverse secondary transform coefficients if coding tools like channel coupling or SPX were used. After the MDCT spectral coefficients are obtained, the IMDCT is applied to blocks of the MDCT spectral coefficients in a conventional manner.
  • the AHT may be used with any audio channel including the coupling channel and the
  • a channel that is encoded using the AHT uses an alternative bit allocation process and two different types of quantization.
  • One type is vector quantization (VQ) and the second type is gain-adaptive quantization (GAQ).
  • VQ vector quantization
  • GAQ gain-adaptive quantization
  • the GAQ technique is discussed in U.S. patent 6,246,345 entitled “Using Gain- Adaptive Quantization and Non-Uniform Symbol Lengths for Improved Audio Coding" by Davidson et al., which issued June 12, 2001 and is incorporated herein by reference in its entirety.
  • the AHT requires a decoder to derive several parameters from information contained in the encoded bit stream.
  • the A/52B document describes how these parameters can be calculated.
  • One set of parameters specify the number of times BFP exponents are carried in a frame and are derived by examining metadata contained in all audio blocks in a frame.
  • Two other sets of parameters identify which BFP mantissas are quantized using GAQ and provide gain-control words for the quantizers and are derived by examining metadata for a channel in an audio block.
  • All of the hybrid transform coefficients for AHT are carried in the first audio block, ABO, of a frame. If the AHT is applied to a coupling channel, the coupling coordinates for the AHT coefficients are distributed across all of the audio blocks in the same manner as for coupled channels without AHT. A process to handle this situation is described below.
  • the spectral extension (SPX) coding technique allows an encoder to reduce the amount of information needed to encode a full-bandwidth channel by excluding high- frequency spectral components from the encoded bit stream and having the decoder synthesize the missing spectral components from lower-frequency spectral components that are contained in the encoded bit stream.
  • the decoder synthesizes missing spectral components by copying lower-frequency MDCT coefficients into higher-frequency MDCT coefficient locations, adding pseudo-random values or noise to the copied transform coefficients, and scaling the amplitude according to a SPX spectral envelope included in the encoded bit stream.
  • the encoder calculates the SPX spectral envelope and inserts it into the encoded bit stream whenever the SPX coding tool is used.
  • the SPX technique is used typically to synthesize the highest bands of spectral components for a channel. It may be used together with channel coupling for a middle range of frequencies. Additional details of processing may be obtained from the A/52B document.
  • the enhanced AC-3 bit stream syntax allows an encoder to generate an encoded bit stream that represents a single program with more than 5.1 channels (channel extension), two or more programs with up to 5.1 channels (program extension), or a combination of programs with up to 5.1 channels and more than 5.1 channels.
  • Program extension is implemented by a multiplex of frames for multiple independent data streams in an encoded bit stream.
  • Channel extension is implemented by a multiplex of frames for one or more dependent data substreams that are associated with an independent data stream.
  • a decoder is informed which program or programs to decode and the decoding process skips over or essentially ignores the streams and substreams representing programs that are not to be decoded.
  • Figs. 5 A to 5C illustrate three examples of bit streams carrying data with program and channel extensions.
  • Fig. 5 A illustrates an exemplary bit stream with channel extension.
  • a single program PI is represented by an independent stream SO and three associated dependent substreams SS0, SSI and SS2.
  • a frame Fn for the independent stream SO is followed immediately by frames Fn for each of the associated dependent substreams SS0 to SS3.
  • Fig. 5B illustrates an exemplary bit stream with program extension.
  • Each of four programs PI, P2, P3 and P4 are represented by independent streams SO, SI, S2 and S3, respectively.
  • a frame Fn for independent stream SO is followed immediately by frames Fn for each of independent streams SI, S2 and S3. These frames are followed by the next frame Fn+1 for each of the independent streams.
  • the enhanced AC-3 bit stream syntax must have at least one independent stream and permits as many as eight independent streams.
  • Fig. 5C illustrates an exemplary bit stream with program extension and channel extension.
  • Program PI is represented by data in independent stream SO
  • program P2 is represented by data in independent stream SI and associated dependent substreams SSO and SSI.
  • a frame Fn for independent stream SO is followed immediately by frame Fn for independent stream SI, which in turn is followed immediately by frames Fn for the associated dependent substreams SSO and SSI. These frames are followed by the next frame Fn+1 for each of the independent streams and dependent substreams.
  • An independent stream without channel extension contains data that may represent up to 5.1 independent audio channels.
  • An independent stream with channel extension or, in other words, an independent stream that has one or more associated dependent substreams, contains data that represents a 5.1 channel downmix of all channels for the program.
  • downmix refers to a combination of channels into a fewer number of channels. This is done for compatibility with decoders that do not decode the dependent substreams.
  • the dependent substreams contain data representing channels that either replace or supplement the channels carried in the associated independent stream.
  • Channel extension permits as many as fourteen channels for a program.
  • bit stream syntax and associate processing may be obtained from the A/52B document.
  • the component 19 parses frames from an encoded bit stream received from the path 1 and extracts data from the frames in response to control signals received from the path 20. The parsing is accomplished by multiple passes over the frame data.
  • the extracted data from one frame is represented by the boxes below the component 19. For example, the box with the label AB0-CH0 represents extracted data for channel 0 in audio block ABO and the box with the label AB5-CH2 represents extracted data for channel 2 in audio block AB5. Only three channels 0 to 2 and three audio blocks 0, 1 and 5 are illustrated to simplify the drawing.
  • the component 19 also passes parameters obtained from frame metadata along the path 20 to the channel processing components 31, 32 and 33.
  • the signal paths and rotary switches to the left of the data boxes represent the logic performed by traditional decoders to process encoded audio data in order by channel.
  • the process channel component 31 receives encoded audio data and metadata through the rotary switch 21 for channel CH0, starting with audio block ABO and concluding with audio block AB5, decodes the data and generates an output signal by applying a synthesis filter bank to the decoded data. The results of its processing is passed along the path
  • the process channel component 32 receives data for channel CHI for audio blocks ABO to AB5 through the rotary switch 22, processes the data and passes its output along the path
  • the process channel component 33 receives data for channel CH2 for audio blocks ABO to AB5 through the rotary switch 23, processes the data and passes its output along the path
  • Fig. 7 parses frames from an encoded bit stream received from the path 1 and extracts data from the frames in response to control signals received from the path 20. In many situations, the parsing is accomplished by a single pass over the frame data.
  • the extracted data from one frame is represented by the boxes below the component 19 in the same manner discussed above for Fig. 6.
  • the component 19 passes parameters obtained from frame metadata along the path 20 to the block processing components 61, 62 and 63.
  • the process block component 61 receives encoded audio data and metadata through the rotary switch 51 for all of the channels in audio block ABO, decodes the data and generates an output signal by applying a synthesis filter bank to the decoded data.
  • the results of its processing for channels CH0, CHI and CH2 are passed through the rotary switch 71 to the appropriate output path 41, 42 and 43, respectively.
  • the process block component 62 receives data for all channels in audio block AB 1 through the rotary switch 52, processes the data and passes its output through the rotary switch 72 to the appropriate output path for each channel.
  • the process block component 63 receives data for all channels in audio block AB5 through the rotary switch 53, processes the data and passes its output through the rotary switch 73 to the appropriate output path for each channel.
  • program fragments are not intended to be practical or optimal implementations but only illustrative examples.
  • order of program statements may be altered by interchanging some of the statements.
  • Statements (1.2) and (1.19) control the decoding process to be performed for each frame in the bit stream, or until the decoding process is stopped by some other means.
  • Statements (1.3) to (1.18) perform processes that decode a frame in the encoded bit stream.
  • Statements (1.3) to (1.5) unpack metadata in the frame, obtain decoding parameters from the unpacked metadata, and determine the location in the bit stream where data begins for the first audio block K in the frame.
  • Statement (1.16) determines the start of the next audio block in the bit stream if any subsequent audio block is in the frame.
  • Statements (1.6) and (1.17) cause the decoding process to be performed for each audio block in the frame.
  • Statements (1.7) to (1.15) perform processes that decode an audio block in the frame.
  • Statements (1.7) to (1.9) unpack metadata in the audio block, obtain decoding parameters from the unpacked metadata, and determine where data begins for the first channel.
  • Statements (1.10) and (1.15) cause the decoding process to be performed for each channel in the audio block.
  • Statements (1.11) to (1.13) unpack and decode exponents, use the decoded exponents to determine the bit allocation to unpack and dequantize each quantized mantissa, and apply the synthesis filter bank to the dequantized mantissas.
  • Statement (1.14) determines the location in the bit stream where data starts for the next channel, if any subsequent channel is in the frame.
  • the audio block in which the extension process begins contains shared parameters needed for SPX in the beginning audio block as well as other audio blocks using SPX in the frame.
  • the shared parameters include an identification of the channels participating in the process, the spectral extension frequency range, and how the SPX spectral envelope for each channel is shared across time and frequency. These parameters are unpacked from the audio block that begins the use of SPX and stored in memory or in computer registers for use in processing SPX in subsequent audio blocks in the frame.
  • a audio block begins SPX if the metadata for that audio block indicates SPX is used and either the metadata for the preceding audio block in the frame indicates SPX is not used or the audio block is the first block in a frame.
  • Each audio block that uses SPX either includes the SPX spectral envelope, referred to as SPX coordinates, that are used for spectral extension processing in that audio block or it includes a "reuse" flag that indicates the SPX coordinates for a previous block are to be used.
  • the SPX coordinates in a block are unpacked and retained for possible reuse by SPX operations in subsequent audio blocks.
  • the following program fragment illustrates one way audio blocks using SPX may be processed.
  • Statement (2.5) unpacks SPX frame parameters from the frame metadata if any are present in that metadata.
  • Statement (2.10) unpacks SPX block parameters from the block metadata if any are present in the block metadata.
  • the block SPX parameters may include SPX coordinates for one or more channels in the block.
  • Statements (2.12) and (2.13) unpack and decode exponents and use the decoded exponents to determine the bit allocation to unpack and dequantize each quantized mantissa.
  • Statement (2.14) determines whether channel C in the current audio block uses SPX. If it does use SPX, statement (2.15) applies SPX processing to extend the bandwidth of the channel C. This process provides the spectral components for channel C that are input to the synthesis filter bank applied in statement (2.17).
  • the first audio block ABO in a frame contains all hybrid transform coefficients for each channel processed by the DCT-II transform.
  • each of the six audio blocks in the frame contains as many as 256 spectral coefficients generated by the MDCT analysis filter bank.
  • an encoded bit stream contains data for the left, center and right channels.
  • audio block ABO contains all of the hybrid transform coefficients for each of the left and right channels and contains as many as 256 MDCT spectral coefficients for the center channel.
  • Audio blocks AB1 through AB5 contain MDCT spectral coefficients for the center channel and no coefficients for the left and right channels.
  • the following program fragment illustrates one way audio blocks with AHT coefficients may be processed.
  • Statement (3.11) determines whether the AHT is in use for the channel C. If it is in use, statement (3.12) determines whether the first audio block ABO is being processed. If the first audio block is being processed, then statements (3.13) to (3.16) obtain all AHT coefficients for the channel C, apply the inverse secondary transform or IDCT-II to the AHT coefficients to obtain the MDCT spectral coefficients, and store them in a buffer. These spectral coefficients correspond to the exponents and dequantized mantissas that are obtained by statements (3.20) and (3.21) for channels for which AHT is not in use. Statement (3.18) obtains the exponents and mantissas of the MDCT spectral coefficients that correspond to the audio block K that is being processed.
  • SPX and the AHT may be used to generate encoded data for the same channels.
  • the logic discussed above separately for spectral extension and hybrid transform processing may be combined to process channels for which SPX is in use, the AHT is in use, or both SPX and the AHT are in use.
  • the following program fragment illustrates one way audio blocks with SPX and AHT coefficients may be processed.
  • Statement (4.5) unpacks SPX frame parameters from the frame metadata if any are present in that metadata.
  • Statement (4.10) unpacks SPX block parameters from the block metadata if any are present in the block metadata.
  • the block SPX parameters may include SPX coordinates for one or more channels in the block.
  • Statement (4.12) determines whether the AHT is in use for channel C. If the AHT is in use for channel C, statement (4.13) determines whether this is the first audio block. If it is the first audio block, statements (4.14) through (4.17) obtain all AHT coefficients for the channel C, apply the inverse secondary transform or IDCT-II to the AHT coefficients to obtain inverse secondary transform coefficients, and store them in a buffer. Statement (4.19) obtains the exponents and mantissas of the inverse secondary transform coefficients that correspond to the audio block K that is being processed. If the AHT is not in use for channel C, statements (4.21) and (4.22) unpack and obtain the exponents and mantissas for the channel C in block K as discussed above for program statements (1.11) and (1.12).
  • Statement (4.24) determines whether channel C in the current audio block uses SPX. If it does use SPX, statement (4.25) applies SPX processing to the inverse secondary transform coefficients to extend the bandwidth, thereby obtaining the MDCT spectral coefficients of the channel C. This process provides the spectral components for channel C that are input to the synthesis filter bank applied in statement (4.27). If SPX processing is not used for channel C, the MDCT spectral coefficients are obtained directly from the inverse secondary transform coefficients.
  • Channel coupling and the AHT may be used to generate encoded data for the same channels.
  • Essentially the same logic discussed above for spectral extension and hybrid transform processing may be used to process bit streams using channel coupling and the AHT because the details of SPX processing discussed above apply to the processing performed for channel coupling.
  • the following program fragment illustrates one way audio blocks with coupling and AHT coefficients may be processed.
  • Statement (5.5) unpacks channel coupling parameters from the frame metadata if any are present in that metadata.
  • Statement (5.10) unpacks channel coupling parameters from the block metadata if any are present in the block metadata. If they are present, coupling coordinates are obtained for the coupled channels in the block.
  • Statement (5.12) determines whether the AHT is in use for channel C. If the AHT is in use, statement (5.13) determines whether it is the first audio block. If it is the first audio block, statements (5.14) through (5.17) obtain all AHT coefficients for the channel C, apply the inverse secondary transform or IDCT-II to the AHT coefficients to obtain inverse secondary transform coefficients, and store them in a buffer. Statement (5.19) obtains the exponents and mantissas of the inverse secondary transform coefficients that correspond to the audio block K that is being processed.
  • statements (5.21) and (5.22) unpack and obtain the exponents and mantissas for the channel C in block K as discussed above for program statements (1.11) and (1.12).
  • Statement (5.24) determines whether channel coupling is in use for channel C. If it is in use, statement (5.25) determines whether channel C is the first channel in the block to use coupling. If it is, the exponents and mantissas for the coupling channel are obtained either from an application of an inverse secondary transform to the coupling channel exponents and mantissas as shown in statements (5.26) through (5.33) or from data in the bit stream as shown in statements (5.35) and (5.36). The data representing the coupling channel mantissas are placed in the bit stream immediately after the data representing mantissas of the channel C. Statement (5.39) derives the coupled channel C from the coupling channel using the appropriate coupling coordinates for the channel C. If channel coupling is not used for channel C, the MDCT spectral coefficients are obtained directly from the inverse secondary transform coefficients.
  • Spectral extension, channel coupling and the AHT may all be used to generate encoded data for the same channels.
  • the logic discussed above for combinations of AHT processing with spectral extension and coupling may be combined to process channels using any combination of the three coding tools by incorporating the additional logic necessary to handle eight possible situations.
  • the processing for channel decoupling is performed before performing SPX processing.
  • FIG. 8 is a schematic block diagram of a device 90 that may be used to implement aspects of the present invention.
  • the processor 92 provides computing resources.
  • RAM 93 is system random access memory (RAM) used by the processor 92 for processing.
  • ROM 94 represents some form of persistent storage such as read only memory (ROM) for storing programs needed to operate the device 90 and possibly for carrying out various aspects of the present invention.
  • I/O control 95 represents interface circuitry to receive and transmit signals by way of the communication channels 1, 16. In the embodiment shown, all major system components connect to the bus 91, which may represent more than one physical or logical bus; however, a bus architecture is not required to implement the present invention.
  • additional components may be included for interfacing to devices such as a keyboard or mouse and a display, and for controlling a storage device having a storage medium such as magnetic tape or disk, or an optical medium.
  • the storage medium may be used to record programs of instructions for operating systems, utilities and applications, and may include programs that implement various aspects of the present invention.
  • Software implementations of the present invention may be conveyed by a variety of machine readable media such as baseband or modulated communication paths throughout the spectrum including from supersonic to ultraviolet frequencies, or storage media that convey information using essentially any recording technology including magnetic tape, cards or disk, optical cards or disc, and detectable markings on media including paper.
  • machine readable media such as baseband or modulated communication paths throughout the spectrum including from supersonic to ultraviolet frequencies, or storage media that convey information using essentially any recording technology including magnetic tape, cards or disk, optical cards or disc, and detectable markings on media including paper.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Spectroscopy & Molecular Physics (AREA)
  • Mathematical Physics (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)
  • Transmission Systems Not Characterized By The Medium Used For Transmission (AREA)
  • Analogue/Digital Conversion (AREA)
PCT/US2010/054480 2009-12-07 2010-10-28 Decoding of multichannel aufio encoded bit streams using adaptive hybrid transformation WO2011071610A1 (en)

Priority Applications (27)

Application Number Priority Date Filing Date Title
BR112012013745-0A BR112012013745B1 (pt) 2009-12-07 2010-10-28 Método para decodificar um quadro de um sinal de áudio digital codificado, aparelho para decodificar um quadro de um sinal de áudio digital codificado e meio de armazenamento que grava o método
KR1020127012464A KR101370522B1 (ko) 2009-12-07 2010-10-28 적응형 하이브리드 변환을 사용한 다채널 오디오 엔코딩된 비트 스트림들의 디코딩
SI201030604T SI2510515T1 (sl) 2009-12-07 2010-10-28 Dekodiranje večkanalnih avdio kodiranih bitnih prenosov s pomočjo adaptivne hibridne transformacije
PL10776017T PL2510515T3 (pl) 2009-12-07 2010-10-28 Dekodowanie wielokanałowych zakodowanych strumieni bitów dźwięku stosując adaptacyjną transformatę hybrydową
ES10776017.5T ES2463840T3 (es) 2009-12-07 2010-10-28 Descodificación de trenes de bits codificados de audio multicanal usando transformación híbrida adaptativa
NZ599981A NZ599981A (en) 2009-12-07 2010-10-28 Decoding of multichannel audio encoded bit streams using adaptive hybrid transformation
UAA201208193A UA100353C2 (uk) 2009-12-07 2010-10-28 Декодування цифрових потоків кодованого багатоканального аудіосигналу з використанням адаптивного гібридного перетворення
RS20140226A RS53288B (en) 2009-12-07 2010-10-28 DECODE CODED BIT FLOWS FROM MULTIPLE AUDIO CHANNELS BY USING ADAPTIVE HYBRID TRANSFORMATIONS
DK10776017.5T DK2510515T3 (da) 2009-12-07 2010-10-28 Dekodning af flerkanalaudio-kodede bitstrømme ved anvendelse af adaptiv hybrid transformation
KR1020137026329A KR101629306B1 (ko) 2009-12-07 2010-10-28 적응형 하이브리드 변환을 사용한 다채널 오디오 엔코딩된 비트 스트림들의 디코딩
CN201410410643.3A CN104217724B (zh) 2009-12-07 2010-10-28 使用自适应混合变换的多声道音频编码比特流的解码
CN201080051553.XA CN102687198B (zh) 2009-12-07 2010-10-28 使用自适应混合变换的多声道音频编码比特流的解码
AP2012006289A AP3301A (en) 2009-12-07 2010-10-28 Decoding of multichannel audio encoded bit streamsusing adaptive hybrid transformation
JP2012541085A JP5547297B2 (ja) 2009-12-07 2010-10-28 適応型ハイブリッド変換を使用した、マルチチャンネルオーディオエンコード化ビットストリームのデコーディング
AU2010328635A AU2010328635B2 (en) 2009-12-07 2010-10-28 Decoding of multichannel aufio encoded bit streams using adaptive hybrid transformation
CA2779453A CA2779453C (en) 2009-12-07 2010-10-28 Decoding of multichannel audio encoded bit streams using adaptive hybrid transformation
EA201270642A EA024310B1 (ru) 2009-12-07 2010-10-28 Способ декодирования цифровых потоков кодированного многоканального аудиосигнала с использованием адаптивного гибридного преобразования
MX2012005723A MX2012005723A (es) 2009-12-07 2010-10-28 Metodo y aparato para decodificar un cuadrado de una señal de audio digital codificada y medio de almacenamiento que graba un programa de instrucciones.
US13/514,286 US8891776B2 (en) 2009-12-07 2010-10-28 Decoding of multichannel audio encoded bit streams using adaptive hybrid transformation
EP10776017.5A EP2510515B1 (en) 2009-12-07 2010-10-28 Decoding of multichannel audio encoded bit streams using adaptive hybrid transformation
IL219304A IL219304A (en) 2009-12-07 2012-04-19 A method and device for decoding a digital audio signal segment encrypted by an adaptive hybrid transformation
ZA2012/03290A ZA201203290B (en) 2009-12-07 2012-05-07 Decoding of multichannel audio encoded bit streams using adaptive hybrid transformation
TNP2012000211A TN2012000211A1 (en) 2009-12-07 2012-05-09 Decoding of multichannel aufio encoded bit streams using adaptive hybrid transformation
MA34912A MA33775B1 (fr) 2009-12-07 2012-05-31 Décodage de flux binaires audio codés utilisant une transformation hybride adaptative
HK12110658.8A HK1170058A1 (en) 2009-12-07 2012-10-25 Decoding of multichannel audio encoded bit streams using adaptive hybrid transformation
HRP20140400AT HRP20140400T1 (hr) 2009-12-07 2014-05-05 Dekodiranje višekanalnog audio kodiranog toka bitova pomoä†u adaptivne hibridne transformacije
US14/512,755 US9620132B2 (en) 2009-12-07 2014-10-13 Decoding of multichannel audio encoded bit streams using adaptive hybrid transformation

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US26742209P 2009-12-07 2009-12-07
US61/267,422 2009-12-07

Related Child Applications (2)

Application Number Title Priority Date Filing Date
US13/514,286 A-371-Of-International US8891776B2 (en) 2009-12-07 2010-10-28 Decoding of multichannel audio encoded bit streams using adaptive hybrid transformation
US14/512,755 Continuation US9620132B2 (en) 2009-12-07 2014-10-13 Decoding of multichannel audio encoded bit streams using adaptive hybrid transformation

Publications (1)

Publication Number Publication Date
WO2011071610A1 true WO2011071610A1 (en) 2011-06-16

Family

ID=43334376

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2010/054480 WO2011071610A1 (en) 2009-12-07 2010-10-28 Decoding of multichannel aufio encoded bit streams using adaptive hybrid transformation

Country Status (37)

Country Link
US (2) US8891776B2 (pt)
EP (3) EP2801975B1 (pt)
JP (2) JP5547297B2 (pt)
KR (2) KR101370522B1 (pt)
CN (2) CN102687198B (pt)
AP (1) AP3301A (pt)
AR (1) AR079878A1 (pt)
AU (1) AU2010328635B2 (pt)
BR (1) BR112012013745B1 (pt)
CA (1) CA2779453C (pt)
CL (1) CL2012001493A1 (pt)
CO (1) CO6460719A2 (pt)
DK (1) DK2510515T3 (pt)
EA (1) EA024310B1 (pt)
EC (1) ECSP12012006A (pt)
ES (1) ES2463840T3 (pt)
GE (1) GEP20146081B (pt)
GT (1) GT201200134A (pt)
HK (1) HK1170058A1 (pt)
HN (1) HN2012000819A (pt)
HR (1) HRP20140400T1 (pt)
IL (1) IL219304A (pt)
MA (1) MA33775B1 (pt)
MX (1) MX2012005723A (pt)
MY (1) MY161012A (pt)
NI (1) NI201200063A (pt)
NZ (1) NZ599981A (pt)
PE (1) PE20130167A1 (pt)
PL (1) PL2510515T3 (pt)
PT (1) PT2510515E (pt)
RS (1) RS53288B (pt)
SI (1) SI2510515T1 (pt)
TN (1) TN2012000211A1 (pt)
TW (1) TWI498881B (pt)
UA (1) UA100353C2 (pt)
WO (1) WO2011071610A1 (pt)
ZA (1) ZA201203290B (pt)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2015504539A (ja) * 2011-11-30 2015-02-12 ドルビー・インターナショナル・アーベー オーディオ・コーデックからの向上したクロマ抽出
JP2015508186A (ja) * 2012-02-23 2015-03-16 ドルビー・インターナショナル・アーベー 高周波オーディオ内容の効率的な復元のための方法およびシステム
US20160005405A1 (en) * 2013-02-14 2016-01-07 Dolby Laboratories Licensing Corporation Methods for Audio Signal Transient Detection and Decorrelation Control
US9460729B2 (en) 2012-09-21 2016-10-04 Dolby Laboratories Licensing Corporation Layered approach to spatial audio coding

Families Citing this family (36)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7711123B2 (en) * 2001-04-13 2010-05-04 Dolby Laboratories Licensing Corporation Segmenting audio signals into auditory events
US8948406B2 (en) * 2010-08-06 2015-02-03 Samsung Electronics Co., Ltd. Signal processing method, encoding apparatus using the signal processing method, decoding apparatus using the signal processing method, and information storage medium
US20120033819A1 (en) * 2010-08-06 2012-02-09 Samsung Electronics Co., Ltd. Signal processing method, encoding apparatus therefor, decoding apparatus therefor, and information storage medium
US9130596B2 (en) * 2011-06-29 2015-09-08 Seagate Technology Llc Multiuse data channel
TWI618050B (zh) 2013-02-14 2018-03-11 杜比實驗室特許公司 用於音訊處理系統中之訊號去相關的方法及設備
TWI618051B (zh) * 2013-02-14 2018-03-11 杜比實驗室特許公司 用於利用估計之空間參數的音頻訊號增強的音頻訊號處理方法及裝置
EP2956935B1 (en) 2013-02-14 2017-01-04 Dolby Laboratories Licensing Corporation Controlling the inter-channel coherence of upmixed audio signals
US8804971B1 (en) * 2013-04-30 2014-08-12 Dolby International Ab Hybrid encoding of higher frequency and downmixed low frequency content of multichannel audio
CA3211308A1 (en) 2013-05-24 2014-11-27 Dolby International Ab Coding of audio scenes
US10499176B2 (en) 2013-05-29 2019-12-03 Qualcomm Incorporated Identifying codebooks to use when coding spatial components of a sound field
US9466305B2 (en) 2013-05-29 2016-10-11 Qualcomm Incorporated Performing positional analysis to code spherical harmonic coefficients
TWM487509U (zh) 2013-06-19 2014-10-01 杜比實驗室特許公司 音訊處理設備及電子裝置
US9530422B2 (en) 2013-06-27 2016-12-27 Dolby Laboratories Licensing Corporation Bitstream syntax for spatial voice coding
EP2830061A1 (en) 2013-07-22 2015-01-28 Fraunhofer Gesellschaft zur Förderung der angewandten Forschung e.V. Apparatus and method for encoding and decoding an encoded audio signal using temporal noise/patch shaping
CN109903776B (zh) 2013-09-12 2024-03-01 杜比实验室特许公司 用于各种回放环境的动态范围控制
CN105659320B (zh) 2013-10-21 2019-07-12 杜比国际公司 音频编码器和解码器
US9489955B2 (en) 2014-01-30 2016-11-08 Qualcomm Incorporated Indicating frame parameter reusability for coding vectors
US9922656B2 (en) 2014-01-30 2018-03-20 Qualcomm Incorporated Transitioning of ambient higher-order ambisonic coefficients
US9620137B2 (en) 2014-05-16 2017-04-11 Qualcomm Incorporated Determining between scalar and vector quantization in higher order ambisonic coefficients
US10770087B2 (en) * 2014-05-16 2020-09-08 Qualcomm Incorporated Selecting codebooks for coding vectors decomposed from higher-order ambisonic audio signals
US9852737B2 (en) 2014-05-16 2017-12-26 Qualcomm Incorporated Coding vectors decomposed from higher-order ambisonics audio signals
CN105280212A (zh) * 2014-07-25 2016-01-27 中兴通讯股份有限公司 混音播放方法及装置
US9747910B2 (en) 2014-09-26 2017-08-29 Qualcomm Incorporated Switching between predictive and non-predictive quantization techniques in a higher order ambisonics (HOA) framework
TWI758146B (zh) 2015-03-13 2022-03-11 瑞典商杜比國際公司 解碼具有增強頻譜帶複製元資料在至少一填充元素中的音訊位元流
US9837086B2 (en) * 2015-07-31 2017-12-05 Apple Inc. Encoded audio extended metadata-based dynamic range control
US10504530B2 (en) 2015-11-03 2019-12-10 Dolby Laboratories Licensing Corporation Switching between transforms
EP3208800A1 (en) * 2016-02-17 2017-08-23 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Apparatus and method for stereo filing in multichannel coding
US10015612B2 (en) 2016-05-25 2018-07-03 Dolby Laboratories Licensing Corporation Measurement, verification and correction of time alignment of multiple audio channels and associated metadata
AU2018208522B2 (en) * 2017-01-10 2020-07-02 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Audio decoder, audio encoder, method for providing a decoded audio signal, method for providing an encoded audio signal, audio stream, audio stream provider and computer program using a stream identifier
US10354667B2 (en) 2017-03-22 2019-07-16 Immersion Networks, Inc. System and method for processing audio data
US10885921B2 (en) * 2017-07-07 2021-01-05 Qualcomm Incorporated Multi-stream audio coding
US10854209B2 (en) * 2017-10-03 2020-12-01 Qualcomm Incorporated Multi-stream audio coding
US10657974B2 (en) * 2017-12-21 2020-05-19 Qualcomm Incorporated Priority information for higher order ambisonic audio data
WO2020123424A1 (en) * 2018-12-13 2020-06-18 Dolby Laboratories Licensing Corporation Dual-ended media intelligence
WO2020207593A1 (en) * 2019-04-11 2020-10-15 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Audio decoder, apparatus for determining a set of values defining characteristics of a filter, methods for providing a decoded audio representation, methods for determining a set of values defining characteristics of a filter and computer program
CN111711493B (zh) * 2020-06-16 2022-03-11 中国电子科技集团公司第三研究所 具有加密解密能力的水下通信设备及发射器和接收器

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5583962A (en) 1991-01-08 1996-12-10 Dolby Laboratories Licensing Corporation Encoder/decoder for multidimensional sound fields
US6246345B1 (en) 1999-04-16 2001-06-12 Dolby Laboratories Licensing Corporation Using gain-adaptive quantization and non-uniform symbol lengths for improved audio coding
US7516064B2 (en) 2004-02-19 2009-04-07 Dolby Laboratories Licensing Corporation Adaptive hybrid transform for signal analysis and synthesis

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6356639B1 (en) * 1997-04-11 2002-03-12 Matsushita Electric Industrial Co., Ltd. Audio decoding apparatus, signal processing device, sound image localization device, sound image control method, audio signal processing device, and audio signal high-rate reproduction method used for audio visual equipment
JPH10340099A (ja) * 1997-04-11 1998-12-22 Matsushita Electric Ind Co Ltd オーディオデコーダ装置及び信号処理装置
US7292901B2 (en) 2002-06-24 2007-11-06 Agere Systems Inc. Hybrid multi-channel/cue coding/decoding of audio signals
US7502743B2 (en) * 2002-09-04 2009-03-10 Microsoft Corporation Multi-channel audio encoding and decoding with multi-channel transform selection
CN1261663C (zh) * 2002-12-31 2006-06-28 深圳市高科智能系统有限公司 无线集中控制门禁/门锁的方法及系统设备
US9454974B2 (en) * 2006-07-31 2016-09-27 Qualcomm Incorporated Systems, methods, and apparatus for gain factor limiting
US7953595B2 (en) * 2006-10-18 2011-05-31 Polycom, Inc. Dual-transform coding of audio signals
KR101325802B1 (ko) * 2007-02-06 2013-11-05 엘지전자 주식회사 디지털 방송 송신기 및 디지털 방송 수신기와 그를 이용한디지털 방송 시스템 및 그 서비스 방법
CN101067931B (zh) * 2007-05-10 2011-04-20 芯晟(北京)科技有限公司 一种高效可配置的频域参数立体声及多声道编解码方法与系统
MX2009013519A (es) * 2007-06-11 2010-01-18 Fraunhofer Ges Forschung Codificador de audio para codificar una señal de audio que tiene una porcion similar a un impulso y una porcion estacionaria, metodos de codificacion, decodificador, metodo de decodificacion, y señal de audio codificada.
CN101816191B (zh) * 2007-09-26 2014-09-17 弗劳恩霍夫应用研究促进协会 用于提取环境信号的装置和方法
WO2009059633A1 (en) * 2007-11-06 2009-05-14 Nokia Corporation An encoder
EP2107556A1 (en) * 2008-04-04 2009-10-07 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Audio transform coding using pitch correction

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5583962A (en) 1991-01-08 1996-12-10 Dolby Laboratories Licensing Corporation Encoder/decoder for multidimensional sound fields
US6246345B1 (en) 1999-04-16 2001-06-12 Dolby Laboratories Licensing Corporation Using gain-adaptive quantization and non-uniform symbol lengths for improved audio coding
US7516064B2 (en) 2004-02-19 2009-04-07 Dolby Laboratories Licensing Corporation Adaptive hybrid transform for signal analysis and synthesis

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
ANDERSEN ROBERT L ET AL: "Introduction to Dolby Digital Plus, an Enhancement to the Dolby Digital Coding System", AES CONVENTION 117; OCTOBER 2004, AES, 60 EAST 42ND STREET, ROOM 2520 NEW YORK 10165-2520, USA, 1 October 2004 (2004-10-01), XP040506945 *
ANONYMOUS: "A/52B, ATSC standard, Digital audio compression standard (AC-3, E-AC-3), revision B", NOT KNOWN,, 14 June 2005 (2005-06-14), XP030001573 *
SAKAMOTO H ET AL: "A Dolby AC-3/MPEG1 audio decoder core suitable for audio/visual system integration", CUSTOM INTEGRATED CIRCUITS CONFERENCE, 1997., PROCEEDINGS OF THE IEEE 1997 SANTA CLARA, CA, USA 5-8 MAY 1997, NEW YORK, NY, USA,IEEE, US, 5 May 1997 (1997-05-05), pages 241 - 244, XP010235299, ISBN: 978-0-7803-3669-8, DOI: DOI:10.1109/CICC.1997.606621 *

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2015504539A (ja) * 2011-11-30 2015-02-12 ドルビー・インターナショナル・アーベー オーディオ・コーデックからの向上したクロマ抽出
US9697840B2 (en) 2011-11-30 2017-07-04 Dolby International Ab Enhanced chroma extraction from an audio codec
JP2015508186A (ja) * 2012-02-23 2015-03-16 ドルビー・インターナショナル・アーベー 高周波オーディオ内容の効率的な復元のための方法およびシステム
US9666200B2 (en) 2012-02-23 2017-05-30 Dolby International Ab Methods and systems for efficient recovery of high frequency audio content
US9984695B2 (en) 2012-02-23 2018-05-29 Dolby International Ab Methods and systems for efficient recovery of high frequency audio content
US9460729B2 (en) 2012-09-21 2016-10-04 Dolby Laboratories Licensing Corporation Layered approach to spatial audio coding
US9495970B2 (en) 2012-09-21 2016-11-15 Dolby Laboratories Licensing Corporation Audio coding with gain profile extraction and transmission for speech enhancement at the decoder
US9502046B2 (en) 2012-09-21 2016-11-22 Dolby Laboratories Licensing Corporation Coding of a sound field signal
US9858936B2 (en) 2012-09-21 2018-01-02 Dolby Laboratories Licensing Corporation Methods and systems for selecting layers of encoded audio signals for teleconferencing
US20160005405A1 (en) * 2013-02-14 2016-01-07 Dolby Laboratories Licensing Corporation Methods for Audio Signal Transient Detection and Decorrelation Control
US9830917B2 (en) * 2013-02-14 2017-11-28 Dolby Laboratories Licensing Corporation Methods for audio signal transient detection and decorrelation control

Also Published As

Publication number Publication date
EA201270642A1 (ru) 2012-12-28
JP2014063187A (ja) 2014-04-10
EP2510515A1 (en) 2012-10-17
AU2010328635B2 (en) 2014-02-13
MY161012A (en) 2017-03-31
EP2706529A2 (en) 2014-03-12
US8891776B2 (en) 2014-11-18
HK1170058A1 (en) 2013-02-15
JP5607809B2 (ja) 2014-10-15
EP2706529A3 (en) 2014-04-02
CN102687198B (zh) 2014-09-24
US9620132B2 (en) 2017-04-11
SI2510515T1 (sl) 2014-06-30
TW201126511A (en) 2011-08-01
CA2779453A1 (en) 2011-06-16
DK2510515T3 (da) 2014-05-19
ZA201203290B (en) 2013-07-31
CN102687198A (zh) 2012-09-19
JP5547297B2 (ja) 2014-07-09
CA2779453C (en) 2015-12-22
AP3301A (en) 2015-06-30
AU2010328635A1 (en) 2012-05-17
GT201200134A (es) 2013-08-29
US20150030161A1 (en) 2015-01-29
CL2012001493A1 (es) 2012-10-19
MX2012005723A (es) 2012-06-13
AR079878A1 (es) 2012-02-29
EP2801975A1 (en) 2014-11-12
HRP20140400T1 (hr) 2014-06-06
KR20120074305A (ko) 2012-07-05
ES2463840T3 (es) 2014-05-29
EP2510515B1 (en) 2014-03-19
MA33775B1 (fr) 2012-11-01
PT2510515E (pt) 2014-05-23
BR112012013745A2 (pt) 2016-03-15
AP2012006289A0 (en) 2012-06-30
JP2013511754A (ja) 2013-04-04
ECSP12012006A (es) 2012-08-31
KR101370522B1 (ko) 2014-03-06
NI201200063A (es) 2013-06-13
IL219304A (en) 2015-05-31
BR112012013745B1 (pt) 2020-10-27
CN104217724A (zh) 2014-12-17
PE20130167A1 (es) 2013-02-16
US20120243692A1 (en) 2012-09-27
EA024310B1 (ru) 2016-09-30
UA100353C2 (uk) 2012-12-10
TN2012000211A1 (en) 2013-12-12
RS53288B (en) 2014-08-29
GEP20146081B (en) 2014-04-25
CN104217724B (zh) 2017-04-05
EP2801975B1 (en) 2017-01-04
PL2510515T3 (pl) 2014-07-31
CO6460719A2 (es) 2012-06-15
IL219304A0 (en) 2012-06-28
KR20130116959A (ko) 2013-10-24
HN2012000819A (es) 2015-03-16
NZ599981A (en) 2014-07-25
KR101629306B1 (ko) 2016-06-10
TWI498881B (zh) 2015-09-01

Similar Documents

Publication Publication Date Title
US9620132B2 (en) Decoding of multichannel audio encoded bit streams using adaptive hybrid transformation
Vernon Design and implementation of AC-3 coders
AU2004211163B2 (en) Conversion of spectral components for encoding and low-complexity transcoding
JP4925671B2 (ja) デジタル信号の符号化/復号化方法及びその装置並びに記録媒体
TWI521502B (zh) 多聲道音訊的較高頻率和降混低頻率內容的混合編碼
JP4685165B2 (ja) 仮想音源位置情報に基づいたチャネル間レベル差量子化及び逆量子化方法
DK2691951T3 (en) TRANSFORMATION WITH REDUCED COMPLEXITY OF AN Low-Frequency
Domazetovic et al. Approach to Implementation of the AC-3 Audio Decoder on the MAS-like DSP Platform
AU2012238001A1 (en) Reduced complexity transform for a low-frequency-effects channel

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 201080051553.X

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 10776017

Country of ref document: EP

Kind code of ref document: A1

DPE1 Request for preliminary examination filed after expiration of 19th month from priority date (pct application filed from 20040101)
WWE Wipo information: entry into national phase

Ref document number: 219304

Country of ref document: IL

WWE Wipo information: entry into national phase

Ref document number: 2010328635

Country of ref document: AU

WWE Wipo information: entry into national phase

Ref document number: 000561-2012

Country of ref document: PE

WWE Wipo information: entry into national phase

Ref document number: 2779453

Country of ref document: CA

WWE Wipo information: entry into national phase

Ref document number: 4025/CHENP/2012

Country of ref document: IN

Ref document number: 12012500906

Country of ref document: PH

WWE Wipo information: entry into national phase

Ref document number: 12078993

Country of ref document: CO

ENP Entry into the national phase

Ref document number: 20127012464

Country of ref document: KR

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 2010328635

Country of ref document: AU

Date of ref document: 20101028

Kind code of ref document: A

WWE Wipo information: entry into national phase

Ref document number: MX/A/2012/005723

Country of ref document: MX

WWE Wipo information: entry into national phase

Ref document number: 2012541085

Country of ref document: JP

WWE Wipo information: entry into national phase

Ref document number: 13514286

Country of ref document: US

WWE Wipo information: entry into national phase

Ref document number: 2012001493

Country of ref document: CL

Ref document number: 1201002125

Country of ref document: TH

Ref document number: 2010776017

Country of ref document: EP

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 12753

Country of ref document: GE

WWE Wipo information: entry into national phase

Ref document number: 201270642

Country of ref document: EA

WWE Wipo information: entry into national phase

Ref document number: A201208193

Country of ref document: UA

REG Reference to national code

Ref country code: BR

Ref legal event code: B01A

Ref document number: 112012013745

Country of ref document: BR

WWE Wipo information: entry into national phase

Ref document number: P-2014/0226

Country of ref document: RS

ENP Entry into the national phase

Ref document number: 112012013745

Country of ref document: BR

Kind code of ref document: A2

Effective date: 20120606