KR20090004778A - Method for processing an audio signal and apparatus for implementing the same - Google Patents
Method for processing an audio signal and apparatus for implementing the same Download PDFInfo
- Publication number
- KR20090004778A KR20090004778A KR1020080065478A KR20080065478A KR20090004778A KR 20090004778 A KR20090004778 A KR 20090004778A KR 1020080065478 A KR1020080065478 A KR 1020080065478A KR 20080065478 A KR20080065478 A KR 20080065478A KR 20090004778 A KR20090004778 A KR 20090004778A
- Authority
- KR
- South Korea
- Prior art keywords
- information
- signal
- decoding
- core
- stereo
- Prior art date
Links
- 230000005236 sound signal Effects 0.000 title claims abstract description 82
- 238000000034 method Methods 0.000 title claims description 54
- 238000003672 processing method Methods 0.000 claims abstract description 17
- 230000010076 replication Effects 0.000 claims description 9
- 230000003595 spectral effect Effects 0.000 claims description 9
- 239000000284 extract Substances 0.000 claims description 5
- 238000000605 extraction Methods 0.000 claims description 5
- 230000007613 environmental effect Effects 0.000 abstract 2
- 238000005070 sampling Methods 0.000 description 5
- 230000015572 biosynthetic process Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000003786 synthesis reaction Methods 0.000 description 2
- 125000004122 cyclic group Chemical group 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/008—Multichannel audio signal coding or decoding using interchannel correlation to reduce redundancy, e.g. joint-stereo, intensity-coding or matrixing
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/04—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
- G10L19/16—Vocoder architecture
- G10L19/167—Audio streaming, i.e. formatting and decoding of an encoded audio signal representation into a data stream for transmission or storage purposes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/439—Processing of audio elementary streams
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Acoustics & Sound (AREA)
- Mathematical Physics (AREA)
- Stereophonic System (AREA)
Abstract
Description
The present invention relates to an audio signal processing method and apparatus, and more particularly, to an audio signal processing method and apparatus capable of encoding or decoding an audio signal.
In general, in order to transmit a digital broadcast signal, a video signal and an audio signal must be transmitted. In this case, a signal corresponding to a mono channel or a stereo channel may be transmitted as an audio signal, and decoding information for upmixing the audio signal to a stereo channel or a multichannel may be transmitted together. In this case, there is a problem in that the bit amount of the audio signal is increased in transmitting the decoding information, and in the decoder, the complexity increases in the process of upmixing to the multichannel.
The present invention was devised to solve the above problems, and provides an audio signal processing method and apparatus capable of generating both a stereo output and a multi-channel output by one codec in upmixing a core signal. The purpose is.
Still another object of the present invention is to determine whether an output mode is a stereo output or a multi output before parsing decoding information for channel extension by using channel extension environment information included in an audio signal bitstream.
In order to achieve the above object, an audio signal processing method includes extracting channel extension environment information from an audio signal bitstream; Extracting at least one of first decoding information and second decoding information from the audio signal bitstream based on the channel extension environment information; And upmixing a core signal using at least one of the first decoding information and the second decoding information, wherein the first decoding information is configured to upgrade the core signal to a stereo signal using a first decoding scheme. Information for mixing, and the second decoding information is information for upmixing the core signal into a stereo signal or a multichannel signal using a second decoding scheme.
According to the present invention, the second decoding information includes first spatial information that is spatial information for upmixing the core signal into the multichannel signal, and the core signal is converted into the core signal using the first spatial information. The method may further include generating second spatial information for upmixing into a stereo signal.
According to the present invention, when the core signal is a stereo signal, the step of upmixing the core signal may be performed using the first spatial information.
According to the present invention, the first decoding method corresponds to a method of generating a stereo channel using a mono signal and a decorrelator, and the second decoding method uses spatial information including a level difference between channels. It may correspond to a method for generating the stereo channel or the multi-channel signal.
According to the present invention, the method further comprises: extracting core channel information from the audio signal bitstream; The method may further include determining whether the core signal is a mono signal or a stereo signal based on the core channel information.
According to the present invention, extracting SBR (Spectral Band Replication) flag information from the audio signal bitstream; And determining whether to use the SBR tool based on the SBR flag information.
According to the present invention, when the SBR (Spectral Band Replication) tool is not used according to the SBR flag information, the step of upmixing the core signal is performed using only the second decoding information, not the first decoding information. Can be.
According to the present invention, the method may further include determining an output mode according to the channel extension environment information. When the output mode is a stereo output mode, the step of upmixing the core signal may be performed in a general manner rather than a binaural mode. May be performed in mode.
According to still another aspect of the present invention, channel extension environment information is extracted from an audio signal bitstream, and at least one of first decoding information and second decoding information is extracted from the audio signal bitstream based on the channel extension environment information. Extraction unit to extract; And an extended decoding unit configured to upmix a core signal using at least one of the first decoding information and the second decoding information, wherein the first decoding information includes a stereo signal using the first decoding scheme. The second decoding information is information for upmixing, and the second decoding information is information for upmixing the core signal into a stereo signal or a multichannel signal using a second decoding scheme.
According to the present invention, the second decoding information includes first spatial information, which is spatial information for upmixing the core signal into the multichannel signal, and the extended decoding unit uses the first spatial information, The apparatus may further include a spatial information generator configured to generate second spatial information for upmixing the core signal to the stereo signal.
According to the present invention, when the core signal is a stereo signal, the extended decoding unit may upmix the core signal using the first spatial information.
According to the present invention, the first decoding method corresponds to a method of generating a stereo channel using a mono signal and a decorrelator, and the second decoding method uses spatial information including a level difference between channels. It may correspond to a method for generating the stereo channel or the multi-channel signal.
According to the present invention, the extractor may further extract core channel information from the audio signal bitstream, and determine whether the core signal is a mono signal or a stereo signal based on the core channel information.
According to the present invention, the extractor may extract SBR (Spectral Band Replication) flag information from the audio signal bitstream and determine whether to use the SBR tool based on the SBR flag information.
According to the present invention, when the SBR (Spectral Band Replication) tool is not used according to the SBR flag information, the extended decoding unit upmixes the core signal using only the second decoding information, not the first decoding information. can do.
According to the present invention, the extractor may determine an output mode according to the channel extension environment information, and when the output mode is a stereo output mode, the extended decoder may upgrade the core signal in a normal mode instead of a binaural mode. You can mix.
According to another aspect of the present invention, the method includes: extracting channel extension environment information from an audio signal bitstream; Determining an output mode based on the channel extension environment information; Generating second spatial information by using first spatial information included in the audio signal bitstream when the output mode is a stereo output mode; And upmixing a core signal using one of the first spatial information and the second spatial information, wherein the first spatial information is used for upmixing the core signal into a stereo signal or a multichannel signal. Provided is information for decoding, and the second spatial information is provided with an audio signal processing method comprising information for upmixing the core signal into a stereo signal.
According to the present invention, when the core signal is a stereo signal, the step of upmixing the core signal may be performed using the first spatial information.
According to the present invention, when the output mode is a stereo output mode, the step of upmixing the core signal may be performed in a normal mode rather than a binaural mode.
According to another aspect of the present invention, an extraction unit for extracting the channel expansion environment information from the audio signal bitstream, and determines the output mode based on the channel expansion environment information; A spatial information generator configured to generate second spatial information by using first spatial information included in the audio signal bitstream when the output mode is a stereo output mode; And an extended decoding unit configured to upmix a core signal using one of the first spatial information and the second spatial information, wherein the first spatial information upmixes the core signal to a stereo signal or a multichannel signal. The second spatial information is provided, and the audio signal processing apparatus includes information for upmixing the core signal into a stereo signal.
According to the present invention, the core signal may be a stereo signal, and the extended decoding unit may upmix the core signal using the first spatial information.
According to the present invention, when the output mode is a stereo output mode, the extended decoding unit may upmix the core signal in a normal mode instead of a binaural mode.
According to an aspect of the present invention, when upmixing a core signal, since both the stereo signal and the multi-channel signal can be upmixed using one codec method, a bitstream corresponding to another codec method is excluded. Therefore, the amount of bits of information required for upmixing can be significantly reduced.
According to another aspect of the present invention, before parsing decoding information (eg, MPEG bitstream) for upmixing, stereo output may be performed based on the channel expansion environment information because it may be determined whether the stereo output mode or the multichannel output mode is used. In mode, using two filter banks can significantly reduce the complexity required for upmixing.
According to another aspect of the invention,
Hereinafter, exemplary embodiments of the present invention will be described in detail with reference to the accompanying drawings. Prior to this, terms or words used in the specification and claims should not be construed as having a conventional or dictionary meaning, and the inventors should properly explain the concept of terms in order to best explain their own invention. Based on the principle that can be defined, it should be interpreted as meaning and concept corresponding to the technical idea of the present invention. Therefore, the embodiments described in the specification and the drawings shown in the drawings are only the most preferred embodiment of the present invention and do not represent all of the technical idea of the present invention, various modifications that can be replaced at the time of the present application It should be understood that there may be equivalents and variations.
1 is a diagram illustrating a configuration of an audio signal processing apparatus according to an embodiment of the present invention. Referring to the drawings, an audio signal processing apparatus according to an exemplary embodiment of the present invention includes an
The
The extended
2 is an example conceptually illustrating an audio signal bitstream. Referring to FIG. 2A, the audio signal bitstream includes
Referring to FIG. 2B, the audio signal bitstream includes
3 is an example of a structure of an audio signal bitstream. Referring to FIG. 3A, the structure of mh_audio_frame () is shown. The syntax of mh_audio_frame () may be configured as follows.
[Table 1: Syntax of mh_audio_frame ()]
First, referring to the syntax, the audio signal version information (mh_audio_version) is extracted, and when the audio signal version is HE_AAC_V2_MPS ("MPEG-4 HE AAC v2 with Baseline MPEG
An example of the structure for mh_audio_ham_frame () is shown in FIG. 2B, and an example of the syntax is as follows.
Table 2: Syntax of mh_audio_ham_frame ()
Here, "n_aus" is the number of access units in the corresponding audio frame,
an [n]: Audio samples for the duration of one audio_ham_frame according to the core sampling rate and n_aus.
au_crc: Each access unit is protected according to a 16 bit CRC. Cyclic Redundancy Check (CRC) is generated by the following polynomial:
An example of the syntax of mh_audio_ham_header () is as follows.
Table 3: Syntax of mh_audio_ham_header ()
Like the syntax, mh_audio_ham_header () includes a sync word (sync_word), a refresh flag (refresh_flag), core channel information (aac_channel_mode), SBR flag information (sbr_flag), first extension channel environment information (ch_extension_config), and the like. The second extension channel environment information (mps_extension_config) may be included instead of the first extension channel environment information. Specifically, when the audio signal bitstream is as shown in FIG. 2A, the first extended channel environment information is included, and as shown in FIG. 2B, the second extended channel environment information. May be included. Hereinafter, each element included in the syntax will be described in order.
A sync word (sync_word) is present to allow an RS (Reed_Solomon) decoder to identify the beginning of an audio frame with a 12-bit long synchronization sequence. The value of sync_word is '0xFF5' in hexadecimal.
The refresh flag (refresh_flag) is a 1-bit flag indicating whether the audio decoder checks whether the audio parameter of the next audio frame is the same as or equal to the audio parameter of the current audio frame, as shown in the table below.
Table 4: Meaning of refresh_flag
The core channel information (aac_channel_mode) is a 1-bit flag indicating whether the core signal is mono or stereo as shown in the following table.
[Table 5: Meaning of core channel information (aac_channel_mode)]
SBR flag information (sbr_flag) is a 1-bit flag indicating whether SBR is used as shown in the following table.
[Table 6: Meaning of SBR flag information (sbr_flag)]
On the other hand, au_start [n] is an unsigned integer, with the most significant bit being given priority and transmitted through a 12-bit field that defines the starting point in the audio frame of the individual access unit by giving the byte number of the first byte of the access unit. The value of au_start of the first access unit is not transmitted but is given by calculating the header size. The decoder derives au_size [n] from au_start [n] and au_start [n + 1] as follows.
[Equation 1]
au_size [n] = au_start [n + 1] -au_start [n] -2;
au_size [n_aus] = audio_frame_size
The output mode when the first extension channel environment information ch_extension_config uses one of the first decoding information PS and the second decoding information MPS and the second decoding information MPS as shown in the following table. Information about whether it is a stereo output or a multichannel (ex: 5.1) output. On the other hand, when the output mode is stereo, since the stereo signal (s) is generated directly from the mono signal (m), it can be referred to as a 2-1-2 configuration, the 2-1-2 configuration is the MPEG surround standard It is not included in the specified bstreeconfig (eg 5-1-5 1 , 5-1-5 2 , 5-2-5, etc.).
[Table 7: Meaning of first extension channel environment information (ch_extension_config)]
When the first extension channel environment information is 'ch_extension_config = 000', when the first decoding information and the second decoding information do not exist, the
When the first extension channel environment information is 'ch_extension_config = 001', only first decoding information exists and may correspond to a case of the
When the first extension channel environment information is 'ch_extension_config = 010', only second decoding information exists, and the output mode is a stereo output mode. In this case, the
When the first extension channel environment information is 'ch_extension_config = 011', both the first decoding information and the second decoding information may be used. In this case, when the second decoding information is used, the output mode is multichannel. Output mode. In this case, when the first decoding information is used, the
When the first extension channel environment information is 'ch_extension_config = 100', only second decoding information is used, and the output mode is a multichannel output mode. In this case, the
As such, referring to the first extended channel environment information, it is possible to know not only whether the second decoding information is present, but also whether the output channel is stereo or multichannel before parsing the second decoding information.
A detailed description of each case of the first extended channel environment information will be described in detail with reference to FIGS. 4 to 6 and 7 to 8.
Meanwhile, as described above, the second extended channel environment information (mps_extension_config) may be included instead of the first extended channel environment information, which is a case where only the second decoding information may be included in the audio signal bitstream. The second extended channel environment information is shown in the following table.
[Table 8: Meaning of second extension channel environment information (mps_extension_config)]
When the second extended channel environment information is 'mps_extension_config = 000', the second decoding information does not exist, and the
When the second extension channel environment information is 'mps_extension_config = 001', the second decoding information is used, and the output mode is a stereo output mode. This case is allowed only when the core signal is a mono signal. In other words, when the core signal is a stereo signal, the stereo output mode is not allowed and the multichannel output mode is forced.
When the second extension channel environment information is 'mps_extension_config = 010', second decoding information is used, and the output mode is a multichannel output mode. In this case, the case where the core signal is a mono signal as well as a stereo signal is allowed.
As such, referring to the second extended channel environment information, it is possible to know whether the output channel is stereo or multichannel even before parsing the second decoding information.
4 to 6 are first, second and third examples showing the detailed configuration of the extended decoding unit. As described above, the first to third examples are cases in which the first channel extension environment information (ch_extension_config) is included in the audio signal bitstream, where both the first decoding information and the second decoding information may exist. Corresponding.
First, referring to FIG. 4, the
Referring to FIG. 5, the
Referring to FIG. 6, the
7 and 8 illustrate a procedure of an audio signal processing method according to an embodiment of the present invention. 7 to 8 also correspond to the case where at least one of the first decoding information and the second decoding information exists in the audio signal bitstream as described above. Referring to FIG. 7, first, an audio signal bitstream is received (step S110). Core channel information, SBR information, first extended channel environment information, and the like are extracted from the audio signal bitstream (S120). Based on the first extended channel environment information, it is determined whether the first decoding information and the second decoding information exist or, if the second decoding information exists, which output mode.
If only the first decoding information exists (YES in step S122), for example, when the first extended channel environment information is '001', the first decoding information is extracted (step S210). The core signal is upmixed using the 1 decoding information (S212).
If both the first decoding information and the second decoding information exist (YES in S124), for example, when the first extended channel environment information is '011', the first decoding information and the second decoding information are determined. Extract all (step S220). When the first decoding information is used among the extracted information (YES in step S222), the core signal is upmixed using the first decoding information (step S212). If the second decoding information is used among the extracted information (NO in step S222), the second decoding information may be a multi-channel output mode, which will be described with reference to FIG. 8. It will be described later.
Referring to FIG. 8, when it corresponds to 'NO' in steps S122 and S124 (step A), and second decoding information does not exist (“No” in step S126), for example, the first extended channel environment If the information is '000', the procedure ends without upmixing the core signal. If only the second decoding information exists (YES in step S126), the second decoding information is extracted (step S230). Then, it is determined whether the output mode is a stereo output mode or a multichannel output mode based on the first extended channel environment information. In case of the stereo output mode (Yes in step S240), for example, when the first extended channel environment information is '010', the second spatial information is used by using the first spatial information included in the audio signal bitstream. In operation S242, the core signal is upmixed into a stereo signal using the second spatial information in operation S244. It can be seen that the output mode is a stereo output (2-1-2 configuration) based on the first extended channel environment information, and can be upmixed using only two synthesis filterbanks corresponding to the stereo output. In this case, the complexity of the decoder can be significantly reduced. On the other hand, the upmixing of the core signal in step S244 is performed in the normal mode, not the binaural mode. On the contrary, when the output mode is not the stereo output mode (No in step S240), for example, when the first extended channel environment information is '100', the core signal is converted into a multi-channel signal using the first spatial information. Upmixing (S250).
9 and 10 are fourth and fifth examples showing detailed configurations of the extended decoding unit. 9 and 10 illustrate a case in which only the second decoding information may be included in the audio signal bitstream, and the second extension channel environment information (mps_extension_config) indicating whether the second decoding information exists and the output mode is It is a case of inclusion. The fourth example shown in FIG. 9 corresponds to a case where 'mps_extension_config = 001' when the second channel extension environment information is shown in Table 8, and the configuration of the
11 is a flowchart illustrating an audio signal processing method according to another embodiment of the present invention. As described above, another embodiment illustrated in FIG. 11 corresponds to a case in which only second decoding information may exist in the audio signal bitstream. Referring to FIG. 11, first, an audio signal bitstream is received (step S310). Core channel information, SBR information, second extended channel environment information, and the like are extracted from the audio signal bitstream (step S320). The presence of the second decoding information and the output mode are determined based on the second extended channel environment information. If the second decoding information does not exist (NO in step S330), for example, when the second extended channel environment information is '000', the extended decoding step is omitted and the procedure is terminated.
When the second decoding information is present (YES in step S330), for example, when the second extended channel environment information is '001' or '010', steps S240 to S250 of FIG. 8 described above are described. The same step as is performed (steps S350 to S380).
As described above, although the present invention has been described by way of limited embodiments and drawings, the present invention is not limited thereto and is intended by those skilled in the art to which the present invention pertains. Of course, various modifications and variations are possible within the scope of equivalents of the claims to be described.
The invention can be applied to the encoding and decoding of audio signals.
1 is a block diagram of an audio signal processing apparatus according to an embodiment of the present invention.
2 is a first example and a second example conceptually illustrating an audio signal bitstream.
3 is an example of the structure of an audio signal bitstream.
4 is a first example showing a detailed configuration of an extended decoding unit.
5 is a second example showing the detailed configuration of an extended decoding unit.
6 is a third example showing the detailed configuration of an extended decoding unit.
7 and 8 are flowcharts of an audio signal processing method according to an embodiment of the present invention.
9 is a fourth example showing the detailed configuration of an extended decoding unit.
10 is a fifth example showing the detailed configuration of an extended decoding unit.
11 is a flowchart of an audio signal processing method according to another embodiment of the present invention.
Claims (22)
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US94818207P | 2007-07-05 | 2007-07-05 | |
US60/948,182 | 2007-07-05 |
Publications (1)
Publication Number | Publication Date |
---|---|
KR20090004778A true KR20090004778A (en) | 2009-01-12 |
Family
ID=40652250
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
KR1020080065478A KR20090004778A (en) | 2007-07-05 | 2008-07-07 | Method for processing an audio signal and apparatus for implementing the same |
Country Status (1)
Country | Link |
---|---|
KR (1) | KR20090004778A (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20120029494A (en) * | 2010-09-16 | 2012-03-27 | 삼성전자주식회사 | Apparatus and method for bandwidth extension for multi-channel audio |
KR20140018929A (en) * | 2011-03-18 | 2014-02-13 | 프라운호퍼 게젤샤프트 쭈르 푀르데룽 데어 안겐반텐 포르슝 에. 베. | Audio encoder and decoder having a flexible configuration functionality |
KR20180023941A (en) * | 2009-10-23 | 2018-03-07 | 삼성전자주식회사 | Apparatus and method for encoding/decoding using phase information and residual signal |
-
2008
- 2008-07-07 KR KR1020080065478A patent/KR20090004778A/en not_active Application Discontinuation
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20180023941A (en) * | 2009-10-23 | 2018-03-07 | 삼성전자주식회사 | Apparatus and method for encoding/decoding using phase information and residual signal |
KR20120029494A (en) * | 2010-09-16 | 2012-03-27 | 삼성전자주식회사 | Apparatus and method for bandwidth extension for multi-channel audio |
KR20140018929A (en) * | 2011-03-18 | 2014-02-13 | 프라운호퍼 게젤샤프트 쭈르 푀르데룽 데어 안겐반텐 포르슝 에. 베. | Audio encoder and decoder having a flexible configuration functionality |
US9524722B2 (en) | 2011-03-18 | 2016-12-20 | Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. | Frame element length transmission in audio coding |
US9773503B2 (en) | 2011-03-18 | 2017-09-26 | Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. | Audio encoder and decoder having a flexible configuration functionality |
US9779737B2 (en) | 2011-03-18 | 2017-10-03 | Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. | Frame element positioning in frames of a bitstream representing audio content |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP7090196B2 (en) | Audio encoders and decoders with program information or substream structure metadata | |
TWI476761B (en) | Audio encoding method and system for generating a unified bitstream decodable by decoders implementing different decoding protocols | |
TWI447707B (en) | Method and apparatus for processing a audio signal | |
JP5006315B2 (en) | Audio signal encoding and decoding method and apparatus | |
AU2005281966B2 (en) | Device and method for reconstructing a multichannel audio signal and for generating a parameter data record therefor | |
US8238561B2 (en) | Method for encoding and decoding multi-channel audio signal and apparatus thereof | |
KR101387902B1 (en) | Encoder and method for encoding multi audio object, decoder and method for decoding and transcoder and method transcoding | |
US20100324915A1 (en) | Encoding and decoding apparatuses for high quality multi-channel audio codec | |
KR101599554B1 (en) | 3 3d binaural filtering system using spectral audio coding side information and the method thereof | |
US20150348558A1 (en) | Audio Bitstreams with Supplementary Data and Encoding and Decoding of Such Bitstreams | |
JP2009514008A5 (en) | ||
CN101243490A (en) | Method and apparatus for encoding and decoding an audio signal | |
WO2014124377A2 (en) | Audio bitstreams with supplementary data and encoding and decoding of such bitstreams | |
TWI504275B (en) | Method and apparatus for signaling and decoding avs1-p2 bitstreams of different versions | |
KR20090004778A (en) | Method for processing an audio signal and apparatus for implementing the same | |
US20110311063A1 (en) | Embedding and extracting ancillary data | |
KR20090039642A (en) | Method of decoding a dmb signal and apparatus of decoding thereof | |
US20070160043A1 (en) | Method, medium, and system encoding and/or decoding audio data | |
RU2802677C2 (en) | Methods and devices for forming or decoding a bitstream containing immersive audio signals | |
KR20080035448A (en) | Method and apparatus for encoding/decoding multi channel audio signal | |
KR20070003574A (en) | Method and apparatus for encoding and decoding an audio signal | |
KR20070041338A (en) | Method and apparatus for decoding and encoding in audio codec | |
KR20070108314A (en) | Method and apparatus for encoding/decoding an audio signal |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WITN | Withdrawal due to no request for examination |