WO2006004048A1 - オーディオ信号符号化装置、オーディオ信号復号化装置、方法、及びプログラム - Google Patents
オーディオ信号符号化装置、オーディオ信号復号化装置、方法、及びプログラム Download PDFInfo
- Publication number
- WO2006004048A1 WO2006004048A1 PCT/JP2005/012221 JP2005012221W WO2006004048A1 WO 2006004048 A1 WO2006004048 A1 WO 2006004048A1 JP 2005012221 W JP2005012221 W JP 2005012221W WO 2006004048 A1 WO2006004048 A1 WO 2006004048A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- signal
- information
- channel
- auxiliary information
- encoding
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S3/00—Systems employing more than two channels, e.g. quadraphonic
Definitions
- Audio signal encoding apparatus Audio signal decoding apparatus, method, and program
- the present invention relates to an audio signal encoding device, decoding device, method, and program.
- Patent Document 1 US Published Patent US2003 / 0035553 "Backwards- compatible Perceptual Coding of Spatial Cues
- Patent Document 2 US Published Patent US2003 / 0219130 "Coherence-based Audio Coding and Synthesis"
- the present invention solves the above-described problem, and an audio signal encoding apparatus for improving encoding efficiency when encoding a multi-channel signal, and decoding the code obtained from the apparatus. It is an object of the present invention to provide an audio signal decoding apparatus that can be used.
- the audio signal encoding device of the present invention represents the original sound signal of each of a plurality of channels, the mixed signal information representing the overall characteristics of the original sound signal, and the feature amount based on the relationship between the individual original sound signals.
- An audio signal encoding device for encoding auxiliary information wherein the mixed signal encoding means for generating the mixed signal information by encoding the mixed signal obtained by mixing the original sound signal, and the original sound signal Based on the above, the feature amount is calculated, and when channel information indicating the playback position of the sound of each channel is given, the relationship of the arrangement of the playback positions indicated by the given channel information
- a different encoding method is determined according to the auxiliary information, and the auxiliary information for generating the auxiliary information is encoded by encoding the feature amount calculated using the determined encoding method.
- the auxiliary information generation means stores in advance a plurality of tables that define quantization points at which different quantization accuracies can be obtained, and the feature value is stored in the channel of the plurality of tables.
- the encoding may be performed by quantizing to a quantization point defined by one according to the arrangement relationship of each reproduction position indicated by the information.
- the auxiliary information generating means may calculate at least one of a level difference and a phase difference between the original sound signals and use it as the feature amount, and further, based on the calculated level difference and phase difference. Then, the direction of the sound image estimated to be perceived by the listener may be calculated as the feature amount.
- the auxiliary information generating means defines a first table that defines a quantization point that is provided symmetrically from the front direction of the listener, and a quantization point that is provided asymmetrically from the left direction of the listener
- the channel information indicates the left front and left rear of the listener
- the feature value is quantized to a quantization point defined by the second table.
- the sign may be given by.
- the auxiliary information generation means may calculate the similarity between the original sound signals and use it as the feature quantity. Also, the auxiliary information generation means may use the cross-correlation value between the original sound signals or the absolute value thereof as the similarity. Further, based on the calculated similarity, at least one of a feeling of spread of sound image and a feeling of distance estimated by the listener may be calculated as the feature amount.
- the audio signal decoding device of the present invention is characterized by the mixed signal information representing the overall characteristics of the original sound signals of each of a plurality of channels and the relationship between the individual original sound signals.
- An audio signal decoding device that decodes auxiliary information representing the amount into a reproduction signal of each of the plurality of channels, and is provided with channel information indicating a reproduction position as viewed from a listener of audio of each channel
- a decoding method switching means for determining a different decoding method in accordance with the relationship between the arrangements of the reproduction positions indicated by the given channel information, and the auxiliary information using the determined decoding method.
- Inter-signal information decoding means for decoding into feature quantities, the mixed signal information, and the Signal synthesizing means for generating a reproduction signal for each of the plurality of channels using the decoded feature quantity.
- the auxiliary information is a value corresponding to the arrangement relationship of each reproduction position indicated by the channel information among a plurality of tables defining quantization points at which different quantization accuracy can be obtained.
- the inter-signal information decoding means stores the plurality of tables in advance, and the channel information of the plurality of tables is encoded. Decode the auxiliary information into the feature amount using one corresponding to the arrangement relationship of the reproduction positions indicated by.
- the feature amount represents at least one of a level difference, a phase difference between the original sound signals, and a direction of a sound image estimated to be perceived by the listener.
- the feature amount represents at least one of a level difference, a phase difference, a similarity, and a direction of sound image estimated to be perceived by a listener, a sense of spread, and a sense of distance. It is good.
- the signal synthesizing unit may generate an audio signal represented by the mixed signal information when the feature value represents at least one of a level difference, a phase difference, and a similarity between the original sound signals.
- the reproduction signal may be generated by giving a level difference, a phase difference, and similarity according to the feature amount.
- the present invention can be realized as a method including steps executed by characteristic means included in such an apparatus as well as being realized as such an audio signal encoding apparatus and audio signal decoding apparatus. It can also be realized as a program that causes a computer to execute these steps. And such a program It goes without saying that the ram can be distributed via a recording medium such as a CD-ROM or a transmission medium such as the Internet.
- the audio signal encoding device and decoding device of the present invention when generating auxiliary information for separating a reproduction signal that approximates an original sound signal from a mixed signal obtained by mixing the original sound signal. It is possible to achieve signal separation to the extent that there is no sense of incongruity and to generate very small auxiliary information by using the characteristics of the direction perception of human sound sources, the sense of spread, and the sense of distance. It becomes possible.
- FIG. 2 is a diagram illustrating an example of a positional relationship between a listener and a sound source indicated by channel information.
- FIG. 3 is a functional block diagram showing an example of a configuration of an auxiliary information generation unit.
- FIG. 6 is a functional block diagram showing another example of the configuration of the auxiliary information generation unit.
- FIG. 7 is a diagram schematically illustrating an example of a table used for quantization of correlation between signals, similarity between signals, and spread feeling estimation value.
- FIG. 10 is a functional block diagram illustrating an example of a configuration of a signal separation processing unit. Explanation of symbols
- FIG. 1 is a block diagram showing an example of a functional configuration of the audio signal encoding device of the present invention.
- This audio signal encoding apparatus encodes the first input signal 201 and the second input signal 202 given by an external force to obtain mixed signal information 206 and is indicated by channel information 207 given from the outside.
- This is a device that obtains auxiliary information 205 using a different encoding method depending on the relationship of the audio reproduction position of each channel, and includes a mixed signal encoding unit 203 and an auxiliary information generation unit 204.
- the mixed signal information 206 and the auxiliary information 205 are information decoded by the audio signal decoding device described later into signals approximating the first input signal 201 and the second input signal 202, respectively, and the channel information 207 Is information indicating which direction force of each decoded signal is reproduced as viewed from the listener.
- FIG. 2 is a diagram illustrating an example of a positional relationship between a sound source for signal reproduction and a listener. This example shows the arrangement direction seen from the listener of each speaker that is the sound source of each channel when 5-channel playback is performed. For example, it is shown that a front L channel speaker and a front R channel speaker are arranged 30 ° to the left and right from the front of the listener. These two speakers are also used for stereo playback.
- the channel information 207 signifies the sound to be reproduced from the front L channel speaker and the front R channel speaker, specifically, counterclockwise with the front of the listener as 0 °. It is expressed by the angle of the sound source such as + 30 ° (front L channel speaker) and 30 ° (front R channel speaker). In addition, in practical terms, the angle of the sound source of each channel is determined in advance using fine angle information such as 30 °, and the same thing can be expressed simply by the channel names such as the front L channel and front R channel. it can.
- the channel information 207 is appropriately supplied to the audio signal encoding apparatus from an external apparatus that knows which channel's audio is encoded.
- the first input signal 201 and the second input signal 202 are stereo originals.
- channel information 207 indicating the front L channel and the front R channel is given.
- the front L channel and the rear L channel are respectively used as the first input signal 201 and the second input signal 202.
- channel information 207 indicating the front L channel and rear L channel is given.
- the auxiliary information generating unit 204 uses the channel information 207 from the first input signal 201, the second input signal 202, the mixed signal generated by the mixed signal encoding unit 203, and the mixed signal information 206, using the channel information 207. 205 is generated.
- the auxiliary information 205 is information for separating from the mixed signal a signal that is as close as possible to the first input signal 201 and the second input signal 202 that are original sound signals before mixing.
- the auxiliary signal 205 can be used to separate the first input signal 201 before mixing from the mixed signal and the same signal as the second input signal 202, but to the extent that there is no difference. These signals may be separated. Even if there is a difference in hearing, the auxiliary information is included in the scope of the present invention as long as the auxiliary information is information for signal separation.
- the auxiliary information generation unit 204 By using the channel information 207, the auxiliary information generation unit 204 generates auxiliary information with a small amount of information that can separate a signal that does not cause a sense of incongruity. For this purpose, the auxiliary information generation unit 204 switches the method of encoding the auxiliary information, specifically, the quantization accuracy of the code ⁇ according to the channel information 207.
- FIG. 3 is a block diagram illustrating a functional configuration of the auxiliary information generation unit according to the first embodiment.
- the auxiliary information generation unit is a unit that generates auxiliary information 205A that is encoded differently according to channel information 207 from the first input signal 201 and the second input signal 202.
- the auxiliary information 205A is calculated by the inter-signal level difference calculation unit 303, the inter-signal phase difference calculation unit 304, and the perceived direction estimation unit 305. This is information obtained by quantizing and signing at least one of the perceived direction estimation values.
- the first input signal 201 and the second input signal 202 are input to the inter-signal level difference calculation unit 303 and the inter-signal phase difference calculation unit 304.
- the inter-signal level difference calculation unit 303 calculates the energy difference between the first input signal 201 and the second input signal 202.
- the signal may be divided into a plurality of frequency bands, and each signal may be calculated, or one may be calculated for all bands. Further, the time unit to be calculated is not particularly limited.
- the energy difference can be expressed as an exponential function value often used in audio expression, for example, dB, but not necessarily limited.
- the inter-signal phase difference calculation unit 304 calculates the cross-correlation between the signals of the first input signal 201 and the second input signal 202, and based on the cross-correlation value, the phase difference that increases the value. Is calculated.
- a phase difference calculation method is known to those skilled in the art.
- the phase that gives the maximum cross-correlation value does not necessarily have to be a phase difference. This is because when the cross-correlation is calculated based on a digital signal, it is a discrete value, so the phase difference can also be obtained as a discrete value.
- the solution is based on the cross-correlation value distribution. It may be the phase difference estimated by interpolation.
- the inter-signal level difference obtained as an output from the inter-signal level difference calculator 303, the inter-signal phase difference obtained as an output from the inter-signal phase difference calculator 304, and the channel information 207 are: This is input to the perceptual direction estimation unit 305.
- the perceptual direction estimation unit 305 is based on the channel information 207, the inter-signal level difference calculation unit 303, and the inter-signal level difference obtained as an output, and the inter-signal phase difference calculation unit 304 outputs the inter-signal phase difference.
- the direction of the sound image perceived by the listener is estimated.
- the direction perceived by the listener when presenting acoustic signals from two speakers is determined by the arrangement of the two speakers and the level difference and phase difference of the 2ch signal.
- spatial sound I Enns browser elutriation, Masayuki Morimoto Toshiyuki Goto, ed., Kajima publishing (198b there ⁇ , spatial Hearing: The Psychophysics of Human sound Localiza tion ", revised edition, MIT Press, 1997)
- the estimation unit 305 estimates the perceived direction of the sound image by the listener, and outputs a perceived direction estimation value representing the estimation result to the code unit 306.
- the encoding unit 306 quantizes at least one of the inter-signal level difference, the inter-signal phase difference, and the perceived direction estimation value with different accuracy depending on the channel information 207 and the perceived direction estimation value. Further, the auxiliary information 205A obtained by performing the sign ⁇ is output.
- the perceptual discrimination characteristics of listeners are bilaterally symmetric with respect to the front direction, the perceptual discrimination characteristics in the front direction are sensitive, and tend to be insensitive toward the front L channel direction (or front R channel direction). There is.
- the listener's perceptual discrimination characteristics are asymmetrical in the front direction, counterclockwise to the back direction, and the perceptual discrimination characteristics in the front direction are responsive and tend to become insensitive as the back channel is applied.
- the sign key unit 306 when the perceived direction estimation value obtained from the perceived direction estimation unit 305 indicates a direction in which the perceptual discrimination characteristic is sensitive, The phase difference and the perceived direction estimate are quantized finely, and conversely, when the perceptual discrimination characteristic shows an insensitive direction, it is quantized roughly as compared with a sensitive direction.
- the encoding unit 306 performs symmetrical quantization in the left-right direction with respect to the perceptual direction, and indicates the front L channel and the back L channel. In this case, quantization that is asymmetric in the front-rear direction with respect to the perceptual direction is performed.
- the sign unit 306 In order to switch the quantization accuracy as described above, for example, the sign unit 306 previously holds a plurality of tables for converting input values into quantized values, from which a channel is converted. One corresponding to the information 207 is used.
- FIG. 4 is a diagram schematically illustrating an example of a table that is stored in advance in the code key unit 306 and is used for quantization of the perceived direction estimation value.
- Each table shows an example of the quantization point of the perceived direction estimation value, (A) is an example for the front L channel and front R channel, and (B) is for the rear L channel and front L channel. It is an example.
- the sign key unit 306 When the channel information 207 indicates the front L channel and the front R channel, the sign key unit 306 generates a perceptual direction estimation value based on the table in FIG. In the vicinity of the sensitive frontal direction, it is quantized with little effort, and the perceptual discrimination characteristics are quantized roughly in the left-right direction, which is relatively insensitive.
- the perceived direction estimation value is calculated based on the table in FIG. 4B in the vicinity of the front direction where the perceptual discrimination characteristics are relatively sensitive. It quantizes with a lot of force, and it quantizes roughly as it moves toward the back direction where the perceptual discrimination characteristics are relatively insensitive.
- FIG. 5 is a diagram schematically showing an example of a table used for quantization of the inter-signal level difference and the inter-signal phase difference.
- Each table shows an example of the quantization point of the inter-signal level difference and inter-signal phase difference for which the predetermined normality is made, and (A) is an example of the table for the front L channel and the front R channel. Yes, (B) is an example of a table for the rear L channel and the front L channel.
- the sign key unit 306 indicates that the perceived direction estimation value is relatively sensitive to the perceptual discrimination characteristics based on the table in FIG. In the case of showing the vicinity in the front direction, the level difference between signals and the phase difference between signals are finely quantized, and the level difference between signals and Roughly quantize the phase difference between signals.
- the perceived direction estimation value is a relatively sensitive frontal direction neighborhood of the perceptual discrimination characteristics.
- the level difference between signals and the phase difference between signals are finely quantized.
- the level difference between signals and the phase difference between signals are roughly quantized as the perceived direction estimation value moves toward the relatively insensitive back direction of the perceptual discrimination characteristics.
- both the tables in FIG. 4 and FIG. 5 show a specific example of a configuration for switching the coding method according to the channel information 207 that is a feature of the present invention. It is not intended to limit the distribution of points to what is shown.
- the present invention includes a case where a table representing another distribution of quantization points reflecting the listener's perceptual discrimination characteristics is used, such as when the channel information 207 indicates a rear L channel and a rear R channel.
- the sign key unit 306 reflects the discrimination ability related to the perceived direction of the sound image of the listener based on the channel information 207 and the perceived direction estimated value obtained from the perceived direction estimating unit 305.
- Quantization accuracy i.e., quantization accuracy that becomes finer in the front direction from the left and right to the back direction
- FIG. 6 is a block diagram illustrating a functional configuration of the auxiliary information generation unit according to the second embodiment.
- the auxiliary information generating unit according to the second embodiment is a unit that generates auxiliary information 205B encoded according to channel information 207 from the first input signal 201 and the second input signal 202, and provides an inter-signal correlation.
- a degree calculation unit 401, a spread feeling estimation unit 402, and a sign key unit 403 are included.
- the auxiliary information 205B includes at least one of the inter-signal correlation calculated by the inter-signal correlation calculating unit 401, the inter-signal similarity, and the expanded feeling estimated value calculated by the expanded feeling estimating unit 402. This is information obtained by quantizing and coding one.
- the first input signal 201 and the second input signal 202 are input to the inter-signal correlation calculation unit 401.
- ⁇ is a term for correcting a phase shift between both ears, and is known to those skilled in the art.
- the signal may be divided into a plurality of frequency bands and calculated for each of them, or one may be calculated for all bands. Further, the time unit to be calculated is not particularly limited.
- the spread feeling estimation unit 402 estimates the degree of spread of the sound image perceived by the listener from the channel information 207 and the similarity between signals obtained as an output from the inter-signal correlation calculation unit 401.
- the extent of the sound image perceived by the listener is expressed by appropriately quantifying the magnitude and smallness as psychological quantities.
- Encoding section 403 quantizes at least one of the correlation between signals, the similarity between signals, and the estimated spread feeling with different accuracy in accordance with the channel information 207, and further The auxiliary information 205B obtained by the conversion is output.
- the sign key 403 quantifies the channel information 207 force in the case of indicating the front L channel and the front R channel and in the case of indicating the front L channel and the back L channel with different accuracy. To implement.
- the sign unit 403 In order to switch the quantization accuracy as described above, for example, the sign unit 403 previously holds a plurality of tables for converting input values into quantized values, from which a channel is converted. One corresponding to the information 207 is used.
- FIG. 7 schematically shows an example of a table used for quantization of inter-signal correlation, inter-signal similarity, and extended S-sense estimation value held in advance in the sign key unit 403.
- FIG. 7 schematically shows an example of a table used for quantization of inter-signal correlation, inter-signal similarity, and extended S-sense estimation value held in advance in the sign key unit 403.
- FIG. 7 schematically shows an example of a table used for quantization of inter-signal correlation, inter-signal similarity, and extended S-sense estimation value held in advance in the sign key unit 403.
- FIG. 7 schematically shows an example of a table used for quantization of inter-signal correlation, inter-signal similarity, and extended S-sense estimation value held in advance in the sign key unit 403.
- FIG. 7 schematically shows an example of a table used for quantization of inter-signal correlation, inter-signal similarity, and extended S-sense estimation value held in advance in the sign key unit 403.
- FIG. 7 schematic
- the sign key unit 403 shows the correlation between signals, the similarity between signals, and the sense of spread. If the estimated values are relatively quantized and the channel information 207 indicates the rear L channel and the front L channel, the correlation between signals, similarity between signals, and The spread feeling estimation value is quantized relatively roughly.
- FIG. 8 is a block diagram illustrating a functional configuration of the auxiliary information generation unit according to the third embodiment.
- the auxiliary information generation unit is a unit that generates auxiliary information 205C encoded according to channel information 207 from the first input signal 201 and the second input signal 202.
- the degree-of-interest calculation unit 401, the distance sensation estimation unit 502, and the sign key unit 503 are included.
- the first input signal 201 and the second input signal 202 are input to the inter-signal correlation calculation unit 401.
- the inter-signal correlation calculation unit 401 the cross-correlation value of the signals of the first input signal 201 and the second input signal 202 and the similarity (coherence) between the signals based on each input signal are described above. Calculate according to Equation 1 above. [0088] When calculating the similarity, the signal may be divided into a plurality of frequency bands and may be calculated for each of them, or one may be calculated for all bands. Further, the time unit to be calculated is not particularly limited.
- the similarity between signals obtained as an output from inter-signal correlation calculation section 401 and channel information 207 are input to distance sense estimation section 502.
- the distance sensation estimation unit 502 estimates the degree of distance perception of the sound image perceived by the listener from the channel information 207 and the similarity between signals obtained as an output from the inter-signal correlation calculation unit 401.
- the sense of distance of the sound image perceived by the listener is expressed by appropriately quantifying the distance and proximity perceived as psychological quantities.
- the distance sensation estimation unit 502 estimates, for example, the distance sensation of the sound image perceived by the listener based on this knowledge, and outputs a distance sensation estimation value representing the estimation result to the sign key unit 503.
- Encoding section 503 quantizes at least one of the inter-signal correlation, inter-signal similarity, and distance sense estimation value with different accuracy according to the above-described channel information 207, and further performs encoding. Auxiliary information 205C obtained by issue is output.
- the encoding unit 503 performs different quantization depending on whether the channel information 207 indicates the front L channel and the front R channel or not when the channel information 207 indicates the front L channel and the back L channel. carry out.
- the sign unit 503 holds, in advance, a plurality of tables for converting input values into quantized values, from which a channel is converted.
- One corresponding to the information 207 is used.
- a table a table similar to the table described in FIG. 7 can be used, and detailed description thereof is omitted here.
- the sign key unit 503 is based on the channel information 207, and the quantization accuracy that reflects the discrimination ability related to the sense of distance to the sound image felt by the listener (that is, is fine in the front direction).
- the quantization accuracy that becomes rougher from the left and right to the back direction is determined, and at least one of the correlation between signals, the similarity between signals, and the sense of distance is quantized with the determined quantization accuracy. Sign.
- the code key method configured in this way it is possible to execute the code key based on the characteristic of the sense of distance of the human sound image, and to efficiently execute the code key.
- the audio signal encoding apparatus is configured by combining the audio signal encoding apparatus according to the first, second, and third embodiments.
- the audio signal encoding apparatus has all the configurations shown in FIG. 3, FIG. 6, and FIG. 8, and is based on two input signal levels, a signal level difference, a signal phase difference, and a signal. Inter-correlation (similarity) is calculated, based on channel information, the direction of perception, spread, and distance are estimated, and the quantization method and quantization table are switched according to the channel information. Encoding is performed.
- Example 4 any two of Examples 1 to 3 may be combined.
- FIG. 9 is a block diagram showing an example of a functional configuration of the audio signal decoding apparatus of the present invention.
- the audio signal decoding apparatus includes a first output signal 105 and a second output signal that approximate the original sound signal from the mixed signal information 206, auxiliary information 205, and channel information 207 generated by the audio signal encoding apparatus.
- 106 includes a mixed signal decoding unit 102 and a signal separation processing unit 103.
- the present invention does not limit the specific method of transferring the mixed signal information 206, the auxiliary information 205, and the channel information 207 from the audio signal encoding device to the audio signal decoding device.
- the mixed signal information 206, the auxiliary information 205, and the channel information 207 are multiplexed and transmitted in one broadcast stream, and the audio signal decoding apparatus receives the broadcast stream and demultiplexes the mixed stream signal.
- mixed signal information 206, auxiliary information 205, and channel information 207 are stored in the recording medium, and the audio signal decoding apparatus receives the mixed signal information 206, auxiliary information from the recording medium. 205 and channel information 207 may be read out.
- transmission of the channel information 207 is omitted by predetermining the prescribed values and the prescribed order between the audio signal encoding device and the audio signal decoding device. Is also possible.
- the mixed signal decoding unit 102 decodes the mixed signal information 206 represented in the code data format into an audio signal format, and sends the decoded audio signal to the signal separation processing unit 103. Output.
- the mixed signal decoding unit 102 performs inverse conversion of the mixed signal code unit 203 in the above-described audio signal encoding device.
- the mixed signal encoding unit 203 stores the mixed signal information 206.
- the mixed signal decoding unit 102 also performs the inverse transformation defined in AAC to obtain the audio signal.
- the format of the audio signal is selected such as a signal format on the time axis, a signal format on the frequency axis, and a format expressed on both the time and frequency axes. In the present invention, the format is limited. do not do.
- the signal separation processing unit 103 Based on the auxiliary information 205 and the channel information 207, the signal separation processing unit 103 outputs the first output signal 105 and the second output signal 106 from the audio signal output from the mixed signal decoding unit 102. Generate and output.
- FIG. 10 is a block diagram showing a functional configuration of the signal separation processing unit 103 according to this embodiment.
- the signal separation processing unit 103 decodes the auxiliary information 205 using different decoding methods depending on the channel information 207, and uses the decoding result to output the first output signal 105 and the second output signal.
- 106 includes a decoding method switching unit 705, an inter-signal information decoding unit 706, and a signal synthesis unit 707.
- the decoding key method switching unit 705 switches the decoding key method to the inter-signal information decoding unit 706 based on the channel information 207. Instruct.
- the inter-signal information decoding unit 706 decodes the auxiliary information 702 into inter-signal information using the decoding method switched according to the instruction from the decoding method switching unit 705.
- This inter-signal information is the inter-signal level difference, the inter-signal phase difference, the inter-signal correlation, etc. described in the first to third embodiments.
- the inter-signal information decoding unit 706 can switch the decoding method by switching the table representing the quantization points, like the code unit in the audio signal encoding device.
- the decoding method may be switched by switching the inverse quantization function or the decoding key procedure itself.
- the signal synthesis unit 707 obtains the inter-signal level difference, inter-signal phase difference, and inter-signal correlation represented by the inter-signal information from the audio signal that is the output signal of the mixed signal decoding unit 704.
- a first output signal 105 and a second output signal 106 are generated.
- two signals obtained by duplicating the audio signal are given half of the inter-signal level difference in the opposite direction, and half of the inter-signal phase difference is given in the opposite direction.
- t which is a well-known method, is used as appropriate, in which the two signals after the level difference and the phase difference are added are mixed according to the degree of correlation between the signals.
- this decoding method is used to generate an n-channel audio signal into an audio signal of more than n channels by simply converting a 1-channel audio signal into a 2-channel audio signal.
- a 1-channel audio signal into a 2-channel audio signal.
- it is effective when obtaining a 6-channel audio signal from a 2-channel audio signal or obtaining a 6-channel audio signal from a 1-channel audio signal.
Abstract
Description
Claims
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/589,818 US20070160236A1 (en) | 2004-07-06 | 2005-07-01 | Audio signal encoding device, audio signal decoding device, and method and program thereof |
JP2006519623A JPWO2006004048A1 (ja) | 2004-07-06 | 2005-07-01 | オーディオ信号符号化装置、オーディオ信号復号化装置、方法、及びプログラム |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2004-199819 | 2004-07-06 | ||
JP2004199819 | 2004-07-06 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2006004048A1 true WO2006004048A1 (ja) | 2006-01-12 |
Family
ID=35782852
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2005/012221 WO2006004048A1 (ja) | 2004-07-06 | 2005-07-01 | オーディオ信号符号化装置、オーディオ信号復号化装置、方法、及びプログラム |
Country Status (4)
Country | Link |
---|---|
US (1) | US20070160236A1 (ja) |
JP (1) | JPWO2006004048A1 (ja) |
CN (1) | CN1922655A (ja) |
WO (1) | WO2006004048A1 (ja) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2007221216A (ja) * | 2006-02-14 | 2007-08-30 | Oki Electric Ind Co Ltd | ミックスダウン方法およびミックスダウン装置 |
JP2007310087A (ja) * | 2006-05-17 | 2007-11-29 | Mitsubishi Electric Corp | 音声符号化装置及び音声復号装置 |
WO2009025023A1 (ja) * | 2007-08-20 | 2009-02-26 | Pioneer Corporation | 音像定位予測装置及び音像定位制御システム並びに音像定位予測方法及び音像定位制御方法 |
KR20140047005A (ko) * | 2012-10-11 | 2014-04-21 | 한국전자통신연구원 | 오디오 데이터 생성 장치 및 방법, 오디오 데이터 재생 장치 및 방법 |
JP2018534625A (ja) * | 2015-11-20 | 2018-11-22 | クアルコム,インコーポレイテッド | 複数のオーディオ信号の符号化 |
JP2022505964A (ja) * | 2018-10-26 | 2022-01-14 | フラウンホーファー-ゲゼルシャフト・ツール・フェルデルング・デル・アンゲヴァンテン・フォルシュング・アインゲトラーゲネル・フェライン | 方向性音量マップベースのオーディオ処理 |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
ES2623551T3 (es) * | 2005-03-25 | 2017-07-11 | Iii Holdings 12, Llc | Dispositivo de codificación de sonido y procedimiento de codificación de sonido |
WO2009050896A1 (ja) * | 2007-10-16 | 2009-04-23 | Panasonic Corporation | ストリーム合成装置、復号装置、方法 |
US20110282674A1 (en) * | 2007-11-27 | 2011-11-17 | Nokia Corporation | Multichannel audio coding |
CN101350931B (zh) | 2008-08-27 | 2011-09-14 | 华为终端有限公司 | 音频信号的生成、播放方法及装置、处理系统 |
CN103718466B (zh) | 2011-08-04 | 2016-08-17 | 杜比国际公司 | 通过使用参量立体声改善fm立体声无线电接收器 |
CN103812824A (zh) * | 2012-11-07 | 2014-05-21 | 中兴通讯股份有限公司 | 音频多编码传输方法及相应装置 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH08275300A (ja) * | 1995-03-30 | 1996-10-18 | Yamaha Corp | 音場制御装置 |
JP2000505266A (ja) * | 1996-07-12 | 2000-04-25 | フラオホッフェル―ゲゼルシャフト ツル フェルデルング デル アンゲヴァンドテン フォルシュング エー.ヴェー. | ステレオ音響スペクトル値の符号化・解読方法 |
JP2000295698A (ja) * | 1999-04-08 | 2000-10-20 | Matsushita Electric Ind Co Ltd | バーチャルサラウンド装置 |
JP2002229598A (ja) * | 2001-02-01 | 2002-08-16 | Matsushita Electric Ind Co Ltd | ステレオ符号化信号復号化装置及び復号化方法 |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE4222623C2 (de) * | 1992-07-10 | 1996-07-11 | Inst Rundfunktechnik Gmbh | Verfahren zum Übertragen oder Speichern von digitalisierten Tonsignalen |
US7583805B2 (en) * | 2004-02-12 | 2009-09-01 | Agere Systems Inc. | Late reverberation-based synthesis of auditory scenes |
US20030035553A1 (en) * | 2001-08-10 | 2003-02-20 | Frank Baumgarte | Backwards-compatible perceptual coding of spatial cues |
US7006636B2 (en) * | 2002-05-24 | 2006-02-28 | Agere Systems Inc. | Coherence-based audio coding and synthesis |
US7447317B2 (en) * | 2003-10-02 | 2008-11-04 | Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V | Compatible multi-channel coding/decoding by weighting the downmix channel |
US7805313B2 (en) * | 2004-03-04 | 2010-09-28 | Agere Systems Inc. | Frequency-based coding of channels in parametric multi-channel coding systems |
-
2005
- 2005-07-01 CN CNA2005800055198A patent/CN1922655A/zh active Pending
- 2005-07-01 WO PCT/JP2005/012221 patent/WO2006004048A1/ja active Application Filing
- 2005-07-01 US US10/589,818 patent/US20070160236A1/en not_active Abandoned
- 2005-07-01 JP JP2006519623A patent/JPWO2006004048A1/ja active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH08275300A (ja) * | 1995-03-30 | 1996-10-18 | Yamaha Corp | 音場制御装置 |
JP2000505266A (ja) * | 1996-07-12 | 2000-04-25 | フラオホッフェル―ゲゼルシャフト ツル フェルデルング デル アンゲヴァンドテン フォルシュング エー.ヴェー. | ステレオ音響スペクトル値の符号化・解読方法 |
JP2000295698A (ja) * | 1999-04-08 | 2000-10-20 | Matsushita Electric Ind Co Ltd | バーチャルサラウンド装置 |
JP2002229598A (ja) * | 2001-02-01 | 2002-08-16 | Matsushita Electric Ind Co Ltd | ステレオ符号化信号復号化装置及び復号化方法 |
Cited By (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2007221216A (ja) * | 2006-02-14 | 2007-08-30 | Oki Electric Ind Co Ltd | ミックスダウン方法およびミックスダウン装置 |
JP2007310087A (ja) * | 2006-05-17 | 2007-11-29 | Mitsubishi Electric Corp | 音声符号化装置及び音声復号装置 |
WO2009025023A1 (ja) * | 2007-08-20 | 2009-02-26 | Pioneer Corporation | 音像定位予測装置及び音像定位制御システム並びに音像定位予測方法及び音像定位制御方法 |
JP5031840B2 (ja) * | 2007-08-20 | 2012-09-26 | パイオニア株式会社 | 音像定位予測装置及び音像定位制御システム並びに音像定位予測方法及び音像定位制御方法 |
US9836269B2 (en) | 2012-10-11 | 2017-12-05 | Electronics And Telecommunications Research Institute | Device and method for generating audio data, and device and method for playing audio data |
JP2015534656A (ja) * | 2012-10-11 | 2015-12-03 | エレクトロニクス アンド テレコミュニケーションズ リサーチ インスチチュートElectronics And Telecommunications Research Institute | オーディオデータ生成装置及び方法、オーディオデータ再生装置及び方法 |
KR20140047005A (ko) * | 2012-10-11 | 2014-04-21 | 한국전자통신연구원 | 오디오 데이터 생성 장치 및 방법, 오디오 데이터 재생 장치 및 방법 |
US10282160B2 (en) | 2012-10-11 | 2019-05-07 | Electronics And Telecommunications Research Institute | Apparatus and method for generating audio data, and apparatus and method for playing audio data |
KR102149411B1 (ko) * | 2012-10-11 | 2020-08-31 | 한국전자통신연구원 | 오디오 데이터 생성 장치 및 방법, 오디오 데이터 재생 장치 및 방법 |
JP2018534625A (ja) * | 2015-11-20 | 2018-11-22 | クアルコム,インコーポレイテッド | 複数のオーディオ信号の符号化 |
JP2019207430A (ja) * | 2015-11-20 | 2019-12-05 | クアルコム,インコーポレイテッド | 複数のオーディオ信号の符号化 |
US10586544B2 (en) | 2015-11-20 | 2020-03-10 | Qualcomm Incorporated | Encoding of multiple audio signals |
US11094330B2 (en) | 2015-11-20 | 2021-08-17 | Qualcomm Incorporated | Encoding of multiple audio signals |
JP2022505964A (ja) * | 2018-10-26 | 2022-01-14 | フラウンホーファー-ゲゼルシャフト・ツール・フェルデルング・デル・アンゲヴァンテン・フォルシュング・アインゲトラーゲネル・フェライン | 方向性音量マップベースのオーディオ処理 |
Also Published As
Publication number | Publication date |
---|---|
JPWO2006004048A1 (ja) | 2008-04-24 |
US20070160236A1 (en) | 2007-07-12 |
CN1922655A (zh) | 2007-02-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2006004048A1 (ja) | オーディオ信号符号化装置、オーディオ信号復号化装置、方法、及びプログラム | |
JP5694279B2 (ja) | 符号化装置 | |
JP5081838B2 (ja) | オーディオ符号化及び復号 | |
RU2618383C2 (ru) | Кодирование и декодирование аудиообъектов | |
Faller | Coding of spatial audio compatible with different playback formats | |
JP5134623B2 (ja) | 複数のパラメータ的に符号化された音源を合成するための概念 | |
JP4589962B2 (ja) | レベル・パラメータを生成する装置と方法、及びマルチチャネル表示を生成する装置と方法 | |
JP6039516B2 (ja) | 多チャンネルオーディオ信号処理装置、多チャンネルオーディオ信号処理方法、圧縮効率向上方法及び多チャンネルオーディオ信号処理システム | |
JP4944902B2 (ja) | バイノーラルオーディオ信号の復号制御 | |
JP5638037B2 (ja) | オーディオソースのパラメトリックジョイント符号化 | |
TWI415111B (zh) | 空間解碼器單元、空間解碼器裝置、音訊系統、消費型電子裝置、產生一對雙耳輸出聲道之方法及電腦可讀媒體 | |
JP2010515099A5 (ja) | ||
MX2011000371A (es) | Uso eficaz de la informacion de fase en la codificacion y decodificacion de audio. | |
KR20090053958A (ko) | 멀티 채널 파라미터 변환 장치 및 방법 | |
WO2005122639A1 (ja) | 音響信号符号化装置および音響信号復号装置 | |
WO2005112002A1 (ja) | オーディオ信号符号化装置及びオーディオ信号復号化装置 | |
JP2009531955A (ja) | フィルタ適応周波数分解能 | |
CN112005560B (zh) | 使用元数据处理音频信号的方法和设备 | |
KR20080078907A (ko) | 양 귀 오디오 신호들의 복호화 제어 | |
WO2006011367A1 (ja) | オーディオ信号符号化装置および復号化装置 | |
KR20190060464A (ko) | 오디오 신호 처리 방법 및 장치 | |
Breebaart et al. | 19th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KM KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NG NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SM SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU LV MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2006519623 Country of ref document: JP |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2007160236 Country of ref document: US Ref document number: 10589818 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 200580005519.8 Country of ref document: CN |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWW | Wipo information: withdrawn in national office |
Country of ref document: DE |
|
WWP | Wipo information: published in national office |
Ref document number: 10589818 Country of ref document: US |
|
122 | Ep: pct application non-entry in european phase |