WO2010070016A1 - Method and apparatus for applying reverb to a multi-channel audio signal using spatial cue parameters - Google Patents

Method and apparatus for applying reverb to a multi-channel audio signal using spatial cue parameters Download PDF

Info

Publication number
WO2010070016A1
WO2010070016A1 PCT/EP2009/067350 EP2009067350W WO2010070016A1 WO 2010070016 A1 WO2010070016 A1 WO 2010070016A1 EP 2009067350 W EP2009067350 W EP 2009067350W WO 2010070016 A1 WO2010070016 A1 WO 2010070016A1
Authority
WO
WIPO (PCT)
Prior art keywords
reverb
channel
parameters
channel signals
reverberator
Prior art date
Application number
PCT/EP2009/067350
Other languages
French (fr)
Inventor
Jonas ENGDEGÅRD
Original Assignee
Dolby Sweden Ab
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Family has litigation
First worldwide family litigation filed litigation Critical https://patents.darts-ip.com/?family=41796192&utm_source=google_patent&utm_medium=platform_link&utm_campaign=public_patent_search&patent=WO2010070016(A1) "Global patent litigation dataset” by Darts-ip is licensed under a Creative Commons Attribution 4.0 International License.
Application filed by Dolby Sweden Ab filed Critical Dolby Sweden Ab
Priority to KR1020117016794A priority Critical patent/KR101342425B1/en
Priority to JP2011541416A priority patent/JP5524237B2/en
Priority to CN2009801511567A priority patent/CN102257562B/en
Priority to EP09801205.7A priority patent/EP2377123B1/en
Priority to US13/132,321 priority patent/US8965000B2/en
Priority to RU2011129154/08A priority patent/RU2509442C2/en
Priority to BRPI0923174-9A priority patent/BRPI0923174B1/en
Publication of WO2010070016A1 publication Critical patent/WO2010070016A1/en

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/008Multichannel audio signal coding or decoding using interchannel correlation to reduce redundancy, e.g. joint-stereo, intensity-coding or matrixing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S3/00Systems employing more than two channels, e.g. quadraphonic
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10KSOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
    • G10K15/00Acoustics not otherwise provided for
    • G10K15/08Arrangements for producing a reverberation or echo sound
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/305Electronic adaptation of stereophonic audio signals to reverberation of the listening space

Definitions

  • the invention relates to methods and systems for applying reverb to a multichannel downmixed audio signal indicative of a larger number of individual audio
  • this is done by upmixingthe input signal and applying reverb to at least some of its individual channels in response to at least one spatial cue parameter (indicative of least one spatial cue for the input signal) so as to apply different reverb impulse responses for each of the individual channels to which reverb is applied.
  • the individual channels are downmixed to generate
  • the input signal is a QMF (quadrature mirror filter) domain MPEG Surround (MPS) encoded signal, and the upmixing and reverb application are performed in the QMF domain in response to MPS spatial cue parameters including at least some of Channel Level Difference (CLD), Channel Prediction Coefficient (CPC), and Inter-channel Cross Correlation (ICC)
  • CLD Channel Level Difference
  • CPC Channel Prediction Coefficient
  • ICC Inter-channel Cross Correlation
  • reverberator (or “reverberator system”) is used to denote a system configured to apply reverb to an
  • audio signal (e.g., to all or some channels of a multi-channel audio signal).
  • system is used in a broad sense to denote a device, system, or subsystem.
  • a subsystem that implements a reverberator may be referred to as a reverberator system (or reverberator), and a system including such a reverberator subsystem (e.g., a decoder
  • reverberator system (or reverberator).
  • production of signals by speakers denotes causing the speakers to produce sound in response to the signals, including by performing any required amplification and/or other processing of the signals.
  • linear combination of values vi, V2, ..., V n , (e.g., n elements of a subset of a set of X individual audio channel signals occuring at a time, t, where n is less than or equal to X) denotes a value equal to aivi + a2V2 + ...+ a n v n , where a ⁇ , a 2 , ..., a n are coefficients.
  • each coefficient can be positive or negative or zero).
  • the expression is used in a broad sense herein, for example to cover the case that one of the coefficients is equal to 1 and the others are equal to zero (e.g., the case that the linear combination aivi + a2V2 + ...+ a n v n is equal to vi(or v2, ..., or V n ).
  • spatial cue parameter of a multichannel audio signal denotes any parameter indicative of at least one spatial cue for the audio signal, where each such "spatial cue” is indicative (e.g., descriptive) of the spatial image of the multichannel signal.
  • spatial cues are level (or intensity) differences between (or ratios of) pairs of the channels of the audio signal, phase differences between such channel pairs, and measures of correlation between such channel pairs.
  • spatial cue parameters are the Channel Level Difference (CLD) parameters and Channel Prediction Coefficient (CPC) parameters which are part of a conventional MPEG Surround (“MPS”) bitstream, and which are employed in MPEG surround coding.
  • CLD Channel Level Difference
  • CPC Channel Prediction Coefficient
  • M M channels
  • N N decoded audio channels
  • a typical, conventional MPS decoder is operable to perform upmixingto generate N decoded audio channels (where N is greater than two) in response to a time-domain, 2- channel, downmixed audio input signal (and MPS spatial cue parameters including Channel Level Difference and Channel Prediction Coefficient parameters).
  • a typical, conventional MPS decoder is operable in a binaural mode to generate a binaural signal in response to a time-domain, 2-channel, downmixed audio input signal and spatial cue parameters, and in at least one other mode to perform upmixing to generate 5 0 (where the notation "x.y” channels denotes “x” full frequency channels and "y” subwoofer channels), 5 1, 7.0, or 7.1 decoded audio channels in response to a time-domain, 2- channel, downmixed audio input signal and spatial cue parameters.
  • the input signal undergoes time domain-to-frequency domain transformation into the QMF (quadrature mirror filter) domain, to generate two channels of QMF domain frequency components
  • QMF quadrature mirror filter
  • N is greater than two, and N is typically equal to 5 or 7
  • L' and R' 2-channel downmixed audio signal
  • MPS spatial cue parameters including Channel Level Difference parameters and Channel Prediction Coefficient parameters
  • the downmixed input signal (L' and R') is indicative of "X" individual audio channels, where X is greater than 2.
  • the downmixed input signal is typically indicative of five individual channels (e.g., left-front, right front, center, left-surround, and right-surround channels)
  • Each of the "left" input signal L' and the "right” input signal R' is a sequence of QMF domain frequency components generated by transforming a 2 channel, time- domain MPS encoded signal (not indicated in Fig 1) in a time doma ⁇ n-to QMF domain transform stage (not shown in Fig 1)
  • the downmixed input signals L' and R' are decoded into IM individual channel signals Sl, S2, . ., SN, in decoder 1 of Fig 1, in response to the MPS spatial cue parameters which are asserted (with the input signals) to the Fig. 1 system.
  • the N sequences of output QMF domain frequency components, Sl, S2, .., SN are typically transformed back into the time domain by a QMF domain to-time domain transform stage (not shown in Fig 1), and can be asserted as output from the system without undergoing post-processing
  • SN undergo postprocessing (in the QMF domain) in post-processor 5 to generate an N-channel audio output signal comprising channels OUTl, OUT2, , OUTN
  • the N sequences of output QMF domain frequency components, OUTl, OUT2, . , OUTN are typically transformed back into the time domain by a QMF doma ⁇ n-to time domain transform stage (not shown in Fig 1), and asserted as output from the system
  • a binaural mode operating in a binaural mode generates 2 channel binaural audio output Sl and S2, and optionally also 2-channel binaural audio output OUTl and 0UT2, in response to a 2-channel downmixed audio signal (L' and R') and MPS spatial cue parameters (including Channel Level Difference parameters and Channel Prediction Coefficient parameters).
  • L' and R' 2-channel downmixed audio signal
  • MPS spatial cue parameters including Channel Level Difference parameters and Channel Prediction Coefficient parameters
  • post-processor 5 When reproduced by a pair of headphones, the 2-chan ⁇ el audio output Sl and S2 is perceived at the listener's eardrums as sound from "X" loudspeakers (where X >2 and X is typically equal to 5 or 7) at any of a wide variety of positions (determined by the coefficients of decoder 1), including positions in front of and behind the listener In the binaural mode, post-processor 5 can apply reverb to the 2-channel output (Sl, S2) of decoder 1 (in this case, post-processor 5 implements an artificial reverberator).
  • X loudspeakers
  • the 2-channel output of post-processor 5 (OUTl and OUT2) is a binaural audio output to which reverb has been applied, and which when reproduced by headphones is perceived at the listener's eardrums as sound from "X" loudspeakers (where X >2 and X is typically equal to 5) at any of a wide variety of positions, including positions in front of and behind the listener
  • Reproduction of signals Sl and S2 (or OUTl and OUT2) generated during binaural mode operation of the Fig. 1 decoder can give the listener the experience of sound that comes from more than two (e g , five) "surround” sources At least some of these sources are virtual. More generally, it is conventional for virtual surround systems to use head- related transfer functions (HRTFs) to generate audio signals (sometimes referred to as virtual surround sound signals) that, when reproduced by a pair of physical speakers (e g , loudspeakers positioned in front of a listener, or headphones) are perceived at the listener's eardrums as sound from more than two sources (e.g., speakers) at any of a wide variety of positions (typically including positions behind the listener).
  • HRTFs head- related transfer functions
  • the MPS decoder of Fig. 1 operating in the binaural mode could be implemented to apply reverb using an artificial reverberator implemented by postprocessor 5.
  • This reverberator could be configured to generate reverb in response to the two-channel output (Sl, S2) of decoder 1 and to apply the reverb to the signals Sl and S2 to generate reverbed two-channel audio OUTl and 0UT2.
  • the reverb would be applied as a post process stereo-to-stereo reverb to the 2-channel signal Sl, S2 from decoder 1, such that the same reverb impulse response is applied to all discrete channels determined by one of the two downmixed audio channels of the binaural audio output of decoder 1 (e.g., to left-front and left-surround channels determined by downmixed channel Sl), and the same reverb impulse response is applied to all discrete channels determined by the other one of the two downmixed audio channels of the binaural audio (e g., to right front and right-surround channels determined by downmixed channel S2).
  • a reverberator has what is known as a Feedback Delay Network-based (FDN-based) structure
  • FDN-based Feedback Delay Network-based
  • An advantage of this structure relative to other reverb structures is the ability to efficiently produce and apply multiple uncorrelated reverb signals to multiple input signals
  • This feature is exploited in the commercially available Dolby Mobile headphone virtualizer which includes a reverberator having FDN based structure and is operable to apply reverb to each channel of a five channel audio signal (having left-front, right-front, center, left- surround, and right surround channels) and to filter each reverbed channel using a different filter pair of a set of five head related transfer function (“HRTF”) filter pairs
  • HRTF head related transfer function
  • the Dolby Mobile headphone virtualizer is also operable in response to a two- channel audio input signal, to generate a two channel "reverbed" audio output (a two- channel virtual surround sound output to which reverb has been applied)
  • a two channel "reverbed” audio output a two- channel virtual surround sound output to which reverb has been applied
  • the reverbed audio output is reproduced by a pair of headphones, it is perceived at the listener's eardrums as H RTF-filtered, reverbed sound from five loudspeakers at left front, right front, center, left rear (surround), and right rear (surround) positions.
  • the virtualizer upmixes a downmixed two-channel audio input (without using any spatial cue parameter received with the audio input) to generate five upmixed audio channels, applies reverb to the upmixed channels, and downmixes the five reverbed channel signals to generate the two-channel reverbed output of the virtualizer.
  • the reverb for each upmixed channel is filtered in a different pair of HRTF filters.
  • the inventor has recognized that it would be desirable to individually determine (and generate) a different reverb signal for each of the discrete channels of an upmix determined from downmixed audio, from each of the discrete channels of the upmix, or to determine and generate a different reverb signal for (and from) each of a set of linear combinations of values of such discrete channels.
  • the inventor has also recognized that with such individual determination of reverb signals for the individual upmix channels (or linear combinations of values of such channels), reverb having a different reverb impulse response can be applied to the upmix channels (or linear combinations)
  • the invention is a method for applying reverb to an M- channel downmixed audio input signal indicative of X individual audio channels, where X is a number greater than M.
  • the method includes the steps of:
  • each of the reverb channel signals at a time, f is a linear combination of at least a subset of values of the X individual audio channels at the time, t;
  • reverb applied to at least one of the reverb channel signals has a different reverb impulse response than does the reverb applied to at least one other one of the reverb channel signals.
  • X Y, but in other embodiments X is not equal to Y.
  • Y is greater than M, and the input signal is upmixed in step (a) in response to the spatial cue parameters to generate the Y reverb channel signals. In other embodiments, Y is equal to M orY is less than M.
  • the input signal is a sequence of values L(t), R(t) indicative of five individual channel signals, Lfront, Rfront, C, Lsur, and R S ur.
  • Each of the five individual channel signals is a sequence of values
  • W is an MPEG Surround upmix matrix of form W 12
  • the input signal is an M-channel, MPEG Surround (“MPS") downmixed signal
  • steps (a) and (b) are performed in the QMF domain
  • the spatial cue parameters are received with the input signal.
  • the spatial cue parameters may be or include Channel Level Difference (CLD) parameters and/or Channel Prediction Coefficient (CPC) parameters of the type comprising part of a conventional MPS bitstream.
  • CLD Channel Level Difference
  • CPC Channel Prediction Coefficient
  • the invention typically includes the step of transforming this time-domain signal into the QMF domain to generate QMF domain frequency components, and performing steps (a) and (b) in the QMF domain on these frequency components.
  • the method also includes a step of generating an N-channel downmixed version of the Y reverbed channel signals (including each of the channel signals to which reverb has been applied and each of the channel signals, if any, to which reverb has not been applied), for example by encoding the reverbed channel signals as an N-channel, downmixed MPS signal.
  • the input downmixed signal is a 2-channel downmixed MPEG Surround (“MPS”) signal indicative of five individual audio channels (left-front, right-front, center, left-surround, and right surround channels), and reverb determined by a different reverb impulse response is applied to each of at least some of these five channels, resulting in improved surround sound quality.
  • MPS MPEG Surround
  • the inventive method also includes a step of applying to the reverbed channel signals corresponding head-related transfer functions (HRTFs), by filtering the reverbed channel signals in an HRTF filter.
  • HRTFs head-related transfer functions
  • Other aspects of the invention are a reverberator configured (e.g., programmed) to perform any embodiment of the inventive method, a virtualizer including such a reverberator, a decoder (e g , an MPS decoder) including such a reverberator, and a computer readable medium (e g , a disc) which stores code for implementing any embodiment of the inventive method
  • FIG 1 is a block diagram of a conventional MPEG Surround decoder system.
  • FIG 2 is a block diagram of a multiple input, multiple output, FDN based reverberator (100) that can be implemented in accordance with an embodiment of the present invention
  • FIG 3 is a block diagram of a reverberator system including reverberator 100 of
  • Fig. 2 conventional MPS processor 102, time doma ⁇ n-to-QMF domain transform filter 99 for transforming a multi-channel input into the QMF domain for processing in reverberator 100 and processor 102, and QMF domain-to-time domain transform filter
  • the invention is a method for applying reverb to an M- channel downmixed audio input signal indicative of X individual audio channels, where X is a number greater than M, and a system configured to perform the method.
  • the method includes the steps of:
  • each of the reverb channel signals at a time, t is a linear combination of at least a subset of values of the X individual audio channels at the time, t, and (b) individually applying reverb to each of at least two of the reverb channel signals (e.g., in the QMF domain), thereby generating Y reverbed channel signals.
  • spatial cue parameters indicative e.g , descriptive
  • QMF quadrature mirror filter
  • the reverb applied to at least one of the reverb channel signals has a different reverb impulse response than does the reverb applied to at least one other one of the reverb channel signals.
  • X Y, but in other embodiments X is not equal to Y.
  • Y is greater than M, and the input signal is upmixed in step (a) in response to the spatial cue parameters to generate the Y reverb channel signals. In other embodiments, Y is equal to IvI or Y is less than M.
  • FIG. 2 is a block diagram of multiple input, multiple output, FDN-based reverberator 100 which can be implemented in a manner to be explained below to perform this method.
  • Each of the reverb channel signals at a time, t is a linear combination of a subset of values of the X individual upmix audio channels at the time, t.
  • matrix B upmixes the input signal to generate the reverb channel signals.
  • M is equal to 2.
  • Element 40 is configured to add the output of gain element gl (i.e., apply feedback from the output of gain element gl) to reverb channel signal Ul.
  • Element 41 is configured to add the output of gain element g2 to reverb channel signal U2.
  • Element 42 is configured to add output of gain element g3 to reverb channel signal U3.
  • Element 43 is configured to add the output of gain element g4 to reverb channel signal U4; scattering matrix 32 (matrix "A"), which is coupled to receive the outputs of addition elements 40, 41, 42, and 43.
  • Matrix 32 is preferably a 4 * 4 unitary matrix configured to assert a filtered version of the output of each of addition elements 40, 41, 42, and 43 to a corresponding one of delay lines, z 'M> , where 0 ⁇ k - 1 ⁇ 3, and is preferably a fully populated matrix in order to provide maximum diffuseness.
  • z-MA ⁇ are labeled respectively as delay lines 50, 51, 52, and 53 in Fig. 2; gain elements, gk, where 0 ⁇ k - 1 ⁇ 3, which apply gain the outputs of delay lines, z 'Mt , thus providing damping factors for controlling the decay time of the reverb applied in each upmix channel.
  • Each gain element, gk is typically combined with a low- pass filter.
  • the gain elements apply different, predetermined gain factors for the different QMF bands.
  • Reverbed channel signals Rl, R2, R3, and R4, respectively, are asserted at the outputs of gain elements gl, g2, g3, and g4; and post-mix matrix 34 (matrix "C"), which is an N * 4 matrix coupled and configured to downmix and/or upmix (and optionally to perform other filtering on) the reverbed channel signals Rl, R2, R3, and R4 asserted at the outputs of gain elements gk, in response to at least a subset (e.g., all or some) of the spatial cue parameters asserted to matrix 30, thereby generating an N-channel, QMF domain, downmixed, reverbed audio output signal comprising channels Sl, S2, ..., and SN.
  • matrix 34 is a constant matrix whose coefficients do not vary with time in response to any spatial cue parameter.
  • the inventive system has Y reverb channels (where Y is less than or greater than four), pre-mix matrix 30 is configured to generate Y discrete reverb channel signals in response to the downmixed, M-channel, input signal and the spatial cue parameters, scattering matrix 32 is replaced by an YxY matrix, and the inventive system has Y delay lines, z 'Mt .
  • a pre-mix matrix (a variation on matrix 30 of Fig. 2) generates two discrete reverb channel signals (e.g., in the quadrature mirror filter or "QMF" domain): one a mix of the front channels; the other a mix of the surround channels.
  • Reverb having a short decay response is generated from (and applied to) one reverb channel signal and reverb having a long decay response is generated from (and applied to) the other reverb channel signal (e.g., to simulate a room with "live end/dead end” acoustics).
  • post-processor 36 optionally is coupled to the outputs of matrix 34 and operable to perform post-processing on the downmixed, reverbed output Sl, S2, ..., SN of matrix 34, to generate an N-channel post-processed audio output signal comprising channels OUTl, 0UT2, ..., and OLJTN.
  • N 2
  • the Fig. 2 system outputs a binaural, downmixed, reverbed audio signal Sl, S2 and/or a binaural, post-processed, downmixed, reverbed audio output signal OUT, 0UT2.
  • L left
  • C center
  • R right
  • LS left-surround
  • RS right-surround
  • post-mix matrix 34 is omitted and the inventive reverberator outputs Y-channel reverbed audio (e.g., upmixed, reverbed audio) in response to an M-channel downmixed audio input.
  • matrix 34 is an identity matrix.
  • the Fig. 2 system has four reverb channels and four delay lines, z 'M " , variations on the system (and other embodiments of the inventive reverberator) implement more than or less than four reverb channels.
  • the inventive reverberator includes one delay line per reverb channel.
  • the input signal is an M-channel
  • the input signal asserted to the inputs of matrix 30 comprises QMF domain signals INl(t,f), IN2(t,f) and INM(t,f), and the Fig. 2 system performs processing (e.g., in matrix 30) and reverb application thereon in the QMF domain.
  • the spatial cue parameters asserted to matrix 30 are typically Channel Level Difference (CLD) parameters and/or Channel Prediction
  • CPC Coefficient
  • ICC Inter-channel Cross Correlation
  • the inventive method would include a preliminary step of transforming this time-domain signal into the QMF domain to generate QMF domain frequency components, and would perform above-described steps (a) and (b) in the QMF domain on these frequency components.
  • the FIG. 3 system includes filter 99 for transforming this time-domain signal into the QMF domain
  • the FIG 3 system includes reverberator 100 (corresponding to and possibly identical to reverberator 100 of Fig 2), conventional MPS processor 102, time domain- to-QMF domain transform filter 99 coupled and configured to transform each of the time domain input channels ll(t), I2(t), .
  • FIG. 3 system also includes QMF domain- to time domain transform filter 101, which is coupled and configured to transform the N channel combined output of reverberator 100 and processor 102 into the time domain Specifically, filter 99 transforms time-domain signals ll(t), I2(t), , and IM(t) respectively into QMF domain signals INl(t,f), !N2(t,f), ..., and INM(t,f), which are asserted to reverberator 100 and processor 102.
  • QMF domain- to time domain transform filter 101 which is coupled and configured to transform the N channel combined output of reverberator 100 and processor 102 into the time domain
  • filter 99 transforms time-domain signals ll(t), I2(t), , and IM(t) respectively into QMF domain signals INl(t,f), !N2(t,f), ..., and INM(t,f), which are asserted to reverberator 100 and processor 102.
  • Each of the N channels output from processor 102 is combined (in an adder) with the corresponding reverbed channel output of reverberator 100 (Sl, S2, ..., or SN indicated in Fig 2, or one of OUTl, 0UT2, ..., or OUTN indicated in Fig. 2 if reverberator 100 of Fig 3 also includes a post processor 36 as shown in Fig 2).
  • Filter 101 of Fig. 3 transforms the combined (reverbed) output of reverberator 100 and processor 102 (N sequences of QMF domain frequency components Sl'(t, f), S2 (t,f), .
  • the input downmixed signal is a 2- channel downmixed MPS signal indicative of five individual audio channels (left-front, right-front, center, left-surround, and right surround channels), and reverb determined by a different reverb impulse response is applied to each of these five channels, resulting in improved surround sound quality.
  • the coefficients of the constant matrices B and C would not change as a function of time in response to spatial cue parameters indicative of the downmixed input audio, and the so-modified Fig. 2 system would operate in a conventional stereo-to-stereo reverb mode.
  • reverb having the same reverb impulse response would be applied to each individual channel in the downmix (i.e., left-front channel content in the downmix would receive reverb having the same impulse response as would right-front channel content in the downmix).
  • the Fig. 2 system can produce and apply reverb to each reverb channel determined by the downmixed input to the system, with individual reverb impulse responses for each of the reverb channels.
  • less reverb is applied in accordance with the invention to a center channel (for clearer speech/dialog) than to at least one other reverb channel so that the impulse response of the reverb applied each of these reverb channels is different.
  • the impulse responses of the reverb applied to different reverb channels are not based on different channel routing to matrix 30 and are instead simply different scale factors applied by pre-mix matrix 30 or post-mix matrix 34 (and/or at least one other system element) to different reverb channels.
  • matrix 30 is a 4 x 2 matrix having time-varying coefficients which depend on current values of coefficients, wij, where / ranges from 1 to 3 and j ranges from 1 to 2.
  • the input signal is a sequence of QMF domain value pairs
  • INl(t,f) L(t)
  • IN2(t,f) R(t)
  • Lf rO nt a sequence of values of five individual channel signals
  • the time-varying coefficients of matrix 30 would depend also on the following four, time-varying channel gain values, in which CLD Xi ls is the current value of the left front/surround CLD parameter, and CLD n IS is the current value of the right front/surround CLD parameter;
  • the time-varying coefficients of matrix 30 would be:
  • the matrix multiplication performed by matrix 30 (having the coefficients shown in Equation 3) can be represented as:
  • matrix multiplication is equivalent to an upmix to five individual channel signals (by the MPEG Surround upmix matrix W defined above) followed by a downmix of these five signals to the four reverb channel signals by matrix Bo.
  • matrix 30 is implemented with the following coefficients:
  • K hf , K Rf , K c , K 15 and K RS are fixed reverb gain values for the different channels, and gif, g ⁇ s , grf, gif, and wii to W32 are as in Equation 2 and Ia, respectively.
  • the four fixed reverb gain values are substantially equal to each other, except that K 0 typically has a slightly lower value than the others (a few decibels lower than the values of the others) in order to apply less reverb to the center channel (e.g., for dryer sounding speech/dialog).
  • Matrix 30 implemented with the coefficients of Equation 4, is equivalent to the product of the MPEG Surround upmix matrix W defined above and the following downmix matrix Bo:
  • reverb channels Ul, U2, U3, and U4 would cause reverb channels Ul, U2, U3, and U4, respectively, to be the left-front upmix channel (feeding branch 1' of the Fig. 2 system), the right-front upmix channel (feeding branch 2' of the Fig. 2 system), the left-surround upmix channel (feeding branch 3' of the Fig. 2 system), and a combined right-surround and center upmix channel (the right-surround channel plus the center channel) feeding branch 4' of the Fig. 2 system.
  • the reverb individually applied to the four branches of the Fig. 2 system would have individually determined impulse responses.
  • matrix 30's coefficients are determined in another manner in response to available spatial cue parameters.
  • matrix 30's coefficients are determined in response to available MPS spatial cue parameters to cause matrix 30 to implement a TTT upmixer operating in a mode other than in a prediction mode (e.g., an energy mode with or without center subtraction). This can be done in a manner that will be apparent to those of ordinary skill in the art given the present description, using the well known upmixing formulas for the relevant cases that are described in the MPEG standard (ISO/IEC 23003-1:2007).
  • matrix 30 is a 4 x 1 matrix having time-varying coefficients:
  • the coefficients are gain factors are derived from the CLD parameters CiD 11 h , CLD n IS , CLD 0 h and CLD 1 , , available as part of a conventional MPS bitstream.
  • discrete reverb channels e.g., upmix channels
  • other spatial cue parameters are employed to upmix a downmixed input signal (e.g., including by control channel weighting).
  • ICC parameters available as part of a conventional MPS bitstream) that describe front-back diffuseness are used to determine coefficients of the pre-mix matrix and thereby to control reverb level.
  • the inventive method also includes a step of applying to the reverbed channel signals corresponding head-related transfer functions (HRTFs), by filtering the 5 reverbed channel signals in an HRTF filter.
  • HRTFs head-related transfer functions
  • matrix 34 of the Fig. 2 system is preferably implemented as the HRTF filter which applies such HRTFs to, and also performs the above-described downmixing operation on, reverbed channels Rl, R2, R3, and R4.
  • Such implementation of matrix 34 would typically perform the same filtering as a 5 ⁇ 4 matrix followed by a 2 * 5 matrix, where the 5 x 4 matrix generates five virtual
  • the HTRF filter would typically perform for each individual QMF band a matrix multiplication by a matrix with complex valued entries.
  • MPS encoded, downmixed input signal are filtered with corresponding HRTFs as follows.
  • the HRTFs in the parametric QMF domain essentially consist of left and right gain parameter values and Inter-channel Phase Difference (IPD) parameter values that characterize the downmixed input signal.
  • IPD Inter-channel Phase Difference
  • the HRTFs are constant gain values (four gain values for each of the left and the right channel, respectively): gg R i Fj r L >
  • fractional delay is applied in at least one reverb channel, and/or reverb is generated and applied differently to different frequency bands of frequency components of audio data in at least one reverb channel.
  • Some such preferred implementations of the inventive reverberator are variations on the Fig. 2 system that are configured to apply fractional delay (in at least one reverb channel) as well as integer sample delay.
  • a fractional delay element is connected in each reverb channel in series with a delay line that applies integer delay equal to an integer number of sample periods (e.g., each fractional delay element is positioned after or otherwise in series with one of delay lines 50, 51, 52, and 53 of Fig. 2).
  • Some of the above-noted preferred implementations of the inventive reverberator are variations on the Fig. 2 system that are configured to apply reverb differently to different frequency bands of the audio data in at least one reverb channel, in order to reduce complexity of the reverberator implementation.
  • the audio input data, INl-INM are QMF domain MPS data
  • the reverb application is performed in the QMF domain
  • the reverb is applied differently to the following four frequency bands of the audio data in each reverb channel:
  • reverb is applied in this band as in the above- described embodiment of Fig. 2, with matrix 30 implemented with the coefficients of Equation 4);
  • One of such eight lowest frequency bands can be used as a complex QMF buffer band, so that complex-valued arithmetic calculations are performed for only seven of the eight lowest QMF frequency bands (so that reverb is applied in this relatively low frequency range as in the above-described embodiment of Fig. 2, with matrix 30 implemented with the coefficients of Equation 4), and real-valued arithmetic calculations are performed for the other 56 QMF frequency bands, with the crossover between complex valued and real valued calculations occurring at the frequency (7 * 44.1kHz)/(64*2) which is approximately equal to 2.4 kHz.
  • reverb is applied in the relatively high frequency range as in the above-described Fig.
  • Reverb is applied in the relatively low frequency range (below 2.4 kHz) as in the Fig. 2 embodiment, e.g., with matrix 30 implemented with the coefficients of Equation 4); 8 kHz- 15 kHz: reverb is applied in this band by a simple delay technique. For example, reverb is applied in a way similar to the manner it is applied the above- described Fig.
  • pre-mix matrix 30 e.g., to apply less reverb to the center channel than to each other channel
  • the two delay branches can be simply fed to left and right outputs, respectively, or can be switched so that echoes from the left front (Lf) and left surround (Ls) channels end up in the right output channel and echoes from the right front (Rf) and right surround (Rs) channels end up in the left output channel
  • the 2 ⁇ 2 pre-mix matrix can have the following coefficients:
  • the inventive system applies reverb to an M-channel downmixed audio input signal indicative of X individual audio channels, where X is a number greater than M, including by generating Y discrete reverb channel signals in response to the downmixed signal but not in response to spatial cue parameters
  • the system individually applies reverb to each of at least two of the reverb channel signals in response to spatial cue parameters indicative of spatial image of the downmixed input signal, thereby generating Y reverbed channel signals
  • the coefficients of a pre mix matrix e.g., a variation on matrix 30 of Fig 2
  • a scattering matrix e.g., a variation on matrix 32 of Fig.
  • a gain stage e.g., a variation on the gain stage comprising elements gl-gk of Fig. 2
  • a post-mix matrix e.g., a variation on matrix 34 of Fig. 2 operates on the reverb channel signals in a manner determined by spatial cue parameters indicative of spatial image of the downmixed input signal, to apply reverb to each of at least two of the reverb channel signals.
  • the inventive reverberator is or includes a general purpose processor coupled to receive or to generate input data indicative of an M- channel downmixed audio input signal, and programmed with software (or firmware) and/or otherwise configured (e.g., in response to control data) to perform any of a variety of operations on the input data, including an embodiment of the inventive method.
  • a general purpose processor would typically be coupled to an input device (e.g , a mouse and/or a keyboard), a memory, and a display device.
  • an input device e.g , a mouse and/or a keyboard
  • 3 system could be implemented in a general purpose processor, with inputs ll(t), I2(t), ..., IM(t), being input data indicative of M channels of downmixed audio data, and outputs Sl(t), S2(t), .., SN(t), being output data indicative of N channels of downmixed, reverbed audio.
  • a conventional digital-to-analog converter could operate on this output data to generate analog versions of the output audio signals for reproduction by speakers (e.g., a pair of headphones)
  • speakers e.g., a pair of headphones

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Mathematical Physics (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Stereophonic System (AREA)

Abstract

A method and system for applying reverb to an M-channel downmixed audio input signal indicative of X individual audio channels, where X is greater than M. Typically, the method includes steps of : in response to spatial cue parameters indicative of spatial image of the downmixed input signal, generating Y discrete reverb channel signals, where each of the reverb channel signals at a time, t, is a linear combination of at least a subset of values of the individual audio channels at the time, t, and individually applying reverb to each of at least two of the reverb channel signals, thereby generating Y reverbed channel signals. Preferably, the reverb applied to at least one of the channel signals has a different reverb impulse response than does the reverb applied to at least one other one of the channel signals., t, is a linear combination of at least a subset of values of the individual audio channels at the time, t, and individually applying reverb to each of at least two of the reverb channel signals, thereby generating Y reverbed channel signals. Preferably, the reverb applied to at least one of the channel signals has a different reverb impulse response than does the reverb applied to at least one other one of the channel signals.

Description

METHOD AND APPARATUS FOR APPLYING REVERB TO A MULTI-CHANNEL AUDIO SIGNAL USING SPATIAL CUE PARAMETERS
BACKGROUND OF THE INVENTION
1. Field of the Invention
The invention relates to methods and systems for applying reverb to a multichannel downmixed audio signal indicative of a larger number of individual audio
10 channels. In some embodiments, this is done by upmixingthe input signal and applying reverb to at least some of its individual channels in response to at least one spatial cue parameter (indicative of least one spatial cue for the input signal) so as to apply different reverb impulse responses for each of the individual channels to which reverb is applied. Optionally, after application of reverb the individual channels are downmixed to generate
15 an N-channel reverbed output signal. In some embodiments the input signal is a QMF (quadrature mirror filter) domain MPEG Surround (MPS) encoded signal, and the upmixing and reverb application are performed in the QMF domain in response to MPS spatial cue parameters including at least some of Channel Level Difference (CLD), Channel Prediction Coefficient (CPC), and Inter-channel Cross Correlation (ICC)
20 parameters.
2. Background of the Invention
Throughout this disclosure including in the claims, the expression "reverberator" (or "reverberator system") is used to denote a system configured to apply reverb to an
25 audio signal (e.g., to all or some channels of a multi-channel audio signal).
Throughout this disclosure including in the claims, the expression "system" is used in a broad sense to denote a device, system, or subsystem. For example, a subsystem that implements a reverberator may be referred to as a reverberator system (or reverberator), and a system including such a reverberator subsystem (e.g., a decoder
30 system that generates X +Y output signals in response to Q + R inputs, in which the reverberator subsystem generates X of the outputs in response to Q of the inputs and the other outputs are generated in another subsystem of the decoder system) may also be referred to as a reverberator system (or reverberator). Throughout this disclosure including in the claims, the expression "reproduction" of signals by speakers denotes causing the speakers to produce sound in response to the signals, including by performing any required amplification and/or other processing of the signals. Throughout this disclosure including in the claims, the expression "linear combination" of values vi, V2, ..., Vn, (e.g., n elements of a subset of a set of X individual audio channel signals occuring at a time, t, where n is less than or equal to X) denotes a value equal to aivi + a2V2 + ...+ anvn , where a±, a2, ..., an are coefficients. In general, there is no restriction on the values of the coefficients (e.g., each coefficient can be positive or negative or zero). The expression is used in a broad sense herein, for example to cover the case that one of the coefficients is equal to 1 and the others are equal to zero (e.g., the case that the linear combination aivi + a2V2 + ...+ anvn is equal to vi(or v2, ..., or Vn).
Throughout this disclosure including in the claims, the expression "spatial cue parameter" of a multichannel audio signal denotes any parameter indicative of at least one spatial cue for the audio signal, where each such "spatial cue" is indicative (e.g., descriptive) of the spatial image of the multichannel signal. Examples of spatial cues are level (or intensity) differences between (or ratios of) pairs of the channels of the audio signal, phase differences between such channel pairs, and measures of correlation between such channel pairs. Examples of spatial cue parameters are the Channel Level Difference (CLD) parameters and Channel Prediction Coefficient (CPC) parameters which are part of a conventional MPEG Surround ("MPS") bitstream, and which are employed in MPEG surround coding.
In accordance with the well known MPEG Surround ("MPS") standard, multiple channels of audio data can be encoded by being downmixed into a smaller number of channels (e.g., M channels, where M is typically equal to 2) and compressed, and such an M-channel downmixed audio signal can be decoded by being decompressed and processed (upmixed) to generate N decoded audio channels (e.g., M = 2 and N = 5).
A typical, conventional MPS decoder is operable to perform upmixingto generate N decoded audio channels (where N is greater than two) in response to a time-domain, 2- channel, downmixed audio input signal (and MPS spatial cue parameters including Channel Level Difference and Channel Prediction Coefficient parameters). A typical, conventional MPS decoder is operable in a binaural mode to generate a binaural signal in response to a time-domain, 2-channel, downmixed audio input signal and spatial cue parameters, and in at least one other mode to perform upmixing to generate 5 0 (where the notation "x.y" channels denotes "x" full frequency channels and "y" subwoofer channels), 5 1, 7.0, or 7.1 decoded audio channels in response to a time-domain, 2- channel, downmixed audio input signal and spatial cue parameters. The input signal undergoes time domain-to-frequency domain transformation into the QMF (quadrature mirror filter) domain, to generate two channels of QMF domain frequency components These frequency components undergo decoding in the QMF domain and the resulting frequency components are typically then transformed back into the time domain to generate the audio output of the decoder. Fig. 1 is a simplified block diagram of elements of a conventional MPS decoder configured to generate N decoded audio channels (where N is greater than two, and N is typically equal to 5 or 7) in response to a 2-channel downmixed audio signal (L' and R') and MPS spatial cue parameters (including Channel Level Difference parameters and Channel Prediction Coefficient parameters) The downmixed input signal (L' and R') is indicative of "X" individual audio channels, where X is greater than 2. The downmixed input signal is typically indicative of five individual channels (e.g., left-front, right front, center, left-surround, and right-surround channels)
Each of the "left" input signal L' and the "right" input signal R' is a sequence of QMF domain frequency components generated by transforming a 2 channel, time- domain MPS encoded signal (not indicated in Fig 1) in a time domaιn-to QMF domain transform stage (not shown in Fig 1)
The downmixed input signals L' and R' are decoded into IM individual channel signals Sl, S2, . ., SN, in decoder 1 of Fig 1, in response to the MPS spatial cue parameters which are asserted (with the input signals) to the Fig. 1 system. The N sequences of output QMF domain frequency components, Sl, S2, .., SN are typically transformed back into the time domain by a QMF domain to-time domain transform stage (not shown in Fig 1), and can be asserted as output from the system without undergoing post-processing Optionally, the signals Sl, S2, . ., SN undergo postprocessing (in the QMF domain) in post-processor 5 to generate an N-channel audio output signal comprising channels OUTl, OUT2, , OUTN The N sequences of output QMF domain frequency components, OUTl, OUT2, . , OUTN, are typically transformed back into the time domain by a QMF domaιn-to time domain transform stage (not shown in Fig 1), and asserted as output from the system The conventional MPS decoder of Fig. 1 operating in a binaural mode generates 2 channel binaural audio output Sl and S2, and optionally also 2-channel binaural audio output OUTl and 0UT2, in response to a 2-channel downmixed audio signal (L' and R') and MPS spatial cue parameters (including Channel Level Difference parameters and Channel Prediction Coefficient parameters). When reproduced by a pair of headphones, the 2-chanπel audio output Sl and S2 is perceived at the listener's eardrums as sound from "X" loudspeakers (where X >2 and X is typically equal to 5 or 7) at any of a wide variety of positions (determined by the coefficients of decoder 1), including positions in front of and behind the listener In the binaural mode, post-processor 5 can apply reverb to the 2-channel output (Sl, S2) of decoder 1 (in this case, post-processor 5 implements an artificial reverberator). The Fig. 1 system could be implemented (in a manner to be described below) so that the 2-channel output of post-processor 5 (OUTl and OUT2) is a binaural audio output to which reverb has been applied, and which when reproduced by headphones is perceived at the listener's eardrums as sound from "X" loudspeakers (where X >2 and X is typically equal to 5) at any of a wide variety of positions, including positions in front of and behind the listener
Reproduction of signals Sl and S2 (or OUTl and OUT2) generated during binaural mode operation of the Fig. 1 decoder can give the listener the experience of sound that comes from more than two (e g , five) "surround" sources At least some of these sources are virtual. More generally, it is conventional for virtual surround systems to use head- related transfer functions (HRTFs) to generate audio signals (sometimes referred to as virtual surround sound signals) that, when reproduced by a pair of physical speakers (e g , loudspeakers positioned in front of a listener, or headphones) are perceived at the listener's eardrums as sound from more than two sources (e.g., speakers) at any of a wide variety of positions (typically including positions behind the listener).
As noted, the MPS decoder of Fig. 1 operating in the binaural mode could be implemented to apply reverb using an artificial reverberator implemented by postprocessor 5. This reverberator could be configured to generate reverb in response to the two-channel output (Sl, S2) of decoder 1 and to apply the reverb to the signals Sl and S2 to generate reverbed two-channel audio OUTl and 0UT2. The reverb would be applied as a post process stereo-to-stereo reverb to the 2-channel signal Sl, S2 from decoder 1, such that the same reverb impulse response is applied to all discrete channels determined by one of the two downmixed audio channels of the binaural audio output of decoder 1 (e.g., to left-front and left-surround channels determined by downmixed channel Sl), and the same reverb impulse response is applied to all discrete channels determined by the other one of the two downmixed audio channels of the binaural audio (e g., to right front and right-surround channels determined by downmixed channel S2).
One type of conventional reverberator has what is known as a Feedback Delay Network-based (FDN-based) structure In operation, such a reverberator applies reverb to a signal by feeding back to the signal a delayed version of the signal. An advantage of this structure relative to other reverb structures is the ability to efficiently produce and apply multiple uncorrelated reverb signals to multiple input signals This feature is exploited in the commercially available Dolby Mobile headphone virtualizer which includes a reverberator having FDN based structure and is operable to apply reverb to each channel of a five channel audio signal (having left-front, right-front, center, left- surround, and right surround channels) and to filter each reverbed channel using a different filter pair of a set of five head related transfer function ("HRTF") filter pairs This virtualizer generates a unique reverb impulse response for each audio channel. The Dolby Mobile headphone virtualizer is also operable in response to a two- channel audio input signal, to generate a two channel "reverbed" audio output (a two- channel virtual surround sound output to which reverb has been applied) When the reverbed audio output is reproduced by a pair of headphones, it is perceived at the listener's eardrums as H RTF-filtered, reverbed sound from five loudspeakers at left front, right front, center, left rear (surround), and right rear (surround) positions. The virtualizer upmixes a downmixed two-channel audio input (without using any spatial cue parameter received with the audio input) to generate five upmixed audio channels, applies reverb to the upmixed channels, and downmixes the five reverbed channel signals to generate the two-channel reverbed output of the virtualizer. The reverb for each upmixed channel is filtered in a different pair of HRTF filters.
US Patent Application Publication No 2008/0071549 Al, published on March 20, 2008, describes another conventional system for applying a form of reverb to a downmixed audio input signal during decoding of the downmixed signal to generate individual channel signals This reference describes a decoder which transforms time- domain downmixed audio input into the QMF domain, applies a form of reverb to the downmixed signal M(t,f) in the QMF domain, adjusts the phase of the reverb to generate a reverb parameter for each upmix channel being determined from the downmixed signal (e g , to generate reverb parameter Lreverb(t, f) for an upmix left channel, and reverb parameter Rreverb(t, f) for an upmix right channel, being determined from the downmixed signal M(t,f)) The downmixed signal is received with spatial cue parameters (e g , an ICC parameter indicative of correlation between left and right components of the downmixed signal, and inter-channel phase difference parameters IPDL and IPDR) The spatial cue parameters are used to generate the reverb parameters (e.g , LreVerb(t, f) and Rrevert>(t, f)) Reverb of lower magnitude is generated from the downmixed signal M(t,f) when the ICC cue indicates that there is more correlation between left and right channel components of the downmixed signal, reverb of greater magnitude is generated from the downmixed signal when the ICC cue indicates that there is less correlation between the left and right channel components of the downmixed signal, and apparently the phase of each reverb parameter is adjusted (in block 206 or 208) in response to the phase indicated by the relevant IPD cue However, the reverb is used only as a decorrelator in a parametric stereo decoder (mono to stereo synthesis) where the decorrelated signal (which is orthogonal to M(t,f)) is used to reconstruct the left-right cross correlation, and the reference does not suggest individually determining (or generating) a different reverb signal, for application to each of discrete channels of an upmix determined from the downmixed audio M(t,f) or to each of a set of linear combinations of values of individual upmix channels determined from the downmixed audio, from each of the discrete channels of the upmix or each of such linear combinations
The inventor has recognized that it would be desirable to individually determine (and generate) a different reverb signal for each of the discrete channels of an upmix determined from downmixed audio, from each of the discrete channels of the upmix, or to determine and generate a different reverb signal for (and from) each of a set of linear combinations of values of such discrete channels The inventor has also recognized that with such individual determination of reverb signals for the individual upmix channels (or linear combinations of values of such channels), reverb having a different reverb impulse response can be applied to the upmix channels (or linear combinations)
Until the present invention, spatial cue parameters received with downmixed audio had not been used both to generate discrete, upmix channels from the downmixed audio (e g , in the QMF domain when the downmixed audio is MPS encoded audio) or linear combinations of values therof, and to generate reverb from each such upmix channel (or linear combination) individually for application to said upmix channel (or linear combination) Nor had reverbed upmix channels that had been generated in this way been recombined to generate reverbed, downmixed audio from input downmixed audio BRIEF DESCRIPTION OF THE INVENTION
In a class of embodiments, the invention is a method for applying reverb to an M- channel downmixed audio input signal indicative of X individual audio channels, where X is a number greater than M. In these embodiments the method includes the steps of:
(a) in response to spatial cue parameters indicative (e.g., descriptive) of the spatial image of the downmixed input signal, generating Y discrete reverb channel signals (e.g., in the quadrature mirror filter or "QMF" domain), where each of the reverb channel signals at a time, f, is a linear combination of at least a subset of values of the X individual audio channels at the time, t; and
(b) individually applying reverb to each of at least two of the reverb channel signals (e.g., in the QMF domain), thereby generating Y reverbed channel signals. Preferably, the reverb applied to at least one of the reverb channel signals has a different reverb impulse response than does the reverb applied to at least one other one of the reverb channel signals. In some embodiments, X = Y, but in other embodiments X is not equal to Y. In some embodiments, Y is greater than M, and the input signal is upmixed in step (a) in response to the spatial cue parameters to generate the Y reverb channel signals. In other embodiments, Y is equal to M orY is less than M.
For example, in one case in which M = 2, X = 5, and Y = 4, the input signal is a sequence of values L(t), R(t) indicative of five individual channel signals, Lfront, Rfront, C, Lsur, and RSur. Each of the five individual channel signals is a sequence of values
(.-^fioiit ""front ^ K, ^u,, )1 = w where W is an MPEG Surround upmix matrix of form W12
Figure imgf000008_0001
and the four reverb channel signals are (gifW±±)L + (gιtWi2)R, (grfW2i)L + (grfW22)R,
(gisWii)L + (g/5wi2)R, and
{grsW2±+ W3i)L + {grSW22 + W32)R, which can be represented as: , where
Figure imgf000009_0001
In some embodiments in which the input signal is an M-channel, MPEG Surround ("MPS") downmixed signal, steps (a) and (b) are performed in the QMF domain, and the spatial cue parameters are received with the input signal. For example, the spatial cue parameters may be or include Channel Level Difference (CLD) parameters and/or Channel Prediction Coefficient (CPC) parameters of the type comprising part of a conventional MPS bitstream. When the input signal is a time-domain, MPS downmixed signal, the invention typically includes the step of transforming this time-domain signal into the QMF domain to generate QMF domain frequency components, and performing steps (a) and (b) in the QMF domain on these frequency components.
Optionally, the method also includes a step of generating an N-channel downmixed version of the Y reverbed channel signals (including each of the channel signals to which reverb has been applied and each of the channel signals, if any, to which reverb has not been applied), for example by encoding the reverbed channel signals as an N-channel, downmixed MPS signal.
In typical embodiments of the inventive method, the input downmixed signal is a 2-channel downmixed MPEG Surround ("MPS") signal indicative of five individual audio channels (left-front, right-front, center, left-surround, and right surround channels), and reverb determined by a different reverb impulse response is applied to each of at least some of these five channels, resulting in improved surround sound quality.
Preferably, the inventive method also includes a step of applying to the reverbed channel signals corresponding head-related transfer functions (HRTFs), by filtering the reverbed channel signals in an HRTF filter. The HRTFs are applied to make the listener perceive the reverb applied in accordance with the invention as being more natural sounding. Other aspects of the invention are a reverberator configured (e.g., programmed) to perform any embodiment of the inventive method, a virtualizer including such a reverberator, a decoder (e g , an MPS decoder) including such a reverberator, and a computer readable medium (e g , a disc) which stores code for implementing any embodiment of the inventive method
BRIEF DESCRIPTION OF THE DRAWINGS
FIG 1 is a block diagram of a conventional MPEG Surround decoder system. FIG 2 is a block diagram of a multiple input, multiple output, FDN based reverberator (100) that can be implemented in accordance with an embodiment of the present invention
FIG 3 is a block diagram of a reverberator system including reverberator 100 of
Fig. 2, conventional MPS processor 102, time domaιn-to-QMF domain transform filter 99 for transforming a multi-channel input into the QMF domain for processing in reverberator 100 and processor 102, and QMF domain-to-time domain transform filter
101 for transforming the combined output of reverberator 100 and processor 102 into the time domain.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
Many embodiments of the present invention are technologically possible It will be apparent to those of ordinary skill in the art from the present disclosure how to implement them. Embodiments of the inventive system, method, and medium will be described with reference to Figs. 2 and 3 In a class of embodiments, the invention is a method for applying reverb to an M- channel downmixed audio input signal indicative of X individual audio channels, where X is a number greater than M, and a system configured to perform the method. In these embodiments the method includes the steps of:
(a) in response to spatial cue parameters indicative (e g , descriptive) of the spatial image of the downmixed input signal, generating Y discrete reverb channel signals (e.g., in the quadrature mirror filter or "QMF" domain), where each of the reverb channel signals at a time, t, is a linear combination of at least a subset of values of the X individual audio channels at the time, t, and (b) individually applying reverb to each of at least two of the reverb channel signals (e.g., in the QMF domain), thereby generating Y reverbed channel signals. Preferably, the reverb applied to at least one of the reverb channel signals has a different reverb impulse response than does the reverb applied to at least one other one of the reverb channel signals. In some embodiments, X = Y, but in other embodiments X is not equal to Y. In some embodiments, Y is greater than M, and the input signal is upmixed in step (a) in response to the spatial cue parameters to generate the Y reverb channel signals. In other embodiments, Y is equal to IvI or Y is less than M.
FIG. 2 is a block diagram of multiple input, multiple output, FDN-based reverberator 100 which can be implemented in a manner to be explained below to perform this method. Reverberator 100 of Fig. 2 includes: pre-mix matrix 30 (matrix "B"), which is a 4 * M matrix coupled and configured to receive and generate four discrete reverb channel signals Ul, U2, U3, and U4 (corresponding to the feeding branches 1', 2', 3', 4', respectively) in response to an M- channel downmixed audio input signal, comprising channels INl, IN2, ..., and INM, which is indicative of five (X = 5) individual upmix audio channels. Each of the reverb channel signals at a time, t, is a linear combination of a subset of values of the X individual upmix audio channels at the time, t. In the case that M is less than four, matrix B upmixes the input signal to generate the reverb channel signals. In a typical embodiment, M is equal to 2. Matrix 30 is coupled also to receive spatial cue parameters which are indicative (e.g., descriptive) of the spatial image of the M-channel downmixed input signal, and is configured to generate four (Y = 4) discrete upmix channel signals, i.e. the discrete reverb channel signals Ul, U2, U3, and U4, in response to the spatial cue parameters; addition elements 40, 41, 42, and 43, coupled to the outputs of matrix 30, to which reverb channel signals Ul, U2, U3, and U4 are asserted. Element 40 is configured to add the output of gain element gl (i.e., apply feedback from the output of gain element gl) to reverb channel signal Ul. Element 41 is configured to add the output of gain element g2 to reverb channel signal U2. Element 42 is configured to add output of gain element g3 to reverb channel signal U3. Element 43 is configured to add the output of gain element g4 to reverb channel signal U4; scattering matrix 32 (matrix "A"), which is coupled to receive the outputs of addition elements 40, 41, 42, and 43. Matrix 32 is preferably a 4 * 4 unitary matrix configured to assert a filtered version of the output of each of addition elements 40, 41, 42, and 43 to a corresponding one of delay lines, z'M> , where 0 < k - 1 < 3, and is preferably a fully populated matrix in order to provide maximum diffuseness. Delay lines Z-Mif Z-M2; Z-M3τ anc| z-MA^ are labeled respectively as delay lines 50, 51, 52, and 53 in Fig. 2; gain elements, gk, where 0 < k - 1 < 3, which apply gain the outputs of delay lines, z'Mt , thus providing damping factors for controlling the decay time of the reverb applied in each upmix channel. Each gain element, gk, is typically combined with a low- pass filter. In some embodiments, the gain elements apply different, predetermined gain factors for the different QMF bands. Reverbed channel signals Rl, R2, R3, and R4, respectively, are asserted at the outputs of gain elements gl, g2, g3, and g4; and post-mix matrix 34 (matrix "C"), which is an N * 4 matrix coupled and configured to downmix and/or upmix (and optionally to perform other filtering on) the reverbed channel signals Rl, R2, R3, and R4 asserted at the outputs of gain elements gk, in response to at least a subset (e.g., all or some) of the spatial cue parameters asserted to matrix 30, thereby generating an N-channel, QMF domain, downmixed, reverbed audio output signal comprising channels Sl, S2, ..., and SN. In variations on the Fig. 2 embodiment, matrix 34 is a constant matrix whose coefficients do not vary with time in response to any spatial cue parameter.
In variations on the Fig. 2 embodiment, the inventive system has Y reverb channels (where Y is less than or greater than four), pre-mix matrix 30 is configured to generate Y discrete reverb channel signals in response to the downmixed, M-channel, input signal and the spatial cue parameters, scattering matrix 32 is replaced by an YxY matrix, and the inventive system has Y delay lines, z'Mt .
For example, in one case in which Y = M =2, the downmixed input signal is indicative of five upmix channels (X = 5): left front, right front, center front, left surround, and right surround channels. In accordance with the invention, in response to spatial cue parameters indicative of the spatial image of the downmixed input signal, a pre-mix matrix (a variation on matrix 30 of Fig. 2) generates two discrete reverb channel signals (e.g., in the quadrature mirror filter or "QMF" domain): one a mix of the front channels; the other a mix of the surround channels. Reverb having a short decay response is generated from (and applied to) one reverb channel signal and reverb having a long decay response is generated from (and applied to) the other reverb channel signal (e.g., to simulate a room with "live end/dead end" acoustics).
With reference again to Fig. 2, post-processor 36 optionally is coupled to the outputs of matrix 34 and operable to perform post-processing on the downmixed, reverbed output Sl, S2, ..., SN of matrix 34, to generate an N-channel post-processed audio output signal comprising channels OUTl, 0UT2, ..., and OLJTN. Typically, N = 2, so that the Fig. 2 system outputs a binaural, downmixed, reverbed audio signal Sl, S2 and/or a binaural, post-processed, downmixed, reverbed audio output signal OUT, 0UT2. For example, the output of matrix 34 of some implementations of the Fig. 2 system is a binaural, virtual surround sound signal, which when reproduced by headphones, is perceived by the listener as sound emitting from left ("L"), center ("C"), and right ("R") front sources (e.g., left, center, and right physical speakers positioned in front of the listener), and left-surround ("LS") and right-surround ("RS") rear sources (e.g., left, and right physical speakers positioned behind the listener).
In some variations on the Fig. 2 system, post-mix matrix 34 is omitted and the inventive reverberator outputs Y-channel reverbed audio (e.g., upmixed, reverbed audio) in response to an M-channel downmixed audio input. In other variations, matrix 34 is an identity matrix. In other variations, the system has Y upmix channels (where Y is a number greater than four) and matrix 34 is an N * Y matrix (e.g,. Y = 7).
Although the Fig. 2 system has four reverb channels and four delay lines, z'M" , variations on the system (and other embodiments of the inventive reverberator) implement more than or less than four reverb channels. Typically, the inventive reverberator includes one delay line per reverb channel. In implementations of the Fig. 2 system in which the input signal is an M-channel,
MPEG Surround ("MPS") downmixed signal, the input signal asserted to the inputs of matrix 30 comprises QMF domain signals INl(t,f), IN2(t,f) and INM(t,f), and the Fig. 2 system performs processing (e.g., in matrix 30) and reverb application thereon in the QMF domain. In such implementations, the spatial cue parameters asserted to matrix 30 are typically Channel Level Difference (CLD) parameters and/or Channel Prediction
Coefficient (CPC) parameters, and/or Inter-channel Cross Correlation (ICC) parameters, of the type comprising part of a conventional MPS bitstream.
In order to provide such QMF domain inputs to matrix 30 in response to a time- domain, M-channel MPS downmixed signal, the inventive method would include a preliminary step of transforming this time-domain signal into the QMF domain to generate QMF domain frequency components, and would perform above-described steps (a) and (b) in the QMF domain on these frequency components.
For example, because the input to the FIG. 3 system is a time-domain MPS downmixed audio signal comprising M channels ll(t), I2(t), ..., and IM(t), the FIG. 3 system includes filter 99 for transforming this time-domain signal into the QMF domain Specifically, the FIG 3 system includes reverberator 100 (corresponding to and possibly identical to reverberator 100 of Fig 2), conventional MPS processor 102, time domain- to-QMF domain transform filter 99 coupled and configured to transform each of the time domain input channels ll(t), I2(t), . ., and IM(t) into the QMF domain (i.e , into a sequence of QMF domain frequency components) for processing in reverberator 100 and conventional processing in processor 102 The FIG. 3 system also includes QMF domain- to time domain transform filter 101, which is coupled and configured to transform the N channel combined output of reverberator 100 and processor 102 into the time domain Specifically, filter 99 transforms time-domain signals ll(t), I2(t), , and IM(t) respectively into QMF domain signals INl(t,f), !N2(t,f), ..., and INM(t,f), which are asserted to reverberator 100 and processor 102. Each of the N channels output from processor 102 is combined (in an adder) with the corresponding reverbed channel output of reverberator 100 (Sl, S2, ..., or SN indicated in Fig 2, or one of OUTl, 0UT2, ..., or OUTN indicated in Fig. 2 if reverberator 100 of Fig 3 also includes a post processor 36 as shown in Fig 2). Filter 101 of Fig. 3 transforms the combined (reverbed) output of reverberator 100 and processor 102 (N sequences of QMF domain frequency components Sl'(t, f), S2 (t,f), . ., SN'(t, f)) into time-domain signals Sl'(t), S2'(t), , SN'(t) In typical embodiments of the invention, the input downmixed signal is a 2- channel downmixed MPS signal indicative of five individual audio channels (left-front, right-front, center, left-surround, and right surround channels), and reverb determined by a different reverb impulse response is applied to each of these five channels, resulting in improved surround sound quality.
If the coefficients of pre-mix matrix 30 (Y x M matrix B, which is a 4 x 2 matrix in the case that Y = 4 and M = 2) were constant coefficients (not time varying coefficients determined in response to spatial cue parameters) and the coefficients of post mix matrix 34 (N * Y matrix C, which is a 2 x 4 matrix in the case that Y = 4 and N = 2) were constant coefficients, the Fig. 2 system could not produce and apply individual reverb with individual impulse responses for different channels in the downmix determined by the M-channel, downmixed, MPS encoded, input to the reverberator (e g , in response to a QMF-domain, MPS-encoded, M-channel downmixed signal INl(t, f), IN2(t, f), . , INM(t, f)) Consider an example in which M = 2, Y = 4, and N = 2, and matrices B and C of Fig 2 (also labeled as matrices 30 and 34 in Fig. 2) were replaced respectively by constant 4 * 2 and 2 * 4 matrices with the following constant coefficients-
Figure imgf000015_0001
In this example, the coefficients of the constant matrices B and C would not change as a function of time in response to spatial cue parameters indicative of the downmixed input audio, and the so-modified Fig. 2 system would operate in a conventional stereo-to-stereo reverb mode. In such conventional reverb mode, reverb having the same reverb impulse response would be applied to each individual channel in the downmix (i.e., left-front channel content in the downmix would receive reverb having the same impulse response as would right-front channel content in the downmix). However, by applying the reverb process in the QMF domain in response to Channel Level Difference (CLD) parameters, Channel Prediction Coefficient (CPC), and/or Inter-channel Cross Correlation (ICC) parameters available as part of the MPS bitstream (and/or in response to other spatial cue parameters) in accordance with the invention, the Fig. 2 system can produce and apply reverb to each reverb channel determined by the downmixed input to the system, with individual reverb impulse responses for each of the reverb channels. In a typical application, less reverb is applied in accordance with the invention to a center channel (for clearer speech/dialog) than to at least one other reverb channel so that the impulse response of the reverb applied each of these reverb channels is different. In such application (and other applications), the impulse responses of the reverb applied to different reverb channels are not based on different channel routing to matrix 30 and are instead simply different scale factors applied by pre-mix matrix 30 or post-mix matrix 34 (and/or at least one other system element) to different reverb channels. For example, in an implementation of the Fig. 2 system configured to apply reverb to a QMF-domain, MPS encoded, stereo downmix of five upmix channels, matrix 30 is a 4 x 2 matrix having time-varying coefficients which depend on current values of coefficients, wij, where / ranges from 1 to 3 and j ranges from 1 to 2.
In this exemplary implementation, M = 2, X = 5, and Y = 4, the input signal is a sequence of QMF domain value pairs, INl(t,f) = L(t), and IN2(t,f) = R(t), indicative of a sequence of values of five individual channel signals, LfrOnt, Rfront, C, Lsur, and RSur- Each of the five individual channel signals is a sequence of values
(Ifront Rfrm C L81111 -S511n )1 = W , where W is an MPEG Surround upmix matrix
\ R of form
Figure imgf000016_0001
In this example, the coefficients wy , would be updated in response to the current values of conventional CPC parameters CPC_1 and CPC_2 and conventional ICC parameter ICC_TTT (the Inter-channel Cross Correlation parameter for the Two-To-Three, or "TTT," upmixer assumed during encoding of the downmixed input signal): wll = (CPC_1 + 2) / (3 * ICCJTT); wl2 = (CPC_2 - 1) / (3 * ICCJTT); w21 = (CPC_1 - 1) / (3 * ICCJTT); (Eq. Ia) w22 = (CPC_2 + 2) / (3 * ICCJTT); w31 = (1- CPC_1) / (3 * ICCJTT); and w32 = (1- CPC_2) / (3 * ICCJTT).
Also using the conventional CLD parameters for the left front/surround channels ( CLD]t ls ) and the right front/surround channels ( CLDn Is ), the time-varying coefficients of matrix 30 would depend also on the following four, time-varying channel gain values, in which CLDXi ls is the current value of the left front/surround CLD parameter, and CLDn IS is the current value of the right front/surround CLD parameter;
Figure imgf000017_0001
1 + 10'
(Eq. 2) io( CLD1, ^/20
Sn =
1 + 10 g,s
1 + 10'
The time-varying coefficients of matrix 30 would be:
Figure imgf000017_0002
Thus, in the exemplary implementation, the four reverb channel signals output from matrix 30 are Ul= igifW±±)l + (gifW±2)R, U2 = (grfW2i)L + (grfW22)R, U3 = (g/sWii)L + {gιswi2)R, and U4 = (g,sW2i+ W3i)L + (graW22 + w32)R- Thus, the matrix multiplication performed by matrix 30 (having the coefficients shown in Equation 3) can be represented as:
, where
Figure imgf000017_0003
This matrix multiplication is equivalent to an upmix to five individual channel signals (by the MPEG Surround upmix matrix W defined above) followed by a downmix of these five signals to the four reverb channel signals by matrix Bo. In a variation on the implementation of matrix 30 having the coefficients shown in Equation 3, matrix 30 is implemented with the following coefficients:
Km-8ή W2X + Kκsg^n KRvg,f W22 + KRsg,,W22
B = B0W = (Eq. 4) K0W3, K0W32
Figure imgf000018_0001
, where Khf , KRf , Kc , K15 and KRS are fixed reverb gain values for the different channels, and gif, gιs, grf, gif, and wii to W32 are as in Equation 2 and Ia, respectively. Typically, the four fixed reverb gain values are substantially equal to each other, except that K0 typically has a slightly lower value than the others (a few decibels lower than the values of the others) in order to apply less reverb to the center channel (e.g., for dryer sounding speech/dialog).
Matrix 30, implemented with the coefficients of Equation 4, is equivalent to the product of the MPEG Surround upmix matrix W defined above and the following downmix matrix Bo:
B = BnW =
Figure imgf000018_0002
, where
K0W32
Kcw,λ Kcw31
K, U= 0 0 0 K^ 0 0 κRS
B0 =
0 0 κc 0 0
0 0 κr 0 0
In the case that matrix 30 is implemented with the coefficients of Equation 3 (or Equation 4), matrix 34 would typically be a constant matrix. Alternatively, matrix 34 would have time-varying coefficients, e.g., in one implementation its coefficients would be C = Bτ , where Bτ is the transpose of matrix 30. Matrix 30 with the coefficients set forth in Equation 3, and matrix 34 (if implemented as the transpose of such matrix), would have the same general form as the constant mix matrices B and C of Equation 1, but with variable coefficients determined by the variable gain values of Equation 2 and above-described variable coefficient values, w/j, of Equation Ia substituted for the constant elements.
Implementing matrix 30 with the variable coefficients of Equation 3 would cause reverb channels Ul, U2, U3, and U4, respectively, to be the left-front upmix channel (feeding branch 1' of the Fig. 2 system), the right-front upmix channel (feeding branch 2' of the Fig. 2 system), the left-surround upmix channel (feeding branch 3' of the Fig. 2 system), and a combined right-surround and center upmix channel (the right-surround channel plus the center channel) feeding branch 4' of the Fig. 2 system. Hence, the reverb individually applied to the four branches of the Fig. 2 system would have individually determined impulse responses.
Alternatively, matrix 30's coefficients are determined in another manner in response to available spatial cue parameters. For example, in some embodiments matrix 30's coefficients are determined in response to available MPS spatial cue parameters to cause matrix 30 to implement a TTT upmixer operating in a mode other than in a prediction mode (e.g., an energy mode with or without center subtraction). This can be done in a manner that will be apparent to those of ordinary skill in the art given the present description, using the well known upmixing formulas for the relevant cases that are described in the MPEG standard (ISO/IEC 23003-1:2007).
In an implementation of the Fig. 2 system configured to apply reverb to a QMF- domain, MPS encoded, single-channel (monaural) downmix of four upmix channels, matrix 30 is a 4 x 1 matrix having time-varying coefficients:
Figure imgf000019_0001
where the coefficients are gain factors are derived from the CLD parameters CiD11 h , CLDn IS , CLD0 h and CLD1 , , available as part of a conventional MPS bitstream. In variations on the Fig. 2 system and other embodiments of the inventive reverberator, discrete reverb channels (e.g., upmix channels) are extracted from a downmixed input signal and routed to individual reverb delay branches in any of many different ways. In various embodiments of the inventive reverberator, other spatial cue parameters are employed to upmix a downmixed input signal (e.g., including by control channel weighting). For example, in some embodiments, ICC parameters (available as part of a conventional MPS bitstream) that describe front-back diffuseness are used to determine coefficients of the pre-mix matrix and thereby to control reverb level.
Preferably, the inventive method also includes a step of applying to the reverbed channel signals corresponding head-related transfer functions (HRTFs), by filtering the 5 reverbed channel signals in an HRTF filter. For example, matrix 34 of the Fig. 2 system is preferably implemented as the HRTF filter which applies such HRTFs to, and also performs the above-described downmixing operation on, reverbed channels Rl, R2, R3, and R4. Such implementation of matrix 34 would typically perform the same filtering as a 5 χ 4 matrix followed by a 2 * 5 matrix, where the 5 x 4 matrix generates five virtual
10 reverbed channel signals (left-front, right-front, center, left-surround and right surround channels) in response to the four reverbed channel signals R1-R4 output from gain elements gl, g2, g3, and g4, and the 2 * 5 matrix applies an appropriate HRTF to each such virtual reverbed channel signal, and downmixes the resulting five channel signals to generate a 2-channel downmixed reverbed output signal. Typically however, matrix 34
15 would be implemented as a single 2 x 4 matrix that performs the described functions of the separate 5 x 4 and 2 x 5 matrices. The HRTFs are applied to make the listener perceive the reverb applied in accordance with the invention as more natural sounding. The HTRF filter would typically perform for each individual QMF band a matrix multiplication by a matrix with complex valued entries.
20 In some embodiments, reverbed channel signals generated from a QMF-domain,
MPS encoded, downmixed input signal are filtered with corresponding HRTFs as follows. In these embodiments, the HRTFs in the parametric QMF domain essentially consist of left and right gain parameter values and Inter-channel Phase Difference (IPD) parameter values that characterize the downmixed input signal. The IPDs optionally are ignored to
25 reduce complexity. Assuming that the IPDs are ignored, the HRTFs are constant gain values (four gain values for each of the left and the right channel, respectively): ggRiFjr L >
thus be applied to the reverbed channel signals Rl, R2, R3, and R4 of Fig. 2 by an implementation of post-mix matrix 34 having the following coefficients:
( Jf_L glIRIF Jf R V
,, „ „ _ &ΗRIF_rf_L SΗRIF_ri_R
■?HRIF Is L &HRIF L R In preferred implementations of the inventive reverberator (which may be implemented, for example, as variations on the Fig. 2 system), fractional delay is applied in at least one reverb channel, and/or reverb is generated and applied differently to different frequency bands of frequency components of audio data in at least one reverb channel.
Some such preferred implementations of the inventive reverberator are variations on the Fig. 2 system that are configured to apply fractional delay (in at least one reverb channel) as well as integer sample delay. For example, in one such implementation a fractional delay element is connected in each reverb channel in series with a delay line that applies integer delay equal to an integer number of sample periods (e.g., each fractional delay element is positioned after or otherwise in series with one of delay lines 50, 51, 52, and 53 of Fig. 2). Fractional delay can be approximated by a phase shift (unity complex multiplication) in each QMF band that corresponds to a fraction of the sample period: f = τ/T, where f is the delay fraction, T is the desired delay for the QMF band, and T is the sample period for the QMF band. It is well known how to apply fractional delay in the context of applying reverb in the QMF domain (see for example, J. Engdegard, et al., "Synthetic Ambience in Parametric Stereo Coding," presented at the 116th Convention of the Audio Engineering Society, in Berlin, Germany, May 8-11, 2004, 12 pages, and U.S. Patent 7,487,097, issued February 3, 2009 to J. Engdegard, et al.).
Some of the above-noted preferred implementations of the inventive reverberator are variations on the Fig. 2 system that are configured to apply reverb differently to different frequency bands of the audio data in at least one reverb channel, in order to reduce complexity of the reverberator implementation. For example, in some implementations in which the audio input data, INl-INM, are QMF domain MPS data, and the reverb application is performed in the QMF domain, the reverb is applied differently to the following four frequency bands of the audio data in each reverb channel:
0 kHz -3 kHz (or 0 kHz -2.4 kHz): reverb is applied in this band as in the above- described embodiment of Fig. 2, with matrix 30 implemented with the coefficients of Equation 4);
3 kHz -8 kHz (or 2.4 kHz - 8 kHz): reverb is applied in this band with real valued arithmatic only. For example, this can be done using the real valued arithmetic techniques described in International Application Publication No. WO 2007/031171 Al, published March 22, 2007. This reference describes a 64 band QMF filterbank in which complex values of the eight lowest frequency bands are audio data are processed and only real values of the upper 56 frequency bands of the audio data are processed. One of such eight lowest frequency bands can be used as a complex QMF buffer band, so that complex-valued arithmetic calculations are performed for only seven of the eight lowest QMF frequency bands (so that reverb is applied in this relatively low frequency range as in the above-described embodiment of Fig. 2, with matrix 30 implemented with the coefficients of Equation 4), and real-valued arithmetic calculations are performed for the other 56 QMF frequency bands, with the crossover between complex valued and real valued calculations occurring at the frequency (7 * 44.1kHz)/(64*2) which is approximately equal to 2.4 kHz. In this exemplary embodiment, reverb is applied in the relatively high frequency range as in the above-described Fig. 2 embodiment but using a simpler implementation of pre-mix matrix 30 to perform real-valued computations only. Reverb is applied in the relatively low frequency range (below 2.4 kHz) as in the Fig. 2 embodiment, e.g., with matrix 30 implemented with the coefficients of Equation 4); 8 kHz- 15 kHz: reverb is applied in this band by a simple delay technique. For example, reverb is applied in a way similar to the manner it is applied the above- described Fig. 2 embodiment but with only two reverb channels with a delay line and low- pass filter in each reverb channel, with matrix elements 32 and 34 omitted, with a simple, 2χ 2 implementation of pre-mix matrix 30 (e.g., to apply less reverb to the center channel than to each other channel), and without feedback from nodes along the reverb channels to the outputs of the pre-mix matrix. The two delay branches can be simply fed to left and right outputs, respectively, or can be switched so that echoes from the left front (Lf) and left surround (Ls) channels end up in the right output channel and echoes from the right front (Rf) and right surround (Rs) channels end up in the left output channel The 2χ 2 pre-mix matrix can have the following coefficients:
B λ ^^ ^
Figure imgf000022_0001
+ Kcw32 ) ' symbols are defined as in Equation 4 above; and
15 - 22.05 kHz: no reverb is applied in this band.
In variations on the embodiments disclosed herein (e.g., the Fig. 2 embodiment, the inventive system applies reverb to an M-channel downmixed audio input signal indicative of X individual audio channels, where X is a number greater than M, including by generating Y discrete reverb channel signals in response to the downmixed signal but not in response to spatial cue parameters In these variations, the system individually applies reverb to each of at least two of the reverb channel signals in response to spatial cue parameters indicative of spatial image of the downmixed input signal, thereby generating Y reverbed channel signals For example, in some such variations the coefficients of a pre mix matrix (e.g., a variation on matrix 30 of Fig 2) are not determined in response to spatial cue parameters, but at least one of a scattering matrix (e g., a variation on matrix 32 of Fig. 2), a gain stage (e.g., a variation on the gain stage comprising elements gl-gk of Fig. 2), and a post-mix matrix (e.g., a variation on matrix 34 of Fig. 2) operates on the reverb channel signals in a manner determined by spatial cue parameters indicative of spatial image of the downmixed input signal, to apply reverb to each of at least two of the reverb channel signals.
In some embodiments, the inventive reverberator is or includes a general purpose processor coupled to receive or to generate input data indicative of an M- channel downmixed audio input signal, and programmed with software (or firmware) and/or otherwise configured (e.g., in response to control data) to perform any of a variety of operations on the input data, including an embodiment of the inventive method. Such a general purpose processor would typically be coupled to an input device (e.g , a mouse and/or a keyboard), a memory, and a display device. For example, the Fig. 3 system could be implemented in a general purpose processor, with inputs ll(t), I2(t), ..., IM(t), being input data indicative of M channels of downmixed audio data, and outputs Sl(t), S2(t), .., SN(t), being output data indicative of N channels of downmixed, reverbed audio. A conventional digital-to-analog converter (DAC) could operate on this output data to generate analog versions of the output audio signals for reproduction by speakers (e.g., a pair of headphones) While specific embodiments of the present invention and applications of the invention have been described herein, it will be apparent to those of ordinary skill in the art that many variations on the embodiments and applications described herein are possible without departing from the scope of the invention described and claimed herein It should be understood that while certain forms of the invention have been shown and described, the invention is not to be limited to the specific embodiments described and shown or the specific methods described.

Claims

CLAIMSWhat is claimed is:
1. A method for applying reverb to an M-channel downmixed audio input signal indicative of X individual audio channels, where X is a number greater than M, said method including the steps of:
(a) in response to spatial cue parameters indicative of a spatial image of the downmixed input signal, generating Y discrete reverb channel signals, where each of the reverb channel signals at a time, t, is a linear combination of at least a subset of values of the X individual audio channels at the time, t; and (b) individually applying reverb to each of at least two of the reverb channel signals, thereby generating Y reverbed channel signals.
2. The method of claim 1, wherein the reverb applied to at least one of the reverb channel signals has a different reverb impulse response than does the reverb applied to at least one other one of the reverb channel signals.
3. The method of any of claims 1 to 2, wherein the input signal is an M-channel, MPEG Surround downmixed signal, and the spatial cue parameters include at least one of Channel Level Difference parameters, Channel Prediction Coefficient parameters, and Inter-channel Cross Correlation parameters.
4. The method of claim 3, wherein the spatial cue parameters include Channel Level Difference parameters, Channel Prediction Coefficient parameters, and Inter- channel Cross Correlation parameters.
5. The method of any of claims 1 to 4, wherein the input signal is a QMF-domain, MPEG Surround downmixed signal comprising M sequences of QMF domain frequency components, and wherein each of steps (a) and (b) is performed in the QMF domain.
6. The method of claim 5, wherein the spatial cue parameters include at least some of Channel Level Difference parameters, Channel Prediction Coefficient parameters, and Inter-channel Cross Correlation parameters.
7. The method of claim 5, wherein the spatial cue parameters include Channel Level Difference parameters, Channel Prediction Coefficient parameters, and Inter- channel Cross Correlation parameters
8. The method of claim 1, wherein the input signal is a time-domain, MPEG
Surround downmixed signal, and also including the step of: before step (a), transforming the time-domain, MPEG Surround downmixed signal into the QMF domain thereby generating M sequences of QMF domain frequency components, and wherein each of steps (a) and (b) is performed in the QMF domain.
9. The method of any of claims l to 8, also includingthe step of downmixing the Y reverbed channel signals, thereby generating an N-channel, downmixed, reverbed audio signal, where N is a number less than Y.
10. The method of claim 9, wherein the downmixing is performed in response to at least a subset of the spatial cue parameters.
11. The method of claim 9, wherein M = 2, Y = 4, and N = 2.
12. The method of any of claims 1 to 10, wherein M = 2 and Y = 4.
13. The method of claim 1, also including the step of applying to the reverbed channel signals corresponding head-related transfer functions by filtering the reverbed channel signals in a head-related transfer function filter.
14. The method of any of claims 1 to 10, wherein M = I.
15. The method of any of claims 1 to 14, wherein Y is greater than M.
16 The method of any of claims 1 to 12, also including the step of downmixing the reverbed channel signals and applying to said reverbed channel signals corresponding head-related transfer functions.
17. A reverberator configured to apply reverb to an M-channel downmixed audio input signal indicative of X individual audio channels, where X is a number greater than M, said reverberator including: a first subsystem, coupled to receive the input signal and spatial cue parameters indicative of a spatial image of said input signal, and configured to generate Y discrete reverb channel signals in response to the input signal, including by applying coefficients determined in response to the spatial cue parameters, such that each of the reverb channel signals at a time, t, is a linear combination of at least a subset of values of the X individual audio channels at the time, t, and a reverb application subsystem coupled to the first subsystem and configured to apply reverb individually to each of at least two of the reverb channel signals, thereby generating a set of Y reverbed channel signals
18 The reverberator of claim 17, wherein the reverb application subsystem includes Y branches, each of the branches configured to apply reverb individually to a different one of the reverb channel signals
19 The reverberator of claim 17, wherein the reverb application subsystem is a feedback delay network including Y branches, each of the branches configured to apply reverb individually to a different one of the reverb channel signals
20. The reverberator of any of claims 17 to 19, wherein the reverb application subsystem is configured to apply the reverb such that the reverb applied to at least one of the reverb channel signals has a different reverb impulse response than does the reverb applied to at least one other one of the reverb channel signals.
21. The reverberator of any of claims 17 to 20, wherein the input signal is an M- channel, MPEG Surround downmixed signal, and the spatial cue parameters include at least some of Channel Level Difference parameters, Channel Prediction Coefficient parameters, and Inter-channel Cross Correlation parameters.
22. The reverberator of any of claims 17 to 21, wherein the spatial cue parameters include Channel Level Difference parameters, Channel Prediction Coefficient parameters, and Inter-channel Cross Correlation parameters.
23. The reverberator of any of claims 17 to 20, wherein the input signal is a QMF- domain, MPEG Surround downmixed signal comprising M sequences of QMF domain frequency components, and the spatial cue parameters include at least some of Channel Level Difference parameters, Channel Prediction Coefficient parameters, and Inter- channel Cross Correlation parameters.
24. The reverberator of claim 23, wherein the spatial cue parameters include Channel Level Difference parameters, Channel Prediction Coefficient parameters, and Inter-channel Cross Correlation parameters
25. The reverberator of claim 17, wherein the downmixed audio input signal is a set of M sequences of QMF domain frequency components, said reverberator also including: a time domain-to-QMF domain transform filter coupled to receive a time-domain, MPEG Surround downmixed signal and configured to generate in response thereto the M sequences of QMF domain frequency components, and wherein the upmix subsystem is coupled and configured to upmix said M sequences of QMF domain frequency components in the QMF domain.
26. The reverberator of any of claims 17 to 25, also including a post-mix subsystem coupled and configured to downmixthe reverbed channel signals, thereby generating an N channel, downmixed, reverbed audio signal, where N is a number less than Y.
27. The reverberator of claim 26, wherein M = 2, Y = 4, and N = 2.
28. The reverberator of any of claims 17 to 26, wherein M = 2 and Y = 4.
29. The reverberator of any of claims 17 to 28, also including: a head-related transfer function filter coupled and configured to apply at least one head-related transfer function to each of the reverbed channel signals.
30. The reverberator of claim 29, wherein M = I.
31. The reverberator of any of claims 17 to 30, also including: a post-mix subsystem coupled and configured to downmixthe reverbed channel signals and apply at least one head-related transfer function to each of the reverbed channel signals, thereby generating an N-channel, downmixed, reverbed audio signal, where N is a number less than Y.
32. The reverberator of any of claims 17 to 31, wherein the reverb application subsystem includes: a set of Y delay and gain elements, having Y outputs at which the reverbed channel signals are asserted and havingY inputs; a set of Y addition elements, each of the addition elements having a first input coupled to a different output of the filter, a second input coupled to receive a different one of the reverbed channel signals, and an output; a scattering matrix having matrix inputs coupled to the outputs of the addition elements, and matrix outputs coupled to the inputs of the delay and gain elements, wherein the scattering matrix is configured to assert a filtered version of the output of each of the addition elements to the input of a corresponding one of the delay and gain elements.
33. The reverberator of claim 32, also including a post-mix subsystem, coupled to the outputs of the delay and gain elements and coupled to receive at least a subset of the spatial cue parameters, and configured to downmixthe reverbed channel signals in response to said at least a subset of the spatial cue parameters, thereby generating an N-channel, downmixed, reverbed audio signal, where N is a number less than Y.
PCT/EP2009/067350 2008-12-19 2009-12-16 Method and apparatus for applying reverb to a multi-channel audio signal using spatial cue parameters WO2010070016A1 (en)

Priority Applications (7)

Application Number Priority Date Filing Date Title
KR1020117016794A KR101342425B1 (en) 2008-12-19 2009-12-16 A method for applying reverb to a multi-channel downmixed audio input signal and a reverberator configured to apply reverb to an multi-channel downmixed audio input signal
JP2011541416A JP5524237B2 (en) 2008-12-19 2009-12-16 Method and apparatus for applying echo to multi-channel audio signals using spatial cue parameters
CN2009801511567A CN102257562B (en) 2008-12-19 2009-12-16 Method and apparatus for applying reverb to a multi-channel audio signal using spatial cue parameters
EP09801205.7A EP2377123B1 (en) 2008-12-19 2009-12-16 Method and apparatus for applying reverb to a multi-channel audio signal using spatial cue parameters
US13/132,321 US8965000B2 (en) 2008-12-19 2009-12-16 Method and apparatus for applying reverb to a multi-channel audio signal using spatial cue parameters
RU2011129154/08A RU2509442C2 (en) 2008-12-19 2009-12-16 Method and apparatus for applying reveberation to multichannel audio signal using spatial label parameters
BRPI0923174-9A BRPI0923174B1 (en) 2008-12-19 2009-12-16 METHOD AND REVERBERATOR TO APPLY REVERBERATION TO AN AUDIO INPUT SIGNAL WITH DOWNMIXING OF CHANNELS

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
SE0802629-6 2008-12-19
SE0802629 2008-12-19
US17285509P 2009-04-27 2009-04-27
US61/172,855 2009-04-27

Publications (1)

Publication Number Publication Date
WO2010070016A1 true WO2010070016A1 (en) 2010-06-24

Family

ID=41796192

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2009/067350 WO2010070016A1 (en) 2008-12-19 2009-12-16 Method and apparatus for applying reverb to a multi-channel audio signal using spatial cue parameters

Country Status (8)

Country Link
US (1) US8965000B2 (en)
EP (1) EP2377123B1 (en)
JP (1) JP5524237B2 (en)
KR (1) KR101342425B1 (en)
CN (1) CN102257562B (en)
BR (1) BRPI0923174B1 (en)
RU (1) RU2509442C2 (en)
WO (1) WO2010070016A1 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103210668A (en) * 2010-09-06 2013-07-17 音尚股份公司 Upmixing method and system for multichannel audio reproduction
CN103430574A (en) * 2011-03-02 2013-12-04 弗兰霍菲尔运输应用研究公司 Apparatus and method for determining a measure for a perceived level of reverberation, audio processor and method for processing a signal
CN103890841A (en) * 2011-11-01 2014-06-25 皇家飞利浦有限公司 Audio object encoding and decoding
CN108665902A (en) * 2017-03-31 2018-10-16 华为技术有限公司 The decoding method and codec of multi-channel signal
KR20190134752A (en) * 2017-04-12 2019-12-04 후아웨이 테크놀러지 컴퍼니 리미티드 Multichannel signal encoding and decoding method, and codec
RU2780508C2 (en) * 2018-01-24 2022-09-26 Л-АКУСТИКС ЮКей ЛТД Method and system for use of time effects in multichannel audio playback system

Families Citing this family (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2011083979A2 (en) 2010-01-06 2011-07-14 Lg Electronics Inc. An apparatus for processing an audio signal and method thereof
US9131313B1 (en) * 2012-02-07 2015-09-08 Star Co. System and method for audio reproduction
US9602927B2 (en) * 2012-02-13 2017-03-21 Conexant Systems, Inc. Speaker and room virtualization using headphones
EP2637427A1 (en) 2012-03-06 2013-09-11 Thomson Licensing Method and apparatus for playback of a higher-order ambisonics audio signal
WO2014161996A2 (en) * 2013-04-05 2014-10-09 Dolby International Ab Audio processing system
KR102150955B1 (en) 2013-04-19 2020-09-02 한국전자통신연구원 Processing appratus mulit-channel and method for audio signals
WO2014171791A1 (en) 2013-04-19 2014-10-23 한국전자통신연구원 Apparatus and method for processing multi-channel audio signal
US9420393B2 (en) 2013-05-29 2016-08-16 Qualcomm Incorporated Binaural rendering of spherical harmonic coefficients
US9319819B2 (en) 2013-07-25 2016-04-19 Etri Binaural rendering method and apparatus for decoding multi channel audio
TWI713018B (en) 2013-09-12 2020-12-11 瑞典商杜比國際公司 Decoding method, and decoding device in multichannel audio system, computer program product comprising a non-transitory computer-readable medium with instructions for performing decoding method, audio system comprising decoding device
CN108347689B (en) 2013-10-22 2021-01-01 延世大学工业学术合作社 Method and apparatus for processing audio signal
ES2837864T3 (en) 2014-01-03 2021-07-01 Dolby Laboratories Licensing Corp Binaural audio generation in response to multichannel audio using at least one feedback delay network
CN104768121A (en) * 2014-01-03 2015-07-08 杜比实验室特许公司 Generating binaural audio in response to multi-channel audio using at least one feedback delay network
KR102574478B1 (en) 2014-04-11 2023-09-04 삼성전자주식회사 Method and apparatus for rendering sound signal, and computer-readable recording medium
CN110827839B (en) * 2014-05-30 2023-09-19 高通股份有限公司 Apparatus and method for rendering higher order ambisonic coefficients
US20170142178A1 (en) * 2014-07-18 2017-05-18 Sony Semiconductor Solutions Corporation Server device, information processing method for server device, and program
WO2016148553A2 (en) * 2015-03-19 2016-09-22 (주)소닉티어랩 Method and device for editing and providing three-dimensional sound
CN112492501B (en) 2015-08-25 2022-10-14 杜比国际公司 Audio encoding and decoding using rendering transformation parameters
CN105916095B (en) * 2016-05-31 2017-08-04 音曼(北京)科技有限公司 The method of feedback delay network tone color optimization
CN107231599A (en) * 2017-06-08 2017-10-03 北京奇艺世纪科技有限公司 A kind of 3D sound fields construction method and VR devices
EP3518556A1 (en) 2018-01-24 2019-07-31 L-Acoustics UK Limited Method and system for applying time-based effects in a multi-channel audio reproduction system
US11544032B2 (en) * 2019-01-24 2023-01-03 Dolby Laboratories Licensing Corporation Audio connection and transmission device

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2007031896A1 (en) * 2005-09-13 2007-03-22 Koninklijke Philips Electronics N.V. Audio coding
WO2007096808A1 (en) * 2006-02-21 2007-08-30 Koninklijke Philips Electronics N.V. Audio encoding and decoding
US20070223749A1 (en) * 2006-03-06 2007-09-27 Samsung Electronics Co., Ltd. Method, medium, and system synthesizing a stereo signal
US20070280485A1 (en) * 2006-06-02 2007-12-06 Lars Villemoes Binaural multi-channel decoder in the context of non-energy conserving upmix rules
US20080037795A1 (en) * 2006-08-09 2008-02-14 Samsung Electronics Co., Ltd. Method, medium, and system decoding compressed multi-channel signals into 2-channel binaural signals

Family Cites Families (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
RU2038704C1 (en) 1993-08-12 1995-06-27 Владимир Анатольевич Ефремов Three-dimensional speaking system
JP4304401B2 (en) * 2000-06-07 2009-07-29 ソニー株式会社 Multi-channel audio playback device
SE0202159D0 (en) 2001-07-10 2002-07-09 Coding Technologies Sweden Ab Efficientand scalable parametric stereo coding for low bitrate applications
EP1500085B1 (en) 2002-04-10 2013-02-20 Koninklijke Philips Electronics N.V. Coding of stereo signals
SE0301273D0 (en) 2003-04-30 2003-04-30 Coding Technologies Sweden Ab Advanced processing based on a complex exponential-modulated filter bank and adaptive time signaling methods
US7949141B2 (en) * 2003-11-12 2011-05-24 Dolby Laboratories Licensing Corporation Processing audio signals with head related transfer function filters and a reverberator
EP1775996A4 (en) * 2004-06-30 2011-08-10 Pioneer Corp Reverberation adjustment device, reverberation adjustment method, reverberation adjustment program, recording medium containing the program, and sound field correction system
CA2572805C (en) 2004-07-02 2013-08-13 Matsushita Electric Industrial Co., Ltd. Audio signal decoding device and audio signal encoding device
US8204261B2 (en) 2004-10-20 2012-06-19 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Diffuse sound shaping for BCC schemes and the like
US20070055510A1 (en) * 2005-07-19 2007-03-08 Johannes Hilpert Concept for bridging the gap between parametric multi-channel audio coding and matrixed-surround multi-channel coding
KR101333031B1 (en) * 2005-09-13 2013-11-26 코닌클리케 필립스 일렉트로닉스 엔.브이. Method of and device for generating and processing parameters representing HRTFs
US7917561B2 (en) 2005-09-16 2011-03-29 Coding Technologies Ab Partially complex modulated filter bank
KR100754220B1 (en) 2006-03-07 2007-09-03 삼성전자주식회사 Binaural decoder for spatial stereo sound and method for decoding thereof
CN101401455A (en) 2006-03-15 2009-04-01 杜比实验室特许公司 Binaural rendering using subband filters
JP4606507B2 (en) 2006-03-24 2011-01-05 ドルビー インターナショナル アクチボラゲット Spatial downmix generation from parametric representations of multichannel signals
JP2007336080A (en) * 2006-06-13 2007-12-27 Clarion Co Ltd Sound compensation device
JP4850628B2 (en) * 2006-08-28 2012-01-11 キヤノン株式会社 Recording device
US8670570B2 (en) * 2006-11-07 2014-03-11 Stmicroelectronics Asia Pacific Pte., Ltd. Environmental effects generator for digital audio signals
US8363843B2 (en) * 2007-03-01 2013-01-29 Apple Inc. Methods, modules, and computer-readable recording media for providing a multi-channel convolution reverb
DE102007018032B4 (en) * 2007-04-17 2010-11-11 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Generation of decorrelated signals
JP2008301427A (en) * 2007-06-04 2008-12-11 Onkyo Corp Multichannel voice reproduction equipment

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2007031896A1 (en) * 2005-09-13 2007-03-22 Koninklijke Philips Electronics N.V. Audio coding
WO2007096808A1 (en) * 2006-02-21 2007-08-30 Koninklijke Philips Electronics N.V. Audio encoding and decoding
US20070223749A1 (en) * 2006-03-06 2007-09-27 Samsung Electronics Co., Ltd. Method, medium, and system synthesizing a stereo signal
US20070280485A1 (en) * 2006-06-02 2007-12-06 Lars Villemoes Binaural multi-channel decoder in the context of non-energy conserving upmix rules
US20080037795A1 (en) * 2006-08-09 2008-02-14 Samsung Electronics Co., Ltd. Method, medium, and system decoding compressed multi-channel signals into 2-channel binaural signals

Cited By (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103210668A (en) * 2010-09-06 2013-07-17 音尚股份公司 Upmixing method and system for multichannel audio reproduction
CN103210668B (en) * 2010-09-06 2016-05-04 杜比国际公司 For upwards mixed method and the system of multi-channel audio regeneration
CN103430574A (en) * 2011-03-02 2013-12-04 弗兰霍菲尔运输应用研究公司 Apparatus and method for determining a measure for a perceived level of reverberation, audio processor and method for processing a signal
CN103430574B (en) * 2011-03-02 2016-05-25 弗劳恩霍夫应用研究促进协会 For determining apparatus and method, audio process and the method for the treatment of signal for the tolerance of reverberation perception level
US9672806B2 (en) 2011-03-02 2017-06-06 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Apparatus and method for determining a measure for a perceived level of reverberation, audio processor and method for processing a signal
CN103890841A (en) * 2011-11-01 2014-06-25 皇家飞利浦有限公司 Audio object encoding and decoding
CN108665902B (en) * 2017-03-31 2020-12-01 华为技术有限公司 Coding and decoding method and coder and decoder of multi-channel signal
US11386907B2 (en) 2017-03-31 2022-07-12 Huawei Technologies Co., Ltd. Multi-channel signal encoding method, multi-channel signal decoding method, encoder, and decoder
US11894001B2 (en) 2017-03-31 2024-02-06 Huawei Technologies Co., Ltd. Multi-channel signal encoding method, multi-channel signal decoding method, encoder, and decoder
EP3588497A4 (en) * 2017-03-31 2020-01-15 Huawei Technologies Co., Ltd. Multi-channel signal encoding and decoding method and codec
CN108665902A (en) * 2017-03-31 2018-10-16 华为技术有限公司 The decoding method and codec of multi-channel signal
KR102281097B1 (en) 2017-03-31 2021-07-22 후아웨이 테크놀러지 컴퍼니 리미티드 Multi-channel signal encoding and decoding methods and codecs
KR20190122839A (en) * 2017-03-31 2019-10-30 후아웨이 테크놀러지 컴퍼니 리미티드 Multi-channel signal encoding and decoding method and codec
EP3917171A1 (en) * 2017-03-31 2021-12-01 Huawei Technologies Co., Ltd. Multi-channel signal encoding method, multi-channel signal decoding method, encoder, and decoder
US11178505B2 (en) 2017-04-12 2021-11-16 Huawei Technologies Co., Ltd. Multi-channel signal encoding method, multi-channel signal decoding method, encoder, and decoder
KR20210094143A (en) * 2017-04-12 2021-07-28 후아웨이 테크놀러지 컴퍼니 리미티드 Multichannel signal encoding and decoding methods, and codec
KR102281822B1 (en) 2017-04-12 2021-07-26 후아웨이 테크놀러지 컴퍼니 리미티드 Multi-channel signal encoding and decoding method, and codec
KR102424184B1 (en) 2017-04-12 2022-07-25 후아웨이 테크놀러지 컴퍼니 리미티드 Multichannel signal encoding and decoding methods, and codec
US11832087B2 (en) 2017-04-12 2023-11-28 Huawei Technologies Co., Ltd. Multi-channel signal encoding method, multi-channel signal decoding method, encoder, and decoder
KR20190134752A (en) * 2017-04-12 2019-12-04 후아웨이 테크놀러지 컴퍼니 리미티드 Multichannel signal encoding and decoding method, and codec
RU2780508C2 (en) * 2018-01-24 2022-09-26 Л-АКУСТИКС ЮКей ЛТД Method and system for use of time effects in multichannel audio playback system

Also Published As

Publication number Publication date
KR101342425B1 (en) 2013-12-17
BRPI0923174B1 (en) 2020-10-06
JP5524237B2 (en) 2014-06-18
KR20110122667A (en) 2011-11-10
US20110261966A1 (en) 2011-10-27
RU2509442C2 (en) 2014-03-10
RU2011129154A (en) 2013-01-27
EP2377123B1 (en) 2014-10-29
JP2012513138A (en) 2012-06-07
CN102257562B (en) 2013-09-11
CN102257562A (en) 2011-11-23
US8965000B2 (en) 2015-02-24
EP2377123A1 (en) 2011-10-19
BRPI0923174A2 (en) 2016-02-16

Similar Documents

Publication Publication Date Title
US8965000B2 (en) Method and apparatus for applying reverb to a multi-channel audio signal using spatial cue parameters
JP4598830B2 (en) Speech coding using uncorrelated signals.
CA2701360C (en) Method and apparatus for generating a binaural audio signal
US11705143B2 (en) Audio decoder and decoding method
JP5081838B2 (en) Audio encoding and decoding
CA2610430C (en) Channel reconfiguration with side information
RU2407226C2 (en) Generation of spatial signals of step-down mixing from parametric representations of multichannel signals
KR101218777B1 (en) Method of generating a multi-channel signal from down-mixed signal and computer-readable medium thereof
JP7383685B2 (en) Improved binaural dialogue
Kermit-Canfield et al. Signal decorrelation using perceptually informed allpass filters
JP2007104601A (en) Apparatus for supporting header transport function in multi-channel encoding
JP2021015310A (en) Audo decoder and decoding method

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 200980151156.7

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 09801205

Country of ref document: EP

Kind code of ref document: A1

DPE1 Request for preliminary examination filed after expiration of 19th month from priority date (pct application filed from 20040101)
WWE Wipo information: entry into national phase

Ref document number: 2290/KOLNP/2011

Country of ref document: IN

WWE Wipo information: entry into national phase

Ref document number: 2011541416

Country of ref document: JP

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 2009801205

Country of ref document: EP

ENP Entry into the national phase

Ref document number: 20117016794

Country of ref document: KR

Kind code of ref document: A

WWE Wipo information: entry into national phase

Ref document number: 2011129154

Country of ref document: RU

Ref document number: 13132321

Country of ref document: US

REG Reference to national code

Ref country code: BR

Ref legal event code: B01A

Ref document number: PI0923174

Country of ref document: BR

ENP Entry into the national phase

Ref document number: PI0923174

Country of ref document: BR

Kind code of ref document: A2

Effective date: 20110620