US7653533B2 - Removing time delays in signal paths - Google Patents

Removing time delays in signal paths Download PDF

Info

Publication number
US7653533B2
US7653533B2 US11/540,920 US54092006A US7653533B2 US 7653533 B2 US7653533 B2 US 7653533B2 US 54092006 A US54092006 A US 54092006A US 7653533 B2 US7653533 B2 US 7653533B2
Authority
US
United States
Prior art keywords
domain
downmix signal
signal
spatial information
downmix
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US11/540,920
Other languages
English (en)
Other versions
US20070094011A1 (en
Inventor
Hee Suk Pang
Dong Soo Kim
Jae Hyun Lim
Hyen O. Oh
Yang Won Jung
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
LG Electronics Inc
Original Assignee
LG Electronics Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from KR1020060078223A external-priority patent/KR20070037986A/ko
Priority claimed from KR1020060078219A external-priority patent/KR20070074442A/ko
Application filed by LG Electronics Inc filed Critical LG Electronics Inc
Priority to US11/540,920 priority Critical patent/US7653533B2/en
Assigned to LG ELECTRONICS, INC. reassignment LG ELECTRONICS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: JUNG, YANG WON, KIM, DONG SOO, LIM, JAE HYUN, OH, HYEN O., PANG, HEE SUK
Publication of US20070094011A1 publication Critical patent/US20070094011A1/en
Application granted granted Critical
Publication of US7653533B2 publication Critical patent/US7653533B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/008Multichannel audio signal coding or decoding using interchannel correlation to reduce redundancy, e.g. joint-stereo, intensity-coding or matrixing
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • HELECTRICITY
    • H03ELECTRONIC CIRCUITRY
    • H03MCODING; DECODING; CODE CONVERSION IN GENERAL
    • H03M7/00Conversion of a code where information is represented by a given sequence or number of digits to a code where the same, similar or subset of information is represented by a different sequence or number of digits
    • H03M7/30Compression; Expansion; Suppression of unnecessary data, e.g. redundancy reduction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S5/00Pseudo-stereo systems, e.g. in which additional channel signals are derived from monophonic signals by means of phase shifting, time delay or reverberation 
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/04Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
    • G10L19/16Vocoder architecture
    • G10L19/167Audio streaming, i.e. formatting and decoding of an encoded audio signal representation into a data stream for transmission or storage purposes
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/04Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
    • G10L19/16Vocoder architecture
    • G10L19/18Vocoders using multiple modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field

Definitions

  • the disclosed embodiments relate generally to signal processing.
  • Multi-channel audio coding captures a spatial image of a multi-channel audio signal into a compact set of spatial parameters that can be used to synthesize a high quality multi-channel representation from a transmitted downmix signal.
  • a downmix signal can become time delayed relative to other downmix signals and/or corresponding spatial parameters due to signal processing (e.g., time-to-frequency domain conversions).
  • the disclosed embodiments include systems, methods, apparatuses, and computer-readable mediums for compensating one or more signals and/or one or more parameters for time delays in one or more signal processing paths.
  • a method of processing an audio signal includes: receiving an audio signal including a downmix signal and spatial information; converting the downmix signal from a first domain to a second domain; and combining the converted downmix signal and the spatial information, wherein at least one of the combined downmix signal and the combined spatial information is delayed by amount of time that includes an elapsed time of the converting.
  • a system for generating a plural-channel audio signal includes a downmix decoder configured for processing an encoded downmix signal.
  • a plural-channel decoder is operatively coupled to the downmix decoder and configured for generating a plural-channel audio signal.
  • a domain converter is included in the plural-channel decoder and configured for converting the downmix signal from a first domain to a second domain.
  • a delay processor included in the plural-channel decoder and configured for compensating the downmix signal for time delay resulting from converting the downmix signal.
  • a system for generating a plural-channel audio signal includes a domain converter configured for converting a downmix signal from a first domain to a second domain.
  • a delay processor is operatively coupled to the domain converter and configured for compensating the converted downmix signal for time delay resulting from the conversion.
  • a plural-channel audio processor is operatively coupled to the delay processor and configured for generating a plural-channel audio signal from the compensated downmix signal and spatial information.
  • FIGS. 1 to 3 are block diagrams of apparatuses for decoding an audio signal according to embodiments of the present invention, respectively;
  • FIG. 4 is a block diagram of a plural-channel decoding unit shown in FIG. 1 to explain a signal processing method
  • FIG. 5 is a block diagram of a plural-channel decoding unit shown in FIG. 2 to explain a signal processing method
  • FIGS. 6 to 10 are block diagrams to explain a method of decoding an audio signal according to another embodiment of the present invention.
  • a domain of the audio signal can be converted in the audio signal processing.
  • the converting of the domain of the audio signal maybe include a T/F(Time/Frequency) domain conversion and a complexity domain conversion.
  • the T/F domain conversion includes at least one of a time domain signal to a frequency domain signal conversion and a frequency domain signal to time domain signal conversion.
  • the complexity domain conversion means a domain conversion according to complexity of an operation of the audio signal processing. Also, the complexity domain conversion includes a signal in a real frequency domain to a signal in a complex frequency domain, a signal in a complex frequency domain to a signal in a real frequency domain, etc. If an audio signal is processed without considering time alignment, audio quality may be degraded. A delay processing can be performed for the alignment.
  • the delay processing can include at least one of an encoding delay and a decoding delay.
  • the encoding delay means that a signal is delayed by a delay accounted for in the encoding of the signal.
  • the decoding delay means a real time delay introduced during decoding of the signal.
  • Downmix input domain means a domain of a downmix signal receivable in a plural-channel decoding unit that generates a plural-channel audio signal.
  • Residual input domain means a domain of a residual signal receivable in the plural-channel decoding unit.
  • Time-series data means data that needs time synchronization with a plural-channel audio signal or time alignment. Some examples of ‘time series data’ includes data for moving pictures, still images, text, etc.
  • Leading means a process for advancing a signal by a specific time.
  • ‘Lagging’ means a process for delaying a signal by a specific time.
  • Spatial information means information for synthesizing plural-channel audio signals.
  • Spatial information can be spatial parameters, including but not limited to: CLD (channel level difference) indicating an energy difference between two channels, ICC (inter-channel coherences) indicating correlation between two channels), CPC (channel prediction coefficients) that is a prediction coefficient used in generating three channels from two channels, etc.
  • CLD channel level difference
  • ICC inter-channel coherences
  • CPC channel prediction coefficients
  • the audio signal decoding described herein is one example of signal processing that can benefit from the present invention.
  • the present invention can also be applied to other types of signal processing (e.g., video signal processing).
  • the embodiments described herein can be modified to include any number of signals, which can be represented in any kind of domain, including but not limited to: time, Quadrature Mirror Filter (QMF), Modified Discreet Cosine Transform (MDCT), complexity, etc.
  • a method of processing an audio signal includes generating a plural-channel audio signal by combining a downmix signal and spatial information.
  • a downmix signal e.g., time domain, QMF, MDCT. Since conversions between domains can introduce time delay in the signal path of a downmix signal, a step of compensating for a time synchronization difference between a downmix signal and spatial information corresponding to the downmix signal is needed.
  • the compensating for a time synchronization difference can include delaying at least one of the downmix signal and the spatial information.
  • the embodiments described herein can be implemented as instructions on a computer-readable medium, which, when executed by a processor (e.g., computer processor), cause the processor to perform operations that provide the various aspects of the present invention described herein.
  • a processor e.g., computer processor
  • the term “computer-readable medium” refers to any medium that participates in providing instructions to a processor for execution, including without limitation, non-volatile media (e.g., optical or magnetic disks), volatile media (e.g., memory) and transmission media.
  • Transmission media includes, without limitation, coaxial cables, copper wire and fiber optics. Transmission media can also take the form of acoustic, light or radio frequency waves.
  • FIG. 1 is a diagram of an apparatus for decoding an audio signal according to one embodiment of the present invention.
  • an apparatus for decoding an audio signal includes a downmix decoding unit 100 and a plural-channel decoding unit 200 .
  • the downmix decoding unit 100 includes a domain converting unit 110 .
  • the downmix decoding unit 100 transmits a downmix signal XQ 1 processed in a QMF domain to the plural-channel decoding unit 200 without further processing.
  • the downmix decoding unit 100 also transmits a time domain downmix signal XT 1 to the plural-channel decoding unit 200 , which is generated by converting the downmix signal XQ 1 from the QMF domain to the time domain using the converting unit 110 .
  • Techniques for converting an audio signal from a QMF domain to a time domain are well-known and have been incorporated in publicly available audio signal processing standards (e.g., MPEG).
  • the plural-channel decoding unit 200 generates a plural-channel audio signal XM 1 using the downmix signal XT 1 or XQ 1 , and spatial information SI 1 or SI 2 .
  • FIG. 2 is a diagram of an apparatus for decoding an audio signal according to another embodiment of the present invention.
  • the apparatus for decoding an audio signal includes a downmix decoding unit 100 a , a plural-channel decoding unit 200 a and a domain converting unit 300 a.
  • the downmix decoding unit 100 a includes a domain converting unit 110 a .
  • the downmix decoding unit 100 a outputs a downmix signal Xm processed in a MDCT domain.
  • the downmix decoding unit 100 a also outputs a downmix signal XT 2 in a time domain, which is generated by converting Xm from the MDCT domain to the time domain using the converting unit 110 a.
  • the downmix signal XT 2 in a time domain is transmitted to the plural-channel decoding unit 200 a .
  • the downmix signal Xm in the MDCT domain passes through the domain converting unit 300 a , where it is converted to a downmix signal XQ 2 in a QMF domain.
  • the converted downmix signal XQ 2 is then transmitted to the plural-channel decoding unit 200 a.
  • the plural-channel decoding unit 200 a generates a plural-channel audio signal XM 2 using the transmitted downmix signal XT 2 or XQ 2 and spatial information SI 3 or SI 4 .
  • FIG. 3 is a diagram of an apparatus for decoding an audio signal according to another embodiment of the present invention.
  • the apparatus for decoding an audio signal includes a downmix decoding unit 100 b , a plural-channel decoding unit 200 b , a residual decoding unit 400 b and a domain converting unit 500 b.
  • the downmix decoding unit 100 b includes a domain converting unit 110 b .
  • the downmix decoding unit 100 b transmits a downmix signal XQ 3 processed in a QMF domain to the plural-channel decoding unit 200 b without further processing.
  • the downmix decoding unit 100 b also transmits a downmix signal XT 3 to the plural-channel decoding unit 200 b , which is generated by converting the downmix signal XQ 3 from a QMF domain to a time domain using the converting unit 110 b.
  • an encoded residual signal RB is inputted into the residual decoding unit 400 b and then processed.
  • the processed residual signal RM is a signal in an MDCT domain.
  • a residual signal can be, for example, a prediction error signal commonly used in audio coding applications (e.g., MPEG).
  • the residual signal RM in the MDCT domain is converted to a residual signal RQ in a QMF domain by the domain converting unit 500 b , and then transmitted to the plural-channel decoding unit 200 b.
  • the processed residual signal can be transmitted to the plural-channel decoding unit 200 b without undergoing a domain converting process.
  • FIG. 3 shows that in some embodiments the domain converting unit 500 b converts the residual signal RM in the MDCT domain to the residual signal RQ in the QMF domain.
  • the domain converting unit 500 b is configured to convert the residual signal RM outputted from the residual decoding unit 400 b to the residual signal RQ in the QMF domain.
  • An audio signal process generates a plural-channel audio signal by decoding an encoded audio signal including a downmix signal and spatial information.
  • the downmix signal and the spatial information undergo different processes, which can cause different time delays.
  • the downmix signal and the spatial information can be encoded to be time synchronized.
  • the downmix signal and the spatial information can be time synchronized by considering the domain in which the downmix signal processed in the downmix decoding unit 100 , 100 a or 100 b is transmitted to the plural-channel decoding unit 200 , 200 a or 200 b.
  • a downmix coding identifier can be included in the encoded audio signal for identifying the domain in which the time synchronization between the downmix signal and the spatial information is matched.
  • the downmix coding identifier can indicate a decoding scheme of a downmix signal.
  • the encoded audio signal can be decoded by an AAC decoder.
  • AAC Advanced Audio Coding
  • the downmix coding identifier can also be used to determine a domain for matching the time synchronization between the downmix signal and the spatial information.
  • a downmix signal can be processed in a domain different from a time-synchronization matched domain and then transmitted to the plural-channel decoding unit 200 , 200 a or 200 b .
  • the decoding unit 200 , 200 a or 200 b compensates for the time synchronization between the downmix signal and the spatial information to generate a plural-channel audio signal.
  • a method of compensating for a time synchronization difference between a downmix signal and spatial information is explained with reference to FIG. 1 and FIG. 4 as follows.
  • FIG. 4 is a block diagram of the plural-channel decoding unit 200 shown in FIG. 1 .
  • the downmix signal processed in the downmix decoding unit 100 can be transmitted to the plural-channel decoding unit 200 in one of two kinds of domains.
  • a downmix signal and spatial information are matched together with time synchronization in a QMF domain. Other domains are possible.
  • a downmix signal XQ 1 processed in the QMF domain is transmitted to the plural-channel decoding unit 200 for signal processing.
  • the transmitted downmix signal XQ 1 is combined with spatial information SI 1 in a plural-channel generating unit 230 to generate the plural-channel audio signal XM 1 .
  • the spatial information SI 1 is combined with the downmix signal XQ 1 after being delayed by a time corresponding to time synchronization in encoding.
  • the delay can be an encoding delay. Since the spatial information SI 1 and the downmix signal XQ 1 are matched with time synchronization in encoding, a plural-channel audio signal can be generated without a special synchronization matching process. That is, in this case, the spatial information ST 1 is not delayed by a decoding delay.
  • the downmix signal XT 1 processed in the time domain is transmitted to the plural-channel decoding unit 200 for signal processing.
  • the downmix signal XQ 1 in a QMF domain is converted to a downmix signal XT 1 in a time domain by the domain converting unit 110 , and the downmix signal XT 1 in the time domain is transmitted to the plural-channel decoding unit 200 .
  • the transmitted downmix signal XT 1 is converted to a downmix signal Xq 1 in the QMF domain by the domain converting unit 210 .
  • At least one of the downmix signal Xq 1 and spatial information SI 2 can be transmitted to the plural-channel generating unit 230 after completion of time delay compensation.
  • the plural-channel generating unit 230 can generate a plural-channel audio signal XM 1 by combining a transmitted downmix signal Xql′ and spatial information SI 2 ′.
  • the time delay compensation should be performed on at least one of the downmix signal Xq 1 and the spatial information SI 2 , since the time synchronization between the spatial information and the downmix signal is matched in the QMF domain in encoding.
  • the domain-converted downmix signal Xq 1 can be inputted to the plural-channel generating unit 230 after being compensated for the mismatched time synchronization difference in a signal delay processing unit 220 .
  • a method of compensating for the time synchronization difference is to lead the downmix signal Xq 1 by the time synchronization difference.
  • the time synchronization difference can be a total of a delay time generated from the domain converting unit 110 and a delay time of the domain converting unit 210 .
  • the spatial information SI 2 is lagged by the time synchronization difference in a spatial information delay processing unit 240 and then transmitted to the plural-channel generating unit 230 .
  • a delay value of substantially delayed spatial information corresponds to a total of a mismatched time synchronization difference and a delay time of which time synchronization has been matched. That is, the delayed spatial information is delayed by the encoding delay and the decoding delay. This total also corresponds to a total of the time synchronization difference between the downmix signal and the spatial information generated in the downmix decoding unit 100 ( FIG. 1 ) and the time synchronization difference generated in the plural-channel decoding unit 200 .
  • the delay value of the substantially delayed spatial information SI 2 can be determined by considering the performance and delay of a filter (e.g., a QMF, hybrid filter bank).
  • a filter e.g., a QMF, hybrid filter bank.
  • a spatial information delay value which considers performance and delay of a filter, can be 961 time samples.
  • the time synchronization difference generated in the downmix decoding unit 100 is 257 time samples and the time synchronization difference generated in the plural-channel decoding unit 200 is 704 time samples.
  • the delay value is represented by a time sample unit, it can be represented by a timeslot unit as well.
  • FIG. 5 is a block diagram of the plural-channel decoding unit 200 a shown in FIG. 2 .
  • the downmix signal processed in the downmix decoding unit 100 a can be transmitted to the plural-channel decoding unit 200 a in one of two kinds of domains.
  • a downmix signal and spatial information are matched together with time synchronization in a QMF domain.
  • Other domains are possible.
  • An audio signal, of which downmix signal and spatial information are matched on a domain different from a time domain, can be processed.
  • the downmix signal XT 2 processed in a time domain is transmitted to the plural-channel decoding unit 200 a for signal processing.
  • a downmix signal Xm in an MDCT domain is converted to a downmix signal XT 2 in a time domain by the domain converting unit 110 a.
  • the converted downmix signal XT 2 is then transmitted to the plural-channel decoding unit 200 a.
  • the transmitted downmix signal XT 2 is converted to a downmix signal Xq 2 in a QMF domain by the domain converting unit 210 a and is then transmitted to a plural-channel generating unit 230 a.
  • the transmitted downmix signal Xq 2 is combined with spatial information S 13 in the plural-channel generating unit 230 a to generate the plural-channel audio signal XM 2 .
  • the spatial information S 13 is combined with the downmix signal Xq 2 after delaying an amount of time corresponding to time synchronization in encoding.
  • the delay can be an encoding delay. Since the spatial information SI 3 and the downmix signal Xq 2 are matched with time synchronization in encoding, a plural-channel audio signal can be generated without a special synchronization matching process. That is, in this case, the spatial information SI 3 is not delayed by a decoding delay.
  • the downmix signal XQ 2 processed in a QMF domain is transmitted to the plural-channel decoding unit 200 a for signal processing.
  • the downmix signal Xm processed in an MDCT domain is outputted from a downmix decoding unit 100 a .
  • the outputted downmix signal Xm is converted to a downmix signal XQ 2 in a QMF domain by the domain converting unit 300 a .
  • the converted downmix signal XQ 2 is then transmitted to the plural-channel decoding unit 200 a.
  • the downmix signal XQ 2 in the QMF domain is transmitted to the plural-channel decoding unit 200 a , at least one of the downmix signal XQ 2 or spatial information SI 4 can be transmitted to the plural-channel generating unit 230 a after completion of time delay compensation.
  • the plural-channel generating unit 230 a can generate the plural-channel audio signal XM 2 by combining a transmitted downmix signal XQ 2 ′ and spatial information SI 4 ′ together.
  • the reason why the time delay compensation should be performed on at least one of the downmix signal XQ 2 and the spatial information SI 4 is because time synchronization between the spatial information and the downmix signal is matched in the time domain in encoding.
  • the domain-converted downmix signal XQ 2 can be inputted to the plural-channel generating unit 230 a after having been compensated for the mismatched time synchronization difference in a signal delay processing unit 220 a.
  • a method of compensating for the time synchronization difference is to lag the downmix signal XQ 2 by the time synchronization difference.
  • the time synchronization difference can be a difference between a delay time generated from the domain converting unit 300 a and a total of a delay time generated from the domain converting unit 110 a and a delay time generated from the domain converting unit 210 a.
  • the spatial information SI 4 is led by the time synchronization difference in a spatial information delay processing unit 240 a and then transmitted to the plural-channel generating unit 230 a.
  • a delay value of substantially delayed spatial information corresponds to a total of a mismatched time synchronization difference and a delay time of which time synchronization has been matched. That is, the delayed spatial information SI 4 ′ is delayed by the encoding delay and the decoding delay.
  • a method of processing an audio signal according to one embodiment of the present invention includes encoding an audio signal of which time synchronization between a downmix signal and spatial information is matched by assuming a specific decoding scheme and decoding the encoded audio signal.
  • the high quality decoding scheme outputs a plural-channel audio signal having audio quality that is more refined than that of the lower power decoding scheme.
  • the lower power decoding scheme has relatively lower power consumption due to its configuration, which is less complicated than that of the high quality decoding scheme.
  • FIG. 6 is a block diagram to explain a method of decoding an audio signal according to another embodiment of the present invention.
  • a decoding apparatus includes a downmix decoding unit 100 c and a plural-channel decoding unit 200 c.
  • a downmix signal XT 4 processed in the downmix decoding unit 100 c is transmitted to the plural-channel decoding unit 200 c , where the signal is combined with spatial information SI 7 or SI 8 to generate a plural-channel audio signal M 1 or M 2 .
  • the processed downmix signal XT 4 is a downmix signal in a time domain.
  • An encoded downmix signal DB is transmitted to the downmix decoding unit 100 c and processed.
  • the processed downmix signal XT 4 is transmitted to the plural-channel decoding unit 200 c , which generates a plural-channel audio signal according to one of two kinds of decoding schemes: a high quality decoding scheme and a low power decoding scheme.
  • the downmix signal XT 4 is transmitted and decoded along a path P 2 .
  • the processed downmix signal XT 4 is converted to a signal XRQ in a real QMF domain by a domain converting unit 240 c.
  • the converted downmix signal XRQ is converted to a signal XQC 2 in a complex QMF domain by a domain converting unit 250 c .
  • the XRQ downmix signal to the XQC 2 downmix signal conversion is an example of complexity domain conversion.
  • the signal XQC 2 in the complex QMF domain is combined with spatial information SI 8 in a plural-channel generating unit 260 c to generate the plural-channel audio signal M 2 .
  • the downmix signal XT 4 is transmitted and decoded along a path P 1 .
  • the processed downmix signal XT 4 is converted to a signal XCQ 1 in a complex QMF domain by a domain converting unit 210 c.
  • the converted downmix signal XCQ 1 is then delayed by a time delay difference between the downmix signal XCQ 1 and spatial information SI 7 in a signal delay processing unit 220 c.
  • the delayed downmix signal XCQ 1 ′ is combined with spatial information SI 7 in a plural-channel generating unit 230 c , which generates the plural-channel audio signal M 1 .
  • the downmix signal XCQ 1 passes through the signal delay processing unit 220 c . This is because a time synchronization difference between the downmix signal XCQ 1 and the spatial information SI 7 is generated due to the encoding of the audio signal on the assumption that a low power decoding scheme will be used.
  • the time synchronization difference is a time delay difference, which depends on the decoding scheme that is used. For example, the time delay difference occurs because the decoding process of, for example, a low power decoding scheme is different than a decoding process of a high quality decoding scheme.
  • the time delay difference is considered until a time point of combining a downmix signal and spatial information, since it may not be necessary to synchronize the downmix signal and spatial information after the time point of combining the downmix signal and the spatial information.
  • the time synchronization difference is a difference between a first delay time occurring until a time point of combining the downmix signal XCQ 2 and the spatial information SI 8 and a second delay time occurring until a time point of combining the downmix signal XCQ 1 ′ and the spatial information SI 7 .
  • a time sample or timeslot can be used as a unit of time delay.
  • the delay time occurring in the domain converting unit 210 c is equal to the. delay time occurring in the domain converting unit 240 c , it is enough for the signal delay processing unit 220 c to delay the downmix signal XCQ 1 by the delay time occurring in the domain converting unit 250 c.
  • the two decoding schemes are included in the plural-channel decoding unit 200 c .
  • one decoding scheme can be included in the plural-channel decoding unit 200 c.
  • the time synchronization between the downmix signal and the spatial information is matched in accordance with the low power decoding scheme.
  • the present invention further includes the case that the time synchronization between the downmix signal and the spatial information is matched in accordance with the high quality decoding scheme.
  • the downmix signal is led in a manner opposite to the case of matching the time synchronization by the low power decoding scheme.
  • FIG. 7 is a block diagram to explain a method of decoding an audio signal according to another embodiment of the present invention.
  • a decoding apparatus includes a downmix decoding unit 100 d and a plural-channel decoding unit 200 d.
  • a downmix signal XT 4 processed in the downmix decoding unit 100 d is transmitted to the plural-channel decoding unit 200 d , where the downmix signal is combined with spatial information SI 7 ′ or SI 8 to generate a plural-channel audio signal M 3 or M 2 .
  • the processed downmix signal XT 4 is a signal in a time domain.
  • An encoded downmix signal DB is transmitted to the downmix decoding unit 100 d and processed.
  • the processed downmix signal XT 4 is transmitted to the plural-channel decoding unit 200 d , which generates a plural-channel audio signal according to one of two kinds of decoding schemes: a high quality decoding scheme and a low power decoding scheme.
  • the downmix signal XT 4 is transmitted and decoded along a path P 4 .
  • the processed downmix signal XT 4 is converted to a signal XRQ in a real QMF domain by a domain converting unit 240 d.
  • the converted downmix signal XRQ is converted to a signal XQC 2 in a complex QMF domain by a domain converting unit 250 d .
  • the XRQ downmix signal to the XCQ 2 downmix signal conversion is an example of complexity domain conversion.
  • the signal XQC 2 in the complex QMF domain is combined with spatial information SI 8 in a plural-channel generating unit 260 d to generate the plural-channel audio signal M 2 .
  • the downmix signal XT 4 is transmitted and decoded along a path P 3 .
  • the processed downmix signal XT 4 is converted to a signal XCQ 1 in a complex QMF domain by a domain converting unit 210 d.
  • the converted downmix signal XCQ 1 is transmitted to a plural-channel generating unit 230 d , where it is combined with the spatial information SI 7 ′ to generate the plural-channel audio signal M 3 .
  • the spatial information SI 7 ′ is the spatial information of which time delay is compensated for as the spatial information SI 7 passes through a spatial information delay processing unit 220 d.
  • the spatial information SI 7 passes through the spatial information delay processing unit 220 d . This is because a time synchronization difference between the downmix signal XCQ 1 and the spatial information SI 7 is generated due to the encoding of the audio signal on the assumption that a low power decoding scheme will be used.
  • the time synchronization difference is a time delay difference, which depends on the decoding scheme that is used. For example, the time delay difference occurs because the decoding process of, for example, a low power decoding scheme is different than a decoding process of a high quality decoding scheme.
  • the time delay difference is considered until a time point of combining a downmix signal and spatial information, since it is not necessary to synchronize the downmix signal and spatial information after the time point of combining the downmix signal and the spatial information.
  • the time synchronization difference is a difference between a first delay time occurring until a time point of combining the downmix signal XCQ 2 and the spatial information SI 8 and a second delay time occurring until a time point of combining the downmix signal XCQ 1 and the spatial information SI 7 ′.
  • a time sample or timeslot can be used as a unit of time delay.
  • the delay time occurring in the domain converting unit 210 d is equal to the delay time occurring in the domain converting unit 240 d , it is enough for the spatial information delay processing unit 220 d to lead the spatial information SI 7 by the delay time occurring in the domain converting unit 250 d.
  • the two decoding schemes are included in the plural-channel decoding unit 200 d .
  • one decoding scheme can be included in the plural-channel decoding unit 200 d.
  • the time synchronization between the downmix signal and the spatial information is matched in accordance with the low power decoding scheme.
  • the present invention further includes the case that the time synchronization between the downmix signal and the spatial information is matched in accordance with the high quality decoding scheme.
  • the downmix signal is lagged in a manner opposite to the case of matching the time synchronization by the low power decoding scheme.
  • FIG. 6 and FIG. 7 exemplarily show that one of the signal delay processing unit 220 c and the spatial information delay unit 220 d is included in the plural-channel decoding unit 200 c or 200 d
  • the present invention includes an embodiment where the spatial information delay processing unit 220 d and the signal delay processing unit 220 c are included in the plural-channel decoding unit 200 c or 200 d .
  • a total of a delay compensation time in the spatial information delay processing unit 220 d and a delay compensation time in the signal delay processing unit 220 c should be equal to the time synchronization difference.
  • FIG. 8 is a block diagram to explain a method of decoding an audio signal according to one embodiment of the present invention.
  • a decoding apparatus includes a downmix decoding unit 100 e and a plural-channel decoding unit 200 e.
  • a downmix signal processed in the downmix decoding unit 100 e can be transmitted to the plural-channel decoding unit 200 e in one of two kinds of domains.
  • time synchronization between a downmix signal and spatial information is matched on a QMF domain with reference to a low power decoding scheme.
  • various modifications can be applied to the present invention.
  • the downmix signal XQ 5 can be any one of a complex QMF signal XCQ 5 and real QMF single XRQ 5 .
  • the XCQ 5 is processed by the high quality decoding scheme in the downmix decoding unit 100 e .
  • the XRQ 5 is processed by the low power decoding scheme in the downmix decoding unit 100 e.
  • a signal processed by a high quality decoding scheme in the downmix decoding unit 100 e is connected to the plural-channel decoding unit 200 e of the high quality decoding scheme
  • a signal processed by the low power decoding scheme in the downmix decoding unit 100 e is connected to the plural-channel decoding unit 200 e of the low power decoding scheme.
  • various modifications can be applied to the present invention.
  • the downmix signal XQ 5 is transmitted and decoded along a path P 6 .
  • the XQ 5 is a downmix signal XRQ 5 in a real QMF domain.
  • the downmix signal XRQ 5 is combined with spatial information SI 10 in a multi-channel generating unit 231 e to generate a multi-channel audio signal M 5 .
  • the downmix signal XQ 5 is transmitted and decoded along a path P 5 .
  • the XQ 5 is a downmix signal XCQ 5 in a complex QMF domain.
  • the downmix signal XCQ 5 is combined with the spatial information S 19 in a multi-channel generating unit 230 e to generate a multi-channel audio signal M 4 .
  • a downmix signal XT 5 processed in the downmix decoding unit 100 e is transmitted to the plural-channel decoding unit 200 e , where it is combined with spatial information SI 11 or SI 12 to generate a plural-channel audio signal M 6 or M 7 .
  • the downmix signal XT 5 is transmitted to the plural-channel decoding unit 200 e , which generates a plural-channel audio signal according to one of two kinds of decoding schemes: a high quality decoding scheme and a low power decoding scheme.
  • the downmix signal XT 5 is transmitted and decoded along a path P 8 .
  • the processed downmix signal XT 5 is converted to a signal XR in a real QMF domain by a domain converting unit 241 e.
  • the converted downmix signal XR is converted to a signal XC 2 in a complex QMF domain by a domain converting unit 250 e .
  • the XR downmix signal to the XC 2 downmix signal conversion is an example of complexity domain conversion.
  • the signal XC 2 in the complex QMF domain is combined with spatial information SI 12 ′ in a plural-channel generating unit 233 e , which generates a plural-channel audio signal M 7 .
  • the spatial information SI 12 ′ is the spatial information of which time delay is compensated for as the spatial information SI 12 passes through a spatial information delay processing unit 240 e.
  • the spatial information SI 12 passes through the spatial information delay processing unit 240 e .
  • a time synchronization difference between the downmix signal XC 2 and the spatial information SI 12 is generated due to the audio signal encoding performed by the low power decoding scheme on the assumption that a domain, of which time synchronization between the downmix signal and the spatial information is matched, is the QMF domain.
  • the delayed spatial information SI 12 ′ is delayed by the encoding delay and the decoding delay.
  • the downmix signal XT 5 is transmitted and decoded along a path P 7 .
  • the processed downmix signal XT 5 is converted to a signal XC 1 in a complex QMF domain by a domain converting unit 240 e.
  • the converted downmix signal XC 1 and the spatial information SI 11 are compensated for a time delay by a time synchronization difference between the downmix signal XC 1 and the spatial information Sl 11 in a signal delay processing unit 250 e and a spatial information delay processing unit 260 e , respectively.
  • time-delay-compensated downmix signal XC 1 ′ is combined with the time-delay-compensated spatial information SI 11 ′ in a plural-channel generating unit 232 e , which generates a plural-channel audio signal M 6 .
  • the downmix signal XC 1 passes through the signal delay processing unit 250 e and the spatial information SI 11 passes through the spatial information delay processing unit 260 e .
  • a time synchronization difference between the downmix signal XC 1 and the spatial information SI 11 is generated due to the encoding of the audio signal under the assumption of a low power decoding scheme, and on the further assumption that a domain, of which time synchronization between the downmix signal and the spatial information is matched, is the QMF domain.
  • FIG. 9 is a block diagram to explain a method of decoding an audio signal according to one embodiment of the present invention.
  • a decoding apparatus includes a downmix decoding unit 100 f and a plural-channel decoding unit 200 f.
  • An encoded downmix signal DB 1 is transmitted to the downmix decoding unit 100 f and then processed.
  • the downmix signal DB 1 is encoded considering two downmix decoding schemes, including a first downmix decoding and a second downmix decoding scheme.
  • the downmix signal DB 1 is processed according to one downmix decoding scheme in downmix decoding unit 100 f .
  • the one downmix decoding scheme can be the first downmix decoding scheme.
  • the processed downmix signal XT 6 is transmitted to the plural-channel decoding unit 200 f , which generates a plural-channel audio signal Mf.
  • the processed downmix signal XT 6 ′ is delayed by a decoding delay in a signal processing unit 210 f .
  • the downmix signal XT 6 ′ can be a delayed by a decoding delay.
  • the reason why the downmix signal XT 6 is delayed is that the downmix decoding scheme that is accounted for in encoding is different from the downmix decoding scheme used in decoding.
  • the delayed downmix signal XT 6 ′ is upsampled in upsampling unit 220 f .
  • the reason why the downmix signal XT 6 ′ is upsampled is that the number of samples of the downmix signal XT 6 ′ is different from the number of samples of the spatial information SI 13 .
  • the order of the delay processing of the downmix signal XT 6 and the upsampling processing of the downmix signal XT 6 ′ is interchangeable.
  • the domain of the upsampled downmix signal UXT 6 is converted in domain processing unit 230 f .
  • the conversion of the domain of the downmix signal UXT 6 can include the F/T domain conversion and the complexity domain conversion.
  • the domain converted downmix signal UXTD 6 is combined with spatial information SI 13 in a plural-channel generating unit 260 d , which generates the plural-channel audio signal Mf.
  • FIG. 10 is a block diagram of an apparatus for decoding an audio signal according to one embodiment of the present invention.
  • an apparatus for decoding an audio signal includes a time series data decoding unit 10 and a plural-channel audio signal processing unit 20 .
  • the plural-channel audio signal processing unit 20 includes a downmix decoding unit 21 , a plural-channel decoding unit 22 and a time delay compensating unit 23 .
  • a downmix bitstream IN 2 which is an example of an encoded downmix signal, is inputted to the downmix decoding unit 21 to be decoded.
  • the downmix bit stream IN 2 can be decoded and outputted in two kinds of domains.
  • the output available domains include a time domain and a QMF domain.
  • a reference number ‘ 50 ’ indicates a downmix signal decoded and outputted in a time domain and a reference number ‘ 51 ’ indicates a downmix signal decoded and outputted in a QMF domain.
  • two kinds of domains are described.
  • the present invention includes downmix signals decoded and outputted on other kinds of domains.
  • the downmix signals 50 and 51 are transmitted to the plural-channel decoding unit 22 and then decoded according to two kinds of decoding schemes 22 H and 22 L, respectively.
  • the reference number ‘ 22 H’ indicates a high quality decoding scheme
  • the reference number ‘ 22 L’ indicates a low power decoding scheme.
  • the downmix signal 50 decoded and outputted in the time domain is decoded according to a selection of one of two paths P 9 and P 10 .
  • the path P 9 indicates a path for decoding by the high quality decoding scheme 22 H and the path P 10 indicates a path for decoding by the low power decoding scheme 22 L.
  • the downmix signal 50 transmitted along the path P 9 is combined with spatial information SI according to the high quality decoding scheme 22 H to generate a plural-channel audio signal MHT.
  • the downmix signal 50 transmitted along the path P 10 is combined with spatial information SI according to the low power decoding scheme 22 L to generate a plural-channel audio signal MLT.
  • the other downmix signal 51 decoded and outputted in the QMF domain is decoded according to a selection of one of two paths P 11 and P 12 .
  • the path P 11 indicates a path for decoding by the high quality decoding scheme 22 H and the path P 12 indicates a path for decoding by the low power decoding scheme 22 L.
  • the downmix signal 51 transmitted along the path P 11 is combined with spatial information SI according to the high quality decoding scheme 22 H to generate a plural-channel audio signal MHQ.
  • the downmix signal 51 transmitted along the path P 12 is combined with spatial information SI according to the low power decoding scheme 22 L to generate a plural-channel audio signal MLQ.
  • At least one of the plural-channel audio signals MHT, MHQ, MLT and MLQ generated by the above-explained methods undergoes a time delay compensating process in the time delay compensating unit 23 and is then outputted as OUT 2 , OUT 3 , OUT 4 or OUT 5 .
  • the time delay compensating process is able to prevent a time delay from occurring in a manner of comparing a time synchronization mismatched plural-channel audio signal MHQ, MLT or MKQ to a plural-channel audio signal MHT on the assumption that a time synchronization between time-series data OUT 1 decoded and outputted in the time series decoding unit 10 and the aforesaid plural-channel audio signal MHT is matched.
  • a time synchronization with the time series data OUT 1 can be matched by compensating for a time delay of one of the rest of the plural-channel audio signals of which time synchronization is mismatched.
  • the embodiment can also perform the time delay compensating process in case that the time series data OUT 1 and the plural-channel audio signal MHT, MHQ, MLT or MLQ are not processed together. For instance, a time delay of the plural-channel audio signal is compensated and is prevented from occurring using a result of comparison with the plural-channel audio signal MLT. This can be diversified in various ways.
  • the present invention provides the following effects or advantages.
  • the present invention prevents audio quality degradation by compensating for the time synchronization difference.
  • the present invention is able to compensate for a time synchronization difference between time series data and a plural-channel audio signal to be processed together with the time series data of a moving picture, a text, a still image and the like.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • Mathematical Physics (AREA)
  • Theoretical Computer Science (AREA)
  • Quality & Reliability (AREA)
  • Stereophonic System (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)
  • Apparatus For Radiation Diagnosis (AREA)
  • Oscillators With Electromechanical Resonators (AREA)
  • Radar Systems Or Details Thereof (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Synchronisation In Digital Transmission Systems (AREA)
US11/540,920 2005-10-24 2006-09-29 Removing time delays in signal paths Active 2028-07-30 US7653533B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US11/540,920 US7653533B2 (en) 2005-10-24 2006-09-29 Removing time delays in signal paths

Applications Claiming Priority (17)

Application Number Priority Date Filing Date Title
US72922505P 2005-10-24 2005-10-24
US75700506P 2006-01-09 2006-01-09
US78674006P 2006-03-29 2006-03-29
US79232906P 2006-04-17 2006-04-17
KR1020060078223A KR20070037986A (ko) 2005-10-04 2006-08-18 다채널 오디오 신호의 처리방법 및 그 장치
KR10-2006-0078219 2006-08-18
KR1020060078222A KR20070037985A (ko) 2005-10-04 2006-08-18 다채널 오디오 신호의 디코딩 방법 및 그 장치
KR10-2006-0078218 2006-08-18
KR1020060078219A KR20070074442A (ko) 2006-01-09 2006-08-18 다채널 오디오 복원 장치 및 방법과 이 장치에서 수행되는프로그램을 기록한 컴퓨터로 읽을 수 있는 기록 매체
KR1020060078225A KR20070037987A (ko) 2005-10-04 2006-08-18 다채널 오디오 신호의 디코딩 방법 및 장치
KR10-2006-0078222 2006-08-18
KR1020060078221A KR20070037984A (ko) 2005-10-04 2006-08-18 다채널 오디오 신호의 디코딩 방법 및 그 장치
KR10-2006-0078223 2006-08-18
KR10-2006-0078225 2006-08-18
KR10-2006-0078221 2006-08-18
KR1020060078218A KR20070037983A (ko) 2005-10-04 2006-08-18 다채널 오디오 신호의 디코딩 방법 및 부호화된 오디오신호 생성방법
US11/540,920 US7653533B2 (en) 2005-10-24 2006-09-29 Removing time delays in signal paths

Publications (2)

Publication Number Publication Date
US20070094011A1 US20070094011A1 (en) 2007-04-26
US7653533B2 true US7653533B2 (en) 2010-01-26

Family

ID=44454038

Family Applications (8)

Application Number Title Priority Date Filing Date
US11/541,397 Expired - Fee Related US7742913B2 (en) 2005-10-24 2006-09-29 Removing time delays in signal paths
US11/540,920 Active 2028-07-30 US7653533B2 (en) 2005-10-24 2006-09-29 Removing time delays in signal paths
US11/541,471 Abandoned US20070092086A1 (en) 2005-10-24 2006-09-29 Removing time delays in signal paths
US11/540,919 Active 2028-05-01 US7761289B2 (en) 2005-10-24 2006-09-29 Removing time delays in signal paths
US11/541,395 Active 2029-01-01 US7840401B2 (en) 2005-10-24 2006-09-29 Removing time delays in signal paths
US11/541,472 Active 2028-09-15 US7716043B2 (en) 2005-10-24 2006-09-29 Removing time delays in signal paths
US12/872,044 Active US8095358B2 (en) 2005-10-24 2010-08-31 Removing time delays in signal paths
US12/872,081 Active US8095357B2 (en) 2005-10-24 2010-08-31 Removing time delays in signal paths

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US11/541,397 Expired - Fee Related US7742913B2 (en) 2005-10-24 2006-09-29 Removing time delays in signal paths

Family Applications After (6)

Application Number Title Priority Date Filing Date
US11/541,471 Abandoned US20070092086A1 (en) 2005-10-24 2006-09-29 Removing time delays in signal paths
US11/540,919 Active 2028-05-01 US7761289B2 (en) 2005-10-24 2006-09-29 Removing time delays in signal paths
US11/541,395 Active 2029-01-01 US7840401B2 (en) 2005-10-24 2006-09-29 Removing time delays in signal paths
US11/541,472 Active 2028-09-15 US7716043B2 (en) 2005-10-24 2006-09-29 Removing time delays in signal paths
US12/872,044 Active US8095358B2 (en) 2005-10-24 2010-08-31 Removing time delays in signal paths
US12/872,081 Active US8095357B2 (en) 2005-10-24 2010-08-31 Removing time delays in signal paths

Country Status (11)

Country Link
US (8) US7742913B2 (fr)
EP (6) EP1952673A1 (fr)
JP (6) JP5399706B2 (fr)
KR (7) KR100875428B1 (fr)
CN (6) CN101297596B (fr)
AU (1) AU2006306942B2 (fr)
BR (1) BRPI0617779A2 (fr)
CA (1) CA2626132C (fr)
HK (1) HK1126071A1 (fr)
TW (6) TWI317246B (fr)
WO (6) WO2007049866A1 (fr)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100198589A1 (en) * 2008-07-29 2010-08-05 Tomokazu Ishikawa Audio coding apparatus, audio decoding apparatus, audio coding and decoding apparatus, and teleconferencing system
US20130282384A1 (en) * 2007-09-25 2013-10-24 Motorola Mobility Llc Apparatus and Method for Encoding a Multi-Channel Audio Signal

Families Citing this family (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7116787B2 (en) * 2001-05-04 2006-10-03 Agere Systems Inc. Perceptual synthesis of auditory scenes
US7644003B2 (en) * 2001-05-04 2010-01-05 Agere Systems Inc. Cue-based audio coding/decoding
US7805313B2 (en) * 2004-03-04 2010-09-28 Agere Systems Inc. Frequency-based coding of channels in parametric multi-channel coding systems
US8204261B2 (en) * 2004-10-20 2012-06-19 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Diffuse sound shaping for BCC schemes and the like
US7720230B2 (en) * 2004-10-20 2010-05-18 Agere Systems, Inc. Individual channel shaping for BCC schemes and the like
DE602005017302D1 (de) * 2004-11-30 2009-12-03 Agere Systems Inc Synchronisierung von parametrischer raumtonkodierung mit extern bereitgestelltem downmix
US7787631B2 (en) * 2004-11-30 2010-08-31 Agere Systems Inc. Parametric coding of spatial audio with cues based on transmitted channels
EP1817767B1 (fr) * 2004-11-30 2015-11-11 Agere Systems Inc. Codage parametrique d'audio spatial avec des informations laterales basees sur des objets
US7903824B2 (en) * 2005-01-10 2011-03-08 Agere Systems Inc. Compact side information for parametric coding of spatial audio
EP1921606B1 (fr) * 2005-09-02 2011-10-19 Panasonic Corporation Dispositif de conformage d'énergie et procédé de conformage d'énergie
US7742913B2 (en) 2005-10-24 2010-06-22 Lg Electronics Inc. Removing time delays in signal paths
CN101578654B (zh) * 2006-07-04 2013-04-24 韩国电子通信研究院 用于恢复多通道音频信号的设备和方法
FR2911031B1 (fr) * 2006-12-28 2009-04-10 Actimagine Soc Par Actions Sim Procede et dispositif de codage audio
FR2911020B1 (fr) * 2006-12-28 2009-05-01 Actimagine Soc Par Actions Sim Procede et dispositif de codage audio
JP5018193B2 (ja) * 2007-04-06 2012-09-05 ヤマハ株式会社 雑音抑圧装置およびプログラム
EP2128856A4 (fr) * 2007-10-16 2011-11-02 Panasonic Corp Dispositif de génération de train, dispositif de décodage et procédé
TWI407362B (zh) * 2008-03-28 2013-09-01 Hon Hai Prec Ind Co Ltd 播放裝置及其音頻輸出方法
US8380523B2 (en) 2008-07-07 2013-02-19 Lg Electronics Inc. Method and an apparatus for processing an audio signal
EP2144231A1 (fr) * 2008-07-11 2010-01-13 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Schéma de codage/décodage audio à taux bas de bits avec du prétraitement commun
EP2144230A1 (fr) * 2008-07-11 2010-01-13 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Schéma de codage/décodage audio à taux bas de bits disposant des commutateurs en cascade
TWI503816B (zh) * 2009-05-06 2015-10-11 Dolby Lab Licensing Corp 調整音訊信號響度並使其具有感知頻譜平衡保持效果之技術
US20110153391A1 (en) * 2009-12-21 2011-06-23 Michael Tenbrock Peer-to-peer privacy panel for audience measurement
US9552818B2 (en) 2012-06-14 2017-01-24 Dolby International Ab Smooth configuration switching for multichannel audio rendering based on a variable number of received channels
EP2757559A1 (fr) * 2013-01-22 2014-07-23 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Appareil et procédé de codage d'objet audio spatial employant des objets cachés pour manipulation de mélange de signaux
EP2959479B1 (fr) 2013-02-21 2019-07-03 Dolby International AB Procédés d'encodage multicanal paramétrique
WO2015036348A1 (fr) 2013-09-12 2015-03-19 Dolby International Ab Alignement temporel de données de traitement basées sur une qmf
US10152977B2 (en) * 2015-11-20 2018-12-11 Qualcomm Incorporated Encoding of multiple audio signals
US9978381B2 (en) * 2016-02-12 2018-05-22 Qualcomm Incorporated Encoding of multiple audio signals
JP6866071B2 (ja) * 2016-04-25 2021-04-28 ヤマハ株式会社 端末装置、端末装置の動作方法およびプログラム
KR101687745B1 (ko) 2016-05-12 2016-12-19 김태서 양방향 데이터통신을 수행하는 교통신호 기반의 광고 시스템 및 그 제어 방법
KR101687741B1 (ko) 2016-05-12 2016-12-19 김태서 교통신호 기반의 능동형 광고 시스템 및 그 제어 방법
EP4336497A3 (fr) * 2018-07-04 2024-03-20 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Codeur multisignal, décodeur multisignal et procédés associés utilisant le blanchiment du signal ou le post-traitement du signal

Citations (87)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS6096079A (ja) 1983-10-31 1985-05-29 Matsushita Electric Ind Co Ltd 多値画像の符号化方法
US4621862A (en) 1984-10-22 1986-11-11 The Coca-Cola Company Closing means for trucks
US4661862A (en) 1984-04-27 1987-04-28 Rca Corporation Differential PCM video transmission system employing horizontally offset five pixel groups and delta signals having plural non-linear encoding functions
JPS6294090A (ja) 1985-10-21 1987-04-30 Hitachi Ltd 符号化装置
US4725885A (en) 1986-12-22 1988-02-16 International Business Machines Corporation Adaptive graylevel image compression system
US4907081A (en) 1987-09-25 1990-03-06 Hitachi, Ltd. Compression and coding device for video signals
EP0372601A1 (fr) 1988-11-10 1990-06-13 Koninklijke Philips Electronics N.V. Codeur pour insérer une information supplémentaire dans un signal audio numérique de format préalablement déterminé, décodeur pour déduire cette information supplémentaire de ce signal numérique, dispositif muni d'un tel codeur, pour enregister un signal numérique sur un support d'information et support d'information obtenu avec ce dispositif
GB2238445A (en) 1989-09-21 1991-05-29 British Broadcasting Corp Digital video coding
EP0599825A2 (fr) 1989-06-02 1994-06-01 Koninklijke Philips Electronics N.V. Système de transmission digitale pour la transmission d'un signal additionnel tel qu'un signal d'effet spatial
WO1995027337A1 (fr) 1994-04-01 1995-10-12 Dolby Laboratories Licensing Corporation Tables de codage a source compacte pour systeme de codage/decodage
US5481643A (en) 1993-03-18 1996-01-02 U.S. Philips Corporation Transmitter, receiver and record carrier for transmitting/receiving at least a first and a second signal component
US5515296A (en) 1993-11-24 1996-05-07 Intel Corporation Scan path for encoding and decoding two-dimensional signals
US5530750A (en) 1993-01-29 1996-06-25 Sony Corporation Apparatus, method, and system for compressing a digital input signal in more than one compression mode
US5563661A (en) 1993-04-05 1996-10-08 Canon Kabushiki Kaisha Image processing apparatus
US5579430A (en) 1989-04-17 1996-11-26 Fraunhofer Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Digital encoding process
KR970014387A (ko) 1995-08-31 1997-03-29 배순훈 무손실 부호화 방법 및 그 장치
US5621856A (en) 1991-08-02 1997-04-15 Sony Corporation Digital encoder with dynamic quantization bit allocation
US5640159A (en) 1994-01-03 1997-06-17 International Business Machines Corporation Quantization method for image data compression employing context modeling algorithm
JPH09275544A (ja) 1996-02-07 1997-10-21 Matsushita Electric Ind Co Ltd デコード装置およびデコード方法
US5682461A (en) 1992-03-24 1997-10-28 Institut Fuer Rundfunktechnik Gmbh Method of transmitting or storing digitalized, multi-channel audio signals
WO1997040630A1 (fr) 1996-04-18 1997-10-30 Nokia Mobile Phones Ltd. Codeur et decodeur de donnees video
US5687157A (en) 1994-07-20 1997-11-11 Sony Corporation Method of recording and reproducing digital audio signal and apparatus thereof
EP0827312A2 (fr) 1996-08-22 1998-03-04 Robert Bosch Gmbh Procédé de changement de configuration de paquets de données
EP0867867A2 (fr) 1997-02-26 1998-09-30 Sony Corporation Procédé et appareil de codage d'information, procédé et appareil de décodage d'information et moyen d'enregistrement d'information
US5890125A (en) * 1997-07-16 1999-03-30 Dolby Laboratories Licensing Corporation Method and apparatus for encoding and decoding multiple audio channels at low bit rates using adaptive selection of encoding method
US5912636A (en) 1996-09-26 1999-06-15 Ricoh Company, Ltd. Apparatus and method for performing m-ary finite state machine entropy coding
JPH11205153A (ja) 1998-01-13 1999-07-30 Kowa Co 振動波の符号化方法及び復号化方法
US5945930A (en) 1994-11-01 1999-08-31 Canon Kabushiki Kaisha Data processing apparatus
EP0948141A2 (fr) 1998-03-30 1999-10-06 Matsushita Electric Industrial Co., Ltd. Dispositif de décodage d'un flux de données audio multicanal
WO1999052326A1 (fr) 1998-04-07 1999-10-14 Ray Milton Dolby Systeme de codage spatial a faible debit binaire et procede correspondant
US5974380A (en) 1995-12-01 1999-10-26 Digital Theater Systems, Inc. Multi-channel audio decoder
EP0957639A2 (fr) 1998-05-13 1999-11-17 Matsushita Electric Industrial Co., Ltd. Dispositif et procédé de décodage d'un signal audio numérique, et support d'enregistrement pour stocker les étapes du procédé
WO2000002357A1 (fr) 1998-07-03 2000-01-13 Dolby Laboratories Licensing Corporation Joint pour un materiau sous forme de feuille et procede servant a reunir ce type de materiau
US6021386A (en) 1991-01-08 2000-02-01 Dolby Laboratories Licensing Corporation Coding method and apparatus for multiple channels of audio information representing three-dimensional sound fields
GB2340351A (en) 1998-07-29 2000-02-16 British Broadcasting Corp Inserting auxiliary data for use during subsequent coding
US6125398A (en) 1993-11-24 2000-09-26 Intel Corporation Communications subsystem for computer-based conferencing system using both ISDN B channels for transmission
WO2000060746A2 (fr) 1999-04-07 2000-10-12 Dolby Laboratories Licensing Corporation Perfectionnements matriciels de codage et de decodage sans perte
US6134518A (en) 1997-03-04 2000-10-17 International Business Machines Corporation Digital audio signal coding using a CELP coder and a transform coder
RU2158970C2 (ru) 1994-03-01 2000-11-10 Сони Корпорейшн Способ кодирования цифрового сигнала и устройство для его осуществления, носитель записи цифрового сигнала, способ декодирования цифрового сигнала и устройство для его осуществления
WO2000079520A1 (fr) 1999-06-21 2000-12-28 Digital Theater Systems, Inc. Amelioration de la qualite sonore de systemes de codage audio a faible debit binaire etablis sans perte de compatibilite du decodeur
KR20010001991A (ko) 1999-06-10 2001-01-05 윤종용 디지털 오디오 데이터의 무손실 부호화 및 복호화장치
JP2001053617A (ja) 1999-08-05 2001-02-23 Ricoh Co Ltd デジタル音響信号符号化装置、デジタル音響信号符号化方法及びデジタル音響信号符号化プログラムを記録した媒体
US6208276B1 (en) 1998-12-30 2001-03-27 At&T Corporation Method and apparatus for sample rate pre- and post-processing to achieve maximal coding gain for transform-based audio encoding and decoding
JP2001188578A (ja) 1998-11-16 2001-07-10 Victor Co Of Japan Ltd 音声符号化方法及び音声復号方法
US6309424B1 (en) 1998-12-11 2001-10-30 Realtime Data Llc Content independent data compression method and system
US6339760B1 (en) 1998-04-28 2002-01-15 Hitachi, Ltd. Method and system for synchronization of decoded audio and video by adding dummy data to compressed audio data
US6421467B1 (en) 1999-05-28 2002-07-16 Texas Tech University Adaptive vector quantization/quantizer
JP2002328699A (ja) 2001-03-02 2002-11-15 Matsushita Electric Ind Co Ltd 符号化装置および復号化装置
JP2003005797A (ja) 2001-06-21 2003-01-08 Matsushita Electric Ind Co Ltd オーディオ信号の符号化方法及び装置、並びに符号化及び復号化システム
US20030009325A1 (en) 1998-01-22 2003-01-09 Raif Kirchherr Method for signal controlled switching between different audio coding schemes
US20030016876A1 (en) 1998-10-05 2003-01-23 Bing-Bing Chai Apparatus and method for data partitioning to improving error resilience
US6556685B1 (en) 1998-11-06 2003-04-29 Harman Music Group Companding noise reduction system with simultaneous encode and decode
KR20030043622A (ko) 2001-11-27 2003-06-02 삼성전자주식회사 좌표 인터폴레이터의 키 및 키 값 데이터의 부호화/복호화장치, 및 좌표 인터폴레이터를 부호화한 비트스트림을기록한 기록 매체
WO2003046889A1 (fr) 2001-11-30 2003-06-05 Koninklijke Philips Electronics N.V. Codage de signal
US20030138157A1 (en) 1994-09-21 2003-07-24 Schwartz Edward L. Reversible embedded wavelet system implementaion
JP2003233395A (ja) 2002-02-07 2003-08-22 Matsushita Electric Ind Co Ltd オーディオ信号の符号化方法及び装置、並びに符号化及び復号化システム
US6631352B1 (en) 1999-01-08 2003-10-07 Matushita Electric Industrial Co. Ltd. Decoding circuit and reproduction apparatus which mutes audio after header parameter changes
US20030195742A1 (en) 2002-04-11 2003-10-16 Mineo Tsushima Encoding device and decoding device
US6636830B1 (en) 2000-11-22 2003-10-21 Vialta Inc. System and method for noise reduction using bi-orthogonal modified discrete cosine transform
WO2003090206A1 (fr) 2002-04-22 2003-10-30 Koninklijke Philips Electronics N.V. Synthese de signaux
WO2003090207A1 (fr) 2002-04-22 2003-10-30 Koninklijke Philips Electronics N.V. Representation parametrique de signaux audio multicanaux
WO2003090028A2 (fr) 2002-04-19 2003-10-30 Droplet Technology, Inc. Systeme de transformee en ondelettes, procede et produit programme informatique associes
US20030236583A1 (en) 2002-06-24 2003-12-25 Frank Baumgarte Hybrid multi-channel/cue coding/decoding of audio signals
WO2004008806A1 (fr) 2002-07-16 2004-01-22 Koninklijke Philips Electronics N.V. Codage audio
WO2004008805A1 (fr) 2002-07-12 2004-01-22 Koninklijke Philips Electronics N.V. Codage audio
EP1396843A1 (fr) 2002-09-04 2004-03-10 Microsoft Corporation Compression audio mixte sans perte
US20040049379A1 (en) 2002-09-04 2004-03-11 Microsoft Corporation Multi-channel audio encoding and decoding
WO2004028142A2 (fr) 2002-09-17 2004-04-01 Vladimir Ceperkovic Codeur-decodeur rapide a rapport de compression eleve et a faible besoin en ressources
JP2004170610A (ja) 2002-11-19 2004-06-17 Kenwood Corp エンコード装置、デコード装置、エンコード方法およびデコード方法
US20040138895A1 (en) 1989-06-02 2004-07-15 Koninklijke Philips Electronics N.V. Decoding of an encoded wideband digital audio signal in a transmission system for transmitting and receiving such signal
WO2004080125A1 (fr) 2003-03-04 2004-09-16 Nokia Corporation Support d'extension audio multivoies
US20040186735A1 (en) 2001-08-13 2004-09-23 Ferris Gavin Robert Encoder programmed to add a data payload to a compressed digital audio frame
US20040199276A1 (en) 2003-04-03 2004-10-07 Wai-Leong Poon Method and apparatus for audio synchronization
WO2004093495A1 (fr) 2003-04-17 2004-10-28 Koninklijke Philips Electronics N.V. Synthese d'un signal audio
US20050058304A1 (en) 2001-05-04 2005-03-17 Frank Baumgarte Cue-based audio coding/decoding
US20050074135A1 (en) 2003-09-09 2005-04-07 Masanori Kushibe Audio device and audio processing method
US20050074127A1 (en) 2003-10-02 2005-04-07 Jurgen Herre Compatible multi-channel coding/decoding
US20050091051A1 (en) 2002-03-08 2005-04-28 Nippon Telegraph And Telephone Corporation Digital signal encoding method, decoding method, encoding device, decoding device, digital signal encoding program, and decoding program
WO2005043511A1 (fr) 2003-10-30 2005-05-12 Koninklijke Philips Electronics N.V. Codage ou decodage de signaux audio
US20050114126A1 (en) 2002-04-18 2005-05-26 Ralf Geiger Apparatus and method for coding a time-discrete audio signal and apparatus and method for decoding coded audio data
US20050137729A1 (en) 2003-12-18 2005-06-23 Atsuhiro Sakurai Time-scale modification stereo audio signals
WO2005059899A1 (fr) 2003-12-19 2005-06-30 Telefonaktiebolaget Lm Ericsson (Publ) Codage de longueur de trames variables a fidelite optimisee
US20050174269A1 (en) 2004-02-05 2005-08-11 Broadcom Corporation Huffman decoder used for decoding both advanced audio coding (AAC) and MP3 audio
US20050216262A1 (en) 2004-03-25 2005-09-29 Digital Theater Systems, Inc. Lossless multi-channel audio codec
WO2006048226A1 (fr) 2004-11-02 2006-05-11 Coding Technologies Ab Codage audio multicanal compatible stereo
WO2006108464A1 (fr) 2005-04-13 2006-10-19 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Groupage adaptatif de parametres assurant une meilleure efficacite de codage
US20070038439A1 (en) * 2003-04-17 2007-02-15 Koninklijke Philips Electronics N.V. Groenewoudseweg 1 Audio signal generation

Family Cites Families (65)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS6294090U (fr) 1985-12-02 1987-06-16
US5243686A (en) * 1988-12-09 1993-09-07 Oki Electric Industry Co., Ltd. Multi-stage linear predictive analysis method for feature extraction from acoustic signals
CA2332407C (fr) 1989-01-27 2002-03-05 Dolby Laboratories Licensing Corporation Procede de definition d'information codee
JP3104400B2 (ja) 1992-04-27 2000-10-30 ソニー株式会社 オーディオ信号符号化装置及び方法
DE4414445A1 (de) * 1994-04-26 1995-11-09 Heidelberger Druckmasch Ag Taktrolle zum Transport von Bogen in eine bogenverarbeitende Maschine
JPH08123494A (ja) 1994-10-28 1996-05-17 Mitsubishi Electric Corp 音声符号化装置、音声復号化装置、音声符号化復号化方法およびこれらに使用可能な位相振幅特性導出装置
KR100209877B1 (ko) * 1994-11-26 1999-07-15 윤종용 복수개의 허프만부호테이블을 이용한 가변장부호화장치 및 복호화장치
JP3371590B2 (ja) 1994-12-28 2003-01-27 ソニー株式会社 高能率符号化方法及び高能率復号化方法
JP3484832B2 (ja) 1995-08-02 2004-01-06 ソニー株式会社 記録装置、記録方法、再生装置及び再生方法
US6047027A (en) 1996-02-07 2000-04-04 Matsushita Electric Industrial Co., Ltd. Packetized data stream decoder using timing information extraction and insertion
US6399760B1 (en) 1996-04-12 2002-06-04 Millennium Pharmaceuticals, Inc. RP compositions and therapeutic and diagnostic uses therefor
US5970152A (en) * 1996-04-30 1999-10-19 Srs Labs, Inc. Audio enhancement system for use in a surround sound environment
KR100206786B1 (ko) * 1996-06-22 1999-07-01 구자홍 디브이디 재생기의 복수 오디오 처리 장치
US5893066A (en) 1996-10-15 1999-04-06 Samsung Electronics Co. Ltd. Fast requantization apparatus and method for MPEG audio decoding
US6131084A (en) 1997-03-14 2000-10-10 Digital Voice Systems, Inc. Dual subframe quantization of spectral magnitudes
US6639945B2 (en) * 1997-03-14 2003-10-28 Microsoft Corporation Method and apparatus for implementing motion detection in video compression
US5924930A (en) * 1997-04-03 1999-07-20 Stewart; Roger K. Hitting station and methods related thereto
US6356639B1 (en) 1997-04-11 2002-03-12 Matsushita Electric Industrial Co., Ltd. Audio decoding apparatus, signal processing device, sound image localization device, sound image control method, audio signal processing device, and audio signal high-rate reproduction method used for audio visual equipment
CN1309252C (zh) * 1997-09-17 2007-04-04 松下电器产业株式会社 将视频数据记录在光盘的设备和方法
US6130418A (en) 1997-10-06 2000-10-10 U.S. Philips Corporation Optical scanning unit having a main lens and an auxiliary lens
US5966688A (en) 1997-10-28 1999-10-12 Hughes Electronics Corporation Speech mode based multi-stage vector quantizer
JP2005063655A (ja) 1997-11-28 2005-03-10 Victor Co Of Japan Ltd オーディオ信号のエンコード方法及びデコード方法
NO306154B1 (no) * 1997-12-05 1999-09-27 Jan H Iien PolstringshÕndtak
AUPP272898A0 (en) * 1998-03-31 1998-04-23 Lake Dsp Pty Limited Time processed head related transfer functions in a headphone spatialization system
US6360204B1 (en) 1998-04-24 2002-03-19 Sarnoff Corporation Method and apparatus for implementing rounding in decoding an audio signal
MY118961A (en) 1998-09-03 2005-02-28 Sony Corp Beam irradiation apparatus, optical apparatus having beam irradiation apparatus for information recording medium, method for manufacturing original disk for information recording medium, and method for manufacturing information recording medium
US6298071B1 (en) * 1998-09-03 2001-10-02 Diva Systems Corporation Method and apparatus for processing variable bit rate information in an information distribution system
US6148283A (en) * 1998-09-23 2000-11-14 Qualcomm Inc. Method and apparatus using multi-path multi-stage vector quantizer
US6757659B1 (en) 1998-11-16 2004-06-29 Victor Company Of Japan, Ltd. Audio signal processing apparatus
JP3323175B2 (ja) 1999-04-20 2002-09-09 松下電器産業株式会社 符号化装置
JP2000352999A (ja) * 1999-06-11 2000-12-19 Nec Corp 音声切替装置
JP2001006291A (ja) * 1999-06-21 2001-01-12 Fuji Film Microdevices Co Ltd オーディオ信号の符号化方式判定装置、及びオーディオ信号の符号化方式判定方法
JP2002093055A (ja) * 2000-07-10 2002-03-29 Matsushita Electric Ind Co Ltd 信号処理装置、信号処理方法、及び光ディスク再生装置
US20020049586A1 (en) 2000-09-11 2002-04-25 Kousuke Nishio Audio encoder, audio decoder, and broadcasting system
JP3566220B2 (ja) 2001-03-09 2004-09-15 三菱電機株式会社 音声符号化装置、音声符号化方法、音声復号化装置及び音声復号化方法
US6504496B1 (en) * 2001-04-10 2003-01-07 Cirrus Logic, Inc. Systems and methods for decoding compressed data
US7583805B2 (en) 2004-02-12 2009-09-01 Agere Systems Inc. Late reverberation-based synthesis of auditory scenes
JP2002335230A (ja) 2001-05-11 2002-11-22 Victor Co Of Japan Ltd 音声符号化信号の復号方法、及び音声符号化信号復号装置
EP1308931A1 (fr) * 2001-10-23 2003-05-07 Deutsche Thomson-Brandt Gmbh Décodage d'un signal audio organisé en trames avec des données d'en-tête de frame
TW569550B (en) 2001-12-28 2004-01-01 Univ Nat Central Method of inverse-modified discrete cosine transform and overlap-add for MPEG layer 3 voice signal decoding and apparatus thereof
BR0302719A (pt) 2002-01-18 2004-03-09 Toshiba Kk Toshiba Corp Método e aparelho de codificação de vìdeo e método e aparelho de decodificação de vìdeo
KR100926469B1 (ko) * 2002-01-31 2009-11-13 톰슨 라이센싱 가변 지연을 제공하는 오디오/비디오 시스템과, 제 1 지연된 디지털 신호에 대해 제 2 디지털 신호를 동기화하기 위한 방법
DE10217297A1 (de) 2002-04-18 2003-11-06 Fraunhofer Ges Forschung Vorrichtung und Verfahren zum Codieren eines zeitdiskreten Audiosignals und Vorrichtung und Verfahren zum Decodieren von codierten Audiodaten
JP2004004274A (ja) * 2002-05-31 2004-01-08 Matsushita Electric Ind Co Ltd 音声信号処理切換装置
KR100486524B1 (ko) * 2002-07-04 2005-05-03 엘지전자 주식회사 비디오 코덱의 지연시간 단축 장치
JP3579047B2 (ja) 2002-07-19 2004-10-20 日本電気株式会社 オーディオ復号装置と復号方法およびプログラム
ES2271654T3 (es) 2002-08-07 2007-04-16 Dolby Laboratories Licensing Corporation Conversion espacial de canales de audio.
JP2004085945A (ja) * 2002-08-27 2004-03-18 Canon Inc 音響出力装置及びそのデータ伝送制御方法
TW567466B (en) 2002-09-13 2003-12-21 Inventec Besta Co Ltd Method using computer to compress and encode audio data
JP2004220743A (ja) 2003-01-17 2004-08-05 Sony Corp 情報記録装置及び情報記録制御方法、並びに情報再生装置及び情報再生制御方法
JP3761522B2 (ja) * 2003-01-22 2006-03-29 パイオニア株式会社 音声信号処理装置および音声信号処理方法
ES2273216T3 (es) 2003-02-11 2007-05-01 Koninklijke Philips Electronics N.V. Codificacion de audio.
US7394903B2 (en) * 2004-01-20 2008-07-01 Fraunhofer-Gesellschaft Zur Forderung Der Angewandten Forschung E.V. Apparatus and method for constructing a multi-channel output signal or for generating a downmix signal
WO2005098821A2 (fr) * 2004-04-05 2005-10-20 Koninklijke Philips Electronics N.V. Codeur a canux multiples
JP4367657B2 (ja) * 2004-04-09 2009-11-18 日本電気株式会社 音声通信方法及び装置
WO2005104563A1 (fr) 2004-04-22 2005-11-03 Mitsubishi Denki Kabushiki Kaisha Dispositif de codage de l'image et dispositif de décodage de l'image
JP2005332449A (ja) 2004-05-18 2005-12-02 Sony Corp 光学ピックアップ装置、光記録再生装置及びチルト制御方法
TWM257575U (en) 2004-05-26 2005-02-21 Aimtron Technology Corp Encoder and decoder for audio and video information
JP2006012301A (ja) * 2004-06-25 2006-01-12 Sony Corp 光記録再生方法、光ピックアップ装置、光記録再生装置、光記録媒体とその製造方法及び半導体レーザ装置
US8204261B2 (en) * 2004-10-20 2012-06-19 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Diffuse sound shaping for BCC schemes and the like
JP2006120247A (ja) 2004-10-21 2006-05-11 Sony Corp 集光レンズ及びその製造方法、これを用いた露光装置、光学ピックアップ装置及び光記録再生装置
US7573912B2 (en) * 2005-02-22 2009-08-11 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschunng E.V. Near-transparent or transparent multi-channel encoder/decoder scheme
CZ300251B6 (cs) 2005-07-20 2009-04-01 Oez S. R. O. Spínací prístroj, zvlášte výkonový jistic
US7742913B2 (en) 2005-10-24 2010-06-22 Lg Electronics Inc. Removing time delays in signal paths
JP4876574B2 (ja) 2005-12-26 2012-02-15 ソニー株式会社 信号符号化装置及び方法、信号復号装置及び方法、並びにプログラム及び記録媒体

Patent Citations (97)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS6096079A (ja) 1983-10-31 1985-05-29 Matsushita Electric Ind Co Ltd 多値画像の符号化方法
US4661862A (en) 1984-04-27 1987-04-28 Rca Corporation Differential PCM video transmission system employing horizontally offset five pixel groups and delta signals having plural non-linear encoding functions
US4621862A (en) 1984-10-22 1986-11-11 The Coca-Cola Company Closing means for trucks
JPS6294090A (ja) 1985-10-21 1987-04-30 Hitachi Ltd 符号化装置
US4725885A (en) 1986-12-22 1988-02-16 International Business Machines Corporation Adaptive graylevel image compression system
US4907081A (en) 1987-09-25 1990-03-06 Hitachi, Ltd. Compression and coding device for video signals
EP0372601A1 (fr) 1988-11-10 1990-06-13 Koninklijke Philips Electronics N.V. Codeur pour insérer une information supplémentaire dans un signal audio numérique de format préalablement déterminé, décodeur pour déduire cette information supplémentaire de ce signal numérique, dispositif muni d'un tel codeur, pour enregister un signal numérique sur un support d'information et support d'information obtenu avec ce dispositif
US5579430A (en) 1989-04-17 1996-11-26 Fraunhofer Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Digital encoding process
EP0599825A2 (fr) 1989-06-02 1994-06-01 Koninklijke Philips Electronics N.V. Système de transmission digitale pour la transmission d'un signal additionnel tel qu'un signal d'effet spatial
US20040138895A1 (en) 1989-06-02 2004-07-15 Koninklijke Philips Electronics N.V. Decoding of an encoded wideband digital audio signal in a transmission system for transmitting and receiving such signal
US5606618A (en) 1989-06-02 1997-02-25 U.S. Philips Corporation Subband coded digital transmission system using some composite signals
GB2238445A (en) 1989-09-21 1991-05-29 British Broadcasting Corp Digital video coding
US6021386A (en) 1991-01-08 2000-02-01 Dolby Laboratories Licensing Corporation Coding method and apparatus for multiple channels of audio information representing three-dimensional sound fields
US5621856A (en) 1991-08-02 1997-04-15 Sony Corporation Digital encoder with dynamic quantization bit allocation
US5682461A (en) 1992-03-24 1997-10-28 Institut Fuer Rundfunktechnik Gmbh Method of transmitting or storing digitalized, multi-channel audio signals
US5530750A (en) 1993-01-29 1996-06-25 Sony Corporation Apparatus, method, and system for compressing a digital input signal in more than one compression mode
US5481643A (en) 1993-03-18 1996-01-02 U.S. Philips Corporation Transmitter, receiver and record carrier for transmitting/receiving at least a first and a second signal component
US5563661A (en) 1993-04-05 1996-10-08 Canon Kabushiki Kaisha Image processing apparatus
US6453120B1 (en) 1993-04-05 2002-09-17 Canon Kabushiki Kaisha Image processing apparatus with recording and reproducing modes for hierarchies of hierarchically encoded video
US6125398A (en) 1993-11-24 2000-09-26 Intel Corporation Communications subsystem for computer-based conferencing system using both ISDN B channels for transmission
US5515296A (en) 1993-11-24 1996-05-07 Intel Corporation Scan path for encoding and decoding two-dimensional signals
US5640159A (en) 1994-01-03 1997-06-17 International Business Machines Corporation Quantization method for image data compression employing context modeling algorithm
RU2158970C2 (ru) 1994-03-01 2000-11-10 Сони Корпорейшн Способ кодирования цифрового сигнала и устройство для его осуществления, носитель записи цифрового сигнала, способ декодирования цифрового сигнала и устройство для его осуществления
WO1995027337A1 (fr) 1994-04-01 1995-10-12 Dolby Laboratories Licensing Corporation Tables de codage a source compacte pour systeme de codage/decodage
US5687157A (en) 1994-07-20 1997-11-11 Sony Corporation Method of recording and reproducing digital audio signal and apparatus thereof
US20030138157A1 (en) 1994-09-21 2003-07-24 Schwartz Edward L. Reversible embedded wavelet system implementaion
US5945930A (en) 1994-11-01 1999-08-31 Canon Kabushiki Kaisha Data processing apparatus
KR970014387A (ko) 1995-08-31 1997-03-29 배순훈 무손실 부호화 방법 및 그 장치
US5974380A (en) 1995-12-01 1999-10-26 Digital Theater Systems, Inc. Multi-channel audio decoder
JPH09275544A (ja) 1996-02-07 1997-10-21 Matsushita Electric Ind Co Ltd デコード装置およびデコード方法
WO1997040630A1 (fr) 1996-04-18 1997-10-30 Nokia Mobile Phones Ltd. Codeur et decodeur de donnees video
EP0827312A2 (fr) 1996-08-22 1998-03-04 Robert Bosch Gmbh Procédé de changement de configuration de paquets de données
US5912636A (en) 1996-09-26 1999-06-15 Ricoh Company, Ltd. Apparatus and method for performing m-ary finite state machine entropy coding
RU2221329C2 (ru) 1997-02-26 2004-01-10 Сони Корпорейшн Способ и устройство кодирования информации, способ и устройство для декодирования информации, носитель для записи информации
EP0867867A2 (fr) 1997-02-26 1998-09-30 Sony Corporation Procédé et appareil de codage d'information, procédé et appareil de décodage d'information et moyen d'enregistrement d'information
US6134518A (en) 1997-03-04 2000-10-17 International Business Machines Corporation Digital audio signal coding using a CELP coder and a transform coder
US5890125A (en) * 1997-07-16 1999-03-30 Dolby Laboratories Licensing Corporation Method and apparatus for encoding and decoding multiple audio channels at low bit rates using adaptive selection of encoding method
JPH11205153A (ja) 1998-01-13 1999-07-30 Kowa Co 振動波の符号化方法及び復号化方法
US20030009325A1 (en) 1998-01-22 2003-01-09 Raif Kirchherr Method for signal controlled switching between different audio coding schemes
US6295319B1 (en) 1998-03-30 2001-09-25 Matsushita Electric Industrial Co., Ltd. Decoding device
EP0948141A2 (fr) 1998-03-30 1999-10-06 Matsushita Electric Industrial Co., Ltd. Dispositif de décodage d'un flux de données audio multicanal
WO1999052326A1 (fr) 1998-04-07 1999-10-14 Ray Milton Dolby Systeme de codage spatial a faible debit binaire et procede correspondant
US6339760B1 (en) 1998-04-28 2002-01-15 Hitachi, Ltd. Method and system for synchronization of decoded audio and video by adding dummy data to compressed audio data
EP0957639A2 (fr) 1998-05-13 1999-11-17 Matsushita Electric Industrial Co., Ltd. Dispositif et procédé de décodage d'un signal audio numérique, et support d'enregistrement pour stocker les étapes du procédé
WO2000002357A1 (fr) 1998-07-03 2000-01-13 Dolby Laboratories Licensing Corporation Joint pour un materiau sous forme de feuille et procede servant a reunir ce type de materiau
GB2340351A (en) 1998-07-29 2000-02-16 British Broadcasting Corp Inserting auxiliary data for use during subsequent coding
US20030016876A1 (en) 1998-10-05 2003-01-23 Bing-Bing Chai Apparatus and method for data partitioning to improving error resilience
US6556685B1 (en) 1998-11-06 2003-04-29 Harman Music Group Companding noise reduction system with simultaneous encode and decode
JP2001188578A (ja) 1998-11-16 2001-07-10 Victor Co Of Japan Ltd 音声符号化方法及び音声復号方法
US6309424B1 (en) 1998-12-11 2001-10-30 Realtime Data Llc Content independent data compression method and system
US6384759B2 (en) 1998-12-30 2002-05-07 At&T Corp. Method and apparatus for sample rate pre-and post-processing to achieve maximal coding gain for transform-based audio encoding and decoding
US6208276B1 (en) 1998-12-30 2001-03-27 At&T Corporation Method and apparatus for sample rate pre- and post-processing to achieve maximal coding gain for transform-based audio encoding and decoding
US6631352B1 (en) 1999-01-08 2003-10-07 Matushita Electric Industrial Co. Ltd. Decoding circuit and reproduction apparatus which mutes audio after header parameter changes
US6611212B1 (en) 1999-04-07 2003-08-26 Dolby Laboratories Licensing Corp. Matrix improvements to lossless encoding and decoding
WO2000060746A2 (fr) 1999-04-07 2000-10-12 Dolby Laboratories Licensing Corporation Perfectionnements matriciels de codage et de decodage sans perte
US6421467B1 (en) 1999-05-28 2002-07-16 Texas Tech University Adaptive vector quantization/quantizer
KR20010001991A (ko) 1999-06-10 2001-01-05 윤종용 디지털 오디오 데이터의 무손실 부호화 및 복호화장치
WO2000079520A1 (fr) 1999-06-21 2000-12-28 Digital Theater Systems, Inc. Amelioration de la qualite sonore de systemes de codage audio a faible debit binaire etablis sans perte de compatibilite du decodeur
JP2001053617A (ja) 1999-08-05 2001-02-23 Ricoh Co Ltd デジタル音響信号符号化装置、デジタル音響信号符号化方法及びデジタル音響信号符号化プログラムを記録した媒体
US6636830B1 (en) 2000-11-22 2003-10-21 Vialta Inc. System and method for noise reduction using bi-orthogonal modified discrete cosine transform
JP2002328699A (ja) 2001-03-02 2002-11-15 Matsushita Electric Ind Co Ltd 符号化装置および復号化装置
US20050058304A1 (en) 2001-05-04 2005-03-17 Frank Baumgarte Cue-based audio coding/decoding
JP2003005797A (ja) 2001-06-21 2003-01-08 Matsushita Electric Ind Co Ltd オーディオ信号の符号化方法及び装置、並びに符号化及び復号化システム
US20040186735A1 (en) 2001-08-13 2004-09-23 Ferris Gavin Robert Encoder programmed to add a data payload to a compressed digital audio frame
KR20030043620A (ko) 2001-11-27 2003-06-02 삼성전자주식회사 좌표 인터폴레이터의 키 값 데이터 부호화/복호화 방법 및장치
KR20030043622A (ko) 2001-11-27 2003-06-02 삼성전자주식회사 좌표 인터폴레이터의 키 및 키 값 데이터의 부호화/복호화장치, 및 좌표 인터폴레이터를 부호화한 비트스트림을기록한 기록 매체
WO2003046889A1 (fr) 2001-11-30 2003-06-05 Koninklijke Philips Electronics N.V. Codage de signal
JP2003233395A (ja) 2002-02-07 2003-08-22 Matsushita Electric Ind Co Ltd オーディオ信号の符号化方法及び装置、並びに符号化及び復号化システム
US20050091051A1 (en) 2002-03-08 2005-04-28 Nippon Telegraph And Telephone Corporation Digital signal encoding method, decoding method, encoding device, decoding device, digital signal encoding program, and decoding program
US20030195742A1 (en) 2002-04-11 2003-10-16 Mineo Tsushima Encoding device and decoding device
US20050114126A1 (en) 2002-04-18 2005-05-26 Ralf Geiger Apparatus and method for coding a time-discrete audio signal and apparatus and method for decoding coded audio data
WO2003090028A2 (fr) 2002-04-19 2003-10-30 Droplet Technology, Inc. Systeme de transformee en ondelettes, procede et produit programme informatique associes
WO2003090206A1 (fr) 2002-04-22 2003-10-30 Koninklijke Philips Electronics N.V. Synthese de signaux
WO2003090207A1 (fr) 2002-04-22 2003-10-30 Koninklijke Philips Electronics N.V. Representation parametrique de signaux audio multicanaux
EP1376538A1 (fr) 2002-06-24 2004-01-02 Agere Systems Inc. Codage et décodage de signaux audiophoniques à canaux multiples hybrides et de repères directionnels
US20030236583A1 (en) 2002-06-24 2003-12-25 Frank Baumgarte Hybrid multi-channel/cue coding/decoding of audio signals
WO2004008805A1 (fr) 2002-07-12 2004-01-22 Koninklijke Philips Electronics N.V. Codage audio
RU2005103637A (ru) 2002-07-12 2005-07-10 Конинклейке Филипс Электроникс Н.В. (Nl) Аудиокодирование
WO2004008806A1 (fr) 2002-07-16 2004-01-22 Koninklijke Philips Electronics N.V. Codage audio
EP1396843A1 (fr) 2002-09-04 2004-03-10 Microsoft Corporation Compression audio mixte sans perte
US20040049379A1 (en) 2002-09-04 2004-03-11 Microsoft Corporation Multi-channel audio encoding and decoding
WO2004028142A2 (fr) 2002-09-17 2004-04-01 Vladimir Ceperkovic Codeur-decodeur rapide a rapport de compression eleve et a faible besoin en ressources
JP2004170610A (ja) 2002-11-19 2004-06-17 Kenwood Corp エンコード装置、デコード装置、エンコード方法およびデコード方法
WO2004080125A1 (fr) 2003-03-04 2004-09-16 Nokia Corporation Support d'extension audio multivoies
US20040199276A1 (en) 2003-04-03 2004-10-07 Wai-Leong Poon Method and apparatus for audio synchronization
US20070038439A1 (en) * 2003-04-17 2007-02-15 Koninklijke Philips Electronics N.V. Groenewoudseweg 1 Audio signal generation
WO2004093495A1 (fr) 2003-04-17 2004-10-28 Koninklijke Philips Electronics N.V. Synthese d'un signal audio
US20050074135A1 (en) 2003-09-09 2005-04-07 Masanori Kushibe Audio device and audio processing method
US20050074127A1 (en) 2003-10-02 2005-04-07 Jurgen Herre Compatible multi-channel coding/decoding
WO2005043511A1 (fr) 2003-10-30 2005-05-12 Koninklijke Philips Electronics N.V. Codage ou decodage de signaux audio
US7519538B2 (en) * 2003-10-30 2009-04-14 Koninklijke Philips Electronics N.V. Audio signal encoding or decoding
US20050137729A1 (en) 2003-12-18 2005-06-23 Atsuhiro Sakurai Time-scale modification stereo audio signals
WO2005059899A1 (fr) 2003-12-19 2005-06-30 Telefonaktiebolaget Lm Ericsson (Publ) Codage de longueur de trames variables a fidelite optimisee
US20050174269A1 (en) 2004-02-05 2005-08-11 Broadcom Corporation Huffman decoder used for decoding both advanced audio coding (AAC) and MP3 audio
US20050216262A1 (en) 2004-03-25 2005-09-29 Digital Theater Systems, Inc. Lossless multi-channel audio codec
WO2006048226A1 (fr) 2004-11-02 2006-05-11 Coding Technologies Ab Codage audio multicanal compatible stereo
WO2006108464A1 (fr) 2005-04-13 2006-10-19 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Groupage adaptatif de parametres assurant une meilleure efficacite de codage

Non-Patent Citations (72)

* Cited by examiner, † Cited by third party
Title
"Text of second working draft for MPEG Surround", ISO/IEC JTC 1/SC 29/WG 11, No. N7387, No. N7387, Jul. 29. 2005, 140 pages.
Bessette B, et al.: Universal Speech/Audio Coding Using Hybrid ACELP/TCX Techniques, 2005, 4 pages.
Boltze Th. et al.; "Audio services and applications." In: Digital Audio Broadcasting. Edited by Hoeg, W. and Lauferback, Th. ISBN 0-470-85013-2. John Wiley & Sons Ltd., 2003. pp. 75-83.
Breebaart, J., AES Convention Paper 'MPEG Spatial audio coding/MPEG surround: Overview and Current Status', 119th Convention, Oct. 7-10, 2005, New York, New York, 17 pages.
Chou, J. et al.: Audio Data Hiding with Application to Surround Sound, 2003, 4 pages.
European Search Report in Application No. 067577510.0-2225, mailed Jun. 8, 2009, 5 pages.
European Search Report in Application No. 06799058.0-2225, mailed Jun. 16, 2009, 6 pages.
Faller C., et al.: Binaural Cue Coding- Part II: Schemes and Applications, 2003, 12 pages, IEEE Transactions on Speech and Audio Processing, vol. 11, No. 6.
Faller C.: Parametric Coding of Spatial Audio. Doctoral thesis No. 3062, 2004, 6 pages.
Faller, C: "Coding of Spatial Audio Compatible with Different Playback Formats", Audio Engineering Society Convention Paper, 2004, 12 pages, San Francisco, CA.
Hamdy K.N., et al.: Low Bit Rate High Quality Audio Coding with Combined Harmonic and Wavelet Representations, 1996, 4 pages.
Heping, D.,: Wideband Audio Over Narrowband Low-Resolution Media, 2004, 4 pages.
Herre, J et al., "Overview of MPEP-4 Audio and its Applications in Mobile Communications", IEEE 2000, pp. 604-613.
Herre, J. et al.: MP3 Surround: Efficient and Compatible Coding of Multi-channel Audio, 2004, 14 pages.
Herre, J. et al: The Reference Model Architecture for MPEG Spatial Audio Coding, 2005, 13 pages, Audio Engineering Society Convention Paper.
Hosoi S., et al.: Audio Coding Using the Best Level Wavelet Packet Transform and Auditory Masking, 1998, 4 pages.
International Search Report corresponding to International Application No. PCT/KR2006/002018 dated Oct. 16, 2006, 1 page.
International Search Report corresponding to International Application No. PCT/KR2006/002019 dated Oct. 16, 2006, 1 page.
International Search Report corresponding to International Application No. PCT/KR2006/002020 dated Oct. 16, 2006, 2 pages.
International Search Report corresponding to International Application No. PCT/KR2006/002021 dated Oct. 16, 2006, 1 page.
International Search Report corresponding to International Application No. PCT/KR2006/002575, dated Jan. 12, 2007, 2 pages.
International Search Report corresponding to International Application No. PCT/KR2006/002578, dated Jan. 12, 2007, 2 pages.
International Search Report corresponding to International Application No. PCT/KR2006/002579, dated Nov. 24, 2006, 1 page.
International Search Report corresponding to International Application No. PCT/KR2006/002581, dated Nov. 24, 2006, 2 pages.
International Search Report corresponding to International Application No. PCT/KR2006/002583, dated Nov. 24, 2006, 2 pages.
International Search Report corresponding to International Application No. PCT/KR2006/003420, dated Jan. 18, 2007, 2 pages.
International Search Report corresponding to International Application No. PCT/KR2006/003424, dated Jan. 31, 2007, 2 pages.
International Search Report corresponding to International Application No. PCT/KR2006/003426, dated Jan. 18, 2007, 2 pages.
International Search Report corresponding to International Application No. PCT/KR2006/003435, dated Dec. 13, 2006, 1 page.
International Search Report corresponding to International Application No. PCT/KR2006/003975, dated Mar. 13, 2007, 2 pages.
International Search Report corresponding to International Application No. PCT/KR2006/004014, dated Jan. 24, 2007, 1 page.
International Search Report corresponding to International Application No. PCT/KR2006/004017, dated Jan. 24, 2007, 1 page.
International Search Report corresponding to International Application No. PCT/KR2006/004020, dated Jan. 24, 2007, 1 page.
International Search Report corresponding to International Application No. PCT/KR2006/004024, dated Jan. 29, 2007, 1 page.
International Search Report corresponding to International Application No. PCT/KR2006/004025, dated Jan. 29, 2007, 1 page.
International Search Report corresponding to International Application No. PCT/KR2006/004027, dated Jan. 29, 2007, 1 page.
International Search Report corresponding to International Application No. PCT/KR2006/004032, dated Jan. 24, 2007, 1 page.
International Search Report in corresponding International Application No. PCT/KR2006/004023, dated Jan. 23, 2007, 1 page.
ISO/IEC 13818-2, Generic Coding of Moving Pictures and Associated Audio, Nov. 1993, Seoul, Korea.
ISO/IEC 14496-3 Information Technology-Coding of Audio-Visual Objects-Part 3: Audio, Second Edition (ISO/IEC), 2001.
Jibra A., et al.: Multi-layer Scalable LPC Audio Format; ISACS 2000, 4 pages, IEEE International Symposium on Circuits and Systems.
Jin C, et al.: Individualization in Spatial-Audio Coding, 2003, 4 pages, IEEE Workshop on Applications of Signal Processing to Audio and Acoustics.
Kostantinides K: An introduction to Super Audio CD and DVD-Audio, 2003, 12 pages, IEEE Signal Processing Magazine.
Liebchem, T.; Reznik, Y.A.: MPEG-4: an Emerging Standard for Lossless Audio Coding, 2004, 10 pages, Proceedings of the Data Compression Conference.
Ming, L.: A novel random access approach for MPEG-1 multicast applications, 2001, 5 pages.
Moon, Han-gil, et al.: A Multi-Channel Audio Compression Method with Virtual Source Location Information for MPEG-4 SAC, IEEE 2005, 7 pages.
Moriya T., et al.,: A Design of Lossless Compression for High-Quality Audio Signals, 2004, 4 pages.
Non-final Office Action in U.S. Appl. No. 12/089,383, mailed Jun. 25, 2009, 16 pages.
Notice of Allowance dated Aug. 25, 2008 by the Korean Patent Office for counterpart Korean Appln. Nos. 2008-7005851, 7005852; and 7005858.
Notice of Allowance dated Dec. 26, 2008 by the Korean Patent Office for counterpart Korean Appln. Nos. 2008-7005836, 7005838, 7005839, and 7005840.
Notice of Allowance dated Jan. 13, 2009 by the Korean Patent Office for a counterpart Korean Appln. No. 2008-7005992.
Notice of Allowance issued in corresponding Korean Application Serial No. 2008-7007453, dated Feb. 27, 2009 (no English translation available).
Office Action dated Jul. 21, 2008 issued by the Taiwan Patent Office, 16 pages.
Oh, E., et al.: Proposed changes in MPEG-4 BSAC multi channel audio coding, 2004, 7 pages, International Organisation for Standardisation.
Oh, H-O et al., "Proposed core experiment on pilot-based coding of spatial parameters for MPEG surround", ISO/IEC JTC 1/SC 29/WG 11, No. M12549, Oct. 13, 2005, 18 pages XP030041219.
Pang, H., "Clipping Prevention Scheme for MPEG Surround", ETRI Journal, vol. 30, No. 4, Aug. 2008, pp. 606-608.
Pang, H., et al., ."Extended Pilot-Based Codling for Lossless Bit Rate Reduction of MPEG Surround", ETRI Journal, vol. 29, No. 1, Feb. 2007.
Puri, A., et al.: MPEG-4: An object-based multimedia coding standard supporting mobile applications, 1998, 28 pages, Baltzer Science Publishers BV.
Quackenbush, S. R. et al., "Noiseless coding of quantized spectral components in MPEG-2 Advanced Audio Coding", Application of Signal Processing to Audio and Acoustics, 1997. 1997 IEEE ASSP Workshop on New Paltz, NY, US held on Oct. 19-22, 1997, New York, NY, US, IEEE, US, (Oct. 19, 1997), 4 pages.
Russian Notice of Allowance in Application No. 2008103314/09(003616), mailed Apr. 27, 2009, 11 pages.
Russian Notice of Allowance in Application No. 2008112226/09(013215), mailed Jun. 5, 2009, 15 pages.
Said, A.: On the Reduction of Entropy Coding Complexity via Symbol Grouping: I-Redundancy Analysis and Optimal Alphabet Partition, 2004, 42 pages, Hewlett-Packard Company.
Schroeder E F et al: DER MPEG-2STANDARD: Generische Codierung fur Bewegtbilder und zugehorige Audio-Information, 1994, 5 pages.
Schuijers, E. et al: Low Complexity Parametric Stereo Coding, 2004, 6 pages, Audio Engineering Society Convention Paper 6073.
Stoll, G.: MPEG Audio Layer II: A Generic Coding Standard for Two and Multichannel Sound for DVB, DAB and Computer Multimedia, 1995, 9 pages, International Broadcasting Convention, XP006528918.
Supplementary European Search Report corresponding to Application No. EP06747465, dated Oct. 10, 2008, 8 pages.
Supplementary European Search Report corresponding to Application No. EP06747467, dated Oct. 10, 2008, 8 pages.
Supplementary European Search Report corresponding to Application No. EP06757755, dated Aug. 1, 2008, 1 page.
Supplementary European Search Report corresponding to Application No. EP06843795, dated Aug. 7, 2008, 1 page.
Ten Kate W. R. Th., et al.: A New Surround-Stereo-Surround Coding Technique, 1992, 8 pages, J. Audio Engineering Society, XP002498277.
Voros P.: High-quality Sound Coding within 2x64 kbit/s Using Instantaneous Dynamic Bit-Allocation, 1988, 4 pages.
Webb J., et al.: Video and Audio Coding for Mobile Applications, 2002, 8 pages, The Application of Programmable DSPs in Mobile Communications.

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130282384A1 (en) * 2007-09-25 2013-10-24 Motorola Mobility Llc Apparatus and Method for Encoding a Multi-Channel Audio Signal
US9570080B2 (en) * 2007-09-25 2017-02-14 Google Inc. Apparatus and method for encoding a multi-channel audio signal
US20100198589A1 (en) * 2008-07-29 2010-08-05 Tomokazu Ishikawa Audio coding apparatus, audio decoding apparatus, audio coding and decoding apparatus, and teleconferencing system
US8311810B2 (en) * 2008-07-29 2012-11-13 Panasonic Corporation Reduced delay spatial coding and decoding apparatus and teleconferencing system

Also Published As

Publication number Publication date
US7840401B2 (en) 2010-11-23
CA2626132C (fr) 2012-08-28
WO2007049866A1 (fr) 2007-05-03
EP1952670A4 (fr) 2012-09-26
TWI317244B (en) 2009-11-11
WO2007049863A8 (fr) 2007-08-02
EP1952674A4 (fr) 2010-09-29
JP5399706B2 (ja) 2014-01-29
WO2007049862A1 (fr) 2007-05-03
CN101297597A (zh) 2008-10-29
JP2009512900A (ja) 2009-03-26
US20070092086A1 (en) 2007-04-26
WO2007049863A3 (fr) 2007-06-14
JP5249038B2 (ja) 2013-07-31
KR20080050442A (ko) 2008-06-05
CN101297596A (zh) 2008-10-29
JP2009513084A (ja) 2009-03-26
EP1952675A4 (fr) 2010-09-29
KR20080050445A (ko) 2008-06-05
TWI310544B (en) 2009-06-01
WO2007049865A1 (fr) 2007-05-03
US20070094010A1 (en) 2007-04-26
KR100888973B1 (ko) 2009-03-17
US20070094011A1 (en) 2007-04-26
KR100928268B1 (ko) 2009-11-24
EP1952674A1 (fr) 2008-08-06
KR20080050443A (ko) 2008-06-05
EP1952672B1 (fr) 2016-04-27
JP2009512899A (ja) 2009-03-26
CN101297598B (zh) 2011-08-17
KR20080096603A (ko) 2008-10-30
US20100324916A1 (en) 2010-12-23
TW200723932A (en) 2007-06-16
KR100875428B1 (ko) 2008-12-22
KR20090018131A (ko) 2009-02-19
JP2009512901A (ja) 2009-03-26
US7761289B2 (en) 2010-07-20
JP2009513085A (ja) 2009-03-26
WO2007049863A2 (fr) 2007-05-03
TWI317245B (en) 2009-11-11
US20070094012A1 (en) 2007-04-26
TWI317243B (en) 2009-11-11
TW200718259A (en) 2007-05-01
US20070094014A1 (en) 2007-04-26
CN101297599A (zh) 2008-10-29
CN101297594A (zh) 2008-10-29
WO2007049864A1 (fr) 2007-05-03
US8095358B2 (en) 2012-01-10
AU2006306942B2 (en) 2010-02-18
KR100888972B1 (ko) 2009-03-17
EP1952673A1 (fr) 2008-08-06
CN101297594B (zh) 2014-07-02
TW200719747A (en) 2007-05-16
TW200723931A (en) 2007-06-16
US8095357B2 (en) 2012-01-10
JP5270357B2 (ja) 2013-08-21
EP1952672A4 (fr) 2010-09-29
BRPI0617779A2 (pt) 2011-08-09
AU2006306942A1 (en) 2007-05-03
TWI317247B (en) 2009-11-11
CN101297595A (zh) 2008-10-29
CN101297596B (zh) 2012-11-07
CN101297597B (zh) 2013-03-27
JP5270358B2 (ja) 2013-08-21
US7716043B2 (en) 2010-05-11
US20100329467A1 (en) 2010-12-30
CA2626132A1 (fr) 2007-05-03
TW200723247A (en) 2007-06-16
EP1952675A1 (fr) 2008-08-06
KR20080040785A (ko) 2008-05-08
JP2009512902A (ja) 2009-03-26
US20070094013A1 (en) 2007-04-26
WO2007049862A8 (fr) 2007-08-02
EP1952672A2 (fr) 2008-08-06
US7742913B2 (en) 2010-06-22
CN101297598A (zh) 2008-10-29
EP1952671A4 (fr) 2010-09-22
KR20080050444A (ko) 2008-06-05
EP1952671A1 (fr) 2008-08-06
KR100888974B1 (ko) 2009-03-17
KR100888971B1 (ko) 2009-03-17
EP1952674B1 (fr) 2015-09-09
WO2007049861A1 (fr) 2007-05-03
HK1126071A1 (en) 2009-08-21
KR101186611B1 (ko) 2012-09-27
JP5249039B2 (ja) 2013-07-31
TWI317246B (en) 2009-11-11
EP1952670A1 (fr) 2008-08-06

Similar Documents

Publication Publication Date Title
US7653533B2 (en) Removing time delays in signal paths
KR100875429B1 (ko) 신호 처리에서 시간 지연을 보상하는 방법
RU2389155C2 (ru) Устранение задержек по времени на трактах обработки сигнала
TWI450603B (zh) 音頻訊號處理方法及其系統與電腦可讀取媒體

Legal Events

Date Code Title Description
AS Assignment

Owner name: LG ELECTRONICS, INC., KOREA, REPUBLIC OF

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PANG, HEE SUK;KIM, DONG SOO;LIM, JAE HYUN;AND OTHERS;REEL/FRAME:018655/0408

Effective date: 20061201

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCF Information on status: patent grant

Free format text: PATENTED CASE

CC Certificate of correction
FPAY Fee payment

Year of fee payment: 4

FPAY Fee payment

Year of fee payment: 8

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 12