WO2004008437A2 - Audio coding - Google Patents

Audio coding Download PDF

Info

Publication number
WO2004008437A2
WO2004008437A2 PCT/IB2003/003152 IB0303152W WO2004008437A2 WO 2004008437 A2 WO2004008437 A2 WO 2004008437A2 IB 0303152 W IB0303152 W IB 0303152W WO 2004008437 A2 WO2004008437 A2 WO 2004008437A2
Authority
WO
WIPO (PCT)
Prior art keywords
frame
time
signal
audio signal
encoded signal
Prior art date
Application number
PCT/IB2003/003152
Other languages
English (en)
French (fr)
Other versions
WO2004008437A3 (en
Inventor
Erik G. P. Schuijers
Adriaan J. Rijnberg
Natasa Topalovic
Original Assignee
Koninklijke Philips Electronics N.V.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Koninklijke Philips Electronics N.V. filed Critical Koninklijke Philips Electronics N.V.
Priority to BR0305556-6A priority Critical patent/BR0305556A/pt
Priority to KR1020057000782A priority patent/KR101001170B1/ko
Priority to US10/520,876 priority patent/US7516066B2/en
Priority to EP03764067.9A priority patent/EP1527441B1/en
Priority to JP2004521016A priority patent/JP4649208B2/ja
Priority to AU2003247040A priority patent/AU2003247040A1/en
Publication of WO2004008437A2 publication Critical patent/WO2004008437A2/en
Publication of WO2004008437A3 publication Critical patent/WO2004008437A3/en

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/04Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/04Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
    • G10L19/06Determination or coding of the spectral characteristics, e.g. of the short-term prediction coefficients
    • G10L19/07Line spectrum pair [LSP] vocoders
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/04Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
    • G10L19/06Determination or coding of the spectral characteristics, e.g. of the short-term prediction coefficients

Definitions

  • the invention relates to coding at least part of an audio signal.
  • LPC Linear Predictive Coding
  • An object of the invention is to provide advantageous coding of at least part of an audio signal.
  • the invention provides a method of encoding, an encoder, an encoded audio signal, a storage medium, a method of decoding, a decoder, a transmitter, a receiver and a system as defined in the independent claims.
  • Advantageous embodiments are defined in the dependent claims.
  • At least part of an audio signal is coded in order to obtain an encoded signal, the coding comprising predictive coding the at least part of the audio signal in order to obtain prediction coefficients which represent temporal properties, such as a temporal envelope, of the at least part of the audio signal, transforming the prediction coefficients into a set of times representing the prediction coefficients, and including the set of times in the encoded signal. Note that times without any amplitude information suffice to represent the prediction coefficients.
  • a temporal shape of a signal or a component thereof can also be directly encoded in the form of a set of amplitude or gain values, it has been the inventor's insight that higher quality can be obtained by using predictive coding to obtain prediction coefficients which represent temporal properties such as a temporal envelope and transforming these prediction coefficients to into a set of times. Higher quality can be obtained because locally (where needed) higher time resolution can be obtained compared to fixed time-axis technique.
  • the predictive coding may be implemented by using the amplitude response of an LPC filter to represent the temporal envelope.
  • Embodiments of the invention can be interpreted as using an LPC spectrum to describe a temporal envelope instead of a spectral envelope and that what is time in the case of a spectral envelope, now is frequency and vice versa, as shown in the bottom part of Fig. 2.
  • the inventors realized that when using overlapping frame analysis/synthesis for the temporal envelope, redundancy in the Line Spectral Representation at the overlap can be exploited. Embodiments of the invention exploit this redundancy in an advantageous manner.
  • the invention and embodiments thereof are in particular advantageous for the coding of a temporal envelope of a noise component in the audio signal in a parametric audio coding schemes such as disclosed in WO 01/69593-A1.
  • a parametric audio coding scheme an audio signal may be dissected into transient signal components, sinusoidal signal components and noise components.
  • the parameters representing the sinusoidal components may be amplitude, frequency and phase.
  • the extension of such parameters with an envelope description is an efficient representation. Note that the invention and embodiments thereof can be applied to the entire relevant frequency band of the audio signal or a component thereof, but also to a smaller frequency band.
  • Fig. 1 shows an example of an LPC spectrum with 8 poles with corresponding 8 Line Spectral Frequencies according to prior art
  • Fig. 2 shows (top) using LPC such that H(z) represents a frequency spectrum
  • Fig. 3 shows a stylized view of exemplary analysis/synthesis windowing
  • Fig. 4 shows an example sequence of LSF times for two subsequent frames
  • Fig. 5 shows matching of LSF times by shifting LSF times in a frame k relative to a previous frame k-1 ;
  • Fig. 6 shows weighting functions as function of overlap
  • Fig. 7 shows a system according to an embodiment of the invention.
  • Fig. 2 shows how a predictive filter such as an LPC filter can be used to describe a temporal envelope of an audio signal or a component thereof.
  • the input signal is first transformed from time domain to frequency domain by e.g. a Fourier Transform. So in fact, the temporal shape is transformed in a spectral shape which is coded by a subsequent conventional LPC filter which is normally used to code a spectral shape.
  • the LPC filter analysis provides prediction coefficients which represent the temporal shape of the input signal. There is a trade-off between time-resolution and frequency resolution. Say that e.g. the LPC spectrum would consist of a number of very sharp peaks (sinusoids).
  • an LPC filter H(z) can generally be described as:
  • the coefficients ⁇ ,,with i running from 1 to m, are the prediction filter coefficients resulting from the LPC analysis.
  • the following procedure can be used. Most of this procedure is valid for a general all-pole filter H(z), so also for frequency domain. Other procedures known for deriving LSFs in the frequency domain can also be used to calculate the time domain equivalents of the LSFs.
  • the polynomial A(z) is split into two polynomials P(z) and Q(z) of order m+1.
  • the polynomial P(z) is formed by adding a reflection coefficient (in lattice filter form) of +1 to A(z), Q(z) is formed by adding a reflection coefficient of -7.
  • the zeros of the polynomials P '(z) and Q '(z) are thus fully characterized by their time t, which runs from 0 to ⁇ over a frame, wherein 0 corresponds to a start of the frame and ⁇ rto an end of that frame, which frame can actually have any practical length, e.g. 10 or 20 ms.
  • the times t resulting from this derivation can be interpreted as time domain equivalents of the line spectral frequencies, which times are further called LSF times herein.
  • LSF times time domain equivalents of the line spectral frequencies
  • Fig. 3 shows a stylized view of an exemplary situation for analysis and synthesis of temporal envelopes.
  • a, not necessarily rectangular, window is used to analyze the segment by LPC. So for each frame, after conversion, a set of N LSF times is obtained.
  • N in principal does not need to be constant, although in many cases this leads to a more efficient representation.
  • the LSF times are uniformly quantized, although other techniques like vector quantization could also be applied here.
  • a derived LSF time is derived which is a weighted average of the LSF times in the pair.
  • a weighted average in this application is to be construed as including the case where only one out of the pair of LSF times is selected. Such a selection can be interpreted as a weighted average wherein the weight of the selected LSF time is one and the weight of the non-selected time is zero. It is also possible that both LSF times of the pair have the same weight.
  • a new set of three derived LSF times is constructed based on the two original sets of three LSF times.
  • a practical approach is to just take the LSF times of frame k-1 (or k), and calculate the LSF times of frame k (or k-1) by simply shifting the LSF times of frame k-1 (or k) to align the frames in time. This shifting is performed in both the encoder and the decoder. In the encoder the LSFs of the right frame k are shifted to match the ones in the left frame k-1. This is necessary to look for pairs and eventually determine the weighted average.
  • the derived time or weighted average is encoded into the bit-stream as a 'representation level' which is an integer value e.g. from 0 until 255 (8 bits) representing 0 until pi.
  • a 'representation level' which is an integer value e.g. from 0 until 255 (8 bits) representing 0 until pi.
  • Huffman coding is applied.
  • For a first frame the first LSF time is coded absolutely (no reference point), all subsequent LSF times (including the weighted ones at the end) are coded differentially to their predecessor. Now, say frame k could make use of the 'trick' using the last 3 LSF times of frame k-1.
  • frame k then takes the last three representation levels of frame k-1 (which are at the end of the region 0 until 255) and shift them back to its own time-axis (at the beginning of the region 0 until 255). All subsequent LSF times in frame k would be encoded differentially to their predecessor starting with the representation level (on the axis of frame k) corresponding to the last LSF in the overlap area. In case frame k could not make use of the 'trick' the first LSF time of frame k would be coded absolutely and all subsequent LSF times of frame k differential to their predecessor.
  • the weight for frame k is calculated as The new LSF times are now calculated as:
  • the first frame in a bit-stream has no history, the first frame of LSF times always need to be coded without exploitation of techniques as mentioned above. This may be done by coding the first LSF time absolutely using Huffman coding, and all subsequent values differentially to their predecessor within a frame using a fixed Huffman table. All frames subsequent to the first frame can in essence make advantage of an above technique. Of course such a technique is not always advantageous. Think for instance of a situation where there are an equal number of LSF times in the overlap area for both frames, but with a very bad match. Calculating a (weighted) mean might then result in perceptual deterioration.
  • the situation where in frame k-1 the number of LSF times is not equal to the number of LSF times in frame k is preferably not defined by an above technique. Therefore for each frame of LSF times an indication, such as a single bit, is included in the encoded signal to indicate whether or not an above technique is used, i.e. should the first number of LSF times be retrieved from the previous frame or are they iri the bit-stream? For example, if the indicator bit is 1 : the weighted LSF times are coded differentially to their predecessor in frame k-1, for frame k the first number of LSF times in the overlap area are derived from the LSFs in frame k-1. If the indicator bit is 0, the first LSF time of frame k is coded absolutely, all following LSFs are coded differentially to their predecessor.
  • the LSF time frames are rather long, e.g. 1440 samples at 44.1kHz; in this case only around 30 bits per second are needed for this extra indication bit.
  • the LSF time data is loss- lessly encoded. So instead of merging the overlap-pairs to single LSF times, the differences of the LSF times in a given frame are encoded with respect to the LSF times in another frame. So in the example of Figure 3 when the values IQ until I N are retrieved of frame k-1, the first three values IQ until from frame k are retrieved by decoding the differences (in the bit- stream) to I N - 2 , IN-I, I N of frame k-1 respectively.
  • Fig. 7 shows a system according to an embodiment of the invention.
  • the system comprises an apparatus 1 for transmitting or recording an encoded signal [S].
  • the apparatus 1 comprises an input unit 10 for receiving at least part of an audio signal S, preferably a noise component of the audio signal.
  • the input unit 10 may be an antenna, microphone, network connection, etc.
  • the apparatus 1 further comprises an encoder 11 for encoding the signal S according to an above described embodiment of the invention (see in particular Figs. 4, 5 and 6) in order to obtain an encoded signal. It is possible that the input unit 10 receives a full audio signal and provides components thereof to other dedicated encoders.
  • the encoded signal is furnished to an output unit 12 which transforms the encoded audio signal in a bit-stream [S] having a suitable format for transmission or storage via a transmission medium or storage medium 2.
  • the system further comprises a receiver or reproduction apparatus 3 which receives the encoded signal [S] in an input unit 30.
  • the input unit 30 furnishes the encoded signal [S] to the decoder 31.
  • the decoder 31 decodes the encoded signal by performing a decoding process which is substantially an inverse operation of the encoding in the encoder 11 wherein a decoded signal S' is obtained which corresponds to the original signal S except for those parts which were lost during the encoding process.
  • the decoder 31 furnishes the decoded signal S' to an output unit 32 that provides the decoded signal S'.
  • the output unit 32 may be reproduction unit such as a speaker for reproducing the decoded signal S'.
  • the output unit 32 may also be a transmitter for further transmitting the decoded signal S' for example over an in-home network, etc.
  • the output unit 32 may include combining means for combining the signal S' with other reconstructed components in order to provide a full audio signal.
  • Embodiments of the invention may be applied in, inter alia, Internet distribution, Solid State Audio, 3G terminals, GPRS and commercial successors thereof.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Spectroscopy & Molecular Physics (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)
PCT/IB2003/003152 2002-07-16 2003-07-11 Audio coding WO2004008437A2 (en)

Priority Applications (6)

Application Number Priority Date Filing Date Title
BR0305556-6A BR0305556A (pt) 2002-07-16 2003-07-11 Método e codificador para codificar pelo menos parte de um sinal de áudio a fim de obter um sinal codificado, sinal codificado representando pelo menos parte de um sinal de áudio, meio de armazenamento, método e decodificador para decodificar um sinal codificado, transmissor, receptor, e, sistema
KR1020057000782A KR101001170B1 (ko) 2002-07-16 2003-07-11 오디오 코딩
US10/520,876 US7516066B2 (en) 2002-07-16 2003-07-11 Audio coding
EP03764067.9A EP1527441B1 (en) 2002-07-16 2003-07-11 Audio coding
JP2004521016A JP4649208B2 (ja) 2002-07-16 2003-07-11 オーディオコーディング
AU2003247040A AU2003247040A1 (en) 2002-07-16 2003-07-11 Audio coding

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP02077870 2002-07-16
EP02077870.0 2002-07-16

Publications (2)

Publication Number Publication Date
WO2004008437A2 true WO2004008437A2 (en) 2004-01-22
WO2004008437A3 WO2004008437A3 (en) 2004-05-13

Family

ID=30011204

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2003/003152 WO2004008437A2 (en) 2002-07-16 2003-07-11 Audio coding

Country Status (9)

Country Link
US (1) US7516066B2 (ru)
EP (1) EP1527441B1 (ru)
JP (1) JP4649208B2 (ru)
KR (1) KR101001170B1 (ru)
CN (1) CN100370517C (ru)
AU (1) AU2003247040A1 (ru)
BR (1) BR0305556A (ru)
RU (1) RU2321901C2 (ru)
WO (1) WO2004008437A2 (ru)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2301021A1 (en) * 2008-07-10 2011-03-30 Voiceage Corporation Device and method for quantizing and inverse quantizing lpc filters in a super-frame
US8255211B2 (en) 2004-08-25 2012-08-28 Dolby Laboratories Licensing Corporation Temporal envelope shaping for spatial audio coding using frequency domain wiener filtering
US9653088B2 (en) 2007-06-13 2017-05-16 Qualcomm Incorporated Systems, methods, and apparatus for signal encoding using pitch-regularizing and non-pitch-regularizing coding
CN107767876A (zh) * 2014-03-24 2018-03-06 株式会社Ntt都科摩 声音编码装置以及声音编码方法

Families Citing this family (41)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7644003B2 (en) * 2001-05-04 2010-01-05 Agere Systems Inc. Cue-based audio coding/decoding
US7116787B2 (en) * 2001-05-04 2006-10-03 Agere Systems Inc. Perceptual synthesis of auditory scenes
US7583805B2 (en) * 2004-02-12 2009-09-01 Agere Systems Inc. Late reverberation-based synthesis of auditory scenes
DE60218068T2 (de) * 2001-11-30 2007-11-22 Koninklijke Philips Electronics N.V. Signalkodierung
US7805313B2 (en) * 2004-03-04 2010-09-28 Agere Systems Inc. Frequency-based coding of channels in parametric multi-channel coding systems
US8204261B2 (en) * 2004-10-20 2012-06-19 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Diffuse sound shaping for BCC schemes and the like
US7720230B2 (en) * 2004-10-20 2010-05-18 Agere Systems, Inc. Individual channel shaping for BCC schemes and the like
US7787631B2 (en) * 2004-11-30 2010-08-31 Agere Systems Inc. Parametric coding of spatial audio with cues based on transmitted channels
DE602005017302D1 (de) * 2004-11-30 2009-12-03 Agere Systems Inc Synchronisierung von parametrischer raumtonkodierung mit extern bereitgestelltem downmix
JP5106115B2 (ja) * 2004-11-30 2012-12-26 アギア システムズ インコーポレーテッド オブジェクト・ベースのサイド情報を用いる空間オーディオのパラメトリック・コーディング
US7903824B2 (en) * 2005-01-10 2011-03-08 Agere Systems Inc. Compact side information for parametric coding of spatial audio
EP1984911A4 (en) * 2006-01-18 2012-03-14 Lg Electronics Inc DEVICE AND METHOD FOR SIGNAL CODING AND DECODING
FR2911031B1 (fr) * 2006-12-28 2009-04-10 Actimagine Soc Par Actions Sim Procede et dispositif de codage audio
CN101231850B (zh) * 2007-01-23 2012-02-29 华为技术有限公司 编解码方法及装置
KR20080073925A (ko) * 2007-02-07 2008-08-12 삼성전자주식회사 파라메트릭 부호화된 오디오 신호를 복호화하는 방법 및장치
CN101266795B (zh) * 2007-03-12 2011-08-10 华为技术有限公司 一种格矢量量化编解码的实现方法及装置
US20090006081A1 (en) * 2007-06-27 2009-01-01 Samsung Electronics Co., Ltd. Method, medium and apparatus for encoding and/or decoding signal
EP2077551B1 (en) * 2008-01-04 2011-03-02 Dolby Sweden AB Audio encoder and decoder
US8380498B2 (en) * 2008-09-06 2013-02-19 GH Innovation, Inc. Temporal envelope coding of energy attack signal by using attack point location
US8276047B2 (en) * 2008-11-13 2012-09-25 Vitesse Semiconductor Corporation Continuously interleaved error correction
EP3998606B8 (en) 2009-10-21 2022-12-07 Dolby International AB Oversampling in a combined transposer filter bank
US9838784B2 (en) 2009-12-02 2017-12-05 Knowles Electronics, Llc Directional audio capture
US8798290B1 (en) 2010-04-21 2014-08-05 Audience, Inc. Systems and methods for adaptive signal equalization
US9558755B1 (en) 2010-05-20 2017-01-31 Knowles Electronics, Llc Noise suppression assisted automatic speech recognition
KR101747917B1 (ko) * 2010-10-18 2017-06-15 삼성전자주식회사 선형 예측 계수를 양자화하기 위한 저복잡도를 가지는 가중치 함수 결정 장치 및 방법
JP5674015B2 (ja) * 2010-10-27 2015-02-18 ソニー株式会社 復号装置および方法、並びにプログラム
US8615394B1 (en) * 2012-01-27 2013-12-24 Audience, Inc. Restoration of noise-reduced speech
US8725508B2 (en) * 2012-03-27 2014-05-13 Novospeech Method and apparatus for element identification in a signal
WO2014118152A1 (en) * 2013-01-29 2014-08-07 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Low-frequency emphasis for lpc-based coding in frequency domain
EP3671738B1 (en) 2013-04-05 2024-06-05 Dolby International AB Audio encoder and decoder
US9536540B2 (en) 2013-07-19 2017-01-03 Knowles Electronics, Llc Speech signal separation and synthesis based on auditory scene analysis and speech modeling
EP2916319A1 (en) 2014-03-07 2015-09-09 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Concept for encoding of information
KR101837153B1 (ko) * 2014-05-01 2018-03-09 니폰 덴신 덴와 가부시끼가이샤 주기성 통합 포락 계열 생성 장치, 주기성 통합 포락 계열 생성 방법, 주기성 통합 포락 계열 생성 프로그램, 기록매체
CN104217726A (zh) * 2014-09-01 2014-12-17 东莞中山大学研究院 一种无损音频压缩编码方法及其解码方法
DE112015004185T5 (de) 2014-09-12 2017-06-01 Knowles Electronics, Llc Systeme und Verfahren zur Wiederherstellung von Sprachkomponenten
US9838700B2 (en) * 2014-11-27 2017-12-05 Nippon Telegraph And Telephone Corporation Encoding apparatus, decoding apparatus, and method and program for the same
CN107210824A (zh) 2015-01-30 2017-09-26 美商楼氏电子有限公司 麦克风的环境切换
KR102125410B1 (ko) * 2015-02-26 2020-06-22 프라운호퍼 게젤샤프트 쭈르 푀르데룽 데어 안겐반텐 포르슝 에. 베. 타깃 시간 도메인 포락선을 사용하여 처리된 오디오 신호를 얻도록 오디오 신호를 처리하기 위한 장치 및 방법
US9820042B1 (en) 2016-05-02 2017-11-14 Knowles Electronics, Llc Stereo separation and directional suppression with omni-directional microphones
CN107871492B (zh) * 2016-12-26 2020-12-15 珠海市杰理科技股份有限公司 音乐合成方法和系统
EP3382700A1 (en) * 2017-03-31 2018-10-03 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Apparatus and method for post-processing an audio signal using a transient location detection

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5749064A (en) * 1996-03-01 1998-05-05 Texas Instruments Incorporated Method and system for time scale modification utilizing feature vectors about zero crossing points
EP0899720A2 (en) * 1997-08-28 1999-03-03 Texas Instruments Inc. Quantization of linear prediction coefficients
WO1999018565A2 (en) * 1997-10-02 1999-04-15 Nokia Mobile Phones Limited Speech coding

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
PL174216B1 (pl) * 1993-11-30 1998-06-30 At And T Corp Sposób redukcji w czasie rzeczywistym szumu transmisji mowy
US5781888A (en) * 1996-01-16 1998-07-14 Lucent Technologies Inc. Perceptual noise shaping in the time domain via LPC prediction in the frequency domain
JP3472974B2 (ja) * 1996-10-28 2003-12-02 日本電信電話株式会社 音響信号符号化方法および音響信号復号化方法
KR20000064913A (ko) * 1997-02-10 2000-11-06 요트.게.아. 롤페즈 저ㄴ송 시스템, 수신기, 및 재구성된 음성 신호 유도 방법
EP1190415B1 (en) 2000-03-15 2007-08-08 Koninklijke Philips Electronics N.V. Laguerre function for audio coding

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5749064A (en) * 1996-03-01 1998-05-05 Texas Instruments Incorporated Method and system for time scale modification utilizing feature vectors about zero crossing points
EP0899720A2 (en) * 1997-08-28 1999-03-03 Texas Instruments Inc. Quantization of linear prediction coefficients
WO1999018565A2 (en) * 1997-10-02 1999-04-15 Nokia Mobile Phones Limited Speech coding

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
KUMARESAN R ET AL: "On representing signals using only timing information" JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, NOV. 2001, ACOUST. SOC. AMERICA THROUGH AIP, USA, vol. 110, no. 5, pages 2421-2439, XP001176748 ISSN: 0001-4966 *
KUMARESAN R ET AL: "On the duality between line-spectral frequencies and zero-crossings of signals" IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, MAY 2001, IEEE, USA, vol. 9, no. 4, pages 458-461, XP002264935 ISSN: 1063-6676 *
WONG J W C ET AL: "Fast time scale modification using envelope-matching technique (EM-TSM)" CIRCUITS AND SYSTEMS, 1998. ISCAS '98. PROCEEDINGS OF THE 1998 IEEE INTERNATIONAL SYMPOSIUM ON MONTEREY, CA, USA 31 MAY-3 JUNE 1998, NEW YORK, NY, USA,IEEE, US, 31 May 1998 (1998-05-31), pages 550-553, XP010289950 ISBN: 0-7803-4455-3 *

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8255211B2 (en) 2004-08-25 2012-08-28 Dolby Laboratories Licensing Corporation Temporal envelope shaping for spatial audio coding using frequency domain wiener filtering
TWI393120B (zh) * 2004-08-25 2013-04-11 Dolby Lab Licensing Corp 用於音訊信號編碼及解碼之方法和系統、音訊信號編碼器、音訊信號解碼器、攜帶有位元流之電腦可讀取媒體、及儲存於電腦可讀取媒體上的電腦程式
US9653088B2 (en) 2007-06-13 2017-05-16 Qualcomm Incorporated Systems, methods, and apparatus for signal encoding using pitch-regularizing and non-pitch-regularizing coding
EP2301021A1 (en) * 2008-07-10 2011-03-30 Voiceage Corporation Device and method for quantizing and inverse quantizing lpc filters in a super-frame
EP2301021A4 (en) * 2008-07-10 2012-08-15 Voiceage Corp DEVICE AND METHOD FOR QUANTIFYING AND REVERSE QUANTIFICATION OF LINEAR PREDICTIVE CODING FILTERS IN A SUPERTRAME
US8332213B2 (en) 2008-07-10 2012-12-11 Voiceage Corporation Multi-reference LPC filter quantization and inverse quantization device and method
US8712764B2 (en) 2008-07-10 2014-04-29 Voiceage Corporation Device and method for quantizing and inverse quantizing LPC filters in a super-frame
US9245532B2 (en) 2008-07-10 2016-01-26 Voiceage Corporation Variable bit rate LPC filter quantizing and inverse quantizing device and method
USRE49363E1 (en) 2008-07-10 2023-01-10 Voiceage Corporation Variable bit rate LPC filter quantizing and inverse quantizing device and method
CN107767876A (zh) * 2014-03-24 2018-03-06 株式会社Ntt都科摩 声音编码装置以及声音编码方法

Also Published As

Publication number Publication date
RU2005104122A (ru) 2005-08-10
BR0305556A (pt) 2004-09-28
US7516066B2 (en) 2009-04-07
AU2003247040A1 (en) 2004-02-02
CN1669075A (zh) 2005-09-14
EP1527441B1 (en) 2017-09-06
EP1527441A2 (en) 2005-05-04
JP2005533272A (ja) 2005-11-04
CN100370517C (zh) 2008-02-20
WO2004008437A3 (en) 2004-05-13
RU2321901C2 (ru) 2008-04-10
KR20050023426A (ko) 2005-03-09
KR101001170B1 (ko) 2010-12-15
JP4649208B2 (ja) 2011-03-09
US20050261896A1 (en) 2005-11-24

Similar Documents

Publication Publication Date Title
EP1527441B1 (en) Audio coding
US8862463B2 (en) Adaptive time/frequency-based audio encoding and decoding apparatuses and methods
RU2389085C2 (ru) Способы и устройства для введения низкочастотных предыскажений в ходе сжатия звука на основе acelp/tcx
US11721349B2 (en) Methods, encoder and decoder for linear predictive encoding and decoding of sound signals upon transition between frames having different sampling rates
US9418666B2 (en) Method and apparatus for encoding and decoding audio/speech signal
KR100388388B1 (ko) 재생위상정보를사용하는음성합성방법및장치
US6081776A (en) Speech coding system and method including adaptive finite impulse response filter
EP0747882A2 (en) Pitch delay modification during frame erasures
EP0747883A2 (en) Voiced/unvoiced classification of speech for use in speech decoding during frame erasures
US20050261897A1 (en) Method and device for robust predictive vector quantization of linear prediction parameters in variable bit rate speech coding
US20070147518A1 (en) Methods and devices for low-frequency emphasis during audio compression based on ACELP/TCX
EP3693964A1 (en) Simultaneous time-domain and frequency-domain noise shaping for tdac transforms
US6138092A (en) CELP speech synthesizer with epoch-adaptive harmonic generator for pitch harmonics below voicing cutoff frequency
US20050091041A1 (en) Method and system for speech coding
US6778953B1 (en) Method and apparatus for representing masked thresholds in a perceptual audio coder
US8880411B2 (en) Critical sampling encoding with a predictive encoder
US11380341B2 (en) Selecting pitch lag
JP2000132193A (ja) 信号符号化装置及び方法、並びに信号復号装置及び方法
US9620139B2 (en) Adaptive linear predictive coding/decoding
US6292774B1 (en) Introduction into incomplete data frames of additional coefficients representing later in time frames of speech signal samples
JPH1049200A (ja) 音声情報圧縮蓄積方法及び装置

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
REEP Request for entry into the european phase

Ref document number: 2003764067

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2003764067

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 3197/CHENP/2004

Country of ref document: IN

WWE Wipo information: entry into national phase

Ref document number: 10520876

Country of ref document: US

WWE Wipo information: entry into national phase

Ref document number: 20038166976

Country of ref document: CN

WWE Wipo information: entry into national phase

Ref document number: 2004521016

Country of ref document: JP

Ref document number: 1020057000782

Country of ref document: KR

ENP Entry into the national phase

Ref document number: 2005104122

Country of ref document: RU

Kind code of ref document: A

WWP Wipo information: published in national office

Ref document number: 1020057000782

Country of ref document: KR

WWP Wipo information: published in national office

Ref document number: 2003764067

Country of ref document: EP