EP1203370B1 - Procede d'amelioration de l'efficacite de codage d'un signal audio - Google Patents

Procede d'amelioration de l'efficacite de codage d'un signal audio Download PDF

Info

Publication number
EP1203370B1
EP1203370B1 EP00944090A EP00944090A EP1203370B1 EP 1203370 B1 EP1203370 B1 EP 1203370B1 EP 00944090 A EP00944090 A EP 00944090A EP 00944090 A EP00944090 A EP 00944090A EP 1203370 B1 EP1203370 B1 EP 1203370B1
Authority
EP
European Patent Office
Prior art keywords
audio signal
coding
signal
coded
predicted
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
EP00944090A
Other languages
German (de)
English (en)
Other versions
EP1203370A1 (fr
Inventor
Juha OJANPERÄ
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nokia Oyj
Original Assignee
Nokia Oyj
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nokia Oyj filed Critical Nokia Oyj
Priority to EP08170594A priority Critical patent/EP2037451A1/fr
Priority to EP05104931A priority patent/EP1587062B1/fr
Publication of EP1203370A1 publication Critical patent/EP1203370A1/fr
Application granted granted Critical
Publication of EP1203370B1 publication Critical patent/EP1203370B1/fr
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/04Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
    • G10L19/08Determination or coding of the excitation function; Determination or coding of the long-term prediction parameters
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/04Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
    • G10L19/16Vocoder architecture
    • G10L19/18Vocoders using multiple modes
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/04Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
    • G10L19/08Determination or coding of the excitation function; Determination or coding of the long-term prediction parameters
    • G10L19/09Long term prediction, i.e. removing periodical redundancies, e.g. by using adaptive codebook or pitch predictor

Definitions

  • audio coding systems produce coded signals from an analog audio signal, such as a speech signal.
  • the coded signals are transmitted to a receiver by means of data transmission methods specific to the data transmission system.
  • an audio signal is produced on the basis of the coded signals.
  • the amount of information to be transmitted is affected e.g. by the bandwidth used for the coded information in the system, as well as by the efficiency with which the coding can be executed.
  • digital samples are produced from the analog signal e.g. at regular intervals of 0.125ms.
  • the samples are typically processed in groups of a fixed size, for example in groups having a duration of approximately 20ms. These groups of samples are also referred to as "frames".
  • a frame is the basic unit in which audio data is processed.
  • the aim of audio coding systems is to produce a sound quality which is as good as possible within the scope of the available bandwidth.
  • the periodicity present in an audio signal can be utilized.
  • the periodicity in speech results e.g. from vibrations in the vocal cords.
  • the period of vibration is in the order of 2ms to 20ms.
  • LTP long-term prediction
  • the part (frame) of the signal to be coded is compared with previously coded parts of the signal.
  • the time delay (lag) between the similar signal and the signal to be coded is examined.
  • a predicted signal representing the signal to be coded is formed on the basis of the similar signal.
  • an error signal is produced, which represents the difference between the predicted signal and the signal to be coded.
  • coding is advantageously performed in such a way that only the lag information and the error signal are transmitted.
  • the correct samples are retrieved from the memory, used to predict the part of the signal to be coded and combined with the error signal on the basis of the lag.
  • the aim is to select coefficients ⁇ k for each frame in such a way that the coding error, i.e. the difference between the actual signal and the signal formed using the preceding samples, is as small as possible.
  • those coefficients are selected to be used in the coding with which the smallest error is achieved using the least squares method.
  • the coefficients are updated frame-by-frame.
  • the patent US 5,528,629 discloses a prior art speech coding system which employs short-term prediction (STP) as well as first order long-term prediction.
  • STP short-term prediction
  • lag information alone provides a good basis for prediction of the signal.
  • the lag is not necessarily an integer multiple of the sampling interval. For example, it may lie between two successive samples of the audio signal.
  • higher order pitch predictors can effectively interpolate between the discrete sampling times, to provide a more accurate representation of the signal.
  • the frequency response of higher order pitch predictors tends to decrease as a function of frequency. This means that higher order pitch predictors provide better modelling of lower frequency components in the audio signal.
  • One purpose of the present invention is to implement a method for improving the coding accuracy and transmission efficiency of audio signals in a data transmission system, in which the audio data is coded to a greater accuracy and transferred with greater efficiency than in methods of prior art.
  • the aim is to predict the audio signal to be coded frame-by-frame as accurately as possible, while ensuring that the amount of information to be transmitted remains low.
  • the present invention achieves considerable advantages when compared to solutions according to prior art.
  • the method according to the invention enables an audio signal to be coded more accurately when compared with prior art methods, while ensuring that the amount of information required to represent the coded signal remains low.
  • the invention also allows coding of an audio signal to be performed in a more flexible manner than in methods according to prior art.
  • the invention may be implemented in such a way as to give preference to the accuracy with which the audio signal is predicted (qualitative maximization), to give preference to the reduction of the amount of information required to represent the encoded audio signal (quantitative minimization), or to provide a trade-off between the two.
  • Using the method according to the invention it is also possible to better take into account the periodicities of different frequencies that exist in the audio signal.
  • Figs. 5a and 5b are examples of data transmission frames generated by the encoder according to a preferred embodiment of the invention.
  • Fig. 1 is a reduced block diagram showing an encoder 1 according to a preferred embodiment of the invention.
  • Fig. 4 is a flow diagram 400 illustrating the method according to the invention.
  • the encoder 1 is, for example, a speech coder of a wireless communication device 2 (Fig. 3) for converting an audio signal into a coded signal to be transmitted in a data transmission system such as a mobile communication network or the Internet network.
  • a decoder 33 is advantageously located in a base station of the mobile communication network.
  • an analog audio signal e.g. a signal produced by a microphone 29 and amplified in an audio block 30 if necessary, is converted in an analog / digital converter 4 into a digital signal.
  • the accuracy of the conversion is e.g. 8 or 12 bits, and the interval (time resolution) between successive samples is e.g. 0.125 ms. It is obvious that the numerical values presented in this description are only examples clarifying, not restricting the invention.
  • the samples obtained from the audio signal are stored in a sample buffer (not shown), which can be implemented in a way known as such e.g. in the memory means 5 of the wireless communication device 2.
  • the samples of a frame to be coded are advantageously transmitted to a transform block 6, where the audio signal is transformed from the time domain to a transform domain (frequency domain), for example by means of a modified discrete cosine transform (MDCT).
  • MDCT modified discrete cosine transform
  • the output of the transform block 6 provides a group of values which represent the properties of the transformed signal in the frequency domain.
  • This transformation is represented by block 404 in the flow diagram of Fig. 4.
  • An alternative implementation for transforming a time domain signal to the frequency domain is a filter bank composed of several band-pass filters.
  • the pass band of each filter is relatively narrow, wherein the magnitudes of the signals at the outputs of the filters represent the frequency spectrum of the signal to be transformed.
  • a lag block 7 determines which preceding sequence of samples best corresponds to the frame to be coded at a given time (block 402). This stage of determining the lag is advantageously conducted in such a way that the lag block 7 compares the values stored in a reference buffer 8 with the samples of the frame to be coded and calculates the error between the samples of the frame to be coded and a corresponding sequence of samples stored in the reference buffer e.g. using a least squares method.
  • the sequence of samples composed of successive samples and having the smallest error is selected as a reference sequence of samples.
  • the lag block 7 transfers information concerning it to a coefficient calculation block 9, in order to conduct pitch predictor coefficient evaluation.
  • the pitch predictor coefficients b(k) for different pitch predictor orders such as 1, 3, 5, and 7, are calculated on the basis of the samples in the reference sequence of samples.
  • the calculated coefficients b(k) are then transferred to the pitch predictor block 10.
  • these stages are shown in blocks 405-411. It is obvious that the orders presented here function only as examples clarifying, not restricting the invention. The invention can also be applied with other orders, and the number of orders available can also differ from the total of four orders presented herein.
  • pitch predictor coefficients After the pitch predictor coefficients have been calculated, they are quantized, wherein quantized pitch predictor coefficients are obtained.
  • the pitch predictor coefficients are preferably quantized in such a way that the reconstructed signal produced in the decoder 33 of the receiver corresponds to the original as closely as possible in error-free data transmission conditions. In quantizing the pitch predictor coefficients, it is advantageous to use the highest possible resolution (smallest possible quantization steps) in order to minimize errors caused by rounding.
  • the stored samples in the reference sequence of samples are transferred to the pitch predictor block 10 where a predicted signal is produced for each pitch predictor order from the samples of the reference sequence, using the calculated and quantized pitch predictor coefficients b(k).
  • Each predicted signal represents the prediction of the signal to be coded, evaluated using the pitch predictor order in question.
  • the predicted signals are further transferred to a second transform block 11, where they are transformed into the frequency domain.
  • the second transform block 11 performs the transformation using two or more different orders, wherein sets of transformed values corresponding to the signals predicted by different pitch predictor orders are produced.
  • the pitch predictor block 10 and the second transform block 11 can be implemented in such a way that they perform the necessary operations for each pitch predictor order, or alternatively a separate pitch predictor block 10 and a separate second transform block 11 can be implemented for each order.
  • the frequency domain transformed values of the predicted signal are compared with the frequency domain transformed representation of the audio signal to be coded, obtained from transform block 6.
  • a prediction error signal is calculated by taking the difference between the frequency spectrum of the audio signal to be coded and the frequency spectrum of the signal predicted using the pitch predictor.
  • the prediction error signal comprises a set of prediction error values corresponding to the difference between the frequency components of the signal to be coded and the frequency components of the predicted signal.
  • a coding error representing e.g. the average difference between the frequency spectrum of the audio signal and the predicted signal is also calculated.
  • the coding error is calculated using a least squares method. Any other appropriate method, including methods based on psychoacoustic modelling of the audio signal, may be used to determine the predicted signal that best represents the audio signal to be coded.
  • a coding efficiency measure (prediction gain) is also calculated in block 12 to determine the information to be transmitted to the transmission channel (block 413).
  • the aim is to minimize the amount of information (bits) to be transmitted (quantitative minimization) as well as the distortions in the signal (qualitative maximization).
  • the coding efficiency measure indicates whether it is possible to transmit the information necessary to decode the signal encoded in the pitch predictor block 10 with a smaller number of bits than necessary to transmit information relating to the original signal. This determination can be implemented, for example, in such a way that a first reference value is defined, representing the amount of information to be transmitted if the information necessary for decoding is produced using a particular pitch predictor.
  • a second reference value representing the amount of information to be transmitted if the information necessary for decoding is formed on the basis of the original audio signal.
  • the coding efficiency measure is advantageously the ratio of the second reference value to the first reference value.
  • the number of bits required to represent the predicted signal depends on, for example, the order of the pitch predictor (i.e. the number of coefficients to be transmitted), the precision with which each coefficient is represented (quantized), as well as the amount and precision of the error information associated with the predicted signal.
  • the number of bits required to transmit information relating to the original audio signal depends on, for example, the precision of the frequency domain representation of the audio signal.
  • the coding efficiency determined in this way is greater than one, it indicates that the information necessary to decode the predicted signal can be transmitted with a smaller number of bits than the information relating to the original signal.
  • the number of bits necessary for the transmission of these different alternatives is determined and the alternative for which the number of bits to be transmitted is smaller is selected (block 414).
  • the pitch predictor order with which the smallest coding error is attained is selected to code the audio signal (block 412). If the coding efficiency measure for the selected pitch predictor is greater than 1, the information relating to the predicted signal is selected for transmission. If the coding efficiency measure is not greater than 1, the information to be transmitted is formed on the basis of the original audio signal. According to this embodiment of the invention, emphasis is placed on minimising the prediction error (qualitative maximization).
  • a coding efficiency measure is calculated for each pitch predictor order.
  • the pitch predictor order that provides the smallest coding error, selected from those orders for which the coding efficiency measure is greater than 1, is then used to code the audio signal. If none of the pitch predictor orders provides a prediction gain (i.e. no coding efficiency measure is greater than 1) then advantageously, the information to be transmitted is formed on the basis of the original audio signal.
  • This embodiment of the invention enables a trade-off between prediction error and coding efficiency.
  • a coding efficiency measure is calculated for each pitch predictor order and the pitch predictor order that provides the highest coding efficiency, selected from those orders for which the coding efficiency measure is greater than 1, is selected to code the audio signal. If none of the pitch predictor orders provides a prediction gain (i.e. no coding efficiency measure is greater than 1) then advantageously, the information to be transmitted is formed on the basis of the original audio signal. Thus, this embodiment of the invention places emphasis on the maximisation of coding efficiency (quantitative minimization).
  • a coding efficiency measure is calculated for each pitch predictor order and the pitch order that provides the highest coding efficiency is selected to code the audio signal, even if the coding efficiency is not greater than 1.
  • Calculation of the coding error and selection of the pitch predictor order is conducted at intervals, preferably separately for each frame, wherein in different frames it is possible to use the pitch predictor order which best corresponds to the properties of the audio signal at a given time.
  • a bit string 501 to be transmitted to the data transmission channel is formed advantageously in the following way (block 415).
  • Information from the calculation block 12 relating to the selected transmission alternative is transferred to selection block 13 (lines D1 and D4 in Fig. 1).
  • selection block 13 the frequency domain transformed values representing the original audio signal are selected to be transmitted to a quantization block 14. Transmission of the frequency domain transformed values of the original audio signal to quantization block 14 is illustrated by line A1 in the block diagram of Fig. 1.
  • the frequency domain transformed signal values are quantized in a way known as such.
  • the quantized values are transferred to a multiplexing block 15, in which the bit string to be transmitted is formed.
  • Figs. 5a and 5b show an example of a bit string structure which can be advantageously applied in connection with the present invention.
  • Information concerning the selected coding method is transferred from the calculation block 12 to multiplexing block 15 (lines D1 and D3), where the bit string is formed according to the transmission alternative.
  • a first logical value e.g. the logical 0 state, is used as coding method information 502 to indicate that frequency domain transformed values representing the original audio signal are transmitted in the bit string in question.
  • the values themselves are transmitted in the bit string, quantized to a given accuracy.
  • the field used for transmission of these values is marked with the reference numeral 503 in Fig. 5a.
  • the number of values transmitted in each bit string depends on the sampling frequency and on the length of the frame examined at a time. In this situation, pitch predictor order information, pitch predictor coefficients, lag and error information are not transmitted because the signal is reconstructed in the receiver on the basis of the frequency domain values of the original audio signal transmitted in the bit string 501.
  • the coding efficiency is greater than one, it is advantageous to encode the audio signal using the selected pitch predictor and the bit string 501 (Fig. 5b) to be transmitted to the data transmission channel is formed advantageously in the following way (block 416).
  • Information relating to the selected transmission alternative is transmitted from the calculation block 12 to the selection block 13. This is illustrated by lines D1 and D4. in the block diagram of Fig. 1.
  • the quantized pitch predictor coefficients are selected to be transferred to the multiplexing block 15. This is illustrated by line B1 in the block diagram of Fig. 1. It is obvious that the pitch predictor coefficients can also be transferred to the multiplexing block 15 in another way than via the selection block 13.
  • the bit string to be transmitted is formed in the multiplexing block 15.
  • Information concerning the selected coding method is transferred from the calculation block 12 to multiplexing block 15 (lines D1 and D3), where the bit string is formed according to the transmission alternative.
  • a second logical value e.g. the logical 1 state, is used as coding method information 502, to indicate that said quantized pitch predictor coefficients are transmitted in the bit string in question.
  • the bits of an order field 504 are set according to the selected pitch predictor order. If there are, for example, four different orders available, two bits (00, 01, 10, 11) are sufficient to indicate which order is selected at a given time.
  • information on the lag is transmitted in the bit string in a lag field 505.
  • the lag is indicated with 11 bits, but it is obvious that other lengths can also be applied within the scope of the invention.
  • the quantized pitch predictor coefficients are added to the bit string in the coefficient field 506. If the selected pitch predictor order is one, only one coefficient is transmitted, if the order is three, three coefficients are transmitted, etc.
  • the number of bits used in the transmission of the coefficients can also vary in different embodiments.
  • the first order coefficient is represented with three bits, the third order coefficients with a total of five bits, the fifth order coefficients with a total of nine bits and the seventh order coefficients with ten bits.
  • the higher the selected order the larger the number of bits required for transmission of the quantized pitch predictor coefficients.
  • This prediction error information is advantageously produced in the calculation block 12 as a difference signal, representing the difference between the frequency spectrum of the audio signal to be coded and the frequency spectrum of the signal that can be decoded (i.e. reconstructed) using the quantized pitch predictor coefficients of the selected pitch predictor in conjunction with the reference sequence of samples.
  • the error signal is transferred e.g. via the first selection block 13 to the quantization block 14 to be quantized.
  • the quantized error signal is transferred from the quantization block 14 to the multiplexing block 15, where the quantized prediction error values are added to the error field 507 of the bit string.
  • the encoder 1 also includes local decoding functionality.
  • the coded audio signal is transferred from the quantization block 14 to inverse quantization block 17.
  • the audio signal is represented by its quantized frequency spectrum values.
  • the quantized frequency spectrum values are transferred to the inverse quantization block 17, where they are inverse quantized in a way known as such, so as to restore the original frequency spectrum of the audio signal as accurately as possible.
  • the inverse quantized values representing the frequency spectrum of the original audio signal are provided as an output from block 17 to summing block 18.
  • the audio signal is represented by pitch predictor information, e.g . pitch predictor order information, quantized pitch predictor coefficients, a lag value and prediction error information in the form of quantized frequency domain values.
  • the prediction error information represents the difference between the frequency spectrum of the audio signal to be coded and the frequency spectrum of the audio signal that can be reconstructed on the basis of the selected pitch predictor and the reference sequence of samples. Therefore, in this case, the quantized frequency domain values that comprise the prediction error information are transferred to the inverse quantization block 17, where they are inverse quantized in such a way as to restore the frequency domain values of the prediction error as accurately as possible.
  • the output of block 17 comprises inverse quantized prediction error values.
  • summing block 18 These values are further provided as an input to summing block 18, where they are summed with the frequency domain values of the signal predicted using the selected pitch predictor. In this way, a reconstructed frequency domain representation of the original audio signal is formed.
  • the frequency domain values of the predicted signal are available from calculation block 12, where they are calculated in connection with determination of the prediction error, and are transferred to summing block 18 as indicated by line C1 in Figure 1.
  • summing block 18 The operation of summing block 18 is gated (switched on and off) according to control information provided by calculation block 12.
  • the transfer of control information enabling this gating operation is indicated by the link between calculation block 12 and summing block 18 (lines D1 and D2 in Figure 1).
  • the gating operation is necessary in order to take into account the different types of inverse quantized frequency domain values provided by inverse quantization block 17. As described above, if the coding efficiency is not greater than 1, the output of block 17 comprises inverse quantized frequency domain values representing the original audio signal. In this case no summing operation is necessary and no information regarding the frequency domain values of any predicted audio signal, constructed in calculation block 12, is required.
  • the operation of summing block 18 is inhibited by the control information supplied from calculation block 12 and the inverse quantized frequency domain values representing the original audio signal pass through summing block 18.
  • the output of block 17 comprises inverse quantized prediction error values.
  • the operation of summing block 18 is enabled by the control information transferred from calculation block 12, causing the inverse quantised prediction error values to be summed with the frequency spectrum of the predicted signal.
  • the necessary control information is provided by the coding method information produced in block 12 in connection with the choice of coding to be applied to the audio signal.
  • quantization can be performed before the calculation of prediction error and coding efficiency values, wherein prediction error and coding efficiency calculations are performed using quantized frequency domain values representing the original signal and the predicted signals.
  • the quantization is performed in quantization blocks positioned in between blocks 6 and 12 and blocks 11 and 12 (not shown).
  • quantization block 14 is not required, but an additional inverse quantization block is required in the path indicated by line C1.
  • the output of summing block 18 is sampled frequency domain data that corresponds to the coded sequence of samples (audio signal). This sampled frequency domain data is further transformed to the time domain in an inverse modified DCT transformer 19 from which the decoded sequence of samples is transferred to the reference buffer 8 to be stored and used in connection with the coding of subsequent frames.
  • the storage capacity of the reference buffer 8 is selected according to the number of samples necessary to attain the coding efficiency demands of the application in question.
  • a new sequence of samples is preferably stored by over-writing the oldest samples in the buffer, i.e. the buffer is a so-called circular buffer.
  • the bit string formed in the encoder 1 is transferred to a transmitter 16, in which modulation is performed in a way known as such.
  • the modulated signal is transferred via the data transmission channel 3 to the receiver e.g. as radio frequency signals.
  • the coded audio signal is transmitted frame by frame, substantially immediately after encoding for a given frame is complete.
  • the audio signal may be encoded, stored in the memory of the transmitting terminal and transmitted at some later time.
  • the signal received from the data transmission channel is demodulated in a way known as such in a receiver block 20.
  • the information contained in the demodulated data frame is determined in the decoder 33.
  • a demultiplexing block 21 of the decoder 33 it is first examined, on the basis of the coding method information 502 of the bit string, whether the received information was formed on the basis of the original audio signal. If the decoder determines that the bit string 501 formed in the encoder 1 does not contain the frequency domain transformed values of the original signal, decoding is advantageously conducted in the following way.
  • the order M to be used in the pitch predictor block 24 is determined from the order field 504 and the lag is determined from the lag field 505.
  • the quantized pitch predictor coefficients received in the coefficient field 506 of the bit string 501, as well as information concerning the order and the lag are transferred to the pitch predictor block 24 of the decoder. This is illustrated by line B2 in Fig. 2.
  • the quantized values of the prediction error signal, received in field 507 of the bit string are inverse quantized in an inverse quantization block 22 and transferred to a summing block 23 of the decoder.
  • the pitch predictor block 24 of the decoder retrieves the samples to be used as a reference sequence from a sample buffer 28, and performs a prediction according to the selected order M, in which the pitch predictor block 24 utilizes the received pitch predictor coefficients.
  • a first reconstructed time domain signal is produced, which is transformed into the frequency domain in a transform block 25.
  • This frequency domain signal is transferred to the summing block 23, wherein a frequency domain signal is produced as a sum of this signal and the inverse quantized prediction error signal.
  • the reconstructed frequency domain signal substantially corresponds to the original coded signal in the frequency domain.
  • This frequency domain signal is transformed to the time domain by means of an inverse modified DCT transform in a inverse transform block 26, wherein a digital audio signal is present at the output of the inverse transform block 26.
  • This signal is converted to an analog signal in a digital / analog converter 27, amplified if necessary and transmitted to other further processing stages in a way known as such. In Fig. 3, this is illustrated by audio block 32.
  • the bit string 501 formed in the encoder 1 comprises the values of the original signal transformed into the frequency domain
  • decoding is advantageously conducted in the following way.
  • the quantized frequency domain transformed values are inverse quantized in the inverse quantization block 22 and transferred via the summing block 23 to the inverse transform block 26.
  • the frequency domain signal is transformed to the time domain by means of an inverse modified DCT transform, wherein a time domain signal corresponding to the original audio signal is produced in digital format. If necessary, this signal is transformed into an analog signal in the digital / analog converter 27.
  • reference A2 illustrates the transmission of control information to the summing block 23.
  • This control information is used in a manner analogous to that described in connection with the local decoder functionality of the encoder.
  • the coding method information provided in field 502 of a received bit string 501 indicates that the bit string contains quantized frequency domain values derived from the audio signal itself, the operation of summing block 23 is inhibited. This allows the quantized frequency domain values of the audio signal to pass through summing block 23 to inverse transform block 26.
  • the operation of summing block 23 is enabled, allowing inverse quantised prediction error data to be summed with the frequency domain representation of the predicted signal produced by transform block 25.
  • the transmitting device is a wireless communication device 2 and the receiving device is a base station 31, wherein the signal transmitted from the wireless communication device 2 is decoded in the decoder 33 of the base station 31, from which the analog audio signal is transmitted to further processing stages in a way known as such.
  • the previously described audio signal coding / decoding stages can be applied in different kinds of data transmission systems, such as mobile communication systems, satellite-TV systems, video on demand systems, etc.
  • a mobile communication system in which audio signals are transmitted in full duplex requires an encoder / decoder pair both in the wireless communication device 2 and in the base station 31 or the like.
  • corresponding functional blocks of the wireless communication device 2 and the base station 31 are primarily marked with the same reference numerals.
  • the encoder 1 and the decoder 33 are shown as separate units in Fig. 3, in practical applications they can be implemented in one unit, a so-called codec, in which all the functions necessary to perform encoding and decoding are implemented.
  • analog / digital conversion and digital / analog conversion are not necessary in the base station.
  • these transformations are conducted in the wireless communication device and in the interface via which the mobile communication network is connected to another telecommunication network, such as a public telephone network. If this telephone network, however, is a digital telephone network, these transformations can also be made e.g. in a digital telephone (not shown) connected to such a telephone network.
  • the previously described encoding stages are not necessarily conducted in connection with transmission, but the coded information can be stored for later transmission.
  • the audio signal applied to the encoder does not necessarily have to be a real-time audio signal, but the audio signal to be coded can be information stored earlier from the audio signal.
  • the best corresponding sequence of samples is determined using the least squares method.
  • E error
  • x () is the input signal in the time domain
  • () is the signal reconstructed from the preceding sequence of samples
  • N is the number of samples in the frame examined.
  • Another alternative for solving the lag ⁇ is to use the normalized correlation method, by utilizing the formula:
  • the lag block 7 has information about the lag, i.e. how much earlier the corresponding sequence of samples appeared in the audio signal.
  • the pitch predictor coefficients b(k) can be calculated for each order M from equation (2), which can be re-expressed in the form:
  • the aim is to utilize the periodicity of the audio signal more effectively than in systems according to prior art. This is achieved by increasing the adaptability of the encoder to changes in the frequency of the audio signal by calculating pitch predictor coefficients for several orders.
  • the pitch predictor order used to code the audio signal can be chosen in such a way as to minimise the prediction error, to maximise the coding efficiency or to provide a trade-off between prediction error and coding efficiency.
  • the selection is performed at certain intervals, preferably independently for each frame.
  • the order and the pitch predictor coefficients can thus vary on a frame-by-frame basis. In the method according to the invention, it is thus possible to increase the flexibility of the coding when compared to coding methods of prior art using a fixed order.
  • the original signal, transformed into the frequency domain can be transmitted instead of the pitch predictor coefficients and the error signal.
  • look-up tables To transmit said pitch predictor coefficients to the receiver, it is possible to use so-called look-up tables. In such a look-up table different coefficient values are stored, wherein instead of the coefficient, the index of this coefficient in the look-up table is transmitted.
  • the look-up table is known to both the encoder 1 and the decoder 33.
  • the use of the look-up table can reduce the number of bits to be transmitted when compared to the transmission of pitch predictor coefficients.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)
  • Transmission Systems Not Characterized By The Medium Used For Transmission (AREA)
  • Reduction Or Emphasis Of Bandwidth Of Signals (AREA)

Claims (36)

  1. Procédé pour le codage d'un signal audio comprenant les étapes suivantes consistant à :
    examiner une partie du signal audio qui doit être codé pour trouver une autre partie du signal audio qui correspond au mieux à la partie du signal audio qui doit être codé, laquelle est détectée comme séquence de référence d'échantillons,
    produire un ensemble de signaux prédits sur la base de ladite séquence de référence d'échantillons en utilisant un ensemble d'ordres de prédicteur de hauteur tonale,
    déterminer une efficacité du codage pour au moins un desdits signaux prédits, et
    utiliser l'efficacité déterminée du codage pour sélectionner un procédé de codage pour la partie du signal audio qui doit être codé.
  2. Procédé selon la revendication 1, caractérisé en ce que les procédés de codage sélectionnables comprennent un procédé dans lequel le signal audio qui doit être codé est codé sur la base d'un signal prédit.
  3. Procédé selon la revendication 2, caractérisé en ce que les procédés de codage sélectionnable comprennent un procédé dans lequel le signal audio qui doit être codé est codé sur la base du signal audio lui-même.
  4. Procédé selon la revendication 1, caractérise en ce qu'une erreur de codage est déterminée pour chacun desdits signaux prédits.
  5. Procédé selon la revendication 4, caractérisé en ce que l'efficacité du codage est définie pour le signal prédit présentant la plus petite desdites erreurs de codage, et en ce que le codage est effectué sur la base du signal prédit ayant la plus petite dite erreur de codage si les informations d'efficacité de codage déterminées indiquent que la quantité des informations codées est inférieure que si le codage est effectué sur la base de la partie du signal audio qui doit être codé.
  6. Procédé selon la revendication 5, caractérisé en ce que la partie du signal audio qui doit être codé est transformée dans le domaine des fréquences pour déterminer le spectre de fréquence du signal audio, et chacun des signaux prédits est transformé dans le domaine des fréquences pour déterminer le spectre de fréquence de chaque signal prédit, et en ce que ladite efficacité du codage est déterminée pour ledit signal prédit présentant l'erreur de codage la plus petite sur la base du spectre de la fréquence du signal audio, et du spectre de fréquence du signal prédit.
  7. Procédé selon la revendication 1, caractérisé en ce qu'une efficacité du codage est déterminée pour chacun desdits signaux prédits et une erreur de codage est déterminée pour ces signaux prédits pour lesquels les informations d'efficacité de codage déterminé indiquent que la quantité des informations codées est inférieure que si le codage est effectué sur la base de la partie du signal audio qui doit être codé et le codage est effectué sur la base du signal prédit qui procure l'erreur de codage la plus petite.
  8. Procédé selon la revendication 1, caractérisé en ce qu'une efficacité du codage est déterminée pour chacun desdits signaux prédits et le codage est effectué sur la base du signal prédit qui procure l'efficacité de codage la plus élevée, si les informations d'efficacité de codage déterminé indiquent que la quantité des informations codées est inférieure que si le codage est effectué sur la base de la partie du signal audio qui doit être codé.
  9. Procédé selon la revendication 1, caractérisé en ce qu'une efficacité du codage est déterminée pour chacun desdits signaux prédits et le codage est effectué sur la base du signal prédit qui procure l'efficacité du codage la plus élevée.
  10. Procédé selon la revendication 7, 8, ou 9, caractérisé en ce que la partie du signal audio qui doit être codé est transformée dans le domaine des fréquences pour déterminer le spectre de fréquence du signal audio, et chaque signal prédit est transformé dans le domaine des fréquences pour déterminer le spectre de fréquence de chaque signal prédit, et en ce que ladite efficacité du codage est déterminée pour chaque signal prédit sur la base du spectre de fréquence du signal audio, et du spectre de fréquence du signal prédit.
  11. Procédé selon la revendication 5, 6, 7, 8 ou 9, caractérisé en ce que les informations d'erreur de prédiction sont déterminées pour chacun desdits signaux prédits.
  12. Procédé selon la revendication 5, 6, 7, 8 ou 9, caractérisé en ce que lesdits signaux prédits sont formés en utilisant un ordre de prédiction différent pour chacun desdits signaux prédits.
  13. Procédé selon la revendication 6 ou 10, caractérisé en ce que lesdites informations d'erreur de prédiction déterminées pour chacun desdits signaux prédits sont calculées comme un spectre de différence représentant l'utilisation dudit spectre de fréquence dudit signal audio et du spectre de fréquence du signal prédit.
  14. Procédé selon la revendication 10 ou 13, caractérisé en ce que la transformation dans le domaine des fréquences est conduite en utilisant une transformée DCT modifiée.
  15. Procédé selon l'une quelconque des revendications 1 à 14, caractérisé en ce que les informations codées (501) du signal prédit comprennent au moins des données se rapportant au procédé de codage (502), des données se rapportant à l'ordre sélectionné (504), un retard (505), des coefficients de prédiction de hauteur tonale (506) et des données se rapportant à l'erreur de prédiction (507).
  16. Procédé selon l'une quelconque des revendications 1 à 15, caractérisé en ce que le signal audio est divisé en trames, dans lesquelles le codage est effectué séparément pour chaque trame formée à partir du signal audio.
  17. Procédé selon l'une quelconque des revendications 1 à 16, caractérisé en ce que le signal audio est un signal de parole.
  18. Procédé selon l'une quelconque des revendications 4 à 7, caractérisé en ce que ladite erreur de codage est déterminée en utilisant un procédé parmi ce qui suit :
    un procédé des moindres carrés ;
    un procédé basé sur la modélisation psychoacoustique du signal audio qui doit être codé.
  19. Procédé selon la revendication 18, caractérisé en ce que si ladite erreur de codage est déterminée en utilisant le procédé des moindres carrés, l'erreur de codage est calculée à partir de l'erreur de prédiction.
  20. Procédé selon l'une quelconque des revendications 1 à 19, caractérisé en ce que ledit signal audio codé est transmis à un dispositif de réception.
  21. Codeur (1) qui comprend un moyen (16, 20) pour coder un signal audio, ledit moyen pour le codage comprend
    un moyen (7) pour examiner une partie du signal audio qui doit être codé pour trouver une autre partie du signal audio qui correspond au mieux à la partie du signal audio qui doit être codé, laquelle est sélectionnée comme séquence de référence d'échantillons,
    un moyen (9, 10) pour utiliser un ensemble d'ordres de prédiction de hauteur tonale pour produire un ensemble de signaux prédits sur la base de ladite séquence de référence d'échantillons,
    un moyen (12) pour déterminer une efficacité de codage pour au moins un desdits signaux prédits, et
    un moyen (12, 13, 14) pour utiliser l'efficacité du codage déterminé pour sélectionner un procédé de codage pour la partie du signal audio qui doit être codé.
  22. Codeur (1) selon la revendication 21, caractérisé en ce qu'il comprend un moyen (4, 6 à 14) pour coder le signal audio sur la base d'un signal prédit.
  23. Codeur (1) selon la revendication 22, caractérisé en ce qu'il comprend un moyen (4, 6, 14) pour coder le signal audio lui-même.
  24. Système de transmission de données qui comprend
    un moyen (16) pour transmettre le signal audio codé.
  25. Système de transmission de données selon la revendication 24, caractérisé en ce qu'il comprend un moyen pour déterminer une erreur de codage pour au moins un desdits signaux prédits.
  26. Système de transmission de données selon la revendication 24, caractérisé en ce qu'il comprend un moyen pour transformer la partie du signal audio qui doit être codé dans le domaine des fréquences, et un moyen pour transformer chaque signal prédit dans le domaine des fréquences.
  27. Système de transmission de données selon la revendication 24, caractérisé en ce qu'il comprend un moyen pour former un train binaire (15) pour transmission au dispositif de réception, ledit train binaire comprend au moins des informations concernant le procédé de codage sélectionné.
  28. Système de transmission de données selon l'une quelconque des revendications 24 à 27, caractérisé en ce qu'il comprend un moyen pour diviser le signal audio en trames.
  29. Système de transmission de données selon l'une quelconque des revendications 24 à 28, caractérisé en ce qu'il comprend un terminal mobile.
  30. Décodeur (33) pour décoder un signal audio codé dans un codeur selon la revendication 21, caractérisé en ce que le décodeur comprend :
    un moyen pour déterminer le procédé de codage du signal audio qui doit être décodé comprenant un moyen pour examiner, sur la base des informations du procédé de codage (502), si les informations reçues ont été formées sur la base du signal audio d'origine, et un moyen pour examiner l'ordre de prédiction de hauteur tonale (M) utilisé dans la phase de codage, et un moyen pour décoder le signal audio en conformité avec le procédé de codage déterminé comprenant un moyen (21) pour recevoir les informations se rapportant au signal prédit, un moyen pour décoder le signal en utilisant les informations codées qui sont formées sur la base du signal audio lui-même, un moyen pour sélectionner l'ordre de prédiction de hauteur tonale pour décoder le signal, et un moyen pour décoder le signal en effectuant une prédiction en conformité avec l'ordre de prédiction de hauteur tonale sélectionné (M).
  31. Décodeur selon la revendication 30, caractérisé en ce que le décodeur comprend un moyen (21) pour déterminer à partir desdites informations reçues au moins les données se rapportant à un ordre sélectionné (504), un retard (505), au moins un coefficient de prédiction de hauteur tonale (506) et des données d'erreur de prédiction (507).
  32. Décodeur selon la revendication 31, caractérisé en ce qu'il comprend un moyen (24, 28) pour produire un signal prédit en utilisant lesdites données se rapportant à un ordre sélectionné (504), un retard (505) et à au moins un coefficient de prédiction de hauteur tonale (506).
  33. Décodeur selon la revendication 31 ou 32, caractérisé en ce qu'il comprend un moyen (23, 24, 28) pour reproduire un signal audio reconstitué en utilisant ledit signal prédit et lesdites données d'erreur de prédiction.
  34. Décodeur selon la revendication 30, caractérisé en ce qu'il comprend un moyen (21) pour recevoir les informations se rapportant au signal audio lui-même.
  35. Décodeur selon la revendication 34, caractérisé en ce qu'il comprend un moyen (22, 23, 26) pour produire un signal audio reconstitué en utilisant lesdites informations reçues se rapportant au signal audio lui-même.
  36. Procédé pour le décodage d'un signal audio qui est codé en conformité avec le procédé selon la revendication 1, caractérisé en ce que le procédé comprend l'étape d'examen sur la base des informations du procédé de codage (502) pour savoir si les informations reçues ont été formées sur la base du signal audio d'origine, dans lequel le signal est décodé en utilisant les informations codées qui sont formées sur la base du signal audio lui-même, dans les autres cas, l'ordre de prédiction de hauteur tonale (M) utilisé dans la phase de codage est examiné, et une prédiction en conformité avec l'ordre de prédiction de hauteur tonale (M) est effectuée pour reconstituer le signal audio.
EP00944090A 1999-07-05 2000-07-05 Procede d'amelioration de l'efficacite de codage d'un signal audio Expired - Lifetime EP1203370B1 (fr)

Priority Applications (2)

Application Number Priority Date Filing Date Title
EP08170594A EP2037451A1 (fr) 1999-07-05 2000-07-05 Procédé pour améliorer l'efficacité de codage d'un signal audio
EP05104931A EP1587062B1 (fr) 1999-07-05 2000-07-05 Procédé pour l'améliorement de l'efficacité de codage d'un signal audio

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
FI991537A FI116992B (fi) 1999-07-05 1999-07-05 Menetelmät, järjestelmä ja laitteet audiosignaalin koodauksen ja siirron tehostamiseksi
FI991537 1999-07-05
PCT/FI2000/000619 WO2001003122A1 (fr) 1999-07-05 2000-07-05 Procede d'amelioration de l'efficacite de codage d'un signal audio

Related Child Applications (1)

Application Number Title Priority Date Filing Date
EP05104931A Division EP1587062B1 (fr) 1999-07-05 2000-07-05 Procédé pour l'améliorement de l'efficacité de codage d'un signal audio

Publications (2)

Publication Number Publication Date
EP1203370A1 EP1203370A1 (fr) 2002-05-08
EP1203370B1 true EP1203370B1 (fr) 2005-06-29

Family

ID=8555025

Family Applications (3)

Application Number Title Priority Date Filing Date
EP00944090A Expired - Lifetime EP1203370B1 (fr) 1999-07-05 2000-07-05 Procede d'amelioration de l'efficacite de codage d'un signal audio
EP05104931A Expired - Lifetime EP1587062B1 (fr) 1999-07-05 2000-07-05 Procédé pour l'améliorement de l'efficacité de codage d'un signal audio
EP08170594A Withdrawn EP2037451A1 (fr) 1999-07-05 2000-07-05 Procédé pour améliorer l'efficacité de codage d'un signal audio

Family Applications After (2)

Application Number Title Priority Date Filing Date
EP05104931A Expired - Lifetime EP1587062B1 (fr) 1999-07-05 2000-07-05 Procédé pour l'améliorement de l'efficacité de codage d'un signal audio
EP08170594A Withdrawn EP2037451A1 (fr) 1999-07-05 2000-07-05 Procédé pour améliorer l'efficacité de codage d'un signal audio

Country Status (13)

Country Link
US (2) US7289951B1 (fr)
EP (3) EP1203370B1 (fr)
JP (2) JP4142292B2 (fr)
KR (2) KR100545774B1 (fr)
CN (2) CN100568344C (fr)
AT (2) ATE298919T1 (fr)
AU (1) AU761771B2 (fr)
BR (1) BRPI0012182B1 (fr)
CA (1) CA2378435C (fr)
DE (2) DE60041207D1 (fr)
ES (1) ES2244452T3 (fr)
FI (1) FI116992B (fr)
WO (1) WO2001003122A1 (fr)

Families Citing this family (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002173892A (ja) 2000-09-27 2002-06-21 Nippon Paper Industries Co Ltd グラビア印刷用塗工紙
FI118067B (fi) 2001-05-04 2007-06-15 Nokia Corp Menetelmä audiosignaalin pakkauksen purkamisessa, pakkauksen purkulaite, ja elektroniikkalaite
DE10138650A1 (de) * 2001-08-07 2003-02-27 Fraunhofer Ges Forschung Verfahren und Vorrichtung zum Verschlüsseln eines diskreten Signals sowie Verfahren und Vorrichtung zur Entschlüsselung
US7933767B2 (en) * 2004-12-27 2011-04-26 Nokia Corporation Systems and methods for determining pitch lag for a current frame of information
US20070213705A1 (en) * 2006-03-08 2007-09-13 Schmid Peter M Insulated needle and system
US7610195B2 (en) * 2006-06-01 2009-10-27 Nokia Corporation Decoding of predictively coded data using buffer adaptation
JP2008170488A (ja) 2007-01-06 2008-07-24 Yamaha Corp 波形圧縮装置、波形伸長装置、プログラムおよび圧縮データの生産方法
EP2077551B1 (fr) 2008-01-04 2011-03-02 Dolby Sweden AB Encodeur audio et décodeur
WO2009132662A1 (fr) * 2008-04-28 2009-11-05 Nokia Corporation Codage/décodage pour une réponse en fréquence améliorée
KR20090122143A (ko) * 2008-05-23 2009-11-26 엘지전자 주식회사 오디오 신호 처리 방법 및 장치
US8380523B2 (en) * 2008-07-07 2013-02-19 Lg Electronics Inc. Method and an apparatus for processing an audio signal
US20100114568A1 (en) * 2008-10-24 2010-05-06 Lg Electronics Inc. Apparatus for processing an audio signal and method thereof
US8364471B2 (en) * 2008-11-04 2013-01-29 Lg Electronics Inc. Apparatus and method for processing a time domain audio signal with a noise filling flag
GB2466672B (en) * 2009-01-06 2013-03-13 Skype Speech coding
GB2466671B (en) 2009-01-06 2013-03-27 Skype Speech encoding
GB2466674B (en) 2009-01-06 2013-11-13 Skype Speech coding
GB2466673B (en) 2009-01-06 2012-11-07 Skype Quantization
GB2466675B (en) 2009-01-06 2013-03-06 Skype Speech coding
KR101614767B1 (ko) * 2009-10-28 2016-04-22 에스케이텔레콤 주식회사 벡터 양자화 기반 2차 예측을 통한 동영상 부호화/복호화 장치, 방법 및 기록 매체
KR101584480B1 (ko) 2010-04-13 2016-01-14 지이 비디오 컴프레션, 엘엘씨 평면 간 예측
CN105933716B (zh) * 2010-04-13 2019-05-28 Ge视频压缩有限责任公司 跨平面预测
KR102166520B1 (ko) 2010-04-13 2020-10-16 지이 비디오 컴프레션, 엘엘씨 샘플 영역 병합
ES2549734T3 (es) 2010-04-13 2015-11-02 Ge Video Compression, Llc Codificación de vídeo que usa subdivisiones multi-árbol de imágenes
CN106454371B (zh) 2010-04-13 2020-03-20 Ge视频压缩有限责任公司 解码器、数组重建方法、编码器、编码方法及存储介质
US9268762B2 (en) 2012-01-16 2016-02-23 Google Inc. Techniques for generating outgoing messages based on language, internationalization, and localization preferences of the recipient
DE102012207750A1 (de) 2012-05-09 2013-11-28 Leibniz-Institut für Plasmaforschung und Technologie e.V. Vorrichtung zur plasmabehandlung von menschlichen, tierischen oder pflanzlichen oberflächen, insbesondere von haut oder schleimhautarealen
EP2887349B1 (fr) * 2012-10-01 2017-11-15 Nippon Telegraph and Telephone Corporation Procédé de codage, dispositif de codage, programme et support d'enregistrement
KR102251833B1 (ko) 2013-12-16 2021-05-13 삼성전자주식회사 오디오 신호의 부호화, 복호화 방법 및 장치
EP2916319A1 (fr) * 2014-03-07 2015-09-09 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Concept pour le codage d'informations

Family Cites Families (43)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US36721A (en) * 1862-10-21 Improvement in breech-loading fire-arms
JPH0683443B2 (ja) * 1985-03-05 1994-10-19 富士通株式会社 フレ−ム内フレ−ム間符号化方式
DE69029120T2 (de) * 1989-04-25 1997-04-30 Toshiba Kawasaki Kk Stimmenkodierer
CA2021514C (fr) 1989-09-01 1998-12-15 Yair Shoham Codage a excitation stochastique avec contrainte
NL9001985A (nl) * 1990-09-10 1992-04-01 Nederland Ptt Werkwijze voor het coderen van een analoog signaal met een herhalend karakter en een inrichting voor het volgens deze werkwijze coderen.
US5528629A (en) 1990-09-10 1996-06-18 Koninklijke Ptt Nederland N.V. Method and device for coding an analog signal having a repetitive nature utilizing over sampling to simplify coding
NL9002308A (nl) 1990-10-23 1992-05-18 Nederland Ptt Werkwijze voor het coderen en decoderen van een bemonsterd analoog signaal met een herhalend karakter en een inrichting voor het volgens deze werkwijze coderen en decoderen.
US5233660A (en) * 1991-09-10 1993-08-03 At&T Bell Laboratories Method and apparatus for low-delay celp speech coding and decoding
US6400996B1 (en) * 1999-02-01 2002-06-04 Steven M. Hoffberg Adaptive pattern recognition based control system and method
US5765127A (en) * 1992-03-18 1998-06-09 Sony Corp High efficiency encoding method
US5842033A (en) * 1992-06-30 1998-11-24 Discovision Associates Padding apparatus for passing an arbitrary number of bits through a buffer in a pipeline system
IT1257065B (it) * 1992-07-31 1996-01-05 Sip Codificatore a basso ritardo per segnali audio, utilizzante tecniche di analisi per sintesi.
FI95086C (fi) 1992-11-26 1995-12-11 Nokia Mobile Phones Ltd Menetelmä puhesignaalin tehokkaaksi koodaamiseksi
CA2116736C (fr) * 1993-03-05 1999-08-10 Edward M. Roney, Iv Selection pour decodeur
JPH06332492A (ja) * 1993-05-19 1994-12-02 Matsushita Electric Ind Co Ltd 音声検出方法および検出装置
IT1270438B (it) 1993-06-10 1997-05-05 Sip Procedimento e dispositivo per la determinazione del periodo del tono fondamentale e la classificazione del segnale vocale in codificatori numerici della voce
US5574825A (en) * 1994-03-14 1996-11-12 Lucent Technologies Inc. Linear prediction coefficient generation during frame erasure or packet loss
JP3277692B2 (ja) 1994-06-13 2002-04-22 ソニー株式会社 情報符号化方法、情報復号化方法及び情報記録媒体
JPH08166800A (ja) * 1994-12-13 1996-06-25 Hitachi Ltd 複数種類の符号化方法を備える音声符号器および復号器
JP3183072B2 (ja) 1994-12-19 2001-07-03 松下電器産業株式会社 音声符号化装置
JPH08190764A (ja) * 1995-01-05 1996-07-23 Sony Corp ディジタル信号処理方法、ディジタル信号処理装置及び記録媒体
FR2729246A1 (fr) * 1995-01-06 1996-07-12 Matra Communication Procede de codage de parole a analyse par synthese
FR2729247A1 (fr) * 1995-01-06 1996-07-12 Matra Communication Procede de codage de parole a analyse par synthese
US5864798A (en) * 1995-09-18 1999-01-26 Kabushiki Kaisha Toshiba Method and apparatus for adjusting a spectrum shape of a speech signal
JP4005154B2 (ja) * 1995-10-26 2007-11-07 ソニー株式会社 音声復号化方法及び装置
TW321810B (fr) * 1995-10-26 1997-12-01 Sony Co Ltd
JPH1091194A (ja) * 1996-09-18 1998-04-10 Sony Corp 音声復号化方法及び装置
JP3707154B2 (ja) * 1996-09-24 2005-10-19 ソニー株式会社 音声符号化方法及び装置
JPH10105194A (ja) * 1996-09-27 1998-04-24 Sony Corp ピッチ検出方法、音声信号符号化方法および装置
DE69712537T2 (de) * 1996-11-07 2002-08-29 Matsushita Electric Industrial Co., Ltd. Verfahren zur Erzeugung eines Vektorquantisierungs-Codebuchs
JPH10149199A (ja) * 1996-11-19 1998-06-02 Sony Corp 音声符号化方法、音声復号化方法、音声符号化装置、音声復号化装置、電話装置、ピッチ変換方法及び媒体
FI964975A (fi) 1996-12-12 1998-06-13 Nokia Mobile Phones Ltd Menetelmä ja laite puheen koodaamiseksi
US6252632B1 (en) * 1997-01-17 2001-06-26 Fox Sports Productions, Inc. System for enhancing a video presentation
US6202046B1 (en) * 1997-01-23 2001-03-13 Kabushiki Kaisha Toshiba Background noise/speech classification method
JP3064947B2 (ja) * 1997-03-26 2000-07-12 日本電気株式会社 音声・楽音符号化及び復号化装置
FI973873A (fi) 1997-10-02 1999-04-03 Nokia Mobile Phones Ltd Puhekoodaus
JP3765171B2 (ja) 1997-10-07 2006-04-12 ヤマハ株式会社 音声符号化復号方式
AU3372199A (en) * 1998-03-30 1999-10-18 Voxware, Inc. Low-complexity, low-delay, scalable and embedded speech and audio coding with adaptive frame loss concealment
US6014618A (en) * 1998-08-06 2000-01-11 Dsp Software Engineering, Inc. LPAS speech coder using vector quantized, multi-codebook, multi-tap pitch predictor and optimized ternary source excitation codebook derivation
US6493665B1 (en) * 1998-08-24 2002-12-10 Conexant Systems, Inc. Speech classification and parameter weighting used in codebook search
US6188980B1 (en) * 1998-08-24 2001-02-13 Conexant Systems, Inc. Synchronized encoder-decoder frame concealment using speech coding parameters including line spectral frequencies and filter coefficients
US6691084B2 (en) * 1998-12-21 2004-02-10 Qualcomm Incorporated Multiple mode variable rate speech coding
EP1095370A1 (fr) * 1999-04-05 2001-05-02 Hughes Electronics Corporation Modelisation spectrale de la phase des composantes d'onde prototype pour un systeme codec interpolatif de la parole a plages de frequence

Also Published As

Publication number Publication date
ATE298919T1 (de) 2005-07-15
DE60021083T2 (de) 2006-05-18
KR20050085977A (ko) 2005-08-29
DE60021083D1 (de) 2005-08-04
AU5832600A (en) 2001-01-22
JP2005189886A (ja) 2005-07-14
JP2003504654A (ja) 2003-02-04
US20060089832A1 (en) 2006-04-27
FI991537A (fi) 2001-01-06
ATE418779T1 (de) 2009-01-15
KR20020019483A (ko) 2002-03-12
EP1587062A1 (fr) 2005-10-19
FI116992B (fi) 2006-04-28
CN1766990A (zh) 2006-05-03
US7289951B1 (en) 2007-10-30
KR100593459B1 (ko) 2006-06-28
EP2037451A1 (fr) 2009-03-18
CN1372683A (zh) 2002-10-02
JP4142292B2 (ja) 2008-09-03
DE60041207D1 (de) 2009-02-05
JP4426483B2 (ja) 2010-03-03
CA2378435A1 (fr) 2001-01-11
AU761771B2 (en) 2003-06-12
ES2244452T3 (es) 2005-12-16
EP1587062B1 (fr) 2008-12-24
KR100545774B1 (ko) 2006-01-24
US7457743B2 (en) 2008-11-25
CN1235190C (zh) 2006-01-04
BRPI0012182B1 (pt) 2017-02-07
WO2001003122A1 (fr) 2001-01-11
EP1203370A1 (fr) 2002-05-08
BR0012182A (pt) 2002-04-16
CN100568344C (zh) 2009-12-09
CA2378435C (fr) 2008-01-08

Similar Documents

Publication Publication Date Title
EP1203370B1 (fr) Procede d'amelioration de l'efficacite de codage d'un signal audio
US5819215A (en) Method and apparatus for wavelet based data compression having adaptive bit rate control for compression of digital audio or other sensory data
KR101061404B1 (ko) 가변 레이트로 오디오를 인코딩 및 디코딩하는 방법
US7610195B2 (en) Decoding of predictively coded data using buffer adaptation
EP0152430A1 (fr) Appareil et procedes de codage, decodage, analyse et synthese d'un signal.
US20070078646A1 (en) Method and apparatus to encode/decode audio signal
JPS60116000A (ja) 音声符号化装置
JP2007504503A (ja) 低ビットレートオーディオ符号化
JP2002330075A (ja) サブバンドadpcm符号化方法、復号方法、サブバンドadpcm符号化装置、復号装置およびワイヤレスマイクロホン送信システム、受信システム
KR20070090261A (ko) Ltp 부호화 시스템에서 피치 래그를 결정하기 위한시스템 및 방법
EP2127088B1 (fr) Quantification audio
CA2551281A1 (fr) Dispositif et procede de codage vocal/musical
JP4359949B2 (ja) 信号符号化装置及び方法、並びに信号復号装置及び方法
JPH10268897A (ja) 信号符号化方法及び装置
US5822722A (en) Wide-band signal encoder
JP2638209B2 (ja) 適応変換符号化の方法及び装置
JP2890523B2 (ja) 適応変換符号化の方法及び装置
Bhutani Comparison of DPCM and Subband Codec performance in the presence of burst errors

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20011228

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE

AX Request for extension of the european patent

Free format text: AL;LT;LV;MK;RO;SI

RIC1 Information provided on ipc code assigned before grant

Ipc: 7G 10L 19/04 A

Ipc: 7G 10L 19:08 J

Ipc: 7G 10L 19:14 J

Ipc: 7H 03M 7/30 J

GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

RIC1 Information provided on ipc code assigned before grant

Ipc: 7H 03M 7/30 J

Ipc: 7G 10L 19/04 A

Ipc: 7H 03M 7:30 J

Ipc: 7G 10L 19:08 J

Ipc: 7G 10L 19:14 J

GRAS Grant fee paid

Free format text: ORIGINAL CODE: EPIDOSNIGR3

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CH

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20050629

Ref country code: LI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20050629

Ref country code: AT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20050629

Ref country code: BE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20050629

REG Reference to a national code

Ref country code: GB

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: CH

Ref legal event code: EP

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CY

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20050705

Ref country code: LU

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20050705

Ref country code: IE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20050705

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MC

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20050731

REF Corresponds to:

Ref document number: 60021083

Country of ref document: DE

Date of ref document: 20050804

Kind code of ref document: P

REG Reference to a national code

Ref country code: IE

Ref legal event code: FG4D

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: DK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20050929

Ref country code: GR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20050929

REG Reference to a national code

Ref country code: SE

Ref legal event code: TRGR

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: PT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20051207

REG Reference to a national code

Ref country code: ES

Ref legal event code: FG2A

Ref document number: 2244452

Country of ref document: ES

Kind code of ref document: T3

REG Reference to a national code

Ref country code: CH

Ref legal event code: PL

ET Fr: translation filed
REG Reference to a national code

Ref country code: IE

Ref legal event code: MM4A

PLBE No opposition filed within time limit

Free format text: ORIGINAL CODE: 0009261

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

26N No opposition filed

Effective date: 20060330

REG Reference to a national code

Ref country code: GB

Ref legal event code: 732E

Free format text: REGISTERED BETWEEN 20150910 AND 20150916

REG Reference to a national code

Ref country code: DE

Ref legal event code: R082

Ref document number: 60021083

Country of ref document: DE

Representative=s name: COHAUSZ & FLORACK PATENT- UND RECHTSANWAELTE P, DE

Ref country code: DE

Ref legal event code: R081

Ref document number: 60021083

Country of ref document: DE

Owner name: NOKIA TECHNOLOGIES OY, FI

Free format text: FORMER OWNER: NOKIA CORP., 02610 ESPOO, FI

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: FI

Payment date: 20150709

Year of fee payment: 16

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: SE

Payment date: 20150713

Year of fee payment: 16

REG Reference to a national code

Ref country code: ES

Ref legal event code: PC2A

Owner name: NOKIA TECHNOLOGIES OY

Effective date: 20151124

REG Reference to a national code

Ref country code: NL

Ref legal event code: PD

Owner name: NOKIA TECHNOLOGIES OY; FI

Free format text: DETAILS ASSIGNMENT: VERANDERING VAN EIGENAAR(S), OVERDRACHT; FORMER OWNER NAME: NOKIA CORPORATION

Effective date: 20151111

REG Reference to a national code

Ref country code: FR

Ref legal event code: PLFP

Year of fee payment: 17

REG Reference to a national code

Ref country code: FR

Ref legal event code: TP

Owner name: NOKIA TECHNOLOGIES OY, FI

Effective date: 20170109

REG Reference to a national code

Ref country code: SE

Ref legal event code: EUG

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: SE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20160706

Ref country code: FI

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20160705

REG Reference to a national code

Ref country code: FR

Ref legal event code: PLFP

Year of fee payment: 18

REG Reference to a national code

Ref country code: FR

Ref legal event code: PLFP

Year of fee payment: 19

REG Reference to a national code

Ref country code: GB

Ref legal event code: 732E

Free format text: REGISTERED BETWEEN 20190124 AND 20190130

REG Reference to a national code

Ref country code: DE

Ref legal event code: R082

Ref document number: 60021083

Country of ref document: DE

Representative=s name: COHAUSZ & FLORACK PATENT- UND RECHTSANWAELTE P, DE

Ref country code: DE

Ref legal event code: R081

Ref document number: 60021083

Country of ref document: DE

Owner name: PROVENANCE ASSET GROUP LLC, PITTSFORD, US

Free format text: FORMER OWNER: NOKIA TECHNOLOGIES OY, ESPOO, FI

REG Reference to a national code

Ref country code: NL

Ref legal event code: PD

Owner name: PROVENANCE ASSET GROUP LLC; US

Free format text: DETAILS ASSIGNMENT: CHANGE OF OWNER(S), ASSIGNMENT; FORMER OWNER NAME: NOKIA TECHNOLOGIES OY

Effective date: 20190325

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: NL

Payment date: 20190719

Year of fee payment: 20

REG Reference to a national code

Ref country code: ES

Ref legal event code: PC2A

Owner name: PROVENANCE ASSET GROUP LLC

Effective date: 20190926

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: FR

Payment date: 20190719

Year of fee payment: 20

Ref country code: DE

Payment date: 20190719

Year of fee payment: 20

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: GB

Payment date: 20190719

Year of fee payment: 20

REG Reference to a national code

Ref country code: DE

Ref legal event code: R071

Ref document number: 60021083

Country of ref document: DE

REG Reference to a national code

Ref country code: NL

Ref legal event code: MK

Effective date: 20200704

REG Reference to a national code

Ref country code: GB

Ref legal event code: PE20

Expiry date: 20200704

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IT

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20190705

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: GB

Free format text: LAPSE BECAUSE OF EXPIRATION OF PROTECTION

Effective date: 20200704

REG Reference to a national code

Ref country code: ES

Ref legal event code: FD2A

Effective date: 20201127

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: ES

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20190706