WO2000013172A1 - Signal processing techniques for time-scale and/or pitch modification of audio signals - Google Patents

Signal processing techniques for time-scale and/or pitch modification of audio signals Download PDF

Info

Publication number
WO2000013172A1
WO2000013172A1 PCT/NZ1999/000143 NZ9900143W WO0013172A1 WO 2000013172 A1 WO2000013172 A1 WO 2000013172A1 NZ 9900143 W NZ9900143 W NZ 9900143W WO 0013172 A1 WO0013172 A1 WO 0013172A1
Authority
WO
WIPO (PCT)
Prior art keywords
frequency
signal
waveform
encoding
maxima
Prior art date
Application number
PCT/NZ1999/000143
Other languages
French (fr)
Inventor
Stephen Marcus Jason Hoek
Original Assignee
Sigma Audio Research Limited
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sigma Audio Research Limited filed Critical Sigma Audio Research Limited
Priority to JP2000568078A priority Critical patent/JP4527287B2/en
Priority to AU54548/99A priority patent/AU5454899A/en
Priority to EP99940754.7A priority patent/EP1127349B1/en
Publication of WO2000013172A1 publication Critical patent/WO2000013172A1/en

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/02Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders
    • G10L19/0212Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders using orthogonal transformation
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
    • G10L25/18Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters the extracted parameters being spectral information of each sub-band

Definitions

  • the present invention relates to encoding and manipulation of digital signals. More particularly, although not exclusively, the present invention relates to time-scale and/or pitch modification of audio signals. As such, the signal analysis and re-synthesis method described herein is not limited to audio signals. It is envisaged that the present invention may find application in the coding of other signals with the (wavelet-like) method described herein. An example of such an application includes image compression. Essentially the present invention may be applied where one wishes to simultaneously analyse different regions of the frequency domain with differing temporal/spatial resolutions
  • Sinusoidal analysis techniques use Short Time Fast Fourier Transforms (FFT) to estimate the frequency of the component sinusoids.
  • FFT Short Time Fast Fourier Transforms
  • the derived signal is then synthesised with a bank of tone generators to produce the desired output.
  • Short Time Fourier Analysis captures information about the frequency content of a signal within a time interval, governed by the Window Function chosen.
  • a significant disadvantage of such techniques is that a single time-domain window is applied to all the frequency content of the signal, so the signal analysis cannot correspond accurately to human perception of the signal content.
  • conventional sinusoidal analysis methods use a local maxima search of the magnitude spectrum to determine the frequency of the constituent sinusoids including consideration of relative phase changes between analysis frames. This technique ignores any side-band information located around each of the local maxima.
  • phase vocoder methods This type of technique uses a Fast Fourier Transform as a large bank of filters and treats the output of each of the filters separately. The relative phase change between two consecutive analyses of the input is used to estimate the frequency of the signal content within each bin. A resulting frequency-domain signal is synthesised from this information, treating each bin as a separate signal. In contrast to sinusoidal analysis techniques, this method retains the spectral energy distribution of the original signal. However, it destroys the relative phase of any transient information. Therefore, the resulting sound is smeared and echo-like.
  • the invention provides for a method of encoding and resynthesising a waveform, the method including: sampling the waveform to obtain a series of discrete samples and constructing therefrom a series of frames, each frame spanning a plurality of samples; multiplying each frame with a windowing, preferably raised cosine, function wherein the peak of the windowing function is centred substantially at a zero point of each frame; applying a Fast Fourier Transform to each frame thereby producing a frequency-domain waveform; convoluting the resultant frequency domain data with a variable kernel function, whose specification varies with frequency; locating local maxima and surrounding minima in the magnitude spectrum of each convolved frame, wherein each local maxima and associated minima define a plurality of regions, each region corresponding to a frequency component of the signal; and analysing each of the regions in the frequency domain representation separately by summing the complex frequency components of bins falling within the defined region into a signal vector; wherein the variable kernel function can be usefully varied to achieve a differing tradeoff
  • the waveform corresponds to a digitised audio frequency waveform wherein the kernel function may be varied to approximate the perceptual characteristics of the human ear.
  • the location of the maxima corresponds to the perceived pitch of the frequency component.
  • the method may further include the step of manipulating the signal while represented as signal vectors.
  • Such manipulation may take the form of modifying pitch or time scale (in an audio signal) or further data reduction adapted for efficient signal storage and/or transmission.
  • the frequency location and phase of analysed signal vectors can be shifted as necessary to achieve a scaling of time and/or pitch.
  • Converting back to the sampled time domain representation of the signal may be achieved by accumulating into the frequency domain an equivalent signal whose components correspond to those signal vectors determined in the analysis of the original signal.
  • an Inverse Fast Fourier Transform may be applied so as to give a time domain signal that may be suitably windowed and accumulated to produce the decoded signal.
  • the form of the convolution function is determined empirically by subjectively assessing the quality of the synthesised output.
  • the application of the kernel function to the frequency domain data is implemented as a single-pole low-pass filter operation on said data, the pole's location being varied with frequency.
  • the pole may be specified by a control function s(f) of the form:
  • the frequency domain filter may be specified by the relation:
  • each signal vector is treated separately; for pitch shifting the frequency of the component is multiplied by a real-valued pitch factor; for both pitch shift and time scale modification the necessary phase shift for glitch free reconstruction is calculated and applied.
  • the method includes the further steps of: zeroing a frequency domain output array, and for each analysed frequency component represented as an analysed signal vector; mapping the real-valued frequency to the two nearest integer-valued frequency bins; and distributing the analysed signal vector between the two bins in proportion to 1 minus the real-valued frequency and the respective bins' locations.
  • the resulting regions may be translated in frequency, so that the location of the maxima is scaled while the surrounding region is translated.
  • each region having a maxima and a first and second associated minima, for pitch shifting of an audio signal, the location of each maxima in the frame is scaled by the pitch shift factor, and associated harmonic information between the first and second minima is translated to respective positions around the scaled maxima.
  • each maxima is retained in the same location in the frequency domain while the band of frequency domain or harmonic information associated with the maxima is stretched or compressed, thereby stretching the amplitude and frequency modulation of the harmonics while preserving the pitch of the input signal.
  • the method may further include the further steps of: resampling the data in each of the frames into a plurality of bins; mapping each bin to a real valued location in an output frame where for a bin x lying within a band with a maximum at a frequency freq max the real valued location in the output frequency domain iny, wherein
  • y is rounded down to the nearest integer z which is less than or equal to y wherein output bins z and z+1 are then added to, in proportion to 1 minus the difference between y and that bins integer location.
  • the invention provides for software adapted to perform the above-mentioned method.
  • the invention provides for hardware adapted to perform the above-mentioned method.
  • Figure 1 illustrates a simplified schematic block diagram of an embodiment of the method of the invention (split over pages 28 to 30);
  • Figure 2 illustrates a simplified schematic block diagram of an embodiment of the alternate method of the invention
  • Figure 3 illustrates a schematic diagram of the process of searching for the maxima/minima
  • Figure 5a and 5b illustrates pitch and time stretching in respect of two maxima.
  • FIG 1 a simplified flowchart illustrates the overall steps in an embodiment of the method of signal processing. For clarity, the schematic is split over pages 15 to 17.
  • An input audio signal is digitised into frames 10. Each of these frames is then processed as follows:
  • Each frame 10 is windowed (20) with (for example) a wide cosine function 30 producing time domain modulated representation of the input signal frame 10.
  • a Fast Fourier Transform 50 is then applied to the frame producing a frequency domain representation of the input signal 60.
  • the frequency domain data 60 is then filtered with a filtering function 71 parameterised by s(f).
  • the filtering function may also be viewed as a low-pass
  • the function s(f) 70 specifies how the behaviour of the filter varies with frequency.
  • the filtering function 71 can be described by the recursive relation:
  • s(f) controls the 'severity' of the filter 71.
  • a different convolution kernel is used for each frequency bin.
  • the real and imaginary components of each bin are convolved separately.
  • the filtering or convolution function 71 has the effect of "blurring" the frequency domain information and therefore the convolving function can be referred to as a blurring function. Blurring or spreading the frequency domain data corresponds to a narrowing of the equivalent window in the time domain frame. Therefore each frequency bin of the fast Fourier Transform is effectively calculated as if a different sized time domain window had been applied before the FFT operation.
  • the effect of the filter does not have to be to blur the data. For example, translating the time domain samples by half the window size would make it necessary to high-pass filter the frequency domain data, to achieve the same equivalent windowing in the time domain.
  • the frequency domain filter 71 is applied to each bin in ascending order and then applied in descending order of frequency bin. This is to ensure that no phase shift is introduced into the frequency domain data.
  • a key aspect of the present invention is that the control function s(f) is chosen, in the case of processing audio frequency data, so as to approximate the excitation response of human cilia located on the basilar membrane in the human ear. In effect, the function s(f) is chosen so as to approximate the time/frequency response of the human ear.
  • control function s(f) is, in the present preferred embodiment, determined empirically by gauging the quality of the output or synthesised waveform under varying circumstances. Although this is a subjective procedure, repeated and varied evaluations of the quality of the synthesised sound have been found to produce a highly satisfactory convolution function.
  • control function s(f) is:
  • the aforementioned steps are analogous to an efficient way to process a signal through a large bank of filters where the bandwidth of each filter is individually controllable by the control function s(f).
  • the convolved frequency domain data 80 is analysed (90) to determine the locations of local maxima and the associated local minima.
  • the data is a local maximum
  • Intensity(f) real(f) 2 + im(f) 2 .
  • each maxima and associated local minima is used to define regions (indicated by the shaded arrows in figure 3) which correspond to an audible harmonic in the original audio frequency signal.
  • the location of the maxima in the frequency domain corresponds to the perceived pitch of the harmonic and the band of the frequency domain information around the maxima represents any associated amplitude or frequency modulations of that harmonic. Since it is important not to lose this information, a summation of the whole band of frequencies around the peak is used to give a signal vector. This way the temporal resolution of the analysis sample will match the bandwidth of any modulations taking place.
  • Each of the regions is processed separately accordingly to the following technique. An accurate estimate of the location of each maxima is determined.
  • the large arrow a (300) is the difference between the smallest intensity of the three intensity arrows (max-1) and the maximum intensity (max).
  • the small arrow b (310) is the difference between the smallest (max-1) and the intermediate intensity (max+1). The ratio of the two is used to offset the integer maximum value.
  • Pitch shifting and time-scale modification are indicated schematically in figure 1 by the numeral 130. At this point alternative applications are indicated by data reduction (133) or transmission/storage (134) steps. These are illustrated as alternative options in figure 1.
  • the manipulated data are re-synthesised according to the following method: For the ith analysed frequency component, vector(i) has a real-valued location y in the frequency domain output. y is rounded down to the nearest integer which is less than or equal to y and
  • Bin[z] Bin[z] + [l-(y-z)] vector®
  • Bin[z+1] Bin[z+1] + (y-z) vector(i) where all operations are carried out on complex numbers.
  • the input time frame is moving by some other number of samples. Therefore, the analysed phase values are already changing as the analysis window moves through the input data.
  • each of the signal vectors defined above has a frequency measurement. This measurement is used to calculate how quickly to spin a vector of magnitude 1, where the vector is a complex number of representation. This vector is multiplied by the signal vector to provide the necessary phase shift for synthesis without affecting the timing of the decay characteristics or other modulations for each region.
  • This phase shift (in radians) is given by:
  • phase( ⁇ ) r —
  • t r reconstruction time step in samples
  • t a analysis time step in samples
  • t w FFT size in samples.
  • One integer array contains the location of the local maximum within a region for all the bins in that region.
  • a corresponding array contains the last phase value (in radians) used to rotate that regions phase.
  • the phase value is stored in the bin with the same index as the location of the maximum.
  • the thirteenth element of the nearest maxima array from the previous analysis frame n gives 16. From the phase array of the previous analysis frame n the phase is given as 57 degrees. A frequency estimate is used to update this phase value and is placed in the position 13 of the next phase array.
  • a frequency domain representation of the signal is constructed from the known signal components. For each signal vector, that vector is added to the frequency domain output array. Since the frequency locations are real valued, the energy from a signal vector is distributed between the nearest two (integer valued) bin locations. The frequency domain representation is then inverse Fourier transformed (150 in figure 1 page 16) to provide a time domain representation of the synthesised signal. Since the signal was analysed with differing temporal resolutions at different frequencies, the synthesised time domain signal is only valid in the region equivalent to the highest temporal analysis resolution used. To this end, the synthesised time domain signal is windowed (160) with a (relatively) small positive cosine window (170), before being added (172) in an overlapping fashion to the final synthesised signal (180).
  • the alternate method is substantially similar to the first method, sharing identically the steps of windowing (420), Fourier transforming (450), filtering (460), minima and maxima detection (490).
  • the major difference between the two methods is after this point.
  • the first method sums the contents of each region into a signal vector (110)
  • the alternate method instead explicitly retains the contents of each region (510).
  • the contents of each region are then translated and scaled in accordance with the pitch shift and time stretch factors respectively (530). For a pitch shift operation, the contents of a region are translated such that the maximum is scaled in frequency.
  • For a time stretch operation the contents of a region are scaled by the time stetch factor, but so that the maximum does not change in frequency.
  • Phase shift compensation is carried out substantially as described above with reference to figure 4a and 4b.
  • the frequency domain data to be synthesised is copied a region at a time from the unaltered output of Fourier transform step.
  • the contents of each region are accumulated into the output frequency domain buffer in the same fashion as the first method.
  • control function s(f) to vary a frequency domain filter at different frequencies. This brings about a windowing effect on the equivalent time-domain data that varies with frequency.
  • this control function is chosen to reflect the response of the human cilia to a range of audio frequencies.
  • a further feature of the present invention resides in the identification and location of the maxima and associated minima.
  • the presently disclosed technique is computationally highly efficient and allows rapid high quality time stretching and pitch shifting of audio signals.
  • the present technique produces a sound with significantly enhanced tonal qualities and it is believed that this is largely achieved through the preservation of the harmonic information in the sidebands of the local frequency maxima.
  • the technique may be implemented in software or alternatively in hardware.
  • the hardware may form part of an audio component such as an audio player.
  • Potential applications of the invention include the sound recording industry where audio signal processing/synthesis is commonly required to meet very high standards of reproduction quality.
  • Alternative applications include those in the entertainment industry and it is anticipated that the technique may find application in sound reproduction/transmission systems where variations in pitch or tempo may be desirable. It is further anticipated that applications may exist in general signal processing, data reduction and/or data transmission and storage. In the latter case, the selection of the particular convolution function may vary.

Abstract

A method of signal processing for time scale and/or pitch modification of audio signals is disclosed. The method involves encoding and resynthesising a wave form whereby the wave form is sampled into a series of frames, each frame is multiplied by a windowing function where the peak of the windowing function is centred at approximately the zero point of each frame. The resulting function is then subjected to a Fast Fourier transform thus producing a frequency-domain wave form. The resultant wave form is convolved with a variable kernel function, the specification of the variable kernel function varying with frequency. Maxima and associated minima in a magnitude spectrum of each convolved frame are located so that each local maxima and associated minima define a plurality of regions. Each region corresponds to a frequency component of the signal. Each of the regions is analysed in the frequency domain representation separately by summing the complex frequency components or bins falling within the defined region to a signal vector. The variable kernel function can be usefully varied to achieve a differing trade of between the frequency and temporal resolution across the frequency range of the signal.

Description

SIGNAL PROCESSING TECHNIQUES FOR TIME-SCALE AND/OR PITCH MODIFICATION OF AUDIO SIGNALS
Field of the Invention The present invention relates to encoding and manipulation of digital signals. More particularly, although not exclusively, the present invention relates to time-scale and/or pitch modification of audio signals. As such, the signal analysis and re-synthesis method described herein is not limited to audio signals. It is envisaged that the present invention may find application in the coding of other signals with the (wavelet-like) method described herein. An example of such an application includes image compression. Essentially the present invention may be applied where one wishes to simultaneously analyse different regions of the frequency domain with differing temporal/spatial resolutions
Background to the Invention
There are a number of existing techniques for time-scale/pitch modification of audio signals which are known in the art. These can be broadly classified as
follows.
(a) Time domain methods:
These techniques attempt to estimate the fundamental period of a musical signal by detecting periodic activity in the audio signal. By this process, an input signal is delayed and multiplied by the undelayed signal, the product of which is then smoothed in a low pass filter to provide an approximate measure of the autocorrelation function. The autocorrelation function is then used to detect a nonperiodic signal or a weak periodic signal which might be hidden in the noise. Once the fundamental period of the musical signal is found the process is repeated and the analysed sections of the signal are overlapped. A significant disadvantage in these techniques is that most audio signals do not have a fundamental period. For example polyphonic instruments, recordings with reverberation and percussion sounds do not have an identifiable fundamental period. Further, when applying such methods, transients in the music are repeated. This leads to notes having multiple starts and ends. Another problem with this technique is that overlapping of the delayed sections of the music can produce an audio effect that is metallic, mechanical or exhibits echo-like nature.
(b) Sinusoidal analysis methods:
These techniques assume that the input signal is made up from pure sinusoids.
The inherent disadvantage of such a method is therefore self-evident.
Sinusoidal analysis techniques use Short Time Fast Fourier Transforms (FFT) to estimate the frequency of the component sinusoids. The derived signal is then synthesised with a bank of tone generators to produce the desired output. Short Time Fourier Analysis captures information about the frequency content of a signal within a time interval, governed by the Window Function chosen. A significant disadvantage of such techniques is that a single time-domain window is applied to all the frequency content of the signal, so the signal analysis cannot correspond accurately to human perception of the signal content. Also, conventional sinusoidal analysis methods use a local maxima search of the magnitude spectrum to determine the frequency of the constituent sinusoids including consideration of relative phase changes between analysis frames. This technique ignores any side-band information located around each of the local maxima. The effect of this is to exclude any signal modulation occurring within a single analysis frame, resulting in a smearing of the sound and almost a complete loss of transients. An example of such a transient, in the audio context, is a guitar pluck, (c) Phase vocoder methods: This type of technique uses a Fast Fourier Transform as a large bank of filters and treats the output of each of the filters separately. The relative phase change between two consecutive analyses of the input is used to estimate the frequency of the signal content within each bin. A resulting frequency-domain signal is synthesised from this information, treating each bin as a separate signal. In contrast to sinusoidal analysis techniques, this method retains the spectral energy distribution of the original signal. However, it destroys the relative phase of any transient information. Therefore, the resulting sound is smeared and echo-like.
In view of the prior art techniques, it would therefore be desirable to analyse and process audio signals so that the resultant output retains the tonal characteristics of the original signal and is capable of accurately capturing transient sounds without smearing or introducing an echo-like character to the output signal.
Accordingly, it is an object of the present invention to provide a technique for processing audio signals which achieves the above mentioned aims, ameliorates at least some of the disadvantages inherent in the prior art or at least provides the public with a useful choice. Further, it is an object of the invention to provide a signal analysis and synthesis method that can also be applied to the coding of signals in general.
Disclosure of the Invention
In one aspect the invention provides for a method of encoding and resynthesising a waveform, the method including: sampling the waveform to obtain a series of discrete samples and constructing therefrom a series of frames, each frame spanning a plurality of samples; multiplying each frame with a windowing, preferably raised cosine, function wherein the peak of the windowing function is centred substantially at a zero point of each frame; applying a Fast Fourier Transform to each frame thereby producing a frequency-domain waveform; convoluting the resultant frequency domain data with a variable kernel function, whose specification varies with frequency; locating local maxima and surrounding minima in the magnitude spectrum of each convolved frame, wherein each local maxima and associated minima define a plurality of regions, each region corresponding to a frequency component of the signal; and analysing each of the regions in the frequency domain representation separately by summing the complex frequency components of bins falling within the defined region into a signal vector; wherein the variable kernel function can be usefully varied to achieve a differing tradeoff between frequency and temporal resolution across the frequency range of the signal.
In a preferred embodiment, the waveform corresponds to a digitised audio frequency waveform wherein the kernel function may be varied to approximate the perceptual characteristics of the human ear.
In the case where the waveform corresponds to an audio signal, the location of the maxima corresponds to the perceived pitch of the frequency component.
The method may further include the step of manipulating the signal while represented as signal vectors. Such manipulation may take the form of modifying pitch or time scale (in an audio signal) or further data reduction adapted for efficient signal storage and/or transmission.
In the case of modifying an audio signal, the frequency location and phase of analysed signal vectors can be shifted as necessary to achieve a scaling of time and/or pitch.
Converting back to the sampled time domain representation of the signal may be achieved by accumulating into the frequency domain an equivalent signal whose components correspond to those signal vectors determined in the analysis of the original signal.
Preferably an Inverse Fast Fourier Transform may be applied so as to give a time domain signal that may be suitably windowed and accumulated to produce the decoded signal.
Preferably the form of the convolution function is determined empirically by subjectively assessing the quality of the synthesised output.
Preferably the application of the kernel function to the frequency domain data is implemented as a single-pole low-pass filter operation on said data, the pole's location being varied with frequency. Preferably, in the case of the analysis of audio signals, the pole may be specified by a control function s(f) of the form:
s(f) = 0.4 + 0.26 arctan(4 ln(0.1/) - 18)
where/is the frequency in hertz (cycles per second).
The frequency domain filter may be specified by the relation:
y 0„, if) = [i - *(/)k if) + - if - 1)
Preferably, for the purposes of manipulating an audio signal, each signal vector is treated separately; for pitch shifting the frequency of the component is multiplied by a real-valued pitch factor; for both pitch shift and time scale modification the necessary phase shift for glitch free reconstruction is calculated and applied.
Preferably the method includes the further steps of: zeroing a frequency domain output array, and for each analysed frequency component represented as an analysed signal vector; mapping the real-valued frequency to the two nearest integer-valued frequency bins; and distributing the analysed signal vector between the two bins in proportion to 1 minus the real-valued frequency and the respective bins' locations.
In an alternative aspect, the resulting regions may be translated in frequency, so that the location of the maxima is scaled while the surrounding region is translated.
For each region, having a maxima and a first and second associated minima, for pitch shifting of an audio signal, the location of each maxima in the frame is scaled by the pitch shift factor, and associated harmonic information between the first and second minima is translated to respective positions around the scaled maxima.
To time stretch or compress the signal, each maxima is retained in the same location in the frequency domain while the band of frequency domain or harmonic information associated with the maxima is stretched or compressed, thereby stretching the amplitude and frequency modulation of the harmonics while preserving the pitch of the input signal.
The method may further include the further steps of: resampling the data in each of the frames into a plurality of bins; mapping each bin to a real valued location in an output frame where for a bin x lying within a band with a maximum at a frequency freqmax the real valued location in the output frequency domain iny, wherein
(scale) Where shift equals the frequency shift and scale equals time expansion ratio.
Preferably, y is rounded down to the nearest integer z which is less than or equal to y wherein output bins z and z+1 are then added to, in proportion to 1 minus the difference between y and that bins integer location.
In a further aspect, the invention provides for software adapted to perform the above-mentioned method.
In a further aspect, the invention provides for hardware adapted to perform the above-mentioned method.
Brief Description of the Drawings
The invention will now be described by way of example only and with reference to the drawings in which: Figure 1 : illustrates a simplified schematic block diagram of an embodiment of the method of the invention (split over pages 28 to 30);
Figure 2: illustrates a simplified schematic block diagram of an embodiment of the alternate method of the invention
(split over pages 31 to 33);
Figure 3 : illustrates a schematic diagram of the process of searching for the maxima/minima; Figure 5a and 5b: illustrates pitch and time stretching in respect of two maxima.
Referring to figure 1, a simplified flowchart illustrates the overall steps in an embodiment of the method of signal processing. For clarity, the schematic is split over pages 15 to 17.
An input audio signal is digitised into frames 10. Each of these frames is then processed as follows:
Each frame 10 is windowed (20) with (for example) a wide cosine function 30 producing time domain modulated representation of the input signal frame 10.
A Fast Fourier Transform 50 is then applied to the frame producing a frequency domain representation of the input signal 60.
The frequency domain data 60 is then filtered with a filtering function 71 parameterised by s(f). The filtering function may also be viewed as a low-pass
y~ if) = & - *(/)k (/) + sif)y0u, if - 1) single pole filter in the present example. The function s(f) 70 specifies how the behaviour of the filter varies with frequency. The filtering function 71 can be described by the recursive relation:
Thus s(f) controls the 'severity' of the filter 71. So in effect, a different convolution kernel is used for each frequency bin. The real and imaginary components of each bin are convolved separately. In the present exemplary embodiment, the filtering or convolution function 71 has the effect of "blurring" the frequency domain information and therefore the convolving function can be referred to as a blurring function. Blurring or spreading the frequency domain data corresponds to a narrowing of the equivalent window in the time domain frame. Therefore each frequency bin of the fast Fourier Transform is effectively calculated as if a different sized time domain window had been applied before the FFT operation.
The effect of the filter does not have to be to blur the data. For example, translating the time domain samples by half the window size would make it necessary to high-pass filter the frequency domain data, to achieve the same equivalent windowing in the time domain.
The frequency domain filter 71 is applied to each bin in ascending order and then applied in descending order of frequency bin. This is to ensure that no phase shift is introduced into the frequency domain data. A key aspect of the present invention is that the control function s(f) is chosen, in the case of processing audio frequency data, so as to approximate the excitation response of human cilia located on the basilar membrane in the human ear. In effect, the function s(f) is chosen so as to approximate the time/frequency response of the human ear.
The form of the control function s(f) is, in the present preferred embodiment, determined empirically by gauging the quality of the output or synthesised waveform under varying circumstances. Although this is a subjective procedure, repeated and varied evaluations of the quality of the synthesised sound have been found to produce a highly satisfactory convolution function.
A preferred form of the control function s(f) is:
s(f) = 0.4 + 0.26 arctan(4 ln(0. If) - 18)
where /is the frequency in hertz (cycles per second).
In effect, the aforementioned steps are analogous to an efficient way to process a signal through a large bank of filters where the bandwidth of each filter is individually controllable by the control function s(f). Once the filter 71 is applied, the convolved frequency domain data 80 is analysed (90) to determine the locations of local maxima and the associated local minima.
To perform this step, it has been found that it is more efficient to use the intensity spectrum. Therefore, for each frequency, the data is a local maximum
if /(/) > /(/ - 1) and /(/) > /(/ + 1) . Local minima exist if 1(f) < I(f - 1)
and I(f) < I(f + ϊ) . Here, Mag(f) = ^real(f)2 + im(ff and
Intensity(f) = real(f)2 + im(f)2.
Referring to figure 2, each maxima and associated local minima is used to define regions (indicated by the shaded arrows in figure 3) which correspond to an audible harmonic in the original audio frequency signal. The location of the maxima in the frequency domain corresponds to the perceived pitch of the harmonic and the band of the frequency domain information around the maxima represents any associated amplitude or frequency modulations of that harmonic. Since it is important not to lose this information, a summation of the whole band of frequencies around the peak is used to give a signal vector. This way the temporal resolution of the analysis sample will match the bandwidth of any modulations taking place. Each of the regions is processed separately accordingly to the following technique. An accurate estimate of the location of each maxima is determined. Referring to figure 3, lower graph, the large arrow a (300) is the difference between the smallest intensity of the three intensity arrows (max-1) and the maximum intensity (max). The small arrow b (310) is the difference between the smallest (max-1) and the intermediate intensity (max+1). The ratio of the two is used to offset the integer maximum value.
Pitch shifting and time-scale modification are indicated schematically in figure 1 by the numeral 130. At this point alternative applications are indicated by data reduction (133) or transmission/storage (134) steps. These are illustrated as alternative options in figure 1.
The manipulated data are re-synthesised according to the following method: For the ith analysed frequency component, vector(i) has a real-valued location y in the frequency domain output. y is rounded down to the nearest integer which is less than or equal to y and
denoted z. Thus z - Int(y).
The output bins z and z+1 are then added to with vector(i), in proportion to 1 minus the difference between and that bins integer location. Bin[z] = Bin[z] + [l-(y-z)] vector® Bin[z+1] = Bin[z+1] + (y-z) vector(i) where all operations are carried out on complex numbers.
To modify the time-scale or pitch of the analysed signal, it is necessary to compensate for any phase shifts so that the synthesised output is consistent (i.e. glitch free). To this end, the output signal in any one frame is moved forward in time by a fixed number of samples. Therefore, for a given pitch measurement it is possible to determine how much the output phase should change so that that the output smoothly joins with the previously synthesised frame.
However, the input time frame is moving by some other number of samples. Therefore, the analysed phase values are already changing as the analysis window moves through the input data.
Therefore the difference between the rate of change of input phase and the required rate of change of output phase is calculated. The difference between these phases is a measure of how fast to rotate the phase of the frequency domain data between analysis and synthesis. Each of the signal vectors defined above has a frequency measurement. This measurement is used to calculate how quickly to spin a vector of magnitude 1, where the vector is a complex number of representation. This vector is multiplied by the signal vector to provide the necessary phase shift for synthesis without affecting the timing of the decay characteristics or other modulations for each region. This phase shift (in radians) is given by:
phase(ι) = r
Where tr = reconstruction time step in samples, ta = analysis time step in samples and tw = FFT size in samples.
Since the measurement of frequency provides a measure of phase difference between one synthesis frame and the next, these differences must be summed cumulatively as synthesis proceeds.
The cumulative sum applies only to one region, therefore regions must be tracked from one synthesis frame to the next.
A convenient data structure has been developed to track regions from one frame to the next and is described with reference to figure 4a and 4b. One integer array contains the location of the local maximum within a region for all the bins in that region. A corresponding array contains the last phase value (in radians) used to rotate that regions phase. The phase value is stored in the bin with the same index as the location of the maximum.
Therefore, when a new frame is analysed and local maxima detected, the location of the maximum is used to index into the integer array. This provides the index of the maximum that existed in the previous frame. This index is then used to access the array containing the last phase value used for the corresponding region in the previous synthesis frame. This is illustrated in figures 3 a and b whereby an analysis frame n is illustrated along with the nearest maxima array and the phase array. Considering the n+1 analysis frame, the first frequency maxima is 7. The corresponding seventh element of the nearest maxima array from the previous frame is 5. The fifth element of the phase array frame from the previous frame n is 12 degrees. This is updated using an estimate of the local maxima and then stored in the phase array for the next frame using position 7. For the second region 410 the thirteenth element of the nearest maxima array from the previous analysis frame n gives 16. From the phase array of the previous analysis frame n the phase is given as 57 degrees. A frequency estimate is used to update this phase value and is placed in the position 13 of the next phase array.
A frequency domain representation of the signal is constructed from the known signal components. For each signal vector, that vector is added to the frequency domain output array. Since the frequency locations are real valued, the energy from a signal vector is distributed between the nearest two (integer valued) bin locations. The frequency domain representation is then inverse Fourier transformed (150 in figure 1 page 16) to provide a time domain representation of the synthesised signal. Since the signal was analysed with differing temporal resolutions at different frequencies, the synthesised time domain signal is only valid in the region equivalent to the highest temporal analysis resolution used. To this end, the synthesised time domain signal is windowed (160) with a (relatively) small positive cosine window (170), before being added (172) in an overlapping fashion to the final synthesised signal (180).
A variation, although equivalent, method of manipulating the information to achieve pitch shifting and time stretching is as follows.
The alternate method is substantially similar to the first method, sharing identically the steps of windowing (420), Fourier transforming (450), filtering (460), minima and maxima detection (490). The major difference between the two methods is after this point. Whereas the first method sums the contents of each region into a signal vector (110), the alternate method instead explicitly retains the contents of each region (510). The contents of each region are then translated and scaled in accordance with the pitch shift and time stretch factors respectively (530). For a pitch shift operation, the contents of a region are translated such that the maximum is scaled in frequency. For a time stretch operation, the contents of a region are scaled by the time stetch factor, but so that the maximum does not change in frequency.
Phase shift compensation is carried out substantially as described above with reference to figure 4a and 4b. To synthesise the output, the frequency domain data to be synthesised is copied a region at a time from the unaltered output of Fourier transform step. The contents of each region are accumulated into the output frequency domain buffer in the same fashion as the first method.
There exist variations in the implementation of these two techniques that will be clear to one skilled in the art. However, the key feature of the present invention resides in using a control function s(f) to vary a frequency domain filter at different frequencies. This brings about a windowing effect on the equivalent time-domain data that varies with frequency. In the case of processing audio frequency signals, this control function is chosen to reflect the response of the human cilia to a range of audio frequencies. Although the shape of this curve is determined empirically, it is possible that other curves may prove suitable for other manipulative techniques and applications.
A further feature of the present invention resides in the identification and location of the maxima and associated minima. The presently disclosed technique is computationally highly efficient and allows rapid high quality time stretching and pitch shifting of audio signals.
Experimentally, it has been shown that the present technique produces a sound with significantly enhanced tonal qualities and it is believed that this is largely achieved through the preservation of the harmonic information in the sidebands of the local frequency maxima. In terms of a practical implementation of the present invention, it is envisaged that the technique may be implemented in software or alternatively in hardware. In the latter case, the hardware may form part of an audio component such as an audio player. Potential applications of the invention include the sound recording industry where audio signal processing/synthesis is commonly required to meet very high standards of reproduction quality. Alternative applications include those in the entertainment industry and it is anticipated that the technique may find application in sound reproduction/transmission systems where variations in pitch or tempo may be desirable. It is further anticipated that applications may exist in general signal processing, data reduction and/or data transmission and storage. In the latter case, the selection of the particular convolution function may vary.
Where in the foregoing description reference has been made to elements or integers having known equivalents, then such equivalents are included as if they were individually set forth.
Although the invention has been described by way of example and with reference to particular embodiments, it is to be understood that modifications and/or improvements may be made without departing from the scope of the appended claims.

Claims

CLAIMS:
1. A method of encoding and re-synthesising a waveform, the method including the steps of: sampling the waveform to obtain a series of discrete samples and constructing therefrom a series of frames, each frame spanning a plurality of samples; multiplying each frame with a windowing function wherein the peak of the windowing function is centred substantially at a zero point of each frame; applying a Fast Fourier Transform to each frame thereby producing a frequency-domain waveform; convoluting the resultant frequency domain data with a variable kernel function, the specification of the variable kernel function varying with frequency; locating local maxima and surrounding minima in the magnitude spectrum of each convolved frame, wherein each local maxima and associated minima define a plurality of regions, each region corresponding to a frequency component of the signal; and analysing each of the regions in the frequency domain representation separately by summing the complex frequency components or bins falling within the defined region into a signal vector; wherein the variable kernel function can be usefully varied to achieve a differing tradeoff between frequency and temporal resolution across the frequency range of the signal.
2. A method of encoding and re-synthesising a waveform as claimed in claim 1 wherein the windowing function is a raised cosine function.
3. A method of encoding and re-synthesising a waveform as claimed in claim 1 wherein the waveform corresponds to a digitised audio frequency waveform wherein the kernel function is varied to approximate the perceptual characteristics of the human ear.
4. A method of encoding and re-synthesising a waveform as claimed in claim 1 wherein the waveform corresponds to an audio signal, and the location of the maxima corresponds to the perceived pitch of the frequency component.
5. A method of encoding and re-synthesising a waveform as claimed in claim 1 further including the step of manipulating the signal while represented as signal vectors.
6. A method of encoding and re-synthesising a waveform as claimed in claim 1 wherein said manipulation takes the form of modifying pitch or time scale (in an audio signal) or further data reduction adapted for efficient signal storage and/or transmission.
7. A method of encoding and re-synthesising a waveform as claimed in claim 1 wherein, in the case of modifying an audio signal, the frequency location and phase of analysed signal vectors are shifted according to a predetermined amount to achieve a scaling of time and/or pitch.
8. A method of encoding and re-synthesising a waveform as claimed in claim 1 wherein converting back to the sampled time domain representation of the signal is achieved by accumulating into the frequency domain an equivalent signal whose components correspond to those signal vectors determined in the analysis of the original signal.
A method of encoding and re-synthesising a waveform as claimed in claim 1 wherein an Inverse Fast Fourier Transform is applied so as to give a time domain signal that may be suitably windowed and accumulated to produce the decoded signal.
10. A method of encoding and re-synthesising a waveform as claimed in claim 1 wherein the form of the convolution function is determined empirically by subjectively assessing the quality of the synthesised output.
11. A method of encoding and re-synthesising a waveform as claimed in claim 1 wherein the application of the kernel function to the frequency domain data is implemented as a single-pole low-pass filter operation on said data, the pole's location being varied with frequency.
12 A method of encoding and re-synthesising a waveform as claimed in claim 11 wherein, in the case of the analysis of audio signals, the pole is specified by a control function s(f) of the form:
s(f) = 0.4 + 0.26arctan(41n(0.1/) - 18)
where /is the frequency in hertz (cycles per second).
13. A method of encoding and re-synthesising a waveform as claimed in claim 1 wherein the frequency domain filter may be specified by the relation :
y~ if) = & - *( >k if) + <f)y 0ΓÇ₧, if - i)
14. A method of encoding and re-synthesising a waveform as claimed in claim 1 wherein, for the purposes of manipulating an audio signal, each signal vector is treated separately; for pitch shifting the frequency of the component is multiplied by a real- valued pitch factor; for both pitch shift and time scale modification the necessary phase shift for glitch free reconstruction is calculated and applied.
15. A method of encoding and re-synthesising a waveform as claimed in claim 1 wherein the method includes the further steps of: zeroing a frequency domain output array, and for each analysed frequency component represented as an analysed signal vector; mapping the real- valued frequency to the two nearest integer-valued frequency bins; and distributing the analysed signal vector between the two bins in proportion to 1 minus the real-valued frequency and the respective bins' locations.
16. A method of encoding and resynthesizing a waveform as claimed in claim 1 wherein the resulting regions in the frequency domain are translated around each maxima to a different frequency, the position of the maxima and the resulting signal being a multiple of the frequency of the maxima so that the location of the maxima is scaled while the surrounding region is translated.
17. A method of encoding and resynthesizing a waveform as claimed in claim 16 wherein for each region, having a maxima and a first and second associated minima, for pitch shifting of an audio signal, the location of each maxima in the frame is scaled and associated harmonic information between the first and second minima and maxima is translated to respective positions around the maxima.
18. A method of encoding and resynthesizing a waveform as claimed in claim 16 or 17 wherein to time stretch the signal, each maxima is retained in the same location in the frequency domain while the band of frequency domain or harmonic information associated with the maxima is compressed, thereby stretching the amplitude and frequency modulation of the harmonics while preserving the pitch of the input signal.
19. A method of encoding and resynthesizing a waveform as claimed in claim including the further steps of: resampling the data in each of the frames into a plurality of bins; mapping each bin to a real valued location in an output frame where for a bin x lying within a band with a maximum at a frequency freqmax the real valued location in the output frequency domain is y, wherein
y = Jreqm^ shift + X fr^
(scale)
where shift equals the frequency shift and scale equals time expansion ratio.
20. A method of encoding and resynthesizing a waveform as claimed in claim 19 wherein, y is rounded down to the nearest integer z which is less than or equal to y wherein output bins z and z+1 are then added to, in proportion to 1 minus the difference between y and that bins integer location.
21. A software application operating in accordance with the method of claims 1 to 20.
22. A device constructed to perform in accordance with the method of claims 1 to 20.
PCT/NZ1999/000143 1998-08-28 1999-08-27 Signal processing techniques for time-scale and/or pitch modification of audio signals WO2000013172A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
JP2000568078A JP4527287B2 (en) 1998-08-28 1999-08-27 A signal processing technique for changing the time scale and / or fundamental frequency of an audio signal
AU54548/99A AU5454899A (en) 1998-08-28 1999-08-27 Signal processing techniques for time-scale and/or pitch modification of audio signals
EP99940754.7A EP1127349B1 (en) 1998-08-28 1999-08-27 Signal processing techniques for time-scale and/or pitch modification of audio signals

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
NZ331639 1998-08-28
NZ33163998 1998-08-28

Publications (1)

Publication Number Publication Date
WO2000013172A1 true WO2000013172A1 (en) 2000-03-09

Family

ID=19926908

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/NZ1999/000143 WO2000013172A1 (en) 1998-08-28 1999-08-27 Signal processing techniques for time-scale and/or pitch modification of audio signals

Country Status (6)

Country Link
US (1) US6266003B1 (en)
EP (1) EP1127349B1 (en)
JP (1) JP4527287B2 (en)
CN (1) CN1128436C (en)
AU (1) AU5454899A (en)
WO (1) WO2000013172A1 (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2003022141A1 (en) * 2001-09-13 2003-03-20 Imagyn Medical Technologies, Inc. A signal processing method and device for signal-to-noise improvement
US7283954B2 (en) 2001-04-13 2007-10-16 Dolby Laboratories Licensing Corporation Comparing audio using characterizations based on auditory events
US7313519B2 (en) 2001-05-10 2007-12-25 Dolby Laboratories Licensing Corporation Transient performance of low bit rate audio coding systems by reducing pre-noise
US7366659B2 (en) 2002-06-07 2008-04-29 Lucent Technologies Inc. Methods and devices for selectively generating time-scaled sound signals
US7461002B2 (en) 2001-04-13 2008-12-02 Dolby Laboratories Licensing Corporation Method for time aligning audio signals using characterizations based on auditory events
EP2017643A1 (en) * 2007-07-17 2009-01-21 Thales Method for optimising radioelectric signal measurements
US7610205B2 (en) 2002-02-12 2009-10-27 Dolby Laboratories Licensing Corporation High quality time-scaling and pitch-scaling of audio signals
US7711123B2 (en) 2001-04-13 2010-05-04 Dolby Laboratories Licensing Corporation Segmenting audio signals into auditory events

Families Citing this family (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB9911737D0 (en) * 1999-05-21 1999-07-21 Philips Electronics Nv Audio signal time scale modification
US6453252B1 (en) * 2000-05-15 2002-09-17 Creative Technology Ltd. Process for identifying audio content
US7421376B1 (en) * 2001-04-24 2008-09-02 Auditude, Inc. Comparison of data signals using characteristic electronic thumbprints
US7240001B2 (en) 2001-12-14 2007-07-03 Microsoft Corporation Quality improvement techniques in an audio encoder
AU2002321917A1 (en) * 2002-08-08 2004-02-25 Cosmotan Inc. Audio signal time-scale modification method using variable length synthesis and reduced cross-correlation computations
WO2004036549A1 (en) * 2002-10-14 2004-04-29 Koninklijke Philips Electronics N.V. Signal filtering
KR100547445B1 (en) * 2003-11-11 2006-01-31 주식회사 코스모탄 Shifting processing method of digital audio signal and audio / video signal and shifting reproduction method of digital broadcasting signal using the same
US7460990B2 (en) 2004-01-23 2008-12-02 Microsoft Corporation Efficient coding of digital media spectral data using wide-sense perceptual similarity
US8744862B2 (en) * 2006-08-18 2014-06-03 Digital Rise Technology Co., Ltd. Window selection based on transient detection and location to provide variable time resolution in processing frame-based data
US7895034B2 (en) * 2004-09-17 2011-02-22 Digital Rise Technology Co., Ltd. Audio encoding system
US7516074B2 (en) * 2005-09-01 2009-04-07 Auditude, Inc. Extraction and matching of characteristic fingerprints from audio signals
JP4839891B2 (en) * 2006-03-04 2011-12-21 ヤマハ株式会社 Singing composition device and singing composition program
CN101479789A (en) * 2006-06-29 2009-07-08 Nxp股份有限公司 Decoding sound parameters
US8046214B2 (en) * 2007-06-22 2011-10-25 Microsoft Corporation Low complexity decoder for complex transform coding of multi-channel sound
US7885819B2 (en) 2007-06-29 2011-02-08 Microsoft Corporation Bitstream syntax for multi-process audio decoding
US8706496B2 (en) * 2007-09-13 2014-04-22 Universitat Pompeu Fabra Audio signal transforming by utilizing a computational cost function
US8249883B2 (en) * 2007-10-26 2012-08-21 Microsoft Corporation Channel extension coding for multi-channel source
EP2293295A3 (en) * 2008-03-10 2011-09-07 Fraunhofer-Gesellschaft zur Förderung der Angewandten Forschung e.V. Device and method for manipulating an audio signal having a transient event
US8249386B2 (en) * 2008-03-28 2012-08-21 Tektronix, Inc. Video bandwidth resolution in DFT-based spectrum analysis
WO2010079377A1 (en) * 2009-01-09 2010-07-15 Universite D'angers Method and an apparatus for deconvoluting a noisy measured signal obtained from a sensor device
ES2374486T3 (en) * 2009-03-26 2012-02-17 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. DEVICE AND METHOD FOR HANDLING AN AUDIO SIGNAL.
EP4120263B1 (en) 2010-01-19 2023-08-09 Dolby International AB Improved subband block based harmonic transposition
KR101924326B1 (en) 2010-09-16 2018-12-03 돌비 인터네셔널 에이비 Cross product enhanced subband block based harmonic transposition
US9093120B2 (en) 2011-02-10 2015-07-28 Yahoo! Inc. Audio fingerprint extraction by scaling in time and resampling
US9159310B2 (en) 2012-10-19 2015-10-13 The Tc Group A/S Musical modification effects
KR101817544B1 (en) * 2015-12-30 2018-01-11 어보브반도체 주식회사 Bluetooth signal receiving method and device using improved carrier frequency offset compensation
WO2018077364A1 (en) 2016-10-28 2018-05-03 Transformizer Aps Method for generating artificial sound effects based on existing sound clips
CN107424616B (en) * 2017-08-21 2020-09-11 广东工业大学 Method and device for removing mask by phase spectrum
CN108281152B (en) * 2018-01-18 2021-01-12 腾讯音乐娱乐科技(深圳)有限公司 Audio processing method, device and storage medium
WO2020003342A1 (en) * 2018-06-25 2020-01-02 日本電気株式会社 Wave-source-direction estimation device, wave-source-direction estimation method, and program storage medium

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0250048A1 (en) * 1986-06-20 1987-12-23 Koninklijke Philips Electronics N.V. Frequency-domain block-adaptive digital filter

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0215915A4 (en) * 1985-03-18 1987-11-25 Massachusetts Inst Technology Processing of acoustic waveforms.
US5179626A (en) * 1988-04-08 1993-01-12 At&T Bell Laboratories Harmonic speech coding arrangement where a set of parameters for a continuous magnitude spectrum is determined by a speech analyzer and the parameters are used by a synthesizer to determine a spectrum which is used to determine senusoids for synthesis
US5297236A (en) * 1989-01-27 1994-03-22 Dolby Laboratories Licensing Corporation Low computational-complexity digital filter bank for encoder, decoder, and encoder/decoder
CN1062963C (en) * 1990-04-12 2001-03-07 多尔拜实验特许公司 Adaptive-block-lenght, adaptive-transform, and adaptive-window transform coder, decoder, and encoder/decoder for high-quality audio
US5327518A (en) * 1991-08-22 1994-07-05 Georgia Tech Research Corporation Audio analysis/synthesis system
DE4316297C1 (en) * 1993-05-14 1994-04-07 Fraunhofer Ges Forschung Audio signal frequency analysis method - using window functions to provide sample signal blocks subjected to Fourier analysis to obtain respective coefficients.
JP3536996B2 (en) * 1994-09-13 2004-06-14 ソニー株式会社 Parameter conversion method and speech synthesis method
WO1997019444A1 (en) * 1995-11-22 1997-05-29 Philips Electronics N.V. Method and device for resynthesizing a speech signal
JP3266819B2 (en) * 1996-07-30 2002-03-18 株式会社エイ・ティ・アール人間情報通信研究所 Periodic signal conversion method, sound conversion method, and signal analysis method

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0250048A1 (en) * 1986-06-20 1987-12-23 Koninklijke Philips Electronics N.V. Frequency-domain block-adaptive digital filter

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7461002B2 (en) 2001-04-13 2008-12-02 Dolby Laboratories Licensing Corporation Method for time aligning audio signals using characterizations based on auditory events
US7283954B2 (en) 2001-04-13 2007-10-16 Dolby Laboratories Licensing Corporation Comparing audio using characterizations based on auditory events
US7711123B2 (en) 2001-04-13 2010-05-04 Dolby Laboratories Licensing Corporation Segmenting audio signals into auditory events
US8195472B2 (en) 2001-04-13 2012-06-05 Dolby Laboratories Licensing Corporation High quality time-scaling and pitch-scaling of audio signals
US8488800B2 (en) 2001-04-13 2013-07-16 Dolby Laboratories Licensing Corporation Segmenting audio signals into auditory events
US7313519B2 (en) 2001-05-10 2007-12-25 Dolby Laboratories Licensing Corporation Transient performance of low bit rate audio coding systems by reducing pre-noise
US6658277B2 (en) 2001-09-13 2003-12-02 Imagyn Medical Technologies, Inc. Signal processing method and device for signal-to-noise improvement
US7060035B2 (en) 2001-09-13 2006-06-13 Conmed Corporation Signal processing method and device for signal-to-noise improvement
WO2003022141A1 (en) * 2001-09-13 2003-03-20 Imagyn Medical Technologies, Inc. A signal processing method and device for signal-to-noise improvement
US7610205B2 (en) 2002-02-12 2009-10-27 Dolby Laboratories Licensing Corporation High quality time-scaling and pitch-scaling of audio signals
US7366659B2 (en) 2002-06-07 2008-04-29 Lucent Technologies Inc. Methods and devices for selectively generating time-scaled sound signals
EP2017643A1 (en) * 2007-07-17 2009-01-21 Thales Method for optimising radioelectric signal measurements
FR2919129A1 (en) * 2007-07-17 2009-01-23 Thales Sa METHOD OF OPTIMIZING RADIO SIGNAL MEASUREMENTS

Also Published As

Publication number Publication date
EP1127349A1 (en) 2001-08-29
US6266003B1 (en) 2001-07-24
JP2002524759A (en) 2002-08-06
JP4527287B2 (en) 2010-08-18
EP1127349B1 (en) 2014-05-28
CN1315033A (en) 2001-09-26
EP1127349A4 (en) 2005-07-13
CN1128436C (en) 2003-11-19
AU5454899A (en) 2000-03-21

Similar Documents

Publication Publication Date Title
EP1127349B1 (en) Signal processing techniques for time-scale and/or pitch modification of audio signals
RU2487429C2 (en) Apparatus for processing audio signal containing transient signal
US5029509A (en) Musical synthesizer combining deterministic and stochastic waveforms
JP4641620B2 (en) Pitch detection refinement
US6182042B1 (en) Sound modification employing spectral warping techniques
US8017855B2 (en) Apparatus and method for converting an information signal to a spectral representation with variable resolution
AU597573B2 (en) Acoustic waveform processing
Virtanen Audio signal modeling with sinusoids plus noise
Beltrán et al. Estimation of the instantaneous amplitude and the instantaneous frequency of audio signals using complex wavelets
Schörkhuber et al. Pitch shifting of audio signals using the constant-q transform
RU2813317C1 (en) Improved harmonic transformation based on block of sub-bands
RU2800676C1 (en) Improved harmonic transformation based on a block of sub-bands
CA2820996A1 (en) Device and method for manipulating an audio signal having a transient event
RU2772356C2 (en) Improved harmonic conversion based on subrange block
Ferreira A new frequency domain approach to time-scale expansion of audio signals
Every et al. Separation of overlapping impulsive sounds by bandwise noise interpolation
AU2012216538B2 (en) Device and method for manipulating an audio signal having a transient event
Le Doux et al. Special session: What do student-generated diagrams say about their understanding?: developmental trajectories of model-based reasoning in engineering students
Hamdy et al. “Department of Electrical Engineering, Stanford University, Palo Alto, CA, USA" Digitronics Development Department, Sony Corporation, Kanagawa, Japan
Schörkhuber Applications of a Constant-Q Transform for Time-and Pitch-Scale Modifications
JPH0816194A (en) Voice signal decoder

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 99810151.6

Country of ref document: CN

AK Designated states

Kind code of ref document: A1

Designated state(s): AE AL AM AT AT AU AZ BA BB BG BR BY CA CH CN CR CU CZ CZ DE DE DK DK DM EE EE ES FI FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SK SL TJ TM TR TT UA UG US UZ VN YU ZA ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): GH GM KE LS MW SD SL SZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
ENP Entry into the national phase

Ref document number: 2000 568078

Country of ref document: JP

Kind code of ref document: A

WWE Wipo information: entry into national phase

Ref document number: 1999940754

Country of ref document: EP

REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

WWP Wipo information: published in national office

Ref document number: 1999940754

Country of ref document: EP