WO2010091013A1 - Bandwidth extension method and apparatus for a modified discrete cosine transform audio coder - Google Patents

Bandwidth extension method and apparatus for a modified discrete cosine transform audio coder Download PDF

Info

Publication number
WO2010091013A1
WO2010091013A1 PCT/US2010/022879 US2010022879W WO2010091013A1 WO 2010091013 A1 WO2010091013 A1 WO 2010091013A1 US 2010022879 W US2010022879 W US 2010022879W WO 2010091013 A1 WO2010091013 A1 WO 2010091013A1
Authority
WO
WIPO (PCT)
Prior art keywords
frequency band
band
transition
adjacent frequency
excitation spectrum
Prior art date
Application number
PCT/US2010/022879
Other languages
French (fr)
Inventor
Tenkasi Ramabadran
Mark Jasiuk
Original Assignee
Motorola, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Motorola, Inc. filed Critical Motorola, Inc.
Priority to JP2011544700A priority Critical patent/JP5597896B2/en
Priority to BRPI1008520A priority patent/BRPI1008520B1/en
Priority to CN201080006565.0A priority patent/CN102308333B/en
Priority to MX2011007807A priority patent/MX2011007807A/en
Priority to KR1020117018182A priority patent/KR101341246B1/en
Priority to EP10704446.3A priority patent/EP2394269B1/en
Publication of WO2010091013A1 publication Critical patent/WO2010091013A1/en

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/038Speech enhancement, e.g. noise reduction or echo cancellation using band spreading techniques
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/04Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
    • G10L19/06Determination or coding of the spectral characteristics, e.g. of the short-term prediction coefficients
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/04Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
    • G10L19/08Determination or coding of the excitation function; Determination or coding of the long-term prediction parameters
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/04Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
    • G10L19/16Vocoder architecture
    • G10L19/18Vocoders using multiple modes
    • G10L19/24Variable rate codecs, e.g. for generating different qualities using a scalable representation such as hierarchical encoding or layered encoding

Definitions

  • the present disclosure is related to audio coders and rendering audible content and more particularly to bandwidth extension techniques for audio coders.
  • Telephonic speech over mobile telephones has usually utilized only a portion of the audible sound spectrum, for example, narrow-band speech within the 300 to 3400 Hz audio spectrum. Compared to normal speech, such narrow-band speech has a muffled quality and reduced intelligibility. Therefore, various methods of extending the bandwidth of the output of speech coders, referred to as “bandwidth extension” or “BWE,” may be applied to artificially improve the perceived sound quality of the coder output.
  • BWE bandwidth extension
  • BWE schemes may be parametric or non-parametric, most known BWE schemes are parametric.
  • the parameters arise from the source-filter model of speech production where the speech signal is considered as an excitation source signal that has been acoustically filtered by the vocal tract.
  • the vocal tract may be modeled by an all-pole filter, for example, using linear prediction (LP) techniques to compute the filter coefficients.
  • LP coefficients effectively parameterize the speech spectral envelope information.
  • Other parametric methods utilize line spectral frequencies (LSF), mel-frequency cepstral coefficients (MFCC), and log-spectral envelope samples (LES) to model the speech spectral envelope.
  • LSF line spectral frequencies
  • MFCC mel-frequency cepstral coefficients
  • LES log-spectral envelope samples
  • MDCT MDCT Transform
  • FIG. 1 is a diagram of an audio signal having a transition band near a high frequency band that is used in the embodiments to estimate the high frequency band signal spectrum.
  • FIG. 2 is a flow chart of basic operation of a coder in accordance with the embodiments.
  • FIG. 3 is a flow chart showing further details of operation of a coder in accordance with the embodiments.
  • FIG. 4 is a block diagram of a communication device employing a coder in accordance with the embodiments.
  • FIG. 5 is a block diagram of a coder in accordance with the embodiments.
  • FIG. 6 is a block diagram of a coder in accordance with an embodiment. DETAILED DESCRIPTION
  • the present disclosure provides a method for bandwidth extension in a coder and includes defining a transition band for a signal having a spectrum within a first frequency band, where the transition band is defined as a portion of the first frequency band, and is located near an adjacent frequency band that is adjacent to the first frequency band.
  • the method analyzes the transition band to obtain a transition band spectral envelope and a transition band excitation spectrum; estimates an adjacent frequency band spectral envelope; generates an adjacent frequency band excitation spectrum by periodic repetition of at least a part of the transition band excitation spectrum with a repetition frequency determined by a pitch frequency of the signal; and combines the adjacent frequency band spectral envelope and the adjacent frequency band excitation spectrum to obtain an adjacent frequency band signal spectrum.
  • a signal processing logic for performing the method is also disclosed.
  • bandwidth extension may be implemented, using at least the quantized MDCT coefficients generated by a speech or audio coder modeling one frequency band, such as 4 to 7 kHz, to predict MDCT coefficients which model another frequency band, such as 7 to 14 kHz.
  • FIG. 1 is a graph 100, which is not to scale, that represents an audio signal 101 over an audible spectrum 102 ranging from 0 to Y kHz.
  • the signal 101 has a low band portion 104, and a high band portion 105 which is not reproduced as part of low band speech.
  • a transition band 103 is selected and utilized to estimate the high band portion 105.
  • the input signal may be obtained in various manners.
  • the signal 101 may be speech received over a digital wireless channel of a communication system, sent to a mobile station.
  • the signal 101 may also be obtained from memory, for example, in an audio playback device from a stored audio file.
  • FIG. 1 is a graph 100, which is not to scale, that represents an audio signal 101 over an audible spectrum 102 ranging from 0 to Y kHz.
  • the signal 101 has a low band portion 104, and a high band portion 105 which is not reproduced as part of low band speech.
  • a transition band 103 is selected and utilized
  • a transition band 103 is defined within a first frequency band 104 of the signal 101.
  • the transition band 103 is defined as a portion of the first frequency band and is located near the adjacent frequency band (such as high band portion 105).
  • the transition band 103 is analyzed to obtain transition band spectral data, and, in 205, the adjacent frequency band signal spectrum is generated using the transition band spectral data.
  • FIG. 3 illustrates further details of operation for one embodiment.
  • a transition band is defined similar to 201.
  • the transition band is analyzed to obtain transition band spectral data that includes the transition band spectral envelope and a transition band excitation spectrum.
  • the adjacent frequency band spectral envelope is estimated.
  • the adjacent frequency band excitation spectrum is then generated, as shown in 307, by periodic repetition of at least a part of the transition band excitation spectrum with a repetition frequency determined by a pitch frequency of the input signal.
  • the adjacent frequency band spectral envelope and the adjacent frequency band excitation spectrum may be combined to obtain a signal spectrum for the adjacent frequency band.
  • FIG. 4 is a block diagram illustrating the components of an electronic device 400 in accordance with the embodiments.
  • the electronic device may be a mobile station, a laptop computer, a personal digital assistant (PDA), a radio, an audio player (such as an MP3 player) or any other suitable device that may receive an audio signal, whether via wire or wireless transmission, and decode the audio signal using the methods and apparatuses of the embodiments herein disclosed.
  • the electronic device 400 will include an input portion 403 where an audio signal is provided to a signal processing logic 405 in accordance with the embodiments.
  • FIG. 4, as well as FIG. 5 and FIG. 6, are for illustrative purposes only, for the purpose of illustrating to one of ordinary skill, the logic necessary for making and using the embodiments herein described. Therefore, the Figures herein are not intended to be complete schematic diagrams of all components necessary for, for example, implementing an electronic device, but rather show only that which is necessary to facilitate an understanding, by one of ordinary skill, how to make and use the embodiments herein described. Therefore, it is also to be understood that various arrangements of logic, and any internal components shown, and any corresponding connectivity there-between, may be utilized and that such arrangements and corresponding connectivity would remain in accordance with the embodiments herein disclosed.
  • logic includes software and/or firmware executing on one or more programmable processors, ASICs, DSPs, hardwired logic or combinations thereof. Therefore, in accordance with the embodiments, any described logic, including for example, signal processing logic 405, may be implemented in any appropriate manner and would remain in accordance with the embodiments herein disclosed.
  • the electronic device 400 may include a receiver, or transceiver, front end portion 401 and any necessary antenna or antennas for receiving a signal. Therefore receiver 401 and/or input logic 403, individually or in combination, will include all necessary logic to provide appropriate audio signals to the signal processing logic 405 suitable for further processing by the signal processing logic 405.
  • the signal processing logic 405 may also include a codebook or codebooks 407 and lookup tables 409 in some embodiments.
  • the lookup tables 409 may be spectral envelope lookup tables.
  • FIG. 5 provides further details of the signal processing logic 405.
  • the signal processing logic 405 includes an estimation and control logic 500, which determines a set of MDCT coefficients to represent the high band portion of an audio signal.
  • An Inverse-MDCT, IMDCT 501 is used to convert the signal to the time- domain which is then combined with the low band portion of the audio signal 503 via a summation operation 505 to obtain a bandwidth extended audio signal.
  • the bandwidth extended audio signal is then output to an audio output logic (not shown).
  • the low band is considered to cover the range from 50 Hz to 7 kHz (nominally referred to as the wideband speech/audio spectrum) and the high band is considered to cover the range from 7 kHz to 14 kHz.
  • the combination of low and high bands, i.e. the range from 50 Hz to 14 kHz, is nominally referred to as the super-wideband speech/audio spectrum.
  • the low and high bands are possible and would remain in accordance with embodiments.
  • the input block 403, which is part of the baseline coder, is shown to provide the following signals: i) the decoded wideband speech/audio signal s w b, ii) the MDCT coefficients corresponding to at least the transition band, and iii) the pitch frequency 606 or the corresponding pitch period/delay.
  • the input block 403, in some embodiments, may provide only the decoded wideband speech/audio signal and the other signals may, in this case, be derived from it at the decoder.
  • a set of quantized MDCT coefficients is selected in 601 to represent a transition band.
  • the frequency band of 4 to 7 kHz may be utilized as a transition band; however other spectral portions may be used and would remain in accordance with the embodiments.
  • the selected transition band MDCT coefficients are used, along with selected parameters computed from the decoded wideband speech/audio (for example up to 7 kHz), to generate an estimated set of MDCT coefficients so as to specify signal content in the adjacent band, for example, from 7-14 kHz.
  • the selected transition band MDCT coefficients are thus provided to transition band analysis logic 603 and transition band energy estimator 615.
  • the energy in the quantized MDCT coefficients, representing the transition band, is computed by the transition band energy estimator logic 615.
  • the output of transition band energy estimator logic 615 is an energy value and is closely related to, although not identical to, the energy in the transition band of the decoded wideband speech/audio signal.
  • the energy value determined in 615 is input to high band energy predictor 611, which is a non-linear energy predictor that computes the energy of the MDCT coefficients modeling the adjacent band, for example the frequency band of 7- 14 kHz.
  • the high band energy predictor 611 may use zero-crossings from the decoded speech, calculated by zero crossings calculator 619, in conjunction with the spectral envelope shape of the transition band spectral portion determined by transition band shape estimator 609. Depending on the zero crossing value and the transition band shape, different non-linear predictors are used thus leading to enhanced predictor performance.
  • a large training database is first divided into a number of partitions based on the zero crossing value and the transition band shape and for each of the partitions so generated, separate predictor coefficients are computed.
  • the output of the zero crossings calculator 619 may be quantized using an 8-level scalar quantizer that quantizes the frame zero-crossings and, likewise, the transition band shape estimator 609 may be an 8-shape spectral envelope vector quantizer (VQ) that classifies the spectral envelope shape.
  • VQ 8-shape spectral envelope vector quantizer
  • the MDCT coefficients representing the signal in that band, are first processed in block 603 by an absolute-value operator.
  • the processed MDCT coefficients which are zero-valued are identified, and the zeroed-out magnitudes are replaced by values obtained through a linear interpolation between the bounding nonzero valued MDCT magnitudes, which have been scaled down (for example, by a factor of 5) prior to applying the linear interpolation operator.
  • the elimination of zero-valued MDCT coefficients as described above reduces the dynamic range of the MDCT magnitude spectrum, and improves the modeling efficiency of the spectral envelope computed from the modified MDCT coefficients.
  • the modified MDCT coefficients are then converted to the dB domain, via 20*loglO(x) operator (not shown).
  • the dB spectrum is obtained by spectral folding about a frequency index corresponding to 7 kHz, to further reduce the dynamic range of the spectral envelope to be computed for the 4-7 kHz frequency band.
  • An Inverse Discrete Fourier Transform (IDFT) is next applied to the dB spectrum thus constructed for the 4-8 kHz frequency band, to compute the first 8 (pseudo-)cepstral coefficients.
  • the dB spectral envelope is then calculated by performing a Discrete Fourier Transform (DFT) operation upon the cepstral coefficients.
  • DFT Discrete Fourier Transform
  • the resulting transition band MDCT spectral envelope is used in two ways. First, it forms an input to the transition band spectral envelope vector quantizer, that is, to transition band shape estimator 609, which returns an index of the pre-stored spectral envelope (one of 8) which is closest to the input spectral envelope. That index, along with an index (one of 8) returned by a scalar quantizer of the zero- crossings computed from the decoded speech, is used to select one of the at most 64 non-linear energy predictors, as previously detailed. Secondly, the computed spectral envelope is used to flatten the spectral envelope of the transition band MDCT coefficients.
  • the flattening may also be implemented in the log domain, in which case the division operation is replaced by a subtraction operation.
  • the MDCT coefficient signs (or polarities) are saved for later reinstatement, because the conversion to log domain requires positive valued inputs.
  • the flattening is implemented in the log domain.
  • the flattened transition-band MDCT coefficients (representing the transition band MDCT excitation spectrum) output by block 603 are then used to generate the MDCT coefficients which model the excitation signal in the band from 7-14 kHz.
  • the range of MDCT indices corresponding to the transition band may be 160 to 279, assuming that the initial MDCT index is 0 and 20 ms frame size at 32 kHz sampling.
  • the MDCT coefficients representing the excitation for indices 280 to 559 corresponding to the 7-14 kHz band are generated, using the following mapping:
  • the value of frequency delay D for a given frame, is computed from the value of long term predictor (LTP) delay for the last subframe of the 20 ms frame which is part of the core codec transmitted information. From this decoded LTP delay, an estimated pitch frequency value for the frame is computed, and the biggest integer multiple of this pitch frequency value is identified, to yield a corresponding integer frequency delay value D (defined in the MDCT index domain) which is less than or equal to 120.
  • LTP long term predictor
  • MDCT coefficients computed from a white noise sequence input may be used to form an estimate of flattened MDCT coefficients in the band from 7-14 kHz. Either way, an estimate of the MDCT coefficients representative of the excitation information in the 7-14 kHz band is formed by the high band excitation generator 605.
  • the 7-14 kHz output by the non-linear energy predictor may be adapted by energy adapter logic 617 based on the decoded wideband signal characteristics to minimize artifacts and enhance the quality of the bandwidth extended output speech.
  • the energy adapter 617 receives the following inputs in addition to the predicted high band energy value: i) the standard deviation ⁇ of the prediction error from high band energy predictor 611, ii) the voicing level v from the voicing level estimator 621, iii) the output d of the onset/plosive detector 623, and iv) the output ss of the steady- state/transition detector 625.
  • the spectral envelope consistent with that energy value is selected from a codebook 407.
  • a codebook of spectral envelopes modeling the spectral envelopes which characterize the MDCT coefficients in the 7- 14 kHz band and classified according to the energy values in that band is trained offline.
  • the envelope corresponding to the energy class closest to the predicted and adapted energy value is selected by high band envelope selector 613.
  • the selected spectral envelope is provided by the high band envelope selector 613 to the high band MDCT generator 607, and is then applied to shape the MDCT coefficients modeling the flattened excitation in the band from 7-14 kHz.
  • the shaped MDCT coefficients corresponding to the 7-14 kHz band representing the high band MDCT spectrum are next applied to an inverse modified cosine transform (IMDCT) 501, to form a time domain signal having content in the 7-14 kHz band.
  • IMDCT inverse modified cosine transform
  • the aforementioned predicted and adapted energy value can serve to facilitate accessing a look-up table 409 that contains a plurality of corresponding candidate spectral envelope shapes.
  • this apparatus can also comprise, if desired, one or more look-up tables 409 that are operably coupled to the signal processing logic 405. So configured, the signal processing logic 405 can readily access the look-up tables 409 as appropriate.
  • the signal processing discussed above may be performed by a mobile station in wireless communication with a base station.
  • the base station may transmit the wideband or narrow-band digital audio signal via conventional means to the mobile station.
  • signal processing logic within the mobile station performs the requisite operations to generate a bandwidth extended version of the digital audio signal that is clearer and more audibly pleasing to a user of the mobile station.
  • a voicing level estimator 621 may be used in conjunction with high band excitation generator 605. For example, a voicing level of 0, indicating unvoiced speech, may be used to determine use of noise excitation. Similarly, a voicing level of 1 indicating voiced speech, may be used to determine use of high band excitation derived from transition band excitation as described above. When the voicing level is in between 0 and 1 indicating mixed- voiced speech, various excitations may be mixed in appropriate proportion as determined by the voicing level and used.
  • the noise excitation may be a pseudo random noise function and as described above, may be considered as filling or patching holes in the spectrum based on the voicing level. A mixed high band excitation is thus suitable for voiced, unvoiced, and mixed-voiced sounds.
  • FIG. 6 shows the Estimation and Control Logic 500 as comprising transition band MDCT coefficient selector logic 601, transition band analysis logic 603, high band excitation generator 605, high band MDCT coefficient generator 607, transition band shape estimator 609, high band energy predictor 611, high band envelope selector 613, transition band energy estimator 615, energy adapter 617, zero- crossings calculator 619, voicing level estimator 621, onset/plosive detector 623, and SS/Transition detector 625.
  • the input 403 provides the decoded wideband speech/audio signal s w b, the MDCT coefficients corresponding to at least the transition band, and the pitch frequency (or delay) for each frame.
  • the transition band MDCT selector logic 601 is part of the baseline coder and provides a set of MDCT coefficients for the transition band to the transition band analysis logic 603 and to the transition band energy estimator 615.
  • a zero-crossing calculator 619 may calculate the number of zero-crossings zc in each frame of the wideband speech s w b as follows:
  • n the sample index
  • N the frame size in samples.
  • the value of the zc parameter calculated as above ranges from 0 to 1. From the zc parameter, a voicing level estimator 621 may estimate the voicing level v as follows.
  • V 0 if zc > ZC high zc - ZC hw
  • a transition-band energy estimator 615 estimates the transition-band energy from the transition band MDCT coefficients.
  • the transition-band is defined here as a frequency band that is contained within the wideband and close to the high band, i.e., it serves as a transition to the high band, (which, in this illustrative example, is about 7000 - 14,000 Hz).
  • One way to calculate the transition-band energy E t b is to sum the energies of the spectral components, i.e. MDCT coefficients, within the transition-band.
  • the coefficients a and ⁇ are selected to minimize the mean squared error between the true and estimated values of the high band energy over a large number of frames from a training speech/audio database.
  • the estimation accuracy can be further enhanced by exploiting contextual information from additional speech parameters such as the zero-crossing parameter zc and the transition-band spectral shape as may be provided by a transition-band shape estimator 609.
  • the zero-crossing parameter is indicative of the speech voicing level.
  • the transition band shape estimator 609 provides a high resolution representation of the transition band envelope shape.
  • a vector quantized representation of the transition band spectral envelope shapes in dB
  • the vector quantizer (VQ) codebook consists of 8 shapes referred to as transition band spectral envelope shape parameters tbs that are computed from a large training database.
  • a corresponding zc-tbs parameter plane may be formed using the zc and tbs parameters to achieve improved performance.
  • the zc-tbs plane is divided into 64 partitions corresponding to 8 scalar quantized levels of zc and the 8 tbs shapes. Some of the partitions may be merged with the nearby partitions for lack of sufficient data points from the training database. For each of the remaining partitions in the zc-tbs plane, separate predictor coefficients are computed.
  • the high band energy predictor 611 can provide additional improvement in estimation accuracy by using higher powers of E t b in estimating E h bo, e.g.,
  • Ehw OCA E t b + &i E t b + on E t b + QTi Ea, + ⁇ .
  • Ehbi is the adapted high band energy in dB
  • Ehw is the estimated high band energy in dB
  • ⁇ > 0 is a proportionality factor
  • is the standard deviation of the estimation error in dB.
  • high band energy predictor 611 additionally determines a measure of unreliability in the estimation of the high band energy level and energy adapter 617 biases the estimated high band energy level to be lower by an amount proportional to the measure of unreliability.
  • the measure of unreliability comprises a standard deviation ⁇ of the error in the estimated high band energy level.
  • Other measures of unreliability may as well be employed without departing from the scope of the embodiments.
  • the probability (or number of occurrences) of energy over-estimation is reduced, thereby reducing the number of artifacts.
  • the amount by which the estimated high band energy is reduced is proportional to how good the estimate is - a more reliable (i.e., low ⁇ value) estimate is reduced by a smaller amount than a less reliable estimate.
  • the ⁇ value corresponding to each partition of the zc-tbs parameter plane is computed from the training speech database and stored for later use in "biasing down" the estimated high band energy.
  • a suitable value of ⁇ for this high band energy predictor, for example, is 1.2.
  • the "bias down” approach described above has an added benefit for voiced frames - namely that of masking any errors in high band spectral envelope shape estimation and thereby reducing the resultant "noisy” artifacts.
  • voiced frames if the reduction in the estimated high band energy is too high, the bandwidth extended output speech no longer sounds like super wide band speech.
  • the estimated high band energy is further adapted in energy adapter 617 depending on its voicing level as
  • Ekb 2 E hbl + (1 -v) • S 1 + v S 2
  • E h b 2 is the voicing-level adapted high band energy in dB
  • v is the voicing level ranging from 0 for unvoiced speech to 1 for voiced speech
  • S ⁇ and Si (Si > Si) are constants in dB.
  • the choice of Si and Si depends on the value of ⁇ used for the "bias down" and is determined empirically to yield the best-sounding output speech. For example, when ⁇ is chosen as 1.2, S and Si may be chosen as 3.0 and -3.0 respectively. Note that other choices for the value of ⁇ may result in different choices for Si and Si - the values of Si and Si may both be positive or negative or of opposite signs.
  • the increased energy level for unvoiced speech emphasizes such speech in the bandwidth extended output compared to the wideband input and also helps to select a more appropriate spectral envelope shape for such unvoiced segments.
  • voicing level estimator 621 outputs a voicing level to energy adapter 617 which further modifies the estimated high band energy level based on wideband signal characteristics by further modifying the estimated high band energy level based on a voicing level.
  • the further modifying may comprise reducing the high band energy level for substantially voiced speech and/or increasing the high band energy level for substantially unvoiced speech.
  • the step of modifying the estimated high band energy level based on the wideband signal characteristics may comprise smoothing the estimated high band energy level (which has been previously modified as described above based on the standard deviation of the estimation ⁇ and the voicing level v), essentially reducing an energy difference between consecutive frames.
  • the voicing-level adapted high band energy E h b 2 may be smoothed using a 3 -point averaging filter as
  • E h b3 is the smoothed estimate and k is the frame index.
  • Smoothing reduces the energy difference between consecutive frames, especially when an estimate is an "outlier", that is, the high band energy estimate of a frame is too high or too low compared to the estimates of the neighboring frames.
  • smoothing helps to reduce the number of artifacts in the output bandwidth extended speech.
  • the 3-point averaging filter introduces a delay of one frame.
  • Other types of filters with or without delay can also be designed for smoothing the energy track.
  • the smoothed energy value E h b3 may be further adapted by energy adapter 617 to obtain the final adapted high band energy estimate E hb .
  • This adaptation can involve either decreasing or increasing the smoothed energy value based on the ss parameter output by the steady-state/transition detector 625 and/or the d parameter output by the onset/plosive detector 623.
  • the step of modifying the estimated high band energy level based on the wideband signal characteristics may include the step of modifying the estimated high band energy level (or previously modified estimated high band energy level) based on whether or not a frame is steady- state or transient.
  • This may include reducing the high band energy level for transient frames and/or increasing the high band energy level for steady-state frames, and may further include modifying the estimated high band energy level based on an occurrence of an onset/plosive.
  • adapting the high band energy value changes not only the energy level but also the spectral envelope shape since the selection of the high band spectrum may be tied to the estimated energy.
  • a frame is defined as a steady-state frame if it has sufficient energy
  • ⁇ > ⁇ i ⁇ 0, are empirically chosen constants in dB to achieve good output speech quality.
  • the values of ⁇ and ⁇ ⁇ depend on the choice of the proportionality constant ⁇ used for the "bias down". For example, when ⁇ is chosen as 1.2, ⁇ as 3.0, and ⁇ as -3.0, ⁇ and ⁇ may be chosen as 1.5 and 6.0 respectively. Notice that in this example we are slightly increasing the estimated high band energy for steady-state frames and decreasing it significantly further for transition frames. Note that other choices for the values of A, and ⁇ may result in different choices for ⁇ and ⁇ ⁇ - the values of ⁇ and ⁇ ⁇ may both be positive or negative or of opposite signs. Further, note that other criteria for identifying steady-state/transition frames may also be used.
  • An onset/plosive presents a special problem because of the following reasons: A) Estimation of high band energy near onset/plosive is difficult; B) Pre-echo type artifacts may occur in the output speech because of the typical block processing employed; and C) Plosive sounds (e.g., [p], [t], and [k]), after their initial energy burst, have characteristics similar to certain sibilants (e.g., [s], [J " ], and [3]) in the wideband but quite different in the high band leading to energy over-estimation and consequent artifacts.
  • k is the frame index.
  • E mm can be set to - ⁇ dB or to the energy of the high band spectral envelope shape with the lowest energy.
  • energy adaptation is done only as long as the voicing level v(k) of the frame exceeds the threshold F 1 .
  • the zero-crossing parameter zc with an appropriate threshold may also be used for this purpose.
  • the step of modifying the estimated high band energy level based on the wideband signal characteristics may comprise the step of modifying the estimated high band energy level (or previously modified estimated high band energy level) based on an occurrence of an onset/plosive.

Abstract

A method includes defining a transition band for a signal having a spectrum within a first frequency band, where the transition band is defined as a portion of the first frequency band, and is located near an adjacent frequency band that is adjacent to the first frequency band. The method analyzes the transition band to obtain a transition band spectral envelope and a transition band excitation spectrum; estimates an adjacent frequency band spectral envelope; generates an adjacent frequency band excitation spectrum by periodic repetition of at least a part of the transition band excitation spectrum with a repetition period determined by a pitch frequency of the signal; and combines the adjacent frequency band spectral envelope and the adjacent frequency band excitation spectrum to obtain an adjacent frequency band signal spectrum. A signal processing logic for performing the method is also disclosed.

Description

BANDWIDTH EXTENSION METHOD AND APPARATUS FOR A MODIFIED DISCRETE COSINE TRANSFORM AUDIO CODER
CROSS-REFERENCE TO RELATED APPLICATIONS
[0001] The present disclosure is related to: U.S. Patent Application No.
11/946,978, Attorney Docket No.: CML04909EV, filed Nov. 29, 2007, entitled METHOD AND APPARATUS TO FACILITATE PROVISION AND USE OF AN ENERGY VALUE TO DETERMINE A SPECTRAL ENVELOPE SHAPE FOR OUT-OF-SIGNAL BANDWIDTH CONTENT; U.S. Patent Application No. 12/024,620, Attorney Docket No.: CML04911EV, filed Feb. 1, 2008, entitled METHOD AND APPARATUS FOR ESTIMATING HIGH-BAND ENERGY IN A BANDWIDTH EXTENSION SYSTEM; U.S. Patent Application No. 12/027,571, Attorney Docket No.: CML06672AUD, filed Feb. 7, 2008, entitled METHOD AND APPARATUS FOR ESTIMATING HIGH-BAND ENERGY IN A BANDWIDTH EXTENSION SYSTEM; all of which are incorporated by reference herein.
FIELD OF THE DISCLOSURE
[0002] The present disclosure is related to audio coders and rendering audible content and more particularly to bandwidth extension techniques for audio coders.
BACKGROUND
[0003] Telephonic speech over mobile telephones has usually utilized only a portion of the audible sound spectrum, for example, narrow-band speech within the 300 to 3400 Hz audio spectrum. Compared to normal speech, such narrow-band speech has a muffled quality and reduced intelligibility. Therefore, various methods of extending the bandwidth of the output of speech coders, referred to as "bandwidth extension" or "BWE," may be applied to artificially improve the perceived sound quality of the coder output.
[0004] Although BWE schemes may be parametric or non-parametric, most known BWE schemes are parametric. The parameters arise from the source-filter model of speech production where the speech signal is considered as an excitation source signal that has been acoustically filtered by the vocal tract. The vocal tract may be modeled by an all-pole filter, for example, using linear prediction (LP) techniques to compute the filter coefficients. The LP coefficients effectively parameterize the speech spectral envelope information. Other parametric methods utilize line spectral frequencies (LSF), mel-frequency cepstral coefficients (MFCC), and log-spectral envelope samples (LES) to model the speech spectral envelope.
[0005] Many current speech/audio coders utilize the Modified Discrete Cosine
Transform (MDCT) representation of the input signal and therefore BWE methods are needed that could be applied to MDCT based speech/audio coders.
BRIEF DESCRIPTION OF THE DRAWINGS
[0006] FIG. 1 is a diagram of an audio signal having a transition band near a high frequency band that is used in the embodiments to estimate the high frequency band signal spectrum.
[0007] FIG. 2 is a flow chart of basic operation of a coder in accordance with the embodiments.
[0008] FIG. 3 is a flow chart showing further details of operation of a coder in accordance with the embodiments.
[0009] FIG. 4 is a block diagram of a communication device employing a coder in accordance with the embodiments.
[0010] FIG. 5 is a block diagram of a coder in accordance with the embodiments.
[0011] FIG. 6 is a block diagram of a coder in accordance with an embodiment. DETAILED DESCRIPTION
[0012] The present disclosure provides a method for bandwidth extension in a coder and includes defining a transition band for a signal having a spectrum within a first frequency band, where the transition band is defined as a portion of the first frequency band, and is located near an adjacent frequency band that is adjacent to the first frequency band. The method analyzes the transition band to obtain a transition band spectral envelope and a transition band excitation spectrum; estimates an adjacent frequency band spectral envelope; generates an adjacent frequency band excitation spectrum by periodic repetition of at least a part of the transition band excitation spectrum with a repetition frequency determined by a pitch frequency of the signal; and combines the adjacent frequency band spectral envelope and the adjacent frequency band excitation spectrum to obtain an adjacent frequency band signal spectrum. A signal processing logic for performing the method is also disclosed.
[0013] In accordance with the embodiments, bandwidth extension may be implemented, using at least the quantized MDCT coefficients generated by a speech or audio coder modeling one frequency band, such as 4 to 7 kHz, to predict MDCT coefficients which model another frequency band, such as 7 to 14 kHz.
[0014] Turning now to the drawings wherein like numerals represent like components, FIG. 1 is a graph 100, which is not to scale, that represents an audio signal 101 over an audible spectrum 102 ranging from 0 to Y kHz. The signal 101 has a low band portion 104, and a high band portion 105 which is not reproduced as part of low band speech. In accordance with the embodiments, a transition band 103 is selected and utilized to estimate the high band portion 105. The input signal may be obtained in various manners. For example, the signal 101 may be speech received over a digital wireless channel of a communication system, sent to a mobile station. The signal 101 may also be obtained from memory, for example, in an audio playback device from a stored audio file. [0015] FIG. 2 illustrates the basic operation of a coder in accordance with the embodiments. In 201 a transition band 103 is defined within a first frequency band 104 of the signal 101. The transition band 103 is defined as a portion of the first frequency band and is located near the adjacent frequency band (such as high band portion 105). In 203 the transition band 103 is analyzed to obtain transition band spectral data, and, in 205, the adjacent frequency band signal spectrum is generated using the transition band spectral data.
[0016] FIG. 3 illustrates further details of operation for one embodiment. In
301 a transition band is defined similar to 201. In 303, the transition band is analyzed to obtain transition band spectral data that includes the transition band spectral envelope and a transition band excitation spectrum. In 305, the adjacent frequency band spectral envelope is estimated. The adjacent frequency band excitation spectrum is then generated, as shown in 307, by periodic repetition of at least a part of the transition band excitation spectrum with a repetition frequency determined by a pitch frequency of the input signal. As shown in 309, the adjacent frequency band spectral envelope and the adjacent frequency band excitation spectrum may be combined to obtain a signal spectrum for the adjacent frequency band.
[0017] FIG. 4 is a block diagram illustrating the components of an electronic device 400 in accordance with the embodiments. The electronic device may be a mobile station, a laptop computer, a personal digital assistant (PDA), a radio, an audio player (such as an MP3 player) or any other suitable device that may receive an audio signal, whether via wire or wireless transmission, and decode the audio signal using the methods and apparatuses of the embodiments herein disclosed. The electronic device 400 will include an input portion 403 where an audio signal is provided to a signal processing logic 405 in accordance with the embodiments.
[0018] It is to be understood that FIG. 4, as well as FIG. 5 and FIG. 6, are for illustrative purposes only, for the purpose of illustrating to one of ordinary skill, the logic necessary for making and using the embodiments herein described. Therefore, the Figures herein are not intended to be complete schematic diagrams of all components necessary for, for example, implementing an electronic device, but rather show only that which is necessary to facilitate an understanding, by one of ordinary skill, how to make and use the embodiments herein described. Therefore, it is also to be understood that various arrangements of logic, and any internal components shown, and any corresponding connectivity there-between, may be utilized and that such arrangements and corresponding connectivity would remain in accordance with the embodiments herein disclosed.
[0019] The term "logic" as used herein includes software and/or firmware executing on one or more programmable processors, ASICs, DSPs, hardwired logic or combinations thereof. Therefore, in accordance with the embodiments, any described logic, including for example, signal processing logic 405, may be implemented in any appropriate manner and would remain in accordance with the embodiments herein disclosed.
[0020] The electronic device 400 may include a receiver, or transceiver, front end portion 401 and any necessary antenna or antennas for receiving a signal. Therefore receiver 401 and/or input logic 403, individually or in combination, will include all necessary logic to provide appropriate audio signals to the signal processing logic 405 suitable for further processing by the signal processing logic 405. The signal processing logic 405 may also include a codebook or codebooks 407 and lookup tables 409 in some embodiments. The lookup tables 409 may be spectral envelope lookup tables.
[0021] FIG. 5 provides further details of the signal processing logic 405. The signal processing logic 405 includes an estimation and control logic 500, which determines a set of MDCT coefficients to represent the high band portion of an audio signal. An Inverse-MDCT, IMDCT 501 is used to convert the signal to the time- domain which is then combined with the low band portion of the audio signal 503 via a summation operation 505 to obtain a bandwidth extended audio signal. The bandwidth extended audio signal is then output to an audio output logic (not shown).
[0022] Further details of some embodiments are illustrated by FIG. 6, although some logic illustrated may not, and need not, be present in all embodiments. For purposes of illustration, in the following, the low band is considered to cover the range from 50 Hz to 7 kHz (nominally referred to as the wideband speech/audio spectrum) and the high band is considered to cover the range from 7 kHz to 14 kHz. The combination of low and high bands, i.e. the range from 50 Hz to 14 kHz, is nominally referred to as the super-wideband speech/audio spectrum. Clearly, other choices for the low and high bands are possible and would remain in accordance with embodiments. Also, for purposes of illustration, the input block 403, which is part of the baseline coder, is shown to provide the following signals: i) the decoded wideband speech/audio signal swb, ii) the MDCT coefficients corresponding to at least the transition band, and iii) the pitch frequency 606 or the corresponding pitch period/delay. The input block 403, in some embodiments, may provide only the decoded wideband speech/audio signal and the other signals may, in this case, be derived from it at the decoder. As illustrated in FIG. 6, from the input block 403, a set of quantized MDCT coefficients is selected in 601 to represent a transition band. For example, the frequency band of 4 to 7 kHz may be utilized as a transition band; however other spectral portions may be used and would remain in accordance with the embodiments.
[0023] Next the selected transition band MDCT coefficients are used, along with selected parameters computed from the decoded wideband speech/audio (for example up to 7 kHz), to generate an estimated set of MDCT coefficients so as to specify signal content in the adjacent band, for example, from 7-14 kHz. The selected transition band MDCT coefficients are thus provided to transition band analysis logic 603 and transition band energy estimator 615. The energy in the quantized MDCT coefficients, representing the transition band, is computed by the transition band energy estimator logic 615. The output of transition band energy estimator logic 615 is an energy value and is closely related to, although not identical to, the energy in the transition band of the decoded wideband speech/audio signal.
[0024] The energy value determined in 615 is input to high band energy predictor 611, which is a non-linear energy predictor that computes the energy of the MDCT coefficients modeling the adjacent band, for example the frequency band of 7- 14 kHz. In some embodiments, to improve the high band energy predictor 611 performance, the high band energy predictor 611 may use zero-crossings from the decoded speech, calculated by zero crossings calculator 619, in conjunction with the spectral envelope shape of the transition band spectral portion determined by transition band shape estimator 609. Depending on the zero crossing value and the transition band shape, different non-linear predictors are used thus leading to enhanced predictor performance. In designing the predictors, a large training database is first divided into a number of partitions based on the zero crossing value and the transition band shape and for each of the partitions so generated, separate predictor coefficients are computed.
[0025] Specifically, the output of the zero crossings calculator 619 may be quantized using an 8-level scalar quantizer that quantizes the frame zero-crossings and, likewise, the transition band shape estimator 609 may be an 8-shape spectral envelope vector quantizer (VQ) that classifies the spectral envelope shape. Thus at each frame at most 64 (i.e., 8x8) nonlinear predictors are provided, and a predictor corresponding to the selected partition is employed at that frame. In most embodiments, fewer than 64 predictors are used, because some of the 64 partitions are not assigned a sufficient number of frames from the training database to warrant their inclusion, and those partitions may be consequently merged with the nearby partitions. A separate energy predictor (not shown), trained over low energy frames, may be used for such low-energy frames in accordance with the embodiments.
[0026] To compute the spectral envelope corresponding to the transition band
(4-7 kHz), the MDCT coefficients, representing the signal in that band, are first processed in block 603 by an absolute-value operator. Next, the processed MDCT coefficients which are zero-valued are identified, and the zeroed-out magnitudes are replaced by values obtained through a linear interpolation between the bounding nonzero valued MDCT magnitudes, which have been scaled down (for example, by a factor of 5) prior to applying the linear interpolation operator. The elimination of zero-valued MDCT coefficients as described above reduces the dynamic range of the MDCT magnitude spectrum, and improves the modeling efficiency of the spectral envelope computed from the modified MDCT coefficients. [0027] The modified MDCT coefficients are then converted to the dB domain, via 20*loglO(x) operator (not shown). In the band from 7 to 8 kHz, the dB spectrum is obtained by spectral folding about a frequency index corresponding to 7 kHz, to further reduce the dynamic range of the spectral envelope to be computed for the 4-7 kHz frequency band. An Inverse Discrete Fourier Transform (IDFT) is next applied to the dB spectrum thus constructed for the 4-8 kHz frequency band, to compute the first 8 (pseudo-)cepstral coefficients. The dB spectral envelope is then calculated by performing a Discrete Fourier Transform (DFT) operation upon the cepstral coefficients.
[0028] The resulting transition band MDCT spectral envelope is used in two ways. First, it forms an input to the transition band spectral envelope vector quantizer, that is, to transition band shape estimator 609, which returns an index of the pre-stored spectral envelope (one of 8) which is closest to the input spectral envelope. That index, along with an index (one of 8) returned by a scalar quantizer of the zero- crossings computed from the decoded speech, is used to select one of the at most 64 non-linear energy predictors, as previously detailed. Secondly, the computed spectral envelope is used to flatten the spectral envelope of the transition band MDCT coefficients. One way in which this may be done is to divide each transition band MDCT coefficient by its corresponding spectral envelope value. The flattening may also be implemented in the log domain, in which case the division operation is replaced by a subtraction operation. In the latter implementation, the MDCT coefficient signs (or polarities) are saved for later reinstatement, because the conversion to log domain requires positive valued inputs. In the embodiments, the flattening is implemented in the log domain.
[0029] The flattened transition-band MDCT coefficients (representing the transition band MDCT excitation spectrum) output by block 603 are then used to generate the MDCT coefficients which model the excitation signal in the band from 7-14 kHz. In one embodiment the range of MDCT indices corresponding to the transition band may be 160 to 279, assuming that the initial MDCT index is 0 and 20 ms frame size at 32 kHz sampling. Given the flattened transition-band MDCT coefficients, the MDCT coefficients representing the excitation for indices 280 to 559 corresponding to the 7-14 kHz band are generated, using the following mapping:
[0030] MDCTexIi) = MDCT^i-D), i = 280,...,559, D <= 120.
[0031] The value of frequency delay D, for a given frame, is computed from the value of long term predictor (LTP) delay for the last subframe of the 20 ms frame which is part of the core codec transmitted information. From this decoded LTP delay, an estimated pitch frequency value for the frame is computed, and the biggest integer multiple of this pitch frequency value is identified, to yield a corresponding integer frequency delay value D (defined in the MDCT index domain) which is less than or equal to 120. This approach ensures the reuse of the flattened transition-band MDCT information thus preserving the harmonic relationship between the MDCT coefficients in the 4-7 kHz band and the MDCT coefficients being estimated for the 7- 14 kHz band. Alternately, MDCT coefficients computed from a white noise sequence input may be used to form an estimate of flattened MDCT coefficients in the band from 7-14 kHz. Either way, an estimate of the MDCT coefficients representative of the excitation information in the 7-14 kHz band is formed by the high band excitation generator 605.
[0032] The predicted energy value of the MDCT coefficients in the band from
7-14 kHz output by the non-linear energy predictor may be adapted by energy adapter logic 617 based on the decoded wideband signal characteristics to minimize artifacts and enhance the quality of the bandwidth extended output speech. For this purpose, the energy adapter 617 receives the following inputs in addition to the predicted high band energy value: i) the standard deviation σ of the prediction error from high band energy predictor 611, ii) the voicing level v from the voicing level estimator 621, iii) the output d of the onset/plosive detector 623, and iv) the output ss of the steady- state/transition detector 625.
[0033] Given the predicted and adapted energy value of the MDCT coefficients in the band from 7-14 kHz, the spectral envelope consistent with that energy value is selected from a codebook 407. Such a codebook of spectral envelopes modeling the spectral envelopes which characterize the MDCT coefficients in the 7- 14 kHz band and classified according to the energy values in that band is trained offline. The envelope corresponding to the energy class closest to the predicted and adapted energy value is selected by high band envelope selector 613.
[0034] The selected spectral envelope is provided by the high band envelope selector 613 to the high band MDCT generator 607, and is then applied to shape the MDCT coefficients modeling the flattened excitation in the band from 7-14 kHz. The shaped MDCT coefficients corresponding to the 7-14 kHz band representing the high band MDCT spectrum are next applied to an inverse modified cosine transform (IMDCT) 501, to form a time domain signal having content in the 7-14 kHz band. This signal is then combined by, for example summation operation 505, with the decoded wideband signal having content up to 7 kHz, that is, low band portion 503, to form the bandwidth extended signal which contains information up to 14 kHz.
[0035] By one approach, the aforementioned predicted and adapted energy value can serve to facilitate accessing a look-up table 409 that contains a plurality of corresponding candidate spectral envelope shapes. To support such an approach, this apparatus can also comprise, if desired, one or more look-up tables 409 that are operably coupled to the signal processing logic 405. So configured, the signal processing logic 405 can readily access the look-up tables 409 as appropriate.
[0036] It is to be understood that the signal processing discussed above may be performed by a mobile station in wireless communication with a base station. For example, the base station may transmit the wideband or narrow-band digital audio signal via conventional means to the mobile station. Once received, signal processing logic within the mobile station performs the requisite operations to generate a bandwidth extended version of the digital audio signal that is clearer and more audibly pleasing to a user of the mobile station.
[0037] Additionally in some embodiments, a voicing level estimator 621 may be used in conjunction with high band excitation generator 605. For example, a voicing level of 0, indicating unvoiced speech, may be used to determine use of noise excitation. Similarly, a voicing level of 1 indicating voiced speech, may be used to determine use of high band excitation derived from transition band excitation as described above. When the voicing level is in between 0 and 1 indicating mixed- voiced speech, various excitations may be mixed in appropriate proportion as determined by the voicing level and used. The noise excitation may be a pseudo random noise function and as described above, may be considered as filling or patching holes in the spectrum based on the voicing level. A mixed high band excitation is thus suitable for voiced, unvoiced, and mixed-voiced sounds.
[0038] FIG. 6 shows the Estimation and Control Logic 500 as comprising transition band MDCT coefficient selector logic 601, transition band analysis logic 603, high band excitation generator 605, high band MDCT coefficient generator 607, transition band shape estimator 609, high band energy predictor 611, high band envelope selector 613, transition band energy estimator 615, energy adapter 617, zero- crossings calculator 619, voicing level estimator 621, onset/plosive detector 623, and SS/Transition detector 625.
[0039] The input 403 provides the decoded wideband speech/audio signal swb, the MDCT coefficients corresponding to at least the transition band, and the pitch frequency (or delay) for each frame. The transition band MDCT selector logic 601 is part of the baseline coder and provides a set of MDCT coefficients for the transition band to the transition band analysis logic 603 and to the transition band energy estimator 615.
[0040] Voicing level estimation: To estimate the voicing level, a zero-crossing calculator 619 may calculate the number of zero-crossings zc in each frame of the wideband speech swb as follows:
[0041] zc = — ^- ∑ \Sgn(swb(n)) - Sgn(swb(n + 1))|
[0042] where
r 1 if s Jn) ≥ O
[0043] Sgn(swb (n)) = \ wb\ \
1 -1 if swb(n) < 0 [0044] where n is the sample index, and N is the frame size in samples. The frame size and percent overlap used in the Estimation and Control Logic 500 are determined by the baseline coder, for example, Ν = 640 at 32 kHz sampling frequency and 50% overlap. The value of the zc parameter calculated as above ranges from 0 to 1. From the zc parameter, a voicing level estimator 621 may estimate the voicing level v as follows.
1 if zc < ZChw
[0045] V = 0 if zc > ZChigh zc - ZC hw
1 - otherwise
ZChigh ~ ZClo
[0046] where, ZCιow and ZChigh represent appropriately chosen low and high thresholds respectively, e.g., ZCιow = 0.125 and ZChigh = 0.30.
[0047] In order to estimate the high band energy, a transition-band energy estimator 615 estimates the transition-band energy from the transition band MDCT coefficients. The transition-band is defined here as a frequency band that is contained within the wideband and close to the high band, i.e., it serves as a transition to the high band, (which, in this illustrative example, is about 7000 - 14,000 Hz). One way to calculate the transition-band energy Etb is to sum the energies of the spectral components, i.e. MDCT coefficients, within the transition-band.
[0048] From the transition-band energy Etb in dB (decibels), the high band energy Ehw in dB is estimated as
Figure imgf000013_0001
+ β
[0050] where, the coefficients a and β are selected to minimize the mean squared error between the true and estimated values of the high band energy over a large number of frames from a training speech/audio database.
[0051] The estimation accuracy can be further enhanced by exploiting contextual information from additional speech parameters such as the zero-crossing parameter zc and the transition-band spectral shape as may be provided by a transition-band shape estimator 609. The zero-crossing parameter, as discussed earlier, is indicative of the speech voicing level. The transition band shape estimator 609 provides a high resolution representation of the transition band envelope shape. For example, a vector quantized representation of the transition band spectral envelope shapes (in dB) may be used. The vector quantizer (VQ) codebook consists of 8 shapes referred to as transition band spectral envelope shape parameters tbs that are computed from a large training database. A corresponding zc-tbs parameter plane may be formed using the zc and tbs parameters to achieve improved performance. As described earlier, the zc-tbs plane is divided into 64 partitions corresponding to 8 scalar quantized levels of zc and the 8 tbs shapes. Some of the partitions may be merged with the nearby partitions for lack of sufficient data points from the training database. For each of the remaining partitions in the zc-tbs plane, separate predictor coefficients are computed.
[0052] The high band energy predictor 611 can provide additional improvement in estimation accuracy by using higher powers of Etb in estimating Ehbo, e.g.,
[0053] Ehw = OCA Etb + &i Etb + on Etb + QTi Ea, + β.
[0054] In this case, five different coefficients, viz., Qr4, or3, a, Qr1, and β, are selected for each partition of the zc-tbs parameter plane. Since the above equations for estimating Ehbo are non-linear, special care must be taken to adjust the estimated high band energy as the input signal level, i.e, energy, changes. One way of achieving this is to estimate the input signal level in dB, adjust Etb up or down to correspond to the nominal signal level, estimate Ehw, and adjust Ehbo down or up to correspond to the actual signal level.
[0055] Estimation of the high band energy is prone to errors. Since over- estimation leads to artifacts, the estimated high band energy is biased to be lower by an amount proportional to the standard deviation of the estimation error of UW That is, the high band energy is adapted in energy adapter 617 as: [0056] Ehbi = Ehb0 - λ σ
[0057] where, Ehbi is the adapted high band energy in dB, Ehw is the estimated high band energy in dB, λ > 0 is a proportionality factor, and σ is the standard deviation of the estimation error in dB. Thus, after determining the estimated high band energy level, the estimated high band energy level is modified based on an estimation accuracy of the estimated high band energy. With reference to FIG. 6, high band energy predictor 611 additionally determines a measure of unreliability in the estimation of the high band energy level and energy adapter 617 biases the estimated high band energy level to be lower by an amount proportional to the measure of unreliability. In one embodiment the measure of unreliability comprises a standard deviation σ of the error in the estimated high band energy level. Other measures of unreliability may as well be employed without departing from the scope of the embodiments.
[0058] By "biasing down" the estimated high band energy, the probability (or number of occurrences) of energy over-estimation is reduced, thereby reducing the number of artifacts. Also, the amount by which the estimated high band energy is reduced is proportional to how good the estimate is - a more reliable (i.e., low σ value) estimate is reduced by a smaller amount than a less reliable estimate. While designing the high band energy predictor 611, the σ value corresponding to each partition of the zc-tbs parameter plane is computed from the training speech database and stored for later use in "biasing down" the estimated high band energy. The σ value of the (<= 64) partitions of the zc-tbs parameter plane, for example, ranges from about 4 dB to about 8 dB with an average value of about 5.9 dB. A suitable value of λ for this high band energy predictor, for example, is 1.2.
[0059] In a prior-art approach, over-estimation of high band energy is handled by using an asymmetric cost function that penalizes over-estimated errors more than under-estimated errors in the design of the high band energy predictor 611. Compared to this prior-art approach, the "bias down" approach described herein has the following advantages: (A) The design of the high band energy predictor 611 is simpler because it is based on the standard symmetric "squared error" cost function; (B) The "bias down" is done explicitly during the operational phase (and not implicitly during the design phase) and therefore the amount of "bias down" can be easily controlled as desired; and (C) The dependence of the amount of "bias down" to the reliability of the estimate is explicit and straightforward (instead of implicitly depending on the specific cost function used during the design phase).
[0060] Besides reducing the artifacts due to energy over-estimation, the "bias down" approach described above has an added benefit for voiced frames - namely that of masking any errors in high band spectral envelope shape estimation and thereby reducing the resultant "noisy" artifacts. However, for unvoiced frames, if the reduction in the estimated high band energy is too high, the bandwidth extended output speech no longer sounds like super wide band speech. To counter this, the estimated high band energy is further adapted in energy adapter 617 depending on its voicing level as
[0061] Ekb2 = Ehbl + (1 -v) S1 + v S2
[0062] where, Ehb2 is the voicing-level adapted high band energy in dB, v is the voicing level ranging from 0 for unvoiced speech to 1 for voiced speech, and S\ and Si (Si > Si) are constants in dB. The choice of Si and Si depends on the value of λ used for the "bias down" and is determined empirically to yield the best-sounding output speech. For example, when λ is chosen as 1.2, S and Si may be chosen as 3.0 and -3.0 respectively. Note that other choices for the value of λ may result in different choices for Si and Si - the values of Si and Si may both be positive or negative or of opposite signs. The increased energy level for unvoiced speech emphasizes such speech in the bandwidth extended output compared to the wideband input and also helps to select a more appropriate spectral envelope shape for such unvoiced segments.
[0063] With reference to FIG. 6, voicing level estimator 621 outputs a voicing level to energy adapter 617 which further modifies the estimated high band energy level based on wideband signal characteristics by further modifying the estimated high band energy level based on a voicing level. The further modifying may comprise reducing the high band energy level for substantially voiced speech and/or increasing the high band energy level for substantially unvoiced speech.
[0064] While the high band energy predictor 611 followed by energy adapter
617 works quite well for most frames, occasionally there are frames for which the high band energy is grossly under- or over-estimated. Some embodiments may therefore provide for such estimation errors and, at least partially, correct them using an energy track smoother logic (not shown) that comprises a smoothing filter. Thus the step of modifying the estimated high band energy level based on the wideband signal characteristics may comprise smoothing the estimated high band energy level (which has been previously modified as described above based on the standard deviation of the estimation σand the voicing level v), essentially reducing an energy difference between consecutive frames.
[0065] For example, the voicing-level adapted high band energy Ehb2 may be smoothed using a 3 -point averaging filter as
[0066] Em = [Ehb2(k-\) + Ehb2(k) + Ehb2(k+l)] I 3
[0067] where, Ehb3 is the smoothed estimate and k is the frame index.
Smoothing reduces the energy difference between consecutive frames, especially when an estimate is an "outlier", that is, the high band energy estimate of a frame is too high or too low compared to the estimates of the neighboring frames. Thus, smoothing helps to reduce the number of artifacts in the output bandwidth extended speech. The 3-point averaging filter introduces a delay of one frame. Other types of filters with or without delay can also be designed for smoothing the energy track.
[0068] The smoothed energy value Ehb3 may be further adapted by energy adapter 617 to obtain the final adapted high band energy estimate Ehb. This adaptation can involve either decreasing or increasing the smoothed energy value based on the ss parameter output by the steady-state/transition detector 625 and/or the d parameter output by the onset/plosive detector 623. Thus, the step of modifying the estimated high band energy level based on the wideband signal characteristics may include the step of modifying the estimated high band energy level (or previously modified estimated high band energy level) based on whether or not a frame is steady- state or transient. This may include reducing the high band energy level for transient frames and/or increasing the high band energy level for steady-state frames, and may further include modifying the estimated high band energy level based on an occurrence of an onset/plosive. By one approach, adapting the high band energy value changes not only the energy level but also the spectral envelope shape since the selection of the high band spectrum may be tied to the estimated energy.
[0069] A frame is defined as a steady-state frame if it has sufficient energy
(that is, it is a speech frame and not a silence frame) and it is close to each of its neighboring frames both in a spectral sense and in terms of energy. Two frames may be considered spectrally close if the Itakura distance between the two frames is below a specified threshold. Other types of spectral distance measures may also be used. Two frames are considered close in terms of energy if the difference in the wideband energies of the two frames is below a specified threshold. Any frame that is not a steady-state frame is considered a transition frame. A steady state frame is able to mask errors in high band energy estimation much better than transient frames. Accordingly, the estimated high band energy of a frame is adapted based on the ss parameter, that is, depending on whether it is a steady-state frame (ss = 1) or transition frame (ss = 0) as
[E hh, + βλ for steady - state frames
[0070] EhbA = \ hK ^ J y J
[ mϊn(Ehb?) - β2 , Ehb2 ) for transition frames
[0071] where, βι > βi ≥ 0, are empirically chosen constants in dB to achieve good output speech quality. The values of β\ and β depend on the choice of the proportionality constant λ used for the "bias down". For example, when λ is chosen as 1.2, δι as 3.0, and δι as -3.0, β\ and βι may be chosen as 1.5 and 6.0 respectively. Notice that in this example we are slightly increasing the estimated high band energy for steady-state frames and decreasing it significantly further for transition frames. Note that other choices for the values of A,
Figure imgf000018_0001
and δ∑ may result in different choices for β\ and β - the values of β\ and β may both be positive or negative or of opposite signs. Further, note that other criteria for identifying steady-state/transition frames may also be used.
[0072] Based on the onset/plosive detector 623 output d, the estimated high band energy level can be adjusted as follows: When d = 1, it indicates that the corresponding frame contains an onset, for example, transition from silence to unvoiced or voiced sound, or a plosive sound. An onset/plosive is detected at the current frame if the wideband energy of the preceding frame is below a certain threshold and the energy difference between the current and preceding frames exceeds another threshold. In another implementation, the transition band energy of the current and preceding frames are used to detect an onset/plosive. Other methods for detecting an onset/plosive may also be employed. An onset/plosive presents a special problem because of the following reasons: A) Estimation of high band energy near onset/plosive is difficult; B) Pre-echo type artifacts may occur in the output speech because of the typical block processing employed; and C) Plosive sounds (e.g., [p], [t], and [k]), after their initial energy burst, have characteristics similar to certain sibilants (e.g., [s], [J"], and [3]) in the wideband but quite different in the high band leading to energy over-estimation and consequent artifacts. High band energy adaptation for an onset/plosive (d = 1) is done as follows:
Figure imgf000019_0001
[0073] Ehb (k) = Ehb 4 (k) - A for k = Kmm + 1,..., Kτ if v(k) > V1
Ehb 4 (k) - A + A τ (k - Kτ ) for k = KT + 1,..., Km if v(k) > V1
[0074] where k is the frame index. For the first Kmm frames starting with the frame (k = 1) at which the onset/plosive is detected, the high band energy is set to the lowest possible value Emm. For example, Emm can be set to -∞ dB or to the energy of the high band spectral envelope shape with the lowest energy. For the subsequent frames (i.e., for the range given by k = Kmm+l to k = KmdiK), energy adaptation is done only as long as the voicing level v(k) of the frame exceeds the threshold F1. Instead of the voicing level parameter, the zero-crossing parameter zc with an appropriate threshold may also be used for this purpose. Whenever the voicing level of a frame within this range becomes less than or equal to F1, the onset energy adaptation is immediately stopped, that is, Ehb(k) is set equal to EhbAK) until the next onset is detected. If the voicing level v(k) is greater than F1, then for k = Kmm + 1 to k = Kτ, the high band energy is decreased by a fixed amount Δ. For k = KT + 1 to k = Kmax, the high band energy is gradually increased from Ehb4(k) - Δ towards EhbAK) by means of the pre-specified sequence Δτik-Kτ) and at k = Imax + 1, Ehb(k) is set equal to Ehb4(k), and this continues until the next onset is detected. Typical values of the parameters used for onset/plosive based energy adaptation, for example, are Kmm = 2, KT = 3, Kmax = 5, Fi = 0.9, Δ = -12 dB, Δτ (1) = 6 dB, and Δτ (2) = 9.5 dB. For d = 0, no further adaptation of the energy is done, that is, Ehb is set equal to £W Thus, the step of modifying the estimated high band energy level based on the wideband signal characteristics may comprise the step of modifying the estimated high band energy level (or previously modified estimated high band energy level) based on an occurrence of an onset/plosive.
[0075] The adaptation of the estimated high band energy as outlined above helps to minimize the number of artifacts in the bandwidth extended output speech and thereby enhance its quality. Although the sequence of operations used to adapt the estimated high band energy has been presented in a particular way, those skilled in the art will recognize that such specificity with respect to sequence is not a requirement, and as such, other sequences may be used and would remain in accordance with the herein disclosed embodiments. Also, the operations described for modifying the high band energy level may selectively be applied in the embodiments.
[0076] Therefore signal processing logic and methods of operation have been disclosed herein for estimating a high band spectral portion, in the range of about 7 to 14 kHz, and determining MDCT coefficients such that an audio output having a spectral portion in the high band may be provided. Other variations that would be equivalent to the herein disclosed embodiments may occur to those of ordinary skill in the art and would remain in accordance with the spirit and scope of embodiments as defined herein by the following claims.

Claims

1. A method comprising: defining a transition band for a signal having a spectrum within a first frequency band, said transition band defined as a portion of said first frequency band, said transition band being located near an adjacent frequency band that is adjacent to said first frequency band; analyzing said transition band to obtain transition band spectral data; and generating an adjacent frequency band signal spectrum using said transition band spectral data.
2. The method of claim 1, wherein generating an adjacent frequency band signal spectrum using said transition band spectral data, comprises: estimating an adjacent frequency band spectral envelope; generating an adjacent frequency band excitation spectrum, using said transition band spectral data; and combining said adjacent frequency band spectral envelope and said adjacent frequency band excitation spectrum to generate said adjacent frequency band signal spectrum.
3. The method of claim 2, wherein analyzing said transition band to obtain transition band spectral data, further comprises: analyzing said transition band to obtain a transition band spectral envelope and a transition band excitation spectrum.
4. The method of claim 3, wherein generating an adjacent frequency band excitation spectrum, using said transition band spectral data, further comprises: generating said adjacent frequency band excitation spectrum by periodic repetition of at least a part of said transition band excitation spectrum with a repetition period determined by a pitch frequency of said signal.
5. The method of claim 2, wherein estimating an adjacent frequency band spectral envelope, further comprises estimating said signal's energy in said adjacent frequency band.
6. The method of claim 2, further comprising combining said spectrum within said first frequency band and said adjacent frequency band signal spectrum to obtain a bandwidth extended signal spectrum and a corresponding bandwidth extended signal.
7. The method of claim 4, wherein generating said adjacent frequency band excitation spectrum, further comprises mixing said adjacent frequency band excitation spectrum generated by periodic repetition of at least a part of said transition band excitation spectrum with a pseudo-noise excitation spectrum within said adjacent frequency band.
8. The method of claim 7, further comprising determining a mixing ratio, for mixing said adjacent frequency band excitation spectrum and said pseudo- noise excitation spectrum, using a voicing level estimated from said signal.
9. The method of claim 8, further comprising Filling any holes in said adjacent frequency band excitation spectrum due to corresponding holes in said transition band excitation spectrum using said pseudo-noise excitation spectrum.
10. A method comprising: defining a transition band for a signal having a spectrum within a first frequency band, said transition band defined as a portion of said first frequency band, said transition band being located near an adjacent frequency band that is adjacent to said first frequency band; analyzing said transition band to obtain a transition band spectral envelope and a transition band excitation spectrum; estimating an adjacent frequency band spectral envelope; generating an adjacent frequency band excitation spectrum by periodic repetition of at least a part of said transition band excitation spectrum with a repetition period determined by a pitch frequency of said signal; and combining said adjacent frequency band spectral envelope and said adjacent frequency band excitation spectrum to obtain an adjacent frequency band signal spectrum.
11. The method of claim 10, wherein estimating an adjacent frequency band spectral envelope, further comprises estimating said signal's energy in said adjacent frequency band.
12. The method of claim 11, further comprising combining said spectrum within said first frequency band and said adjacent frequency band signal spectrum to obtain a bandwidth extended signal spectrum and a corresponding bandwidth extended signal.
13. The method of claim 12, wherein generating said adjacent frequency band excitation spectrum, further comprises mixing said adjacent frequency band excitation spectrum generated by periodic repetition of at least a part of said transition band excitation spectrum with a pseudo-noise excitation spectrum within said adjacent frequency band.
14. The method of claim 11, further comprising determining a mixing ratio, for mixing said adjacent frequency band excitation spectrum and said pseudo- noise excitation spectrum, using a voicing level estimated from said signal.
15. The method of claim 11, further comprising Filling any holes in said adjacent frequency band excitation spectrum due to corresponding holes in said transition band excitation spectrum using said pseudo-noise excitation spectrum.
16. A device comprising: signal processing logic operative to: define a transition band for a signal having a spectrum within a first frequency band, said transition band defined as a portion of said first frequency band, said transition band being located near an adjacent frequency band that is adjacent to said first frequency band;
analyze said transition band to obtain a transition band spectral envelope and a transition band excitation spectrum;
estimate an adjacent frequency band spectral envelope;
generate an adjacent frequency band excitation spectrum by periodic repetition of at least a part of said transition band excitation spectrum with a repetition period determined by a pitch frequency of said signal; and combine said adjacent frequency band spectral envelope and said adjacent frequency band excitation spectrum to obtain an adjacent frequency band signal spectrum.
17. The device of claim 16, wherein said signal processing logic is further operative to estimate said signal's energy in said adjacent frequency band.
18. The device of claim 17, wherein said signal processing logic is further operative to combine said spectrum within said first frequency band and said adjacent frequency band signal spectrum to obtain a bandwidth extended signal spectrum and a corresponding bandwidth extended signal.
19. he device of claim 17, wherein said signal processing logic is further operative to mix said adjacent frequency band excitation spectrum generated by periodic repetition of at least a part of said transition band excitation spectrum with a pseudo-noise excitation spectrum within said adjacent frequency band.
20. The device of claim 19, wherein said signal processing logic is further operative to determine a mixing ratio, for mixing said adjacent frequency band excitation spectrum and said pseudo-noise excitation spectrum, using a voicing level estimated from said signal.
21. The device of claim 20, wherein said signal processing logic is further operative to fill any holes in said adjacent frequency band excitation spectrum due to corresponding holes in said transition band excitation spectrum using said pseudo-noise excitation spectrum.
PCT/US2010/022879 2009-02-04 2010-02-02 Bandwidth extension method and apparatus for a modified discrete cosine transform audio coder WO2010091013A1 (en)

Priority Applications (6)

Application Number Priority Date Filing Date Title
JP2011544700A JP5597896B2 (en) 2009-02-04 2010-02-02 Bandwidth expansion method and apparatus for modified discrete cosine transform speech coder
BRPI1008520A BRPI1008520B1 (en) 2009-02-04 2010-02-02 bandwidth extension device and method
CN201080006565.0A CN102308333B (en) 2009-02-04 2010-02-02 Bandwidth extension method and apparatus for a modified discrete cosine transform audio coder
MX2011007807A MX2011007807A (en) 2009-02-04 2010-02-02 Bandwidth extension method and apparatus for a modified discrete cosine transform audio coder.
KR1020117018182A KR101341246B1 (en) 2009-02-04 2010-02-02 Bandwidth extension method and apparatus for a modified discrete cosine transform audio coder
EP10704446.3A EP2394269B1 (en) 2009-02-04 2010-02-02 Audio bandwidth extension method and device

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US12/365,457 2009-02-04
US12/365,457 US8463599B2 (en) 2009-02-04 2009-02-04 Bandwidth extension method and apparatus for a modified discrete cosine transform audio coder

Publications (1)

Publication Number Publication Date
WO2010091013A1 true WO2010091013A1 (en) 2010-08-12

Family

ID=42101566

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2010/022879 WO2010091013A1 (en) 2009-02-04 2010-02-02 Bandwidth extension method and apparatus for a modified discrete cosine transform audio coder

Country Status (8)

Country Link
US (1) US8463599B2 (en)
EP (1) EP2394269B1 (en)
JP (2) JP5597896B2 (en)
KR (1) KR101341246B1 (en)
CN (1) CN102308333B (en)
BR (1) BRPI1008520B1 (en)
MX (1) MX2011007807A (en)
WO (1) WO2010091013A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102419980A (en) * 2010-09-27 2012-04-18 富士通株式会社 Voice-band extending apparatus and voice-band extending method
CN103971693A (en) * 2013-01-29 2014-08-06 华为技术有限公司 Forecasting method for high-frequency band signal, encoding device and decoding device

Families Citing this family (49)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1569200A1 (en) * 2004-02-26 2005-08-31 Sony International (Europe) GmbH Identification of the presence of speech in digital audio data
US8688441B2 (en) * 2007-11-29 2014-04-01 Motorola Mobility Llc Method and apparatus to facilitate provision and use of an energy value to determine a spectral envelope shape for out-of-signal bandwidth content
US8433582B2 (en) * 2008-02-01 2013-04-30 Motorola Mobility Llc Method and apparatus for estimating high-band energy in a bandwidth extension system
US20090201983A1 (en) * 2008-02-07 2009-08-13 Motorola, Inc. Method and apparatus for estimating high-band energy in a bandwidth extension system
US8463412B2 (en) * 2008-08-21 2013-06-11 Motorola Mobility Llc Method and apparatus to facilitate determining signal bounding frequencies
US9947340B2 (en) * 2008-12-10 2018-04-17 Skype Regeneration of wideband speech
EP2360687A4 (en) * 2008-12-19 2012-07-11 Fujitsu Ltd Voice band extension device and voice band extension method
JP4932917B2 (en) * 2009-04-03 2012-05-16 株式会社エヌ・ティ・ティ・ドコモ Speech decoding apparatus, speech decoding method, and speech decoding program
JP5754899B2 (en) 2009-10-07 2015-07-29 ソニー株式会社 Decoding apparatus and method, and program
WO2011045926A1 (en) * 2009-10-14 2011-04-21 パナソニック株式会社 Encoding device, decoding device, and methods therefor
JPWO2011121955A1 (en) * 2010-03-30 2013-07-04 パナソニック株式会社 Audio equipment
JP5609737B2 (en) 2010-04-13 2014-10-22 ソニー株式会社 Signal processing apparatus and method, encoding apparatus and method, decoding apparatus and method, and program
JP5850216B2 (en) 2010-04-13 2016-02-03 ソニー株式会社 Signal processing apparatus and method, encoding apparatus and method, decoding apparatus and method, and program
US9047875B2 (en) * 2010-07-19 2015-06-02 Futurewei Technologies, Inc. Spectrum flatness control for bandwidth extension
JP2012032713A (en) * 2010-08-02 2012-02-16 Sony Corp Decoding apparatus, decoding method and program
JP6075743B2 (en) 2010-08-03 2017-02-08 ソニー株式会社 Signal processing apparatus and method, and program
WO2012037515A1 (en) 2010-09-17 2012-03-22 Xiph. Org. Methods and systems for adaptive time-frequency resolution in digital data coding
JP5707842B2 (en) 2010-10-15 2015-04-30 ソニー株式会社 Encoding apparatus and method, decoding apparatus and method, and program
EP2674942B1 (en) * 2011-02-08 2017-10-25 LG Electronics Inc. Method and device for audio bandwidth extension
US9015042B2 (en) * 2011-03-07 2015-04-21 Xiph.org Foundation Methods and systems for avoiding partial collapse in multi-block audio coding
US9009036B2 (en) 2011-03-07 2015-04-14 Xiph.org Foundation Methods and systems for bit allocation and partitioning in gain-shape vector quantization for audio coding
WO2012122303A1 (en) 2011-03-07 2012-09-13 Xiph. Org Method and system for two-step spreading for tonal artifact avoidance in audio coding
TWI562133B (en) 2011-05-13 2016-12-11 Samsung Electronics Co Ltd Bit allocating method and non-transitory computer-readable recording medium
US9251800B2 (en) 2011-11-02 2016-02-02 Telefonaktiebolaget L M Ericsson (Publ) Generation of a high band extension of a bandwidth extended audio signal
RU2725416C1 (en) 2012-03-29 2020-07-02 Телефонактиеболагет Лм Эрикссон (Пабл) Broadband of harmonic audio signal
CN103928029B (en) * 2013-01-11 2017-02-08 华为技术有限公司 Audio signal coding method, audio signal decoding method, audio signal coding apparatus, and audio signal decoding apparatus
US9601125B2 (en) * 2013-02-08 2017-03-21 Qualcomm Incorporated Systems and methods of performing noise modulation and gain adjustment
JP6157926B2 (en) * 2013-05-24 2017-07-05 株式会社東芝 Audio processing apparatus, method and program
CN104217727B (en) 2013-05-31 2017-07-21 华为技术有限公司 Signal decoding method and equipment
FR3007563A1 (en) * 2013-06-25 2014-12-26 France Telecom ENHANCED FREQUENCY BAND EXTENSION IN AUDIO FREQUENCY SIGNAL DECODER
FR3008533A1 (en) 2013-07-12 2015-01-16 Orange OPTIMIZED SCALE FACTOR FOR FREQUENCY BAND EXTENSION IN AUDIO FREQUENCY SIGNAL DECODER
CN108364657B (en) * 2013-07-16 2020-10-30 超清编解码有限公司 Method and decoder for processing lost frame
JP6531649B2 (en) 2013-09-19 2019-06-19 ソニー株式会社 Encoding apparatus and method, decoding apparatus and method, and program
CN104517611B (en) 2013-09-26 2016-05-25 华为技术有限公司 A kind of high-frequency excitation signal Forecasting Methodology and device
US10083708B2 (en) * 2013-10-11 2018-09-25 Qualcomm Incorporated Estimation of mixing factors to generate high-band excitation signal
KR101498113B1 (en) * 2013-10-23 2015-03-04 광주과학기술원 A apparatus and method extending bandwidth of sound signal
RU2764260C2 (en) 2013-12-27 2022-01-14 Сони Корпорейшн Decoding device and method
FR3017484A1 (en) * 2014-02-07 2015-08-14 Orange ENHANCED FREQUENCY BAND EXTENSION IN AUDIO FREQUENCY SIGNAL DECODER
BR112016020988B1 (en) 2014-03-14 2022-08-30 Telefonaktiebolaget Lm Ericsson (Publ) METHOD AND ENCODER FOR ENCODING AN AUDIO SIGNAL, AND, COMMUNICATION DEVICE
CN106537500B (en) * 2014-05-01 2019-09-13 日本电信电话株式会社 Periodically comprehensive envelope sequence generator, periodically comprehensive envelope sequence generating method, recording medium
PL3509063T3 (en) * 2014-05-01 2020-08-24 Nippon Telegraph And Telephone Corporation Encoder, decoder, coding method, decoding method, coding program, decoding program and recording medium
JP2016038435A (en) * 2014-08-06 2016-03-22 ソニー株式会社 Encoding device and method, decoding device and method, and program
US9536537B2 (en) 2015-02-27 2017-01-03 Qualcomm Incorporated Systems and methods for speech restoration
US10847170B2 (en) 2015-06-18 2020-11-24 Qualcomm Incorporated Device and method for generating a high-band signal from non-linearly processed sub-ranges
US9837089B2 (en) * 2015-06-18 2017-12-05 Qualcomm Incorporated High-band signal generation
KR20180056032A (en) 2016-11-18 2018-05-28 삼성전자주식회사 Signal processing processor and controlling method thereof
US20190051286A1 (en) * 2017-08-14 2019-02-14 Microsoft Technology Licensing, Llc Normalization of high band signals in network telephony communications
US11100941B2 (en) * 2018-08-21 2021-08-24 Krisp Technologies, Inc. Speech enhancement and noise suppression systems and methods
CN112180762B (en) * 2020-09-29 2021-10-29 瑞声新能源发展(常州)有限公司科教城分公司 Nonlinear signal system construction method, apparatus, device and medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030187663A1 (en) * 2002-03-28 2003-10-02 Truman Michael Mead Broadband frequency translation for high frequency regeneration
EP1367566A2 (en) * 1997-06-10 2003-12-03 Coding Technologies Sweden AB Source coding enhancement using spectral-band replication
US6708145B1 (en) * 1999-01-27 2004-03-16 Coding Technologies Sweden Ab Enhancing perceptual performance of sbr and related hfr coding methods by adaptive noise-floor addition and noise substitution limiting
US20040247037A1 (en) * 2002-08-21 2004-12-09 Hiroyuki Honma Signal encoding device, method, signal decoding device, and method
US20050165611A1 (en) * 2004-01-23 2005-07-28 Microsoft Corporation Efficient coding of digital media spectral data using wide-sense perceptual similarity

Family Cites Families (60)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4771465A (en) * 1986-09-11 1988-09-13 American Telephone And Telegraph Company, At&T Bell Laboratories Digital speech sinusoidal vocoder with transmission of only subset of harmonics
JPH02166198A (en) 1988-12-20 1990-06-26 Asahi Glass Co Ltd Dry cleaning agent
US5765127A (en) * 1992-03-18 1998-06-09 Sony Corp High efficiency encoding method
US5245589A (en) * 1992-03-20 1993-09-14 Abel Jonathan S Method and apparatus for processing signals to extract narrow bandwidth features
JP2779886B2 (en) * 1992-10-05 1998-07-23 日本電信電話株式会社 Wideband audio signal restoration method
US5455888A (en) * 1992-12-04 1995-10-03 Northern Telecom Limited Speech bandwidth extension method and apparatus
JPH07160299A (en) * 1993-12-06 1995-06-23 Hitachi Denshi Ltd Sound signal band compander and band compression transmission system and reproducing system for sound signal
DE69619284T3 (en) * 1995-03-13 2006-04-27 Matsushita Electric Industrial Co., Ltd., Kadoma Device for expanding the voice bandwidth
JP2956548B2 (en) * 1995-10-05 1999-10-04 松下電器産業株式会社 Voice band expansion device
JPH0916198A (en) * 1995-06-27 1997-01-17 Japan Radio Co Ltd Excitation signal generating device and excitation signal generating method in low bit rate vocoder
JP3522954B2 (en) * 1996-03-15 2004-04-26 株式会社東芝 Microphone array input type speech recognition apparatus and method
US5794185A (en) * 1996-06-14 1998-08-11 Motorola, Inc. Method and apparatus for speech coding using ensemble statistics
US5949878A (en) * 1996-06-28 1999-09-07 Transcrypt International, Inc. Method and apparatus for providing voice privacy in electronic communication systems
JPH10124088A (en) * 1996-10-24 1998-05-15 Sony Corp Device and method for expanding voice frequency band width
US6453287B1 (en) * 1999-02-04 2002-09-17 Georgia-Tech Research Corporation Apparatus and quality enhancement algorithm for mixed excitation linear predictive (MELP) and other speech coders
JP2000305599A (en) * 1999-04-22 2000-11-02 Sony Corp Speech synthesizing device and method, telephone device, and program providing media
US7330814B2 (en) * 2000-05-22 2008-02-12 Texas Instruments Incorporated Wideband speech coding with modulated noise highband excitation system and method
SE0001926D0 (en) * 2000-05-23 2000-05-23 Lars Liljeryd Improved spectral translation / folding in the subband domain
DE10041512B4 (en) * 2000-08-24 2005-05-04 Infineon Technologies Ag Method and device for artificially expanding the bandwidth of speech signals
WO2002029782A1 (en) * 2000-10-02 2002-04-11 The Regents Of The University Of California Perceptual harmonic cepstral coefficients as the front-end for speech recognition
US6990446B1 (en) * 2000-10-10 2006-01-24 Microsoft Corporation Method and apparatus using spectral addition for speaker recognition
US6889182B2 (en) * 2001-01-12 2005-05-03 Telefonaktiebolaget L M Ericsson (Publ) Speech bandwidth extension
ATE319162T1 (en) * 2001-01-19 2006-03-15 Koninkl Philips Electronics Nv BROADBAND SIGNAL TRANSMISSION SYSTEM
SE522553C2 (en) * 2001-04-23 2004-02-17 Ericsson Telefon Ab L M Bandwidth extension of acoustic signals
US6895375B2 (en) * 2001-10-04 2005-05-17 At&T Corp. System for bandwidth extension of Narrow-band speech
US6988066B2 (en) * 2001-10-04 2006-01-17 At&T Corp. Method of bandwidth extension for narrow-band speech
JP3579047B2 (en) 2002-07-19 2004-10-20 日本電気株式会社 Audio decoding device, decoding method, and program
KR100917464B1 (en) * 2003-03-07 2009-09-14 삼성전자주식회사 Method and apparatus for encoding/decoding digital data using bandwidth extension technology
US20050004793A1 (en) * 2003-07-03 2005-01-06 Pasi Ojala Signal adaptation for higher band coding in a codec utilizing band split coding
US20050065784A1 (en) * 2003-07-31 2005-03-24 Mcaulay Robert J. Modification of acoustic signals using sinusoidal analysis and synthesis
DE502004003788D1 (en) * 2003-09-03 2007-06-21 Phoenix Conveyor Belt Sys Gmbh DEVICE FOR MONITORING A CONVEYOR SYSTEM
US7461003B1 (en) * 2003-10-22 2008-12-02 Tellabs Operations, Inc. Methods and apparatus for improving the quality of speech signals
JP2005136647A (en) * 2003-10-30 2005-05-26 New Japan Radio Co Ltd Bass booster circuit
KR100587953B1 (en) * 2003-12-26 2006-06-08 한국전자통신연구원 Packet loss concealment apparatus for high-band in split-band wideband speech codec, and system for decoding bit-stream using the same
CA2454296A1 (en) * 2003-12-29 2005-06-29 Nokia Corporation Method and device for speech enhancement in the presence of background noise
DE602004020765D1 (en) * 2004-09-17 2009-06-04 Harman Becker Automotive Sys Bandwidth extension of band-limited tone signals
KR100708121B1 (en) 2005-01-22 2007-04-16 삼성전자주식회사 Method and apparatus for bandwidth extension of speech
RU2376657C2 (en) * 2005-04-01 2009-12-20 Квэлкомм Инкорпорейтед Systems, methods and apparatus for highband time warping
US20060224381A1 (en) * 2005-04-04 2006-10-05 Nokia Corporation Detecting speech frames belonging to a low energy sequence
US7813931B2 (en) * 2005-04-20 2010-10-12 QNX Software Systems, Co. System for improving speech quality and intelligibility with bandwidth compression/expansion
US8249861B2 (en) * 2005-04-20 2012-08-21 Qnx Software Systems Limited High frequency compression integration
TWI317933B (en) * 2005-04-22 2009-12-01 Qualcomm Inc Methods, data storage medium,apparatus of signal processing,and cellular telephone including the same
US8311840B2 (en) * 2005-06-28 2012-11-13 Qnx Software Systems Limited Frequency extension of harmonic signals
KR101171098B1 (en) * 2005-07-22 2012-08-20 삼성전자주식회사 Scalable speech coding/decoding methods and apparatus using mixed structure
US7953605B2 (en) * 2005-10-07 2011-05-31 Deepen Sinha Method and apparatus for audio encoding and decoding using wideband psychoacoustic modeling and bandwidth extension
EP1772855B1 (en) * 2005-10-07 2013-09-18 Nuance Communications, Inc. Method for extending the spectral bandwidth of a speech signal
US7490036B2 (en) * 2005-10-20 2009-02-10 Motorola, Inc. Adaptive equalizer for a coded speech signal
US20070109977A1 (en) * 2005-11-14 2007-05-17 Udar Mittal Method and apparatus for improving listener differentiation of talkers during a conference call
US7546237B2 (en) * 2005-12-23 2009-06-09 Qnx Software Systems (Wavemakers), Inc. Bandwidth extension of narrowband speech
US7835904B2 (en) * 2006-03-03 2010-11-16 Microsoft Corp. Perceptual, scalable audio compression
US7844453B2 (en) * 2006-05-12 2010-11-30 Qnx Software Systems Co. Robust noise estimation
US20080004866A1 (en) * 2006-06-30 2008-01-03 Nokia Corporation Artificial Bandwidth Expansion Method For A Multichannel Signal
US8260609B2 (en) * 2006-07-31 2012-09-04 Qualcomm Incorporated Systems, methods, and apparatus for wideband encoding and decoding of inactive frames
ATE446572T1 (en) 2006-08-22 2009-11-15 Harman Becker Automotive Sys METHOD AND SYSTEM FOR PROVIDING AN EXTENDED BANDWIDTH AUDIO SIGNAL
US8639500B2 (en) * 2006-11-17 2014-01-28 Samsung Electronics Co., Ltd. Method, medium, and apparatus with bandwidth extension encoding and/or decoding
US8229106B2 (en) * 2007-01-22 2012-07-24 D.S.P. Group, Ltd. Apparatus and methods for enhancement of speech
US8688441B2 (en) 2007-11-29 2014-04-01 Motorola Mobility Llc Method and apparatus to facilitate provision and use of an energy value to determine a spectral envelope shape for out-of-signal bandwidth content
US8433582B2 (en) * 2008-02-01 2013-04-30 Motorola Mobility Llc Method and apparatus for estimating high-band energy in a bandwidth extension system
US20090201983A1 (en) * 2008-02-07 2009-08-13 Motorola, Inc. Method and apparatus for estimating high-band energy in a bandwidth extension system
US8463412B2 (en) * 2008-08-21 2013-06-11 Motorola Mobility Llc Method and apparatus to facilitate determining signal bounding frequencies

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1367566A2 (en) * 1997-06-10 2003-12-03 Coding Technologies Sweden AB Source coding enhancement using spectral-band replication
US6708145B1 (en) * 1999-01-27 2004-03-16 Coding Technologies Sweden Ab Enhancing perceptual performance of sbr and related hfr coding methods by adaptive noise-floor addition and noise substitution limiting
US20030187663A1 (en) * 2002-03-28 2003-10-02 Truman Michael Mead Broadband frequency translation for high frequency regeneration
US20040247037A1 (en) * 2002-08-21 2004-12-09 Hiroyuki Honma Signal encoding device, method, signal decoding device, and method
US20050165611A1 (en) * 2004-01-23 2005-07-28 Microsoft Corporation Efficient coding of digital media spectral data using wide-sense perceptual similarity

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
EPPS AND W H HOLMES J: "SPEECH ENHANCEMENT USING STC-BASED BANDWIDTH EXTENSION", 19981001, 1 October 1998 (1998-10-01), pages P711, XP007000515 *

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102419980A (en) * 2010-09-27 2012-04-18 富士通株式会社 Voice-band extending apparatus and voice-band extending method
CN103971693A (en) * 2013-01-29 2014-08-06 华为技术有限公司 Forecasting method for high-frequency band signal, encoding device and decoding device
EP2937861A4 (en) * 2013-01-29 2016-08-03 Huawei Tech Co Ltd Prediction method and coding/decoding device for high frequency band signal
CN106847297A (en) * 2013-01-29 2017-06-13 华为技术有限公司 The Forecasting Methodology of high-frequency band signals, coding/decoding apparatus
US9704500B2 (en) 2013-01-29 2017-07-11 Huawei Technologies Co., Ltd. Method for predicting high frequency band signal, encoding device, and decoding device
KR101837191B1 (en) 2013-01-29 2018-03-09 후아웨이 테크놀러지 컴퍼니 리미티드 Prediction method and coding/decoding device for high frequency band signal
KR20180026812A (en) * 2013-01-29 2018-03-13 후아웨이 테크놀러지 컴퍼니 리미티드 Prediction method and coding/decoding device for high frequency band signal
US10089997B2 (en) 2013-01-29 2018-10-02 Huawei Technologies Co.,Ltd. Method for predicting high frequency band signal, encoding device, and decoding device
KR101980057B1 (en) 2013-01-29 2019-05-17 후아웨이 테크놀러지 컴퍼니 리미티드 Prediction method and coding/decoding device for high frequency band signal
US10636432B2 (en) 2013-01-29 2020-04-28 Huawei Technologies Co., Ltd. Method for predicting high frequency band signal, encoding device, and decoding device
CN106847297B (en) * 2013-01-29 2020-07-07 华为技术有限公司 Prediction method of high-frequency band signal, encoding/decoding device
EP3779980A3 (en) * 2013-01-29 2021-07-07 Huawei Technologies Co., Ltd. Method for predicting high frequency band signal, encoding device, and decoding device

Also Published As

Publication number Publication date
JP2014016622A (en) 2014-01-30
BRPI1008520B1 (en) 2020-05-05
US8463599B2 (en) 2013-06-11
BRPI1008520A2 (en) 2016-03-08
KR20110111463A (en) 2011-10-11
JP2012514763A (en) 2012-06-28
KR101341246B1 (en) 2013-12-12
MX2011007807A (en) 2011-09-21
EP2394269A1 (en) 2011-12-14
CN102308333B (en) 2014-03-19
CN102308333A (en) 2012-01-04
US20100198587A1 (en) 2010-08-05
JP5597896B2 (en) 2014-10-01
EP2394269B1 (en) 2017-04-05

Similar Documents

Publication Publication Date Title
EP2394269B1 (en) Audio bandwidth extension method and device
US10885926B2 (en) Classification between time-domain coding and frequency domain coding for high bit rates
US8942988B2 (en) Efficient temporal envelope coding approach by prediction between low band signal and high band signal
US9653088B2 (en) Systems, methods, and apparatus for signal encoding using pitch-regularizing and non-pitch-regularizing coding
RU2389085C2 (en) Method and device for introducing low-frequency emphasis when compressing sound based on acelp/tcx
JP5722437B2 (en) Method, apparatus, and computer readable storage medium for wideband speech coding
US20070147518A1 (en) Methods and devices for low-frequency emphasis during audio compression based on ACELP/TCX
RU2756435C2 (en) Optimized scale coefficient for expanding frequency range in audio frequency signal decoder
US20100070270A1 (en) CELP Post-processing for Music Signals
US9899032B2 (en) Systems and methods of performing gain adjustment
KR102426029B1 (en) Improved frequency band extension in an audio signal decoder
US8909539B2 (en) Method and device for extending bandwidth of speech signal
EP3055861B1 (en) Estimation of mixing factors to generate high-band excitation signal
US20150332706A1 (en) Apparatus and method for generating a frequency enhanced signal using shaping of the enhancement signal
US20140019125A1 (en) Low band bandwidth extended
Atti et al. Super-wideband bandwidth extension for speech in the 3GPP EVS codec
WO2023147650A1 (en) Time-domain superwideband bandwidth expansion for cross-talk scenarios

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 201080006565.0

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 10704446

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 4545/CHENP/2011

Country of ref document: IN

REEP Request for entry into the european phase

Ref document number: 2010704446

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2010704446

Country of ref document: EP

ENP Entry into the national phase

Ref document number: 2011544700

Country of ref document: JP

Kind code of ref document: A

WWE Wipo information: entry into national phase

Ref document number: MX/A/2011/007807

Country of ref document: MX

ENP Entry into the national phase

Ref document number: 20117018182

Country of ref document: KR

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

REG Reference to national code

Ref country code: BR

Ref legal event code: B01A

Ref document number: PI1008520

Country of ref document: BR

ENP Entry into the national phase

Ref document number: PI1008520

Country of ref document: BR

Kind code of ref document: A2

Effective date: 20110804