WO2001002929A2 - Coded domain noise control - Google Patents
Coded domain noise control Download PDFInfo
- Publication number
- WO2001002929A2 WO2001002929A2 PCT/US2000/018165 US0018165W WO0102929A2 WO 2001002929 A2 WO2001002929 A2 WO 2001002929A2 US 0018165 W US0018165 W US 0018165W WO 0102929 A2 WO0102929 A2 WO 0102929A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- parameter
- adjusted
- gain
- codebook
- generate
- Prior art date
Links
- 230000006835 compression Effects 0.000 claims abstract description 59
- 238000007906 compression Methods 0.000 claims abstract description 59
- 230000005236 sound signal Effects 0.000 claims abstract description 21
- 238000004891 communication Methods 0.000 claims abstract description 19
- 238000000034 method Methods 0.000 claims description 171
- 239000013598 vector Substances 0.000 claims description 105
- 230000015572 biosynthetic process Effects 0.000 claims description 53
- 238000003786 synthesis reaction Methods 0.000 claims description 53
- 230000005284 excitation Effects 0.000 claims description 41
- 230000003595 spectral effect Effects 0.000 claims description 38
- 230000004044 response Effects 0.000 claims description 33
- 238000012545 processing Methods 0.000 claims description 28
- 239000000872 buffer Substances 0.000 claims description 16
- 238000001914 filtration Methods 0.000 claims description 10
- 230000007774 longterm Effects 0.000 claims description 6
- 238000010295 mobile communication Methods 0.000 claims 2
- 238000013139 quantization Methods 0.000 description 115
- 230000009467 reduction Effects 0.000 description 52
- 230000006870 function Effects 0.000 description 39
- 238000010586 diagram Methods 0.000 description 35
- 230000008569 process Effects 0.000 description 33
- 230000005540 biological transmission Effects 0.000 description 23
- 238000013459 approach Methods 0.000 description 22
- 230000000694 effects Effects 0.000 description 18
- 230000001413 cellular effect Effects 0.000 description 17
- 230000003139 buffering effect Effects 0.000 description 16
- 230000036961 partial effect Effects 0.000 description 15
- 238000012937 correction Methods 0.000 description 14
- 238000012986 modification Methods 0.000 description 14
- 230000004048 modification Effects 0.000 description 14
- 230000003044 adaptive effect Effects 0.000 description 13
- 238000005259 measurement Methods 0.000 description 10
- 238000012546 transfer Methods 0.000 description 10
- 238000005457 optimization Methods 0.000 description 9
- 230000002829 reductive effect Effects 0.000 description 8
- 230000001629 suppression Effects 0.000 description 7
- 230000002238 attenuated effect Effects 0.000 description 6
- 238000001514 detection method Methods 0.000 description 6
- 230000003534 oscillatory effect Effects 0.000 description 6
- 238000012935 Averaging Methods 0.000 description 5
- 238000010606 normalization Methods 0.000 description 5
- IVEKVTHFAJJKGA-BQBZGAKWSA-N (2s)-2-amino-5-[[(2r)-1-ethoxy-1-oxo-3-sulfanylpropan-2-yl]amino]-5-oxopentanoic acid Chemical compound CCOC(=O)[C@H](CS)NC(=O)CC[C@H](N)C(O)=O IVEKVTHFAJJKGA-BQBZGAKWSA-N 0.000 description 4
- 230000006978 adaptation Effects 0.000 description 4
- 230000006399 behavior Effects 0.000 description 4
- 230000008901 benefit Effects 0.000 description 4
- 230000008859 change Effects 0.000 description 4
- 238000009432 framing Methods 0.000 description 4
- 230000002265 prevention Effects 0.000 description 4
- 238000001228 spectrum Methods 0.000 description 4
- 230000009466 transformation Effects 0.000 description 4
- 230000001934 delay Effects 0.000 description 3
- 230000015654 memory Effects 0.000 description 3
- 230000003278 mimic effect Effects 0.000 description 3
- 239000000654 additive Substances 0.000 description 2
- 230000000996 additive effect Effects 0.000 description 2
- 238000005311 autocorrelation function Methods 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 2
- 230000003831 deregulation Effects 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000000670 limiting effect Effects 0.000 description 2
- 239000000203 mixture Substances 0.000 description 2
- 238000005070 sampling Methods 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 230000003321 amplification Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 239000004020 conductor Substances 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 230000006735 deficit Effects 0.000 description 1
- 230000003111 delayed effect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000008030 elimination Effects 0.000 description 1
- 238000003379 elimination reaction Methods 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 239000002360 explosive Substances 0.000 description 1
- 231100001261 hazardous Toxicity 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 238000012423 maintenance Methods 0.000 description 1
- 238000002715 modification method Methods 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 238000003199 nucleic acid amplification method Methods 0.000 description 1
- 230000002250 progressing effect Effects 0.000 description 1
- 230000000717 retained effect Effects 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000002459 sustained effect Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 238000000844 transformation Methods 0.000 description 1
- 230000001131 transforming effect Effects 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
- 239000002699 waste material Substances 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M9/00—Arrangements for interconnection not involving centralised switching
- H04M9/08—Two-way loud-speaking telephone systems with means for conditioning the signal, e.g. for suppressing echoes for one or both directions of traffic
- H04M9/082—Two-way loud-speaking telephone systems with means for conditioning the signal, e.g. for suppressing echoes for one or both directions of traffic using echo cancellers
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0316—Speech enhancement, e.g. noise reduction or echo cancellation by changing the amplitude
- G10L21/0364—Speech enhancement, e.g. noise reduction or echo cancellation by changing the amplitude for improving intelligibility
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04B—TRANSMISSION
- H04B3/00—Line transmission systems
- H04B3/02—Details
- H04B3/20—Reducing echo effects or singing; Opening or closing transmitting path; Conditioning for transmission in one direction or the other
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L1/00—Arrangements for detecting or preventing errors in the information received
- H04L1/0001—Systems modifying transmission characteristics according to link quality, e.g. power backoff
- H04L1/0014—Systems modifying transmission characteristics according to link quality, e.g. power backoff by adapting the source coding
Definitions
- the present invention relates to coded domain enhancement of compressed speech and in particular to coded domain noise contol
- GSM 06 Digital cellular telecommunication system (Phase 2), Full rate speech, Part 2 Transcoding", ETS 300 580-2, March 1998
- GSM 06 60 "Digital cellular telecommunications system (Phase 2), Enhanced Full Rate (EFR) speech transcoding", June 1998
- GSM 08 62 "Digital cellular telecommunications system (Phase 2+), Inband Tandem Free Operation (TFO) of Speech Codecs", ETSI, March 2000
- Network enhancement of coded speech would normally require decoding, linear processing and re-encoding of the processed signal. Such a method is very expensive. Moreover, the encoding process is often an order of magnitude more computationally intensive than the speech enhancement methods.
- Speech compression is increasingly used in telecommunications, especially in cellular telephony and voice over packet networks.
- Past network speech enhancement techniques which operate in the linear domain have several shortcomings. For example, past network speech enhancement techniques which operate in the linear domain require decoding of compressed speech, performing the necessary enhancements and re-encoding of the speech. This processing can be computationally intensive, is especially prone to additional quantization noise, and can cause additional delay.
- PSTN Public Switched Telephone Network
- Telephony customers expect a comfortable listening level to maximize comprehension of their conversation.
- the transmitted speech level from a telephone instrument depends on the speaker's volume and the position of the speaker relative to the microphone. If volume control is available on the telephone instrument, the listener could manually adjust it to a desirable level. However, for historical reasons, most telephone instruments do not have volume controls. Also, direct volume control by the listener does not address the need to maintain appropriate levels for network equipment. Furthermore, as technology is progressing towards the era of hands-free telephony especially in the case of mobile phones in vehicles, manual adjustment is considered cumbersome and potentially hazardous to the vehicle operators.
- Speech transmission between the mobile stations (handsets) and the base station is in compressed or coded form.
- Speech coding techniques such as the GSM FR [1] and EFR [2] are used to compress the speech.
- the devices used to compress speech are called vocoders.
- the coded speech requires less than 2 bits per sample. This situation is depicted in Figure 1. Between the base stations, the speech is transmitted in an uncoded form (using PCM companding which requires 8 bits per sample).
- coded speech and uncoded speech are defined as follows:
- Uncoded speech refers to the digital speech signal samples typically used in telephony; these samples are either in linear 13-bits per sample form or companded
- Coded speech refers to the compressed speech signal parameters (also referred to as coded parameters) which use a bit rate typically well below 64kbps such as 13 kbps in the case of the GSM FR and 12.2 kbps in the case of GSM EFR; the compression methods are more extensive than the simple PCM companding scheme; examples of compression methods are linear predictive coding, code-excited linear prediction and multi-band excitation coding. Tandem-Free Operation (TFO) in GSM
- TFO Tandem-Free Operation
- the TFO standard applies to mobile-to- mobile calls.
- the speech signal is conveyed between mobiles in a compressed form after a brief negotiation period.
- the elimination of tandem codecs is known to improve speech quality in the case where the original signal is clean.
- the key point to note is that the speech transmission remains coded between the mobile handsets and is depicted in Figure 2.
- Environmental background noise is a major impairment that affects telephony applications. Such additive noise can be especially severe in the case of cellular telephones operated in noisy environments. Telephony service providers use noise reduction equipment in their networks to improve the quality of speech so as to encourage longer talk times and increase customer satisfaction. Although noise could be handled at the source in the case of digital cellular handsets, few handset models provide such features due to cost and power limitations. Where such features are provided, they may still not meet the service provider's requirements. Hence service providers consider network speech enhancement equipment to be essential for their competitiveness in the face of deregulation and greater customer expectations. The explosive increase in the use of cellular telephones, which are often operated in the presence of severe background noise conditions, has also increased the use of noise reduction equipment in the network.
- the noisy signal is decomposed into different frequency bands, e.g. using the discrete Fourier transform.
- a silence detector is used to demarcate gaps in speech. During such silence segments, the noise spectrum (i.e. the noise power in each frequency band) is estimated. At all times, the noisy signal power in each frequency band is also estimated. These power estimates provide information such as the signal-to-noise ratio in each frequency band during the time of measurement. Based on these power estimates, the magnitude of each frequency component is attenuated. The phase information is not changed. The resulting magnitude and phase information are recombined. Using the inverse discrete Fourier transform, a noise-reduced signal is reconstructed.
- the delay introduced by the decoding and re-encoding processes is undesirable.
- a vocoder tandem i.e. two encoder/decoder pairs placed in series
- the proposed techniques are capable of performing noise reduction directly on the coded speech (i.e. by direct modification of the coded parameters). Low computational complexity and delay are achieved. Tandeming effects are avoided or minimized, resulting in better perceived quality after noise reduction.
- Speech compression which falls under the category of lossy source coding, is commonly referred to as speech coding.
- Speech coding is performed to minimize the bandwidth necessary for speech transmission. This is especially important in wireless telephony where bandwidth is scarce. In the relatively bandwidth abundant packet networks, speech coding is still important to minimize network delay and jitter. This is because speech communication, unlike data, is highly intolerant of delay. Hence a smaller packet size eases the transmission through a packet network.
- Table 1 The four ETSI GSM standards of concern are listed in Table 1.
- a set of consecutive digital speech samples is referred to as a speech frame.
- the GSM coders operate on a frame size of 20ms (160 samples at 8kHz sampling rate). Given a speech frame, a speech encoder determines a small set of parameters for a speech synthesis model. With these speech parameters and the speech synthesis model, a speech frame can be reconstructed that appears and sounds very similar to the original speech frame. The reconstruction is performed by the speech decoder. In the GSM vocoders listed above, the encoding process is much more computationally intensive than the decoding process.
- the speech parameters determined by the speech encoder depend on the speech synthesis model used.
- the GSM coders in Table 1 utilize linear predictive coding (LPC) models.
- LPC linear predictive coding
- a block diagram of a simplified view of a generic LPC speech synthesis model is shown in Figure 6. This model can be used to generate speech-like signals by specifying the model parameters appropriately.
- the parameters include the time-varying filter coefficients, pitch periods, codebook vectors and the gain factors.
- the synthetic speech is generated as follows.
- An appropriate codebook vector, c(n) is first scaled by the codebook gain
- n denotes sample time.
- a pitch synthesis filter whose parameters include the pitch gain, g p , and the pitch
- T The result is sometimes referred to as the total excitation vector, u(n) .
- the pitch synthesis filter provides the harmonic quality of voiced speech.
- the total excitation vector is then filtered by the LPC synthesis filter which specifies the broad spectral shape of the speech frame.
- the parameters are usually updated more than once.
- the codebook vector, codebook gain and the pitch synthesis filter parameters are determined every subframe (5ms).
- the LPC synthesis filter parameters are determined twice per frame (every 10ms) in EFR and once per frame in FR.
- a window e.g. Hamming window
- the GSM FR vocoder As an example of the arrangement of coded parameters in the bit-stream transmitted by the encoder, the GSM FR vocoder is considered.
- a frame is defined as 160 samples of speech sampled at 8kHz, i.e. a frame is 20ms long. With A-law PCM companding, 160 samples would require 1280 bits for transmission.
- the encoder compresses the 160 samples into 260 bits.
- the arrangement of the various coded parameters in the 260 bits of each frame is shown in Figure 7.
- the first 36 bits of each coded frame consists of the log-area ratios which correspond to LPC synthesis filter.
- the remaining 224 bits can be grouped into 4 subframes of 56 bits each. Within each subframe, the coded parameter bits contain the pitch synthesis filter related parameters followed by the codebook vector and gain related parameters. Speech Synthesis Transfer Function and Typical Coded Parameters
- the codebook vector, c(n) is filtered by H(z) to result in the synthesized
- LPC-based vocoders use parameters similar to the above set, parameters that may be converted to the above forms, or parameters that are related to the above forms.
- the LPC coefficients in LPC-based vocoders may be represented using log-area ratios (e.g. the GSM FR) or line spectral frequencies (e.g. GSM EFR); both of these forms can be converted to LPC coefficients.
- An example of a case where a parameter is related to the above form is the block maximum parameter in the GSM FR vocoder; the block maximum can be considered to be directly proportional to the codebook gain in the model described by equation (1A).
- coded parameter modification methods is mostly limited to the generic speech decoder model, it is relatively straightforward to tailor these methods for any LPC-based vocoder, and possibly even other models.
- the invention is useful in a communication system for transmitting digital signals using a compression code comprising a predetermined plurality of parameters including a first parameter.
- the parameters represent an audio signal having a plurality of audio characteristics including a noise characteristic.
- the compression code is decodable by a plurality of decoding steps.
- the noise characteristic can be managed by reading at least the first parameter, and by generating an adjusted first parameter in response to the compression code and the first parameter.
- the first parameter is replaced with the adjusted first parameter.
- the reading, generating and replacing are preferably performed by a processor.
- the invention also is useful in a communication system for transmitting digital signals comprising code samples further comprising first bits using a compression code and second bits using a linear code.
- the code samples represent an audio signal having a plurality of audio characteristics including a noise characteristic.
- the noise characteristic can be managed without decoding the compression code by adjusting the first bits and second bits in response to the second bits.
- Figure 1 is a schematic block diagram of a system for speech transmission in a GSM digital cellular network.
- FIG. 2 is a schematic block diagram of a system for speech transmission in a GSM network under tandem-free operation (TFO).
- FIG. 3 is a graph illustrating transmission of speech under tandem-free operation (TFO).
- Figure 4 is a schematic block diagram of a traditional noise reduction approach using spectral subtraction.
- Figure 5 is a schematic block diagram illustrating noise reduction of coded speech using a traditional approach.
- Figure 6 is a schematic block diagram of a generic LPC speech synthesis model or speech decoder model.
- Figure 7 is a schematic block diagram illustrating an arrangement of coded parameters in a bit-stream for GSM FR.
- Figure 8 is a schematic block diagram distinguishing coded domain digital speech parameters from linear domain digital speech samples.
- Figure 9 is a graph illustrating GSM full rate codec quantization levels for block maxima.
- Figure 10a is a schematic block diagram of a backward adaptive standard deviation based quantizer.
- Figure 10b is a schematic block diagram of a backward adaptive differential based quantizer.
- Figure 11 is a schematic block diagram of an adaptive differential quantizer using a linear predictor.
- Figure 12 is a schematic block diagram of a GSM enhanced full rate codebook gain (speech level related parameter) quantizer.
- Figure 13 is a graph illustrating GSM enhanced full rate codec quantization levels for a gain correction factor.
- Figure 14 is a schematic block diagram of one technique for coded domain ALC.
- Figure 15 is a flow diagram illustrating a technique for overflow/underflow prevention.
- Figure 16 is a schematic block diagram of a preferred form of ALC system using feedback of the realized gain in ALC algorithms requiring past gain values.
- Figure 17 is a schematic block diagram of one form of a coded domain ALC device.
- Figure 18 is a schematic block diagram of a system for instantaneous scalar requantization for a GSM FR codec.
- Figure 19 is a schematic block diagram of a system for differential scalar requantization for a GSM EFR codec.
- Figure 20a is a graph showing a step in desired gain.
- Figure 20b is a graph showing actual realized gain superimposed on the desired gain with a quantizer in the feedback loop.
- Figure 20c is a graph showing actual realized gain superimposed on the desired gain resulting from placing a quantizer outside the feedback loop shown in
- Figure 21 is a schematic block diagram of an ALC device showing a quantizer placed outside the feedback loop.
- Figure 22 is a schematic block diagram of a simplified version of the ALC device shown in Figure 21.
- Figure 23a is a schematic block diagram of a coded domain ALC implementation for ALC algorithms using feedback of past gain values with a quantizer in the feedback loop.
- Figure 23b is a schematic block diagram of a coded domain ALC implementation for ALC algorithms using feedback of past gain values with a quantizer outside the feedback loop.
- Figure 24 is a graph showing spacing between adjacent Rj values in an EFR codec, and more specifically showing EFR Codec SLRPs: (R, + 1 - R, ) against i.
- Figure 25a is a diagram of a compressed speech frame of an EFR encoder illustrating the times at which various bits are received and the earliest possible decoding of samples as a buffer is filled from left to right.
- Figure 25b is a diagram of a compressed speech frame of an FR encoder illustrating the times at which various bits are received and the earliest possible decoding of samples as a buffer is filled from left to right.
- Figure 26 is a schematic block diagram illustrating a single-band linear domain noise reduction technique.
- Figure 27 is a schematic block diagram of a differential scalar quantization technique.
- Figure 28 is a schematic block diagram of a system for differential requentization of a differentially quantized parameter.
- Figure 29 is a graph illustrating reverberations caused by differential quantization.
- Figure 30 is a schematic block diagram of a system for reverberation-free differential requantization.
- Figure 31 is a simplified schematic block diagram of a simplified reverberation-free differential requantization system.
- Figure 32 is schematic block diagram of a dual-source view of speech synthesis.
- Figure 33 is a schematic block diagram of a preferred form of network noise reduction.
- Figure 34 is a graph illustrating magnitude frequency response of comb filters.
- Figure 35 is a graph illustrating increase in spectral peakresponse of a comb filter due to pitch gain control.
- Figure 36 is a schematic block diagram of one preferred form of a coded domain noise reduction system using codebook gain attenuation.
- Figure 37 is a flow diagram of a preferred form of coded domain noise reduction methodology according to the invention.
- Figure 38 is a schematic block diagram of a system for coded domain noise reduction by modification of the codebook vector parameter.
- Figure 39 is a graph illustrating a spectral interpretation of line spectral frequencies.
- speech signals are digitally sampled prior to transmission.
- Such digital (i.e. discrete-time discrete- valued) signals are herein referred to in this specification as being in the linear domain.
- the adjustment of the speech levels in such linear domain signals is accomplished by multiplying every sample of the signal by an appropriate gain factor to attain the desired target speech level.
- Digital speech signals that are typically carried in telephony networks usually undergo a basic form of compression such as pulse code modulation (PCM) before transmission.
- PCM pulse code modulation
- Such compression schemes are very inexpensive in terms of computations and delay. It is a relatively simple matter for an ALC or NR device to convert the compressed digital samples to the linear domain, process the linear samples, and then compress the processed samples before transmission. As such, these signals can effectively be considered to be in the linear domain.
- compressed or coded speech will refer to speech that is compressed using advanced compression techniques that require significant computational complexity.
- linear code and compression code have the following meanings:
- Linear code By a linear code, we mean a compression technique that results in one coded parameter or coded sample for each sample of the audio signal. Examples of linear codes are PCM (A-law and ⁇ -law) ADPCM (adaptive differential pulse code modulation), and delta modulation.
- PCM A-law and ⁇ -law
- ADPCM adaptive differential pulse code modulation
- Compression code By a compression code, we mean a technique that results in fewer than one coded parameter for each sample of the audio signal. Typically, compression codes result in a small set of coded parameters for each block or frame of audio signal samples. Examples of compression codes are linear predictive coding based vocoders such as the GSM vocoders (HR, FR, EFR).
- Speech compression which falls under the category of lossy source coding, is commonly referred to as speech coding.
- Speech coding is performed to minimize the bandwidth necessary for speech transmission. This is especially important in wireless telephony where bandwidth is a scarce resource.
- speech coding is still important to minimize network delay and jitter. This is because speech communication, unlike data, is highly intolerant of delay. Hence a smaller packet size eases the transmission through a packet network.
- Several industry standard speech codecs coder-decoder pairs
- Table 1 for reference.
- speech coding a set of consecutive digital speech samples is referred to as a speech frame. Given a speech frame, a speech encoder determines a small set of parameters for a speech synthesis model.
- a speech frame can be reconstructed that appears and sounds very similar to the original speech frame.
- the reconstruction is performed by the speech decoder.
- MIPs instructions per second
- the speech parameters determined by the speech encoder depend on the speech synthesis model used.
- the coders in Table 1 utilize linear predictive coding (LPC) models. (To be more specific, these coders belong to the class of code-excited linear prediction or CELP coders.)
- LPC linear predictive coding
- FIG. 6 A block diagram of a simplified view of the LPC speech synthesis model is shown in Figure 6. This model can be used to generate speech-like signals by specifying the model parameters appropriately.
- the parameters include the time-varying filter coefficients, pitch periods, excitation vectors and gain factors. Basically, the excitation vector, c(n), is first scaled by the gain factor, G. The result is
- pitch synthesis filter whose parameters include the pitch gain, g .
- the gain factor g may be modified to
- some parameters are concerned with the spectral and/or waveform shapes of the speech signal for that frame. These parameters typically include the LPC coefficients and the pitch information in the case of the LPC speech synthesis model.
- SLRPs speech level related parameters
- the first three GSM codecs in Table 1 will now be discussed. All of the first three coders process speech sampled at 8kHz and assume that the samples are obtained as 13-bit linear PCM values.
- the frame length is 160 samples (20ms). Furthermore, they divide each frame into four subframes of 40 samples each.
- the SLRPs for these codecs are listed in Table 2. Table 2. Speech Level Related Parameters in GSM Speech Codecs
- the SLRP may be specified each subframe (e.g. the GSM FR and EFR codecs) or once per frame (e.g. the GSM HR codec).
- the quantized and corresponding unquantized parameters are related through
- the quantization function is a many-to-one transformation and is not
- Figure 8 distinguishes the coded domain from the linear domain.
- the coded domain refers to the output of speech encoders or the input of the speech decoders, which should be identical if there are no channel errors.
- the coded domain includes both the speech parameters and the methods used to quantize or dequantize these parameters.
- the speech parameters that are determined by the encoder undergo a quantization process prior to transmission. This quantization is critical to achieving bit rates lower than that required by the original digital speech signal. The quantization process often involves the use of look-up tables.
- Processing of speech in the coded domain involves directly modifying the quantized speech parameters to a different set of quantized values allowed by the quantizer for each of the parameters.
- the parameters being modified are the SLRPs.
- NR noise reduction
- the quantization of a single speech parameter is termed scalar quantization.
- Vector quantization is usually applied to a set of parameters that are related to each other in some way, such as the LPC coefficients.
- Scalar quantization is generally applied to a parameter that is relatively independent of the other parameters.
- the quantization process is independent of the past and future values of the parameter. Only the current value of the parameter is used in the quantization process.
- the parameter to be quantized is compared to a set of permitted quantization levels.
- the quantization level that best matches the given parameter in terms of some closeness measure is chosen to represent that parameter.
- the permitted quantization levels are stored in a look-up table at both the encoder and the decoder.
- the index into the table of the chosen quantization level is transmitted by the encoder to the decoder.
- the quantization level may be determined using a mathematical formula.
- the quantization levels are usually spaced non-uniformly in the case of
- Adaptive quantizers may utilize forward adaptation or backward adaptation.
- forward adaptation schemes extra side information regarding the dynamic range has to be transmitted periodically to the decoder in addition to the quantization table index. Thus, such schemes are usually not used in speech coders.
- Backward adaptive quantizers are preferred because they do not require transmission of any side information. Two general types of backward adaptive quantizers are commonly used: standard deviation based and differential. These are depicted in
- ⁇ (n) is transmitted to the dequantizer where the inverse process norm is performed.
- a quantized version of the normalization factor is used at both the quantizer and dequantizer.
- decisions to expand or compress the quantization intervals may be based simply on the previous parameter input only.
- the correlation between current and previous parameter values is used to advantage. When the correlation is high, a significant reduction in the quantization dynamic range can be achieved by quantizing the prediction error, r(n).
- the prediction error is the difference between the actual and predicted parameter values. The same predictor
- differential quantization scheme can also be represented as in Figure 10 when a linear predictor, P(z), is used. Note that if we approximate the transfer function P(z)/[1-P(z)] by the linear predictor,
- g c (n) denotes the gain factor that is used to scale the
- the quantization of this parameter utilizes the scheme shown in Figure 11 but is rather indirect.
- the actual 'gain' parameter that is transmitted is actually a correction factor between g c (n) and the predicted gain, g c '(n).
- a 32-level non-uniform quantization is performed on ⁇ sc (n) to obtain ⁇ (n).
- Equation (3) E is a
- the decoder thus, can obtain the predicted gain in the same manner as the encoder using (3) once the current subframe
- R(n) denotes the prediction error given by
- the actual information transmitted from the encoder to the decoder are the bits representing the look-up table index of the quantized R(n) parameter,
- the quantization of the SLRP at the encoder is performed indirectly by using the mean-removed excitation vector energy each subframe.
- E(n) denotes the mean-
- second line of equation (5) is the mean excitation vector energy, E, (n) , i.e.
- excitation vector (c( ) ⁇ is preferred at the decoder prior to the
- the decoder decodes the excitation vector and computes E, (n) using equation
- Figure 13 Note that the vertical axis in Figure 13 which represents the quantization levels is plotted on a logarithmic scale.
- a post-filtering process is performed on the signal decoded using the LPC-based model. This post-filtering helps to reduce quantization noise but does not change the overall power level of the signal. Thus, in partial decoding of CELP-coded speech, the post-filtering process can be avoided for economy.
- silence suppression scheme is often used in cellular telephony and voice over packet networks. In these schemes, coded speech frames are transmitted only during voice activity and very little transmission is performed during silence. The decoders automatically insert some comfort noise during the silence periods to mimic the background noise from the other end.
- One example of such a scheme used in GSM cellular networks is called discontinuous transmission (DTX).
- DTX discontinuous transmission
- the decoder in the ALC device can completely avoid decoding the signal during silence. In such cases, the determination of voice and double-talk activities can also be simplified in
- the coded speech bits for each channel will be carried through the wireline network between base stations at 64 kbits/sec.
- This bitstream can be divided into 8-bit samples.
- the 2 least significant bits of each sample will contain the coded speech bits while the upper 6 bits will contain the bits corresponding to the appropriate PCM samples.
- the conversion of the PCM information to linear speech is very inexpensive and provides a somewhat noisy version of the linear speech signal. It is possible to use this noisy linear domain speech signal to perform the necessary voice activity, double-talk and speech level measurements as is usually done in linear domain ALC algorithms. Thus, in this case, only a minimal amount of interpretation of the PCM samples is necessary.
- the iterative techniques of Figure 15 can be incorporated in the Gain Determination block. Basically, after deciding on a desired gain value, the realized gain value after requantization of the SLRP may be computed. The realized gain is checked to see if overflow or underflow problems could occur. This could be accomplished, for example, by determining what the new speech level would be by multiplying the realized gain by the original speech level. Alternatively, a speech decoder could be used in the ALC device to see whether overflow/underflow actually occurs.
- the gain that is fed back should be the realized gain after the SLRP requantization process, not the desired gain.
- a preferred approach is shown in Figure 16. If the desired gain was used in the feedback loop instead of the realized gain, the controller would not be tracking the actual decoded speech signal level, resulting in erroneous level control.
- these methods preferably include the integration of the gain determination and SLRP requantization techniques.
- FIG. 17 illustrates the general configuration of an ALC device that uses joint gain determination and SLRP requantization. The details will depend on the particular ALC device.
- the desired gain determined by the ALC device will be denoted by g(n) .
- the realized gain after SLRP requantization will be denoted by
- overflow and underflow prevention are desired, then the iterative scheme described in Figure 15 may be used.
- the partial decoding of the speech samples using the requantized SLRP may be performed to the extent necessary. This, of course, involves additional complexity in the algorithm.
- the decoded samples can then be directly inspected to ensure that overflow or underflow has not taken place. Note that for a given received ⁇ (n) , there are m possible realized gain values.
- the SLRP quantization table values are uniformly spaced (either linearly or logarithmically), then it is possible to simplify the scalar requantization process. This simplification is achieved by allowing only a discrete set of desired gain values in the ALC device. These desired gain values preferably have the same spacing as the SLRP quantization values, with OdB being one of the gains. This ensures that the desired and realized gain values will always be aligned so that equation (8) would not have to be evaluated for each table value. Hence the requantization is greatly simplified.
- the original quantization index of the SLRP is simply increased or decreased by a value corresponding to the desired gain value divided by the SLRP quantization table spacing. For instance, suppose that the SLRP quantization table spacing is denoted by
- the discrete set of permitted desired gain values would be 1 + ⁇ ..., -2, -, 0, , 2, ... ⁇ if the SLRP quantization table values are uniformly spaced linearly, and 0+ ⁇ ..., -2, -, 0, , 2, ... ⁇ if the SLRP quantization table values are uniformly spaced
- ⁇ would be the average spacing between adjacent quantization table values, where the average is performed appropriately using either linear or logarithmic distances between the values.
- This codec's SLRP is the block maximum, x max , which is
- the Q and Q "1 blocks represent the SLRP requantization
- the index of the block maximum is first
- the index of the requantized x max is then substituted for the original value in the
- This requantization technique forms the basic component of all the techniques described in Figures 14-17 when implementing coded domain ALC for the GSM FR standard.
- GSM EFR codec will be used as an example for illustrating the implementation of coded domain ALC using this requantization technique.
- FIG. 19 shows a general coded domain ALC technique with only the components relevant to ALC being shown. Note that (G(n) denotes the original logarithmic gain value determined by the encoder. In the case of the EFR codec, G(n) is equal to
- E(n) defined in equation (5) and R(n) is as defined in equation (4).
- the ALC device determines the desired gain, ⁇ G (n) .
- the SLRP, R(n) is modified by the ALC
- the device to R ALC (n) based on the desired gain.
- the realized gain, ⁇ R(n) is the
- the actual realized gain is essentially an amplified version of the SLRP realized gain due to the decoding process, under steady-state conditions.
- steady-state it is meant that ⁇ G(n) is kept constant for a period of time that is sufficiently long so
- ⁇ R(n) is either steady or oscillates in a regular manner about a particular level.
- This method for differential scalar requantization basically attempts to mimic the operation of the encoder at the ALC device. If the presence of the quantizers at the encoder and the ALC device is ignored, then both the encoder and the ALC device
- G alc (n) G(n) + ⁇ G(n) + quantization error
- the feedback of the SLRP realized gain, ⁇ R(n) , in the ALC device can cause
- each element could contain one of 32 possible values.
- the non-linear system in the ALC device can be in any one of over a million possible states at any given time. This is mentioned because the behavior of this non-linear system is heavily influenced by its initial conditions.
- Figure 20(a) shows the step in the desired gain.
- Figure 20(b) shows the actual realized gain superimposed on the desired gain.
- the reverberations in the SLRP realized gain shown in Figure 20(b) cause a modulation of the speech signal and can result in audible distortions. Thus, depending on the ALC specifications, such reverberations may be undesirable.
- the reverberations can be eliminated by 'moving' the quantizer outside the feedback loop as shown in Figure 20. (In this embodiment, the computation of is unnecessary but is included for comparison to Figure 19.) Placing the quantizer outside the feedback loop results in the actual realized gain shown in Figure 20(c), superimposed on the desired gain. It should be noted that, although reverberations are eliminated, the average error (i.e. the average difference between the desired and actual realized gains) is higher than that shown in Figure 20(b). Specifically, in these examples, the average error during steady state operation of the requantizer with and without the quantizer in the feedback loop are 0.39dB and 1.03dB, respectively.
- the ALC apparatus of Figure 21 can be simplified as shown in Figure 22, resulting in savings in computation. This is done by replacing the linear system
- Some ALC algorithms may utilize past gain values to determine current and future gain values.
- the gain that is fed back should be the actual realized gain after the SLRP requantization process, not the desired gain. This was discussed above in conjunction with Figure 16.
- any of the methods described earlier that have quantizers within the feedback loop may be used.
- any of the methods described earlier that have quantizers outside the feedback loop may be used. If, however, both accuracy and avoidance of reverberations are necessary as is often the case in ALC, then a different approach is necessary.
- the current method avoids reverberations in the actual realized gains by placing the quantizers outside the feedback loop as in Figures 21, 22, or 23(b). Additionally, the average error between desired and actual realized gains is minimized by restricting the desired gain values to belong to the set of possible actual realized
- the ALC device computes the
- the ALC algorithm should preferably be designed such that the desired gain, , is selected from this set. Such restrictions can be easily imposed on a large variety of ALC algorithms since most of them already operate using a finite set of possible desired gain values.
- desired gain value is independent of the current original SLRP value, R(n) .
- the 32 R ; values in the EFR codec can be divided into
- R(n) will fall into the lower or higher regions only for very short durations such as at the transitions between speech and silence. Hence reverberations cannot be sustained in these regions.
- the desired gain value can be selected to be a multiple of
- the EFR encoder compresses a 20ms speech frame into 244 bits.
- the earliest point at which the first sample can be decoded is after the reception of bit 91 as shown in Figure 25(a). This represents a buffering delay of approximately 7.46ms. It turns out that sufficient information is received to decode not just the first sample but the entire first subframe at this point. Similarly, the entire first subframe can be decoded after about 7.11ms of buffering delay in the FR decoder.
- each subframe has an associated SLRP in both the EFR and FR coding schemes. This is generally true for most other codecs where the encoder operates at a subframe level.
- ALC and NR in the coded domain can be performed subframe-by-subframe rather than frame-by-frame.
- the new SLRP computed by the ALC device can replace the original SLRP in the received bitstream.
- the delay incurred before the SLRP can be decoded is determined by the position of the bits corresponding to the SLRP in the received bitstream. In the case of the FR and EFR codecs, the position of the SLRP bits for the first subframe determines this delay.
- the ALC algorithm must be designed to determine the gain for the current subframe based on previous subframes only. In this way, almost no buffering delay will be necessary to modify the SLRP. As soon as the bits corresponding to the
- SLRP in a given subframe will first be decoded. Then the new SLRP will be computed based on the original SLRP and information from the previous subframes only. The original SLRP bits will be replaced with the new SLRP bits. There is no need to wait until all the bits necessary to decode the current subframe are received. Hence, the buffering delay incurred by the algorithm will depend on the processing delay which is small. Information about the speech level is derived from the current subframe only after replacement of the SLRP for the current subframe. Those skilled in communications recognize that the same principles apply to NR algorithms.
- the subframe excitation vector is also needed to decode the SLRP and the more complex differential requantization techniques have to be used for requantizing the SLRP. Even in this case, significant reduction in the delay is attained by performing the speech level update based on the current subframe after the SLRP is replaced for the current subframe.
- the received bitstream can be divided into 8-bit samples.
- the 2 least significant bits of each sample will contain the coded speech bits while the upper 6 bits will contain the bits corresponding to the appropriate PCM samples.
- a noisy version of the linear speech samples is available to the ALC device in this case. It is possible to use this noisy linear domain speech signal to perform the necessary voice activity, double- talk and speech level measurements as is usually done in linear domain ALC algorithms.
- this noisy linear domain speech signal is necessary voice activity, double- talk and speech level measurements as is usually done in linear domain ALC algorithms.
- only a minimal amount of decoding of the coded domain speech parameters is necessary. Only parameters that are required for the determination and requantization of the SLRP would have to be decoded.
- Partial decoding of the speech signal is unnecessary as the noisy linear domain speech samples can be relied upon to measure the speech level as well as perform voice activity and double-talk detection.
- An object of the present invention is to derive methods to perform noise reduction in the coded domain via methods that are less computationally intensive than using linear domain techniques of similar quality that require re-encoding of the processed signal.
- the flexibility available in the coded domain to modify parameters to effect desired changes in the signal characteristics may be limited due to quantization.
- a survey of the different speech parameters and the corresponding quantization methods used by industry standard speech coders was performed. The modification of the different speech parameters will be considered, in turn, and possible methods for utilizing them to achieve noise reduction will be discussed.
- the short-time power or energy of a speech signal is a useful means for inferring the amplitude variations of the signal.
- a preferred method utilizes a recursive averaging technique. In this technique, the short-time power, P(n), of a discrete-time signal s(n) is defined as
- the root-mean-square (RMS) short-time power may be more desirable.
- the square-root operation is avoided by using an approximation to the RMS power by averaging the magnitude of s(n) rather than its square as follows:
- the power in an analysis window of size N may, for example, be averaged as follows:
- VAD algorithms are essential for many speech processing applications. A wide variety of VAD methods have been developed. Distinguishing speech from background noise relies on the a few basic assumptions about speech. Most VAD algorithms make use of some or all of these assumptions in different ways to distinguish between speech and silence or background noise.
- the first assumption is that the speech signal level is usually greater than the background noise level. This is often the most important criterion used and many VAD algorithms are based solely on this assumption. Using this assumption, the presence of speech can be detected by comparing signal power measurements to thresholds values.
- a second assumption is that speech is non-stationary while noise is relatively stationary. Using this assumption, many schemes can be devised based on steadiness of the signal spectrum or the amount of variation in the signal pitch measurements.
- a single-band noise reduction system is the most basic noise reduction system conceivable.
- the sum of the speech and background noise power is the noise power.
- Both power measures may be performed using recursive averaging filters as given in equation (11).
- the total power measure is continuously updated.
- the noise power measure is updated only during the absence of speech as determined by the VAD.
- the noise suppression is effected by a gain, g (n) , given by (16)
- equation (15) may actually result in a negative value for the desired signal power due to estimation errors.
- additional heuristics are used to ensure that is always non-negative.
- a serious blemish associated with the single-band noise suppression technique is the problem of noise modulation by the speech signal.
- the noise When speech is absent, the noise may be totally suppressed. However, noise can be heard at every speech burst. Hence the effect is that the noise follows the speech and the amount of noise is roughly proportional to the loudness of the speech burst.
- This annoying artifact can be overcome to a limited extent (but not eliminated) by limiting the lowest possible gain to a small but non-zero value such as 0.1.
- the modulation of the noise may be less annoying with this solution.
- parameters are relatively independent of the other parameters and are usually quantized separately. Furthermore, they usually have a good range of quantized values (unlike the codebook excitation).
- the preferred embodiment uses these two parameters to achieve noise reduction.
- the requantization process often involves searching through a table of quantized gain values and finding the value that minimizes the squared distance.
- a slightly more complex situation arises when a gain parameter (or any other parameter to be modified) is quantized using a differential scalar quantization scheme. Even in this case, the cost of such re-encoding is still usually several orders of magnitude lower.
- the quantization of a single speech parameter is termed scalar quantization.
- vector quantization When a set of parameters are quantized together, the process is called vector quantization.
- Vector quantization is usually applied to a set of parameters that are related to each other in some way such as the LPC coefficients.
- Scalar quantization is generally applied to a parameter that is relatively independent of the other parameters
- the quantization process is independent of the past and future values of the parameter. Only the current value of the parameter is used in the quantization process.
- the parameter to be quantized is compared to a set of permitted quantization levels.
- the quantization level that best matches the given parameter in terms of some closeness measure is chosen to represent that parameter.
- the permitted quantization levels are stored in a look-up table at both the encoder and the decoder. The index into the table of the chosen quantization level is transmitted by the encoder to the decoder.
- the prediction error is the difference between the actual
- Vector quantization involves the joint quantization of a set of parameters. In its simplest form, the vector is compared to a set of allowed vectors from a table. As in scalar quantization, usually a mean squared error measure is used to select the closest vector from the quantization table. A weighted mean squared error measure is often used to emphasize the components of the vector that are known to be perceptually more important.
- LPC coefficients In the case of LPC coefficients, the range of the coefficients is unconstrained at least theoretically. This as well as stability problems due to slight e ⁇ ors in representation have resulted in first transforming the LPC coefficients to a more suitable parameter domain prior to quantization.
- the transformations allow the LPC coefficients to be represented with a set of parameters that have a known finite range and prevent instability or at least reduce its likelihood. Available methods include log-area ratios and inverse sine functions.
- a more computationally complex representation of the LPC coefficients is the line spectrum pair (LSP) representation.
- LSPs provide a pseudo-frequency representation of the LPC coefficients and have been found to be capable of improving coding efficiency by more than other transformation techniques as well as having other desirable properties such as a simple way to guarantee stability of the LP synthesis filter.
- Gain parameters and pitch periods are sometimes quantized this way.
- the GSM EFR coder quantizes the codebook gain differentially. A general technique for differential requantization will now be discussed.
- G(n) is the parameter to be requantized and that the linear predictor used in the quantization scheme is denoted P(z) as shown in Figure 28.
- the quantized difference, R(n) is the actual coded domain parameter normally transmitted from the encoder to the decoder. This parameter is preferably intercepted by the network
- Such a system preferably includes an output, R new (n) , which is
- R new (n) R(n) + ⁇ R(n)
- G new (n) G(n) + ⁇ G(n) + quantization error
- G(n) can cause undesirable oscillatory effects if G(n) is not changing for long periods of time. This can have undesirable consequences to the speech signal especially if G(n) is a gain parameter.
- the G(n) corresponds to the logarithm of the codebook gain. During silent periods, G(n) may remain at the same quantized level for long durations. During such silence, if attenuation of the signal is attempted by the network device by modifying G(n) by an appropriate amount ⁇ G(n) , quasi-periodic modulation of the noise could occur resulting in a soft but
- the reverberations can be eliminated by 'moving' the quantizer outside the feedback loop as shown in Figure 30. (In Figure 30, the computation of is unnecessary but is included for comparison to Figure 28.) Placing the quantizer outside the feedback loop results in the actual realized gain shown in Figure 29(c), superimposed on the desired gain. It should be noted that, although reverberations are eliminated, the average e ⁇ or (i.e. the average difference between the desired and actual realized gains) is higher than that shown in Figure 29(b). Specifically, for this example, the average error during steady state operation of the requantizer with and without the quantizer in the feedback loop are 0.39dB and 1.03dB, respectively.
- a commonly used subframe period of duration 5ms will be assumed.
- a subframe will consist of 40 samples.
- a sample index will be denoted using n, and the subframe index using . Since the coded parameters are updated at most once per subframe and apply to all the samples in the subframe, there will be no confusion if these coded parameters are simply indexed using m. Other variables that are updated or apply to an entire subframe will also be indexed in this manner.
- the individual samples within a subframe will be normally indexed using n. However, if more than one subframe is spanned by an equation, then it will make sense to index a sample, such as a speech sample, as s(n, m).
- the speech synthesis model that is used in hybrid, parametric, time domain coding techniques can be thought of as time varying system with an overall transfer
- FCB fixed codebook
- b'(n) is obtained from the LTP buffer.
- the most recently computed subframe of LP excitation samples, u(n), are preferably shifted into the left end of the LTP buffer. These samples are also used to excite the LP synthesis filter to reconstruct the coded speech.
- the two sources of the LP synthesis filter excitation, u(n), have been explicitly identified. These two excitation sources, denoted as b(n) and c(n), are called the pitch excitation and codebook excitation, respectively.
- the LTP is also often called the adaptive codebook, due to its ever-changing buffer contents, in contrast to the FCB. Obviously, the LTP output is not independent of the FCB output.
- spectral subtraction concepts preferably are not directly applied to the two sources.
- the two sources have different characteristics. This difference in characteristic is taken advantage of to derive a noise reduction technique.
- the encoder are modified. This modification will be achieved by multiplying these gain factors by the noise reduction gain factors, y p and ⁇ c , respectively, to generate
- a prefe ⁇ ed network noise reduction device is shown in Figure 33.
- a decoder 20 is termed the reference decoder and performs decoding of the coded speech received from the encoder, such as the speech encoder 10 shown in Figure 14.
- the decoding performed by decoder 20 may be complete or partial, depending on the particular codec. For the current embodiment, it is assumed that it performs complete decoding, producing the noisy speech output y(n). However, as described above, the embodiment also will operate with partial decoding. Essentially, decoding which does not substantially affect, for example, the power of the noise characteristic, can be avoided, thereby saving time.
- the bottom half of Figure 33 shows a destination decoder 120.
- This decoder mimics the actual decoder at the destination, such as the receiving handset. It produces the estimated clean speech output on a conductor 148. Note that, although drawn separately for clarity, some of the parts of the reference decoder and destination decoder model can be shared. For instance, the fixed codebook (FCB) signal is identical for both decoders.
- FCB fixed codebook
- decoders 20 and 120 may be substituted for the following blocks of Figure 14: Partial or Complete Decoding block;
- Multiply function having inputs SLRP and gain; SLRP Requantization; and
- the Voice Activity function referred to in Figure 14 is incorporated into the Figure 33 embodiment.
- the speech decoder 12 shown in Figure 33 may be the same type of speech decoder shown in Figure 14. More specifically, the Figure 33 decoders are useful in a communication system 8 using various compression code parameters, such as the parameters described in Figure 7, including codebook gain, pitch gain and codebook RPE pulses.
- Such parameters represent an audio signal having various audio characteristics, including a noise characteristic and signal to noise ratio (SNR).
- SNR signal to noise ratio
- Decoders 20 and 120 may be implemented by a processor generally indicated by 150 which may include a noise reduction controller 160 which includes a VAD function.
- Processor 150 may comprise a microprocessor, a microcontroller or a digital signal processor, as well as other logic units capable of logical and arithmetic operations. Decoders 20 and 120 may be implemented by software, hardware or some combination of software and hardware.
- Processor 150 responds to the compression code of the digital signals sent by encoder 10 on a network 11. Decoders 20 and 120 each read certain compression code parameters of the type described in Figure 7, such as codebook gain and pitch gain. Processor 150 is responsive to the compression code to perform the partial decoding , if any, needed to measure the power of the noise characteristic. The decoding results in the decoded signals in the linear domain which simplify the task of measuring the noise power.
- the reference decoder 20 receives the compression coded digital signals on terminals 13. Decoder 20 includes a fixed codebook (FCB) function 22 which generates codebook vectors C'(n) that are multiplied or scaled by codebook gain g c in a multiply function 24.
- the codebook gain is read by processor 150 from the compressed code signals received at terminals 13.
- the multiply function generates scaled codebook vectors c(n) which are supplied to a pitch synthesis filter 26.
- Processor 150 calculates the power P c of the scaled codebook vectors as shown in equation 31. The power is used to adjust the pitch gain. Processor 150 reduces the codebook gain to attenuate the scaled codebook vector contribution to the noise characteristic.
- Filter 26 includes a long term predictor (LTP) buffer 28 responsive to the scaled codebook vectors c(n) to generate sample vectors. The samples are scaled by the pitch gain g p in a multiply function 30 to generate scaled samples b ref (n) that are processed by an adder function 32. Processor 150 increases the pitch gain to increase the contribution of the scaled samples in order to manage the noise characteristic as indicated in equations 30-33.
- LTP long term predictor
- Processor 150 determines the power of the scaled samples Pbref • A similar power P is generated by decoder 120. The two powers are used to adjust the pitch gain as indicated by equations 30 and 33.
- Filter 26 generates a total codebook excitation vector or LPC excitation vector u(n) at its output. Processor calculates the power P u of vector u(n) and uses the power to adjust the pitch gain as indicated in equation 32.
- the vector u(n) excites an LPC synthesis filter 34 like the one shown in Figure 6.
- the output of filter 34 is returned to controller 160.
- Decoder 120 includes many functions which are identical to the functions described in connection with decoder 20. The like functions bear numbers which are indexed by 100. For example, codebook 22 is identical to codebook 122. Decoder 120 includes multiplier functions 140 and 142 which are not included in decoder 20.
- Multiplier function 140 receives ⁇ p as an input which is defined in equation 33.
- Multipler function 142 receives ⁇ c as an input which is defined in equation
- decoder 120 uses a pitch synthesis filter 144 which is different from pitch synthesis filter 26.
- processor adjusts the codebook gain and/or pitch gain to manage the noise characteristic of the signals received at terminals 13.
- the adjusted gain values are quantized in the manner previously described and the quantized parameters are transmitted on an output network 15 through a terminal 16.
- the basic single-band noise suppressor discussed above can be implemented
- P w (m) and P y (m) are the noise power and total power estimate, respectively
- E is the maximum loss that can be applied by the single-band noise suppressor. It is usually set to a small value such as 0.1.
- Such a DC gain control system will suffer from severe noise modulation because the noise power fluctuates in sync with the speech signal. This can be perceptually annoying and one way to compensate for this is by trading off the amount of noise suppression for the amount of noise modulation.
- a coded domain noise reduction method may be derived that is superior to the
- the y c will be allowed to rise and fall appropriately.
- the LTP excitation output contributes to a large amount of the resulting signal power and has a better SNR relative to the FCB excitation output.
- y c should preferably be driven close to zero or some maximum loss
- ⁇ ⁇ should preferably be large during silence gaps and small
- the first weight may
- a voice activity detector can be used to demarcate the silence segments from the speech segments in the reference decoder's output signal, y(n).
- noise power, P w can be estimated during silence gaps in the decoded speech signal
- the power, P y , of the signal, y(n), can also be
- g p (m) which also reflects the amount of co ⁇ elation in the speech, may be used.
- the pitch gain is the result of an optimization procedure at the encoder that determines the pitch synthesis filter. In essence, this procedure finds a past sequence from the LTP buffer that has the best correlation with the sequence to be embodied. Therefore, if the co ⁇ elation is high, then the pitch gain would also be correspondingly high. As such, the remaining weight may be specified to be inversely proportional to the pitch gain: ⁇ T — A - g p (m)
- the parameter ⁇ is preferably empirically determined. It is quite common to have parameters that require to be tuned based on perceptual tests in speech enhancement algorithms.
- g p is also large.
- pitch synthesis filter used in the coder will be considered.
- the pitch synthesis filter is basically a comb filter.
- the pitch gain and pitch period are used to specify the pitch synthesis filter, there is no DC gain factor available to simultaneously control the amount of gain at both the spectral peaks and the valleys.
- some encoders allow pitch gains greater than one. Theoretically, this results will result in an unstable comb filter.
- Due to the manner in which the optimization procedure attempts to match the synthetic signal to the original speech signal no actual instability results. Another way to look at this is to think of the FCB output as being designed in such a manner that never actually results in instability.
- Amplification to compensate for power loss may be
- ⁇ p l could be compared with y p 2
- y p is preferably chosen as
- CDNR coded domain noise reduction
- the first embodiment is suitable for
- Figure 36 shows a novel implementation of CDNR.
- a silence detector also refe ⁇ ed to as a voice activity detector
- the background noise power is estimated.
- the total power of the signal is estimated.
- the dequantized codebook gain parameter is attenuated, and then quantized again. This new quantized codebook gain parameter substitutes the original one in the bit-stream.
- P y ( ) and P w (n) are the noisy uncoded speech power and the noise
- Power estimates may be performed in a variety of ways.
- One example
- the codebook gain parameters are defined every subframe. If this is the case, the formulae are evaluated using the power estimates computed during the last sample of the corresponding subframe. In both the above approaches, the attenuation factor depends on the signal-to-noise ratio of the uncoded speech. In formula (35), a suitable value for ⁇ are in the range from 1 to 1.5. In formula (36), a
- suitable value for ⁇ is 0.8.
- the decoding of signals may be complete or partial depending on the vocoder being used for the encode and decode operations. Some examples of situations where partial decoding suffices are listed below:
- CELP code-excited linear prediction
- the CDNR device may be placed between the base station and the switch (known as the A-interface) or between the two switches. Since the 6 MSBs of each 8-bit sample of the speech signal corresponds to the PCM code as shown in Figure 3, it is possible to avoid decoding the coded speech altogether in this situation. A simple table-lookup is sufficient to convert the 8-bit companded samples to 13-bit linear speech samples using A-law companding tables. This provides an economical way to obtain a version of the speech signal without invoking the appropriate decoder. Note that the speech signal obtained in this manner is somewhat noisy, but has been found to be adequate for the measurement of the power estimates.
- the buffering delay is reduced by almost 13ms.
- the coded LPC synthesis filter parameters are modified based on information available at the end of the first subframe of the frame. In other words, the entire frame is affected by the echo likelihood computed based on the first subframe. In experiments conducted, no noticeable artifacts were found due to this 'early' decision.
- bit-stream When applying the novel coded domain processing techiques described in this report for removing or reducing noise, some are all of the bits corresponding to the coded parameters are modified in the bit-stream. This may affect other e ⁇ or- correction or detection bits that may also be embedded in the bit-stream. For instance, a speech encoder may embed some checksums in the bit-stream for the decoder to verify to ensure that an error-free frame is received. Such checksums as well as any parity check bits, error correction or detection bits, and framing bits are updated in accordance with the appropriate standard, if necessary.
- Figure 38 shows a technique for coded domain noise reduction by modification of the codebook vector parameter.
- noise reduction is performed in two stages. The first stage involves modification of the codebook gain as discussed earlier.
- the codebook vector is optimized to minimize the noise.
- several codebook vector patterns are attempted that vary from the original received codebook vector.
- the partial decoding is performed and the noise power is estimated.
- the best codebook vector pattern is determined as the one that minimizes the noise power. In practice, a fixed number of iterations or trials are performed.
- the codebook vector pattern for each subframe has 40 positions, of which 13 contain non-zero pulses.
- the positions of the 13 non-zero pulses are not modified. Only their amplitudes are varied in each trial.
- the non-zero pulses are denoted by
- each pulse may be one of the following amplitude
- modified codebook vector modified codebook gain parameter, and the remainder of the original parameters, partially decode the signal.
- LSPs line spectral pairs
- LSFs frequencies
- LPC parameters also are represented by log area ratios in the GSM FR vocoder.
- LSFs may be directly modified for speech enhancement purposes.
- a technique that directly adapts the LSFs to attain a desired frequency response for use in a coded domain noise reduction system is described in the following. This general technique may be applied to modify the LSFs, for example, received from a GSM EFR encoder.
- the adaptive technique may be used to alter the spectral shape of the LPC synthesis filter
- A(z) can be recovered as A(z) - —[P(z) + Q(z)] ⁇
- each polynomial can be thought of as the transfer functions of a (p + 1) th order predictor derived from a lattice structure.
- A(z) is minimum phase, the two important properties of P(z) and Q(z) can be
- LSFs have a pseudo-frequency interpretation that is often useful in the design of quantization techniques.
- Figure 39 shows a randomly generated set of LSFs and the frequency response of the co ⁇ esponding linear predictor which has 10 coefficients.
- the solid vertical lines are
- ⁇ is an appropriate step-size parameter.
- the value of ⁇ is set to 0.00002.
- the frequency response of the LPC synthesis filter can be modified to have a desired frequency response.
- the desired frequency response of the LPC synthesis filter can be computed based on, for example, standard noise reduction techniques such as spectral subtraction.
- the compression code parameters are modified to reduce the effects of noise. More specifically, the LPC coefficients or one of their representations (e.g., line spectral frequencies or log-arc ratios) are modified to atenuate the noise in spectral regions affected by noise.
Landscapes
- Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Quality & Reliability (AREA)
- Computer Networks & Wireless Communication (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Compression, Expansion, Code Conversion, And Decoders (AREA)
- Transmission Systems Not Characterized By The Medium Used For Transmission (AREA)
- Cable Transmission Systems, Equalization Of Radio And Reduction Of Echo (AREA)
- Reduction Or Emphasis Of Bandwidth Of Signals (AREA)
Abstract
Description
Claims
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP00946954A EP1208413A2 (en) | 1999-07-02 | 2000-06-30 | Coded domain noise control |
CA002378035A CA2378035A1 (en) | 1999-07-02 | 2000-06-30 | Coded domain noise control |
AU60636/00A AU6063600A (en) | 1999-07-02 | 2000-06-30 | Coded domain noise control |
JP2001508667A JP2003504669A (en) | 1999-07-02 | 2000-06-30 | Coding domain noise control |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14213699P | 1999-07-02 | 1999-07-02 | |
US60/142,136 | 1999-07-02 |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2001002929A2 true WO2001002929A2 (en) | 2001-01-11 |
WO2001002929A3 WO2001002929A3 (en) | 2001-07-19 |
Family
ID=22498680
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2000/018293 WO2001003317A1 (en) | 1999-07-02 | 2000-06-30 | Coded domain adaptive level control of compressed speech |
PCT/US2000/018165 WO2001002929A2 (en) | 1999-07-02 | 2000-06-30 | Coded domain noise control |
PCT/US2000/018104 WO2001003316A1 (en) | 1999-07-02 | 2000-06-30 | Coded domain echo control |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2000/018293 WO2001003317A1 (en) | 1999-07-02 | 2000-06-30 | Coded domain adaptive level control of compressed speech |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2000/018104 WO2001003316A1 (en) | 1999-07-02 | 2000-06-30 | Coded domain echo control |
Country Status (5)
Country | Link |
---|---|
EP (3) | EP1190494A1 (en) |
JP (3) | JP2003503760A (en) |
AU (3) | AU6203300A (en) |
CA (3) | CA2378035A1 (en) |
WO (3) | WO2001003317A1 (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1301018A1 (en) * | 2001-10-02 | 2003-04-09 | Alcatel | Apparatus and method for modifying a digital signal in the coded domain |
EP1521242A1 (en) * | 2003-10-01 | 2005-04-06 | Siemens Aktiengesellschaft | Speech coding method applying noise reduction by modifying the codebook gain |
EP1553560A1 (en) * | 2002-07-16 | 2005-07-13 | Sony Corporation | Transmission device, transmission method, reception device, reception method, transmission/reception device, communication device, communication method, recording medium, and program |
US8874437B2 (en) | 2005-03-28 | 2014-10-28 | Tellabs Operations, Inc. | Method and apparatus for modifying an encoded signal for voice quality enhancement |
US11031023B2 (en) | 2017-07-03 | 2021-06-08 | Pioneer Corporation | Signal processing device, control method, program and storage medium |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3946074B2 (en) * | 2002-04-05 | 2007-07-18 | 日本電信電話株式会社 | Audio processing device |
US7613607B2 (en) | 2003-12-18 | 2009-11-03 | Nokia Corporation | Audio enhancement in coded domain |
US8078659B2 (en) * | 2005-10-31 | 2011-12-13 | Telefonaktiebolaget L M Ericsson (Publ) | Reduction of digital filter delay |
US7852792B2 (en) * | 2006-09-19 | 2010-12-14 | Alcatel-Lucent Usa Inc. | Packet based echo cancellation and suppression |
JP4915575B2 (en) * | 2007-05-28 | 2012-04-11 | パナソニック株式会社 | Audio transmission system |
JP4915576B2 (en) * | 2007-05-28 | 2012-04-11 | パナソニック株式会社 | Audio transmission system |
JP4915577B2 (en) * | 2007-05-28 | 2012-04-11 | パナソニック株式会社 | Audio transmission system |
US8032365B2 (en) * | 2007-08-31 | 2011-10-04 | Tellabs Operations, Inc. | Method and apparatus for controlling echo in the coded domain |
WO2012106926A1 (en) | 2011-07-25 | 2012-08-16 | 华为技术有限公司 | A device and method for controlling echo in parameter domain |
TWI469135B (en) * | 2011-12-22 | 2015-01-11 | Univ Kun Shan | Adaptive differential pulse code modulation (adpcm) encoding and decoding method |
JP6011188B2 (en) * | 2012-09-18 | 2016-10-19 | 沖電気工業株式会社 | Echo path delay measuring apparatus, method and program |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4969192A (en) * | 1987-04-06 | 1990-11-06 | Voicecraft, Inc. | Vector adaptive predictive coder for speech and audio |
US5097507A (en) * | 1989-12-22 | 1992-03-17 | General Electric Company | Fading bit error protection for digital cellular multi-pulse speech coder |
US5680508A (en) * | 1991-05-03 | 1997-10-21 | Itt Corporation | Enhancement of speech coding in background noise for low-rate speech coder |
Family Cites Families (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0683114B2 (en) * | 1985-03-08 | 1994-10-19 | 松下電器産業株式会社 | Eco-Cancer |
US5140543A (en) * | 1989-04-18 | 1992-08-18 | Victor Company Of Japan, Ltd. | Apparatus for digitally processing audio signal |
JP3353257B2 (en) * | 1993-08-30 | 2002-12-03 | 日本電信電話株式会社 | Echo canceller with speech coding and decoding |
US5828995A (en) * | 1995-02-28 | 1998-10-27 | Motorola, Inc. | Method and apparatus for intelligible fast forward and reverse playback of time-scale compressed voice messages |
JPH0954600A (en) * | 1995-08-14 | 1997-02-25 | Toshiba Corp | Voice-coding communication device |
JPH0993132A (en) * | 1995-09-27 | 1997-04-04 | Toshiba Corp | Device and method for coding decoding |
JPH10143197A (en) * | 1996-11-06 | 1998-05-29 | Matsushita Electric Ind Co Ltd | Reproducing device |
US5943645A (en) * | 1996-12-19 | 1999-08-24 | Northern Telecom Limited | Method and apparatus for computing measures of echo |
JP3283200B2 (en) * | 1996-12-19 | 2002-05-20 | ケイディーディーアイ株式会社 | Method and apparatus for converting coding rate of coded audio data |
US6064693A (en) * | 1997-02-28 | 2000-05-16 | Data Race, Inc. | System and method for handling underrun of compressed speech frames due to unsynchronized receive and transmit clock rates |
JP3317181B2 (en) * | 1997-03-25 | 2002-08-26 | ヤマハ株式会社 | Karaoke equipment |
US6112177A (en) * | 1997-11-07 | 2000-08-29 | At&T Corp. | Coarticulation method for audio-visual text-to-speech synthesis |
CN1737903A (en) * | 1997-12-24 | 2006-02-22 | 三菱电机株式会社 | Method and apparatus for speech decoding |
-
2000
- 2000-06-30 EP EP00946994A patent/EP1190494A1/en not_active Withdrawn
- 2000-06-30 CA CA002378035A patent/CA2378035A1/en not_active Abandoned
- 2000-06-30 AU AU62033/00A patent/AU6203300A/en not_active Abandoned
- 2000-06-30 JP JP2001508064A patent/JP2003503760A/en active Pending
- 2000-06-30 AU AU60671/00A patent/AU6067100A/en not_active Abandoned
- 2000-06-30 JP JP2001508667A patent/JP2003504669A/en active Pending
- 2000-06-30 AU AU60636/00A patent/AU6063600A/en not_active Abandoned
- 2000-06-30 CA CA002378012A patent/CA2378012A1/en not_active Abandoned
- 2000-06-30 JP JP2001508063A patent/JP2003533902A/en active Pending
- 2000-06-30 CA CA002378062A patent/CA2378062A1/en not_active Abandoned
- 2000-06-30 WO PCT/US2000/018293 patent/WO2001003317A1/en not_active Application Discontinuation
- 2000-06-30 EP EP00948555A patent/EP1190495A1/en not_active Withdrawn
- 2000-06-30 WO PCT/US2000/018165 patent/WO2001002929A2/en not_active Application Discontinuation
- 2000-06-30 WO PCT/US2000/018104 patent/WO2001003316A1/en not_active Application Discontinuation
- 2000-06-30 EP EP00946954A patent/EP1208413A2/en active Pending
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4969192A (en) * | 1987-04-06 | 1990-11-06 | Voicecraft, Inc. | Vector adaptive predictive coder for speech and audio |
US5097507A (en) * | 1989-12-22 | 1992-03-17 | General Electric Company | Fading bit error protection for digital cellular multi-pulse speech coder |
US5680508A (en) * | 1991-05-03 | 1997-10-21 | Itt Corporation | Enhancement of speech coding in background noise for low-rate speech coder |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1301018A1 (en) * | 2001-10-02 | 2003-04-09 | Alcatel | Apparatus and method for modifying a digital signal in the coded domain |
EP1553560A1 (en) * | 2002-07-16 | 2005-07-13 | Sony Corporation | Transmission device, transmission method, reception device, reception method, transmission/reception device, communication device, communication method, recording medium, and program |
EP1553560A4 (en) * | 2002-07-16 | 2005-12-21 | Sony Corp | Transmission device, transmission method, reception device, reception method, transmission/reception device, communication device, communication method, recording medium, and program |
US7515661B2 (en) | 2002-07-16 | 2009-04-07 | Sony Corporation | Transmission device, transmission method, reception device, reception method, transmission/reception device, communication method, recording medium, and program |
US7688921B2 (en) | 2002-07-16 | 2010-03-30 | Sony Corporation | Transmitting apparatus and transmitting method, receiving apparatus and receiving method, transceiver apparatus, communication apparatus and method, recording medium, and program |
US7688922B2 (en) | 2002-07-16 | 2010-03-30 | Sony Corporation | Transmitting apparatus and transmitting method, receiving apparatus and receiving method, transceiver apparatus, communication apparatus and method, recording medium, and program |
EP1521242A1 (en) * | 2003-10-01 | 2005-04-06 | Siemens Aktiengesellschaft | Speech coding method applying noise reduction by modifying the codebook gain |
US8874437B2 (en) | 2005-03-28 | 2014-10-28 | Tellabs Operations, Inc. | Method and apparatus for modifying an encoded signal for voice quality enhancement |
US11031023B2 (en) | 2017-07-03 | 2021-06-08 | Pioneer Corporation | Signal processing device, control method, program and storage medium |
Also Published As
Publication number | Publication date |
---|---|
JP2003504669A (en) | 2003-02-04 |
WO2001003316A1 (en) | 2001-01-11 |
EP1190494A1 (en) | 2002-03-27 |
JP2003533902A (en) | 2003-11-11 |
AU6067100A (en) | 2001-01-22 |
CA2378012A1 (en) | 2001-01-11 |
JP2003503760A (en) | 2003-01-28 |
EP1208413A2 (en) | 2002-05-29 |
CA2378062A1 (en) | 2001-01-11 |
AU6203300A (en) | 2001-01-22 |
WO2001002929A3 (en) | 2001-07-19 |
AU6063600A (en) | 2001-01-22 |
WO2001003317A1 (en) | 2001-01-11 |
EP1190495A1 (en) | 2002-03-27 |
CA2378035A1 (en) | 2001-01-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
RU2325707C2 (en) | Method and device for efficient masking of deleted shots in speech coders on basis of linear prediction | |
US7613607B2 (en) | Audio enhancement in coded domain | |
US6654716B2 (en) | Perceptually improved enhancement of encoded acoustic signals | |
WO2001002929A2 (en) | Coded domain noise control | |
US5913187A (en) | Nonlinear filter for noise suppression in linear prediction speech processing devices | |
JPH08278799A (en) | Noise load filtering method | |
US20030065507A1 (en) | Network unit and a method for modifying a digital signal in the coded domain | |
JP2003533902A5 (en) | ||
JP3074680B2 (en) | Post-noise shaping filter for speech decoder. | |
EP1544848B1 (en) | Audio enhancement in coded domain | |
Chandran et al. | Compressed domain noise reduction and echo suppression for network speech enhancement | |
US20050102136A1 (en) | Speech codecs | |
Lee | An enhanced ADPCM coder for voice over packet networks | |
CN100369108C (en) | Audio enhancement in coded domain |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A2 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CR CU CZ DE DK DM DZ EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG US UZ VN YU ZA ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A2 Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
AK | Designated states |
Kind code of ref document: A3 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CR CU CZ DE DK DM DZ EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG US UZ VN YU ZA ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A3 Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2378035 Country of ref document: CA |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2000946954 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 60636/00 Country of ref document: AU |
|
REG | Reference to national code |
Ref country code: DE Ref legal event code: 8642 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 10019617 Country of ref document: US |
|
WWP | Wipo information: published in national office |
Ref document number: 2000946954 Country of ref document: EP |
|
WWW | Wipo information: withdrawn in national office |
Ref document number: 2000946954 Country of ref document: EP |