WO2011049514A1 - Method and background estimator for voice activity detection - Google Patents
Method and background estimator for voice activity detection Download PDFInfo
- Publication number
- WO2011049514A1 WO2011049514A1 PCT/SE2010/051116 SE2010051116W WO2011049514A1 WO 2011049514 A1 WO2011049514 A1 WO 2011049514A1 SE 2010051116 W SE2010051116 W SE 2010051116W WO 2011049514 A1 WO2011049514 A1 WO 2011049514A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- noise
- background
- frames
- update
- input signal
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 37
- 230000000694 effects Effects 0.000 title claims abstract description 17
- 238000001514 detection method Methods 0.000 title claims description 16
- 206010019133 Hangover Diseases 0.000 claims description 27
- 230000003044 adaptive effect Effects 0.000 claims description 10
- 238000004458 analytical method Methods 0.000 claims description 10
- 238000011084 recovery Methods 0.000 description 12
- 230000004048 modification Effects 0.000 description 9
- 238000012986 modification Methods 0.000 description 9
- 238000012360 testing method Methods 0.000 description 8
- 230000006870 function Effects 0.000 description 5
- 230000006872 improvement Effects 0.000 description 5
- 230000007774 longterm Effects 0.000 description 5
- 230000003595 spectral effect Effects 0.000 description 5
- 238000004364 calculation method Methods 0.000 description 4
- 101150014198 epsP gene Proteins 0.000 description 4
- CBGUOGMQLZIXBE-XGQKBEPLSA-N clobetasol propionate Chemical compound C1CC2=CC(=O)C=C[C@]2(C)[C@]2(F)[C@@H]1[C@@H]1C[C@H](C)[C@@](C(=O)CCl)(OC(=O)CC)[C@@]1(C)C[C@@H]2O CBGUOGMQLZIXBE-XGQKBEPLSA-N 0.000 description 3
- 229940069205 cormax Drugs 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 230000000903 blocking effect Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 230000008450 motivation Effects 0.000 description 2
- 230000009467 reduction Effects 0.000 description 2
- 230000035945 sensitivity Effects 0.000 description 2
- 101100366889 Caenorhabditis elegans sta-2 gene Proteins 0.000 description 1
- 230000006978 adaptation Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 238000004422 calculation algorithm Methods 0.000 description 1
- 238000012512 characterization method Methods 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 238000013100 final test Methods 0.000 description 1
- 238000009499 grossing Methods 0.000 description 1
- 238000013101 initial test Methods 0.000 description 1
- 230000002265 prevention Effects 0.000 description 1
- 230000001629 suppression Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/78—Detection of presence or absence of voice signals
- G10L25/84—Detection of presence or absence of voice signals for discriminating voice from noise
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/06—Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
- G10L15/063—Training
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/20—Speech recognition techniques specially adapted for robustness in adverse environments, e.g. in noise, of stress induced speech
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/78—Detection of presence or absence of voice signals
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/06—Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
- G10L15/063—Training
- G10L2015/0635—Training updating or merging of old and new templates; Mean values; Weighting
- G10L2015/0636—Threshold criteria for the updating
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/78—Detection of presence or absence of voice signals
- G10L2025/783—Detection of presence or absence of voice signals based on threshold decision
- G10L2025/786—Adaptive threshold
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/03—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
- G10L25/06—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters the extracted parameters being correlation coefficients
Definitions
- the embodiments of the present invention relates to a method and a background estimator of a voice activity detector.
- Background noise estimates are used as a characterization of the background noise and is of use in applications such as: Noise suppression, Voice Activity Detectors, SNR (Signal-to-Noise Ratio) estimates.
- the background noise estimate should be able to track changes in the input noise characteristics and it should also be able to handle step changes such as sudden changes in the noise characteristics and/ or level while still avoiding using non-noise segments to update the background noise estimate.
- discontinuous transmission In speech coding systems used for conversational speech it is common to use discontinuous transmission (DTX) to increase the efficiency of the encoding. It is also possible to use variable bit rate (VBR) encoding to reduce the bit rate.
- VBR variable bit rate
- conversational speech contains large amounts of pauses embedded in the speech, e.g. while one person is talking the other one is listening. So with discontinuous transmission (DTX) the speech encoder is only active about 50 percent of the time on average and the rest is encoded using comfort noise.
- AMR Adaptive Multi Rate
- VAD Voice Activity Detector
- FIG. 1 shows an overview block diagram of a generalized VAD 180, which takes the input signal 100, divided into data frames, 5-30 ms depending on the implementation, as input and produces VAD decisions as output 160.
- a VAD decision 160 is a decision for each frame whether the frame contains speech or noise which is also referred to as VAD_flag.
- the generic VAD 180 comprises a feature extractor 120 which extracts the main feature used for VAD decisions from the input signal, one such example is subband energy used as a frequency representation of each frame of the input signal.
- a background estimator 130 provides subband energy estimates of the background signal (estimated over earlier input frames).
- An operation controller 1 10 collects characteristics of the input signal, such as long term noise level, long term speech level for long term SNR calculation and long term noise level variation as input signals to a primary voice detector.
- a preliminary decision, "vad_prim” 150 is made by a primary voice activity detector 140 and is basically just a comparison of the features for the current frame and background features (estimated from previous input frames), where a difference larger than a threshold causes an active primary decision.
- a hangover addition block 170 is used to extend the primary decision based on past primary decisions to form the final decision, "vad_flag" 160. The reason for using hangover is mainly to reduce/remove the risk of mid speech and backend clipping of speech bursts. However, the hangover can also be used to avoid clipping in music passages.
- the operation controller 1 10 may adjust the threshold(s) for the primary voice activity detector 140 and the length of the hangover addition 170 according to the characteristics of the input signal.
- the background estimation can be done by two basically different principles, either by using the primary decision i.e. with decision (or decision metric) feedback indicated by dash-doted line in figure 1 or by using some other characteristics of the input signal i.e. without decision feedback. It is also possible to use combinations of the two strategies.
- VADs There are a number of different features that can be used but one feature utilized in VADs is the frequency characteristics of the input signal. Calculating the energy in frequency subbands for the input signal is one popular way of representing the input frequency characteristics. In this way one of the background noise features is the vector with the energy values for each subband. These are values that characterize the background noise in the input signal in the frequency domain.
- the first way is to use an AR-process (Autoregressive process) per frequency bin to handle the update. Basically for this type of update the step size of the update is proportional to the observed difference between current input and the current background estimate.
- the second way is to use multiplicative scaling of current estimate with the restriction that the estimate never is bigger than the current input or smaller than a minimum value. This means that the estimate is increased for each frame until it is higher than the current input. In that situation the current input is used as estimate.
- the third way is to use minimum technique where the estimate is the minimum value during a sliding time window of prior frames. This basically gives a minimum estimate which is scaled, using a compensation factor, to get and
- a method for updating a background noise estimate for an input signal in a background estimator in a VAD is provided.
- the input signal for a current frame is received and it is determined whether the current frame of the input signal comprises non- noise. Further, an additional determination is performed whether the current frame of the non-noise input comprises noise by analyzing characteristics at least related to correlation and energy level of the input signal, and background noise estimate is updated if it is determined that the current frame comprises noise.
- a background estimator in a VAD for updating a background noise estimate for an input signal is provided.
- the background estimator comprises an input section configured to receive the input signal for a current frame.
- the background estimator further comprises a processor configured to determine whether the current frame of the input signal comprises non-noise, to perform an additional determination whether the current frame of the non-noise input comprises noise by analyzing characteristics at least related to correlation and energy level of the input signal, and to update background noise estimate if it is determined that the current frame comprises noise.
- a better noise tracking for background noise estimates especially for non-stationary noise is achieved.
- VAD functionality seen as a reduction in false speech frames reported in non-stationary noise.
- an improved deadlock recovery of background noise estimation for stationary noise types may be provided. From a system point of view the reduction in excessive activity would result in better capacity.
- a method and a background estimator of a voice activity detector of e.g. an encoder of a transmitter in user equipments are provided which are configured to implement the solution of the embodiments of the present invention.
- FIG 1 illustrates a generic Voice Activity Detector (VAD) with background estimation according to prior art.
- VAD Voice Activity Detector
- Figure 2 is a flowchart illustrating a background update procedure for a background noise estimator to be implemented in a transmitter according to prior art.
- Figure 3 is a flowchart illustrating a background update procedure for a background noise estimator to be implemented in a transmitter according to embodiments of the present invention.
- Figure 4 is another flowchart illustrating a method according to embodiments of the present invention.
- Figure 5 illustrates schematically a background estimator according to embodiments of the present invention.
- Figure 6 illustrates improved noise tracking for mixed speech (-26dBov) and noise babble 64 (-36dBov) input according to embodiments of the present invention.
- Figure 7 illustrates improved noise tracking for mixed speech (-26dBov) and pink noise (-46dBov) input according to embodiments of the present invention.
- the AR (Autoregressive) -process is used for background noise estimation where downwards adjustments of the noise estimates are always allowed.
- Figure 2 shows a basic flowchart of the decision logic for such a background estimator according to prior art.
- the update process of the background estimate starts with a frequency analysis to derive subband levels from the current input frame. Also other features used for the decision logic are calculated in this step, such as examples of features related to the noise estimation, total energy Etot, correlation, including pitch and voicing
- a vad_flag i.e. the decision whether voice is detected by the voice activity detector, is also calculated in this step. 2.1n this step, calculation of a potentially new noise estimate, tmpN is performed. This estimate is only based on the current input frames and the background noise estimate from the last frame. Already at this point the current noise estimate can be reduced if the currently estimated background estimate is higher than the potentially new noise estimate. In the pseudo code below that corresponds to that tmpN[i] is lower than bckr[i].
- hangover counter For active speech signals a hangover counter is activated if needed. Note that it is common also for background update procedures to use a hangover period and this is done to avoid using large noise like segments of a speech signal for background estimation. 5. If the hangover counter is not zero, the background estimation is still in hangover and there will not be any background noise update during this frame. If the hangover period is over, the hangover counter is zero. It may be possible to increase the noise estimate.
- the final steps before ending the noise update procedure is to update feature state history for usage in an evaluation of the next frame.
- an additional determination is performed whether the current frame of the non-noise input comprises noise. This is performed by analyzing characteristics at least related to correlation and energy level of the input signal, and the background noise estimate is updated if it is determined that the current frame comprises noise.
- the flowchart of figure 3 comprises additional or modified steps denoted “non-noise input?" denoted 3, "Noise input?" denoted 4a, "Background update (up)” denoted 4b, "High energy step” denoted 7, and “deadlock recovery?” denoted 8 and Background update reduced step (up) denoted 10a.
- the other blocks have the same functionality as the corresponding blocks in figure 2.
- the improved decision logic combines existing and new features to improve the non-noise detection in block 3 and adds the second noise input detection step in block 4a which also allows for an additional background update (see step 4b) although it was determined in block 5 that one still is in background noise update hangover.
- the additional noise input detection step in block 4a introduces an extra check of frames which are identified as potential voice frames in the "non-noise input" if they really are voice. If it is now determined that the frames are noise, then an increase in the noise estimate is allowed to be used to update the background in block 4b. Basically this allows better tracking of noise estimates close to speech bursts and some times even within speech bursts.
- the logic of the "Background update (up)" block denoted 4b allows an increase of the noise estimate but with a smaller step size compared to the "normal” noise increase used in the block of figure 2.
- noise (4a and 4b) although it is determined in block 5 that the hangover period for background noise update is still ongoing. It is possible to sharpen the requirements for normal (i.e. when it is determined in block 5 that sufficient time has passed since non-noise input was present) noise update without increasing the risk of ending up in noise estimate deadlock in the "high energy step?" block denoted 7.
- Noise estimate deadlock implies that it is not allowed to further increase the noise estimate. It is desirable to sharpen these requirements as it prevents some unwanted regular noise updates which e.g.
- E f lm , u> is a smoothed minimum energy tracker that is updated every frame. This is mainly used as a basis for other features.
- E t — E ⁇ Um LP is the difference in energy for current frame compared to smoothed minimum energy tracker.
- N lol - E f low ]P is the difference in energy for current noise estimate compared to smoothed minimum energy tracker.
- N bg is a counter for the number of consecutive possible background frames, based on E f low LP and the total energy E, . Note that this feature will not create a deadlock for stationary noise.
- N con . is a correlation event counter which counts the number of consecutive frames since the last frame that indicated correlation.
- SNR mm is a decision metric from a subband SNR VAD. In the improved background noise update logic this is used as a weighted spectral difference feature.
- the correlation event counter N corr is used in an improved non-noise detector as it is only in long speech/music pauses that the feature N cnrr will reach high values. This can be used to decrease the sensitivity of the non-noise detector when there has been a long pause since the last correlation event. This will allow the background noise estimator to better track the noise level in the case of noise only input.
- the feature E t - E f low LP can be used to detect when such energy steps occur and temporary block noise update from tracking the input. Note that for a step to a new level the feature E t - E f law LP will eventually recover since E f law LP only is based on the input energy and will adapt to the new level after a certain delay.
- the additional noise detector step can be seen as a combination of secondary noise update and alternative deadlock recovery. Two additional conditions are allowed for background update outside the normal update procedure. The first uses the features N corr , E, - E f low LP , N lgl - E f lm , u , , and N bg .
- N con ensures that a number of frames have been correlation free
- E t - E f low LP ensures that the current energy is close to the current estimated noise level
- N lol - E f , paragraph, command LP ensures that the two noise estimates are close (this is needed since E f low LP is allowed to track the input energy also in music)
- N bg that that the input level has been reasonably low (close to ⁇ law LP ) for a number of frames.
- the second uses the features N con . and SNR sum . Where N corr as before ensures a number of correlation free frames and SNR sum is used as a weighted spectral difference measure to decide when the input is noise like. Any of these two conditions can allow
- E f low ,p is as mentioned above a smoothed function of a minimum estimate of the frame energy that is slowly increased until a new minimum is found.
- E f low is an unsmoothed value which is increased with a small value S f low if the current frame energy E t is lower than the modified E f low . Then E f low is set to E t .
- the new value for E f /0H is then used to update the smoothed value through using an AR-process:
- E f jow_LP I 1 - ⁇ x) E fjow_u> + a E j jow ⁇ Note that after smoothing E f Jow LP is no longer a strict minimum estimate.
- the embodiments of the invention improve the decision logic for blocking the normal noise update process but also adds an alternative for updating the background estimate. This is done so that the background noise estimator achieves better tracking of non-stationary input noise and to avoid deadlock for the stationary noise types such as pink and white noise and still maintain /improve the ability of not tracking music or front ends of speech bursts.
- a frequency analysis and feature calculation is performed as explained in conjunction with block 1 of figure 2.
- the noise level estimate may be updated as in block 2 of figure 2.
- the determination whether the input frames comprises non-noise input is performed in block 3. .
- the input to the VAD is needed to be modified. This is done in block 3 according to the embodiments by introducing a counter for counting the number of frames since the 1116
- the feature of detecting sudden increases in input energy is introduced in block 3 based on (EtotJJp or E f low lP ) which later is used in the feature (Etot-EtotJJp or
- EtotJJp 0.01 EtotJ + 0.99 EtotJJp;
- Etot_l is increased every frame but can never be higher than the current input energy. This metric is further low pass filtered to form EtotJJp.
- the condition (Etot-EtotJJp > 10) prevents normal noise update from being used on frames with high energy compared to the current smoothed minimum estimate.
- This embodiment prevents non_sta, tmp_pc, and noise_char features to stop a background update if there has not been a harmonic or correlation event within the last 80 frames.
- bg_cnt + 1 ; //increment counter of pause frames
- bg_cnt forms a combined energy based pause detector and pause burst length counter that ensures the current frame energy is not far from its long term estimate. This is used to ensure that non-speech frames are not used for a background update without the risk of ending up in a deadlock.
- bckr[i] bckr[i] + O l * (tmpN[i] - bckr[i]);
- bckr[i] bckr[i] + 05 * (tmpN[i] - bckr[i]);
- modification block with update using a reduced stepsize which corresponds to blocks 8 and 10a of figure 3.
- This pseudo code corresponds partly to the functionality of the modified blocks 7 and the blocks 1 1 and 10 in figure 3.
- the second modification block of the pseudo code above allows for reduced step size update if there has not been correlation in 20 frames and the difference between Etot and totalNoise is less 25 dB. Also the deadlock recovery is only allowed to use reduced step size update.
- This pseudo code corresponds partly to the functionality of blocks 8, 1 1 and 10a of the blocks in figure 3.
- the pseudo code block ends with the increment of the deadlock recovery counter if none of the above noise adjustments have been possible, corresponding to block 9 in figure 3.
- the third modification block of the pseudo code above contains the additional noise detection test in block 4a and an added background noise update possibility in block 4b. Note that this pseudo code block is executed when normal noise estimate is prohibited due to hangover. There are two alternatives, and both alternatives depend on the correlation counter harm_cor_cnt. In the first alternative, more than 20
- correlation free frames are required in addition to low energy differences using the new metrics totalNoise-Etot_l_lp and Etot - Etot_l_lp combined with the low complex pause length counter bg_cnt.
- more than 80 correlation free frames are required in addion to a low snr_sum.
- snr_sum is the decision metric used in the VAD and in this case it is used as a spectral difference between the current frame and the current background noise estimate. With snr_sum as a spectral difference measure no weight is put on a decrease in energy for a subband compared to the background estimate. For this spectral difference only an increase of subband energy has any weight.
- the feature E T - E F LOW LP is compared to an adaptive threshold.
- the feature E T - E F KM LP is also compared to an adaptive threshold.
- An alternative, would be to use hysteresis in the decision threshold for E T - E F LOW LP , that is different adaptive thresholds are used depending on if one is looking for a speech burst ( N B > 0 ) or a speech pause ( N BG -1 ).
- All the above threshold adaptations can be based on input features such as Input energy variation, estimated SNR, background level, or combinations thereof.
- the additional noise test function in block 4a is applied to all frames, not just the frames for non-noise or hangover.
- a method for updating a background noise estimate of an input signal in a background estimator of a VAD comprises receiving 401 the input signal for a current frame. It should be noted that the reception is shared between other blocks of the VAD and the background estimator can receive other input signals needed to perform the background estimate. Further, the method of the embodiment further comprises determining 402 whether the current frame of the input signal comprises non-noise or that one still is in background noise hangover from such frame(s) as in block 5 of figure 3.
- the background estimate is updated. If it is determined that we are not in hangover, then the background estimate is updated. If it is determined that one is in hangover, then an additional determination whether the current frame input comprises noise is performed 403 by analyzing characteristics at least related to correlation and energy level of the input signal. The additional determination 403 corresponds to block 4a I figure 3. Then the background noise estimate is updated 404 if it is determined that the current frame comprises noise which corresponds to block 4b in figure 3.
- the additional determination whether the current frame of the non-noise input comprises noise further comprises at least one of: detection of correlation and counting the number of frames from a frame last indicated a correlation event, if the energy level of the input signal is within in a first range from a smooth minimum energy level and if the total noise is within a second range from the smooth minimum energy level according to embodiments.
- the detection of correlation and counting the number of frames from a frame last indicated a correlation event are performed to reduce the step size of the update of the background noise estimate and to determine when an update of the background noise estimate should be performed according to one embodiment.
- the analysis of if the energy level of the input signal is within in a first range from the smooth minimum energy level is used to prevent from updating background noise estimate based on frames with high energy compared to the smooth minimum energy level and to determine when an update of the
- the analysis of if the total noise is within a second range from the current estimated noise level is used to determine when an update of the background noise estimate should be performed in block 4b of figure 3.
- the first and second ranges may be fixed ranges or adaptive ranges.
- the additional determination performed in block 4a of figure 3 is applied to all frames not only to the frames that are considered to comprise background update hangover frames in block 5 of figure 3.
- a background estimator 500 in a VAD for updating a background noise estimate for an input signal 501 comprises an input section 502 configured to receive the input signal 501 for a current frame and other signals used for estimating the background noise.
- the background estimator 500 further comprises a processor 503, a memory 504 and an output section 505.
- the processor 503 is configured to determine whether the current frame of the input signal comprises non- noise, to perform an additional determination 4a whether the current frame of the non-noise input comprises noise by analyzing characteristics at least related to correlation and energy level of the input signal, and to update background noise P T/SE2010/051116
- the memory 504 is configured to store software code portions for performing the functions of the processor 503 and background noise estimates and other data relating to noise and signal energy estimates.
- the additional determination 4a whether the current frame of the non-noise input comprises noise further may comprise at least one of: detection of correlation and counting the number of frames from a frame last indicated a correlation event, if the energy level of the input signal is within in a first range from a smooth minimum energy level and if the total noise is within a second range from the smooth minimum energy level.
- the processor 503 may be configured to reduce the step size of the update of the background noise estimate and to determine when an update of the background noise estimate should be performed based on detection of correlation and the number of frames from a frame last indicated a correlation event.
- the processor 503 is configured to use analysis of if the energy level of the input signal is within in a first range from the smooth minimum energy level to prevent from updating background noise estimate based on frames with high energy compared to the smooth minimum energy level and to determine when an update of the background noise estimate should be performed.
- the processor 503 may be configured to determine when an update of the background noise estimate should be performed based on analysis of if the total noise is within a second range from the current estimated noise level.
- the first and second ranges may be fixed or adaptive ranges.
- processor 503 is according to one embodiment configured to apply the additional determination on non-noise frames or frames in hangover.
- significance thresholds may be used to determine the energy levels of subbands of the input signal.
- Figure 6 shows the improvement for speech mixed with babble noise with 64 concurrent speakers with 10 dB SNR.
- Figure 6 clearly shows that the improved decision logic allows for more updates in the speech pauses. Also for the initial segment with noise only the original decision logic is not able to track the input noise but instead shows a decreasing trend due to the always update downwards policy.
- Figure 7 shows the improvement for speech mixed with pink noise input with 20dB SNR. The figure clearly shows that the original solution does not even allow the noise tracking to start. For the improved logic there is only a small delay before the tracking starts and also here the tracking is allowed to work even in the speech pauses.
Landscapes
- Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Artificial Intelligence (AREA)
- Noise Elimination (AREA)
- Telephonic Communication Services (AREA)
- Telephone Function (AREA)
Abstract
Description
Claims
Priority Applications (8)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2010800579955A CN102667927B (en) | 2009-10-19 | 2010-10-18 | Method and background estimator for voice activity detection |
JP2012535162A JP5712220B2 (en) | 2009-10-19 | 2010-10-18 | Method and background estimator for speech activity detection |
AU2010308597A AU2010308597B2 (en) | 2009-10-19 | 2010-10-18 | Method and background estimator for voice activity detection |
EP10825285.9A EP2491559B1 (en) | 2009-10-19 | 2010-10-18 | Method and background estimator for voice activity detection |
CA2778342A CA2778342C (en) | 2009-10-19 | 2010-10-18 | Method and background estimator for voice activity detection |
US13/502,962 US9202476B2 (en) | 2009-10-19 | 2010-10-18 | Method and background estimator for voice activity detection |
IN3221DEN2012 IN2012DN03221A (en) | 2009-10-19 | 2012-04-13 | |
US14/945,495 US9418681B2 (en) | 2009-10-19 | 2015-11-19 | Method and background estimator for voice activity detection |
Applications Claiming Priority (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US25285809P | 2009-10-19 | 2009-10-19 | |
US61/252,858 | 2009-10-19 | ||
US26258309P | 2009-11-19 | 2009-11-19 | |
US61/262,583 | 2009-11-19 | ||
US37675210P | 2010-08-25 | 2010-08-25 | |
US61/376,752 | 2010-08-25 |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/502,962 A-371-Of-International US9202476B2 (en) | 2009-10-19 | 2010-10-18 | Method and background estimator for voice activity detection |
US14/945,495 Continuation US9418681B2 (en) | 2009-10-19 | 2015-11-19 | Method and background estimator for voice activity detection |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2011049514A1 true WO2011049514A1 (en) | 2011-04-28 |
Family
ID=43900543
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/SE2010/051116 WO2011049514A1 (en) | 2009-10-19 | 2010-10-18 | Method and background estimator for voice activity detection |
Country Status (9)
Country | Link |
---|---|
US (2) | US9202476B2 (en) |
EP (2) | EP2491559B1 (en) |
JP (1) | JP5712220B2 (en) |
CN (1) | CN102667927B (en) |
AU (1) | AU2010308597B2 (en) |
CA (1) | CA2778342C (en) |
IN (1) | IN2012DN03221A (en) |
PT (1) | PT2491559E (en) |
WO (1) | WO2011049514A1 (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102820035A (en) * | 2012-08-23 | 2012-12-12 | 无锡思达物电子技术有限公司 | Self-adaptive judging method of long-term variable noise |
CN103578477A (en) * | 2012-07-30 | 2014-02-12 | 中兴通讯股份有限公司 | Denoising method and device based on noise estimation |
WO2014035328A1 (en) * | 2012-08-31 | 2014-03-06 | Telefonaktiebolaget L M Ericsson (Publ) | Method and device for voice activity detection |
WO2015094083A1 (en) * | 2013-12-19 | 2015-06-25 | Telefonaktiebolaget L M Ericsson (Publ) | Estimation of background noise in audio signals |
KR20160003192A (en) * | 2013-05-30 | 2016-01-08 | 후아웨이 테크놀러지 컴퍼니 리미티드 | Media data transmission method, device and system |
WO2016018186A1 (en) | 2014-07-29 | 2016-02-04 | Telefonaktiebolaget L M Ericsson (Publ) | Estimation of background noise in audio signals |
CN110114827A (en) * | 2016-11-17 | 2019-08-09 | 弗劳恩霍夫应用研究促进协会 | The device and method of audio signal are decomposed using variable thresholding |
Families Citing this family (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3493205B1 (en) * | 2010-12-24 | 2020-12-23 | Huawei Technologies Co., Ltd. | Method and apparatus for adaptively detecting a voice activity in an input audio signal |
US9351176B2 (en) * | 2012-01-09 | 2016-05-24 | Qualcomm Incorporated | Phase and amplitude tracking in the presence of a walking pilot signal |
CN103730110B (en) * | 2012-10-10 | 2017-03-01 | 北京百度网讯科技有限公司 | A kind of method and apparatus of detection sound end |
MY178710A (en) * | 2012-12-21 | 2020-10-20 | Fraunhofer Ges Forschung | Comfort noise addition for modeling background noise at low bit-rates |
US9997172B2 (en) * | 2013-12-02 | 2018-06-12 | Nuance Communications, Inc. | Voice activity detection (VAD) for a coded speech bitstream without decoding |
CN105336344B (en) * | 2014-07-10 | 2019-08-20 | 华为技术有限公司 | Noise detection method and device |
US9576589B2 (en) * | 2015-02-06 | 2017-02-21 | Knuedge, Inc. | Harmonic feature processing for reducing noise |
CN106297819B (en) * | 2015-05-25 | 2019-09-06 | 国家计算机网络与信息安全管理中心 | A kind of noise cancellation method applied to Speaker Identification |
CN106328169B (en) * | 2015-06-26 | 2018-12-11 | 中兴通讯股份有限公司 | A kind of acquisition methods, activation sound detection method and the device of activation sound amendment frame number |
CN106448696A (en) * | 2016-12-20 | 2017-02-22 | 成都启英泰伦科技有限公司 | Adaptive high-pass filtering speech noise reduction method based on background noise estimation |
CN108762221B (en) * | 2018-07-09 | 2021-05-11 | 西安电子科技大学 | Deadlock free control method for automatic manufacturing system with uncontrollable event |
KR102280692B1 (en) * | 2019-08-12 | 2021-07-22 | 엘지전자 주식회사 | Intelligent voice recognizing method, apparatus, and intelligent computing device |
CN111554314B (en) * | 2020-05-15 | 2024-08-16 | 腾讯科技(深圳)有限公司 | Noise detection method, device, terminal and storage medium |
CN112489692B (en) * | 2020-11-03 | 2024-10-18 | 北京捷通华声科技股份有限公司 | Voice endpoint detection method and device |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1239456A1 (en) * | 1991-06-11 | 2002-09-11 | QUALCOMM Incorporated | Variable rate vocoder |
WO2004012097A1 (en) * | 2002-07-26 | 2004-02-05 | Motorola, Inc. | Method for fast dynamic estimation of background noise |
WO2008143569A1 (en) * | 2007-05-22 | 2008-11-27 | Telefonaktiebolaget Lm Ericsson (Publ) | Improved voice activity detector |
Family Cites Families (32)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE3623147A1 (en) | 1986-07-10 | 1988-01-21 | Metallgesellschaft Ag | DEVICE FOR GAS DEDUSTING |
JPS63237100A (en) * | 1987-03-26 | 1988-10-03 | 沖電気工業株式会社 | Voice detector |
US5410632A (en) * | 1991-12-23 | 1995-04-25 | Motorola, Inc. | Variable hangover time in a voice activity detector |
JPH06332492A (en) * | 1993-05-19 | 1994-12-02 | Matsushita Electric Ind Co Ltd | Method and device for voice detection |
US6001131A (en) * | 1995-02-24 | 1999-12-14 | Nynex Science & Technology, Inc. | Automatic target noise cancellation for speech enhancement |
US5727072A (en) * | 1995-02-24 | 1998-03-10 | Nynex Science & Technology | Use of noise segmentation for noise cancellation |
US5659622A (en) * | 1995-11-13 | 1997-08-19 | Motorola, Inc. | Method and apparatus for suppressing noise in a communication system |
FI100840B (en) * | 1995-12-12 | 1998-02-27 | Nokia Mobile Phones Ltd | Noise attenuator and method for attenuating background noise from noisy speech and a mobile station |
US5819217A (en) * | 1995-12-21 | 1998-10-06 | Nynex Science & Technology, Inc. | Method and system for differentiating between speech and noise |
DE69716266T2 (en) * | 1996-07-03 | 2003-06-12 | British Telecommunications P.L.C., London | VOICE ACTIVITY DETECTOR |
US6070137A (en) * | 1998-01-07 | 2000-05-30 | Ericsson Inc. | Integrated frequency-domain voice coding using an adaptive spectral enhancement filter |
US6415253B1 (en) * | 1998-02-20 | 2002-07-02 | Meta-C Corporation | Method and apparatus for enhancing noise-corrupted speech |
US6801895B1 (en) * | 1998-12-07 | 2004-10-05 | At&T Corp. | Method and apparatus for segmenting a multi-media program based upon audio events |
US6275798B1 (en) * | 1998-09-16 | 2001-08-14 | Telefonaktiebolaget L M Ericsson | Speech coding with improved background noise reproduction |
US6424938B1 (en) * | 1998-11-23 | 2002-07-23 | Telefonaktiebolaget L M Ericsson | Complex signal activity detection for improved speech/noise classification of an audio signal |
US6381570B2 (en) * | 1999-02-12 | 2002-04-30 | Telogy Networks, Inc. | Adaptive two-threshold method for discriminating noise from speech in a communication signal |
US6618701B2 (en) * | 1999-04-19 | 2003-09-09 | Motorola, Inc. | Method and system for noise suppression using external voice activity detection |
US6490556B2 (en) * | 1999-05-28 | 2002-12-03 | Intel Corporation | Audio classifier for half duplex communication |
US7171357B2 (en) * | 2001-03-21 | 2007-01-30 | Avaya Technology Corp. | Voice-activity detection using energy ratios and periodicity |
FR2825826B1 (en) * | 2001-06-11 | 2003-09-12 | Cit Alcatel | METHOD FOR DETECTING VOICE ACTIVITY IN A SIGNAL, AND ENCODER OF VOICE SIGNAL INCLUDING A DEVICE FOR IMPLEMENTING THIS PROCESS |
US20020198708A1 (en) * | 2001-06-21 | 2002-12-26 | Zak Robert A. | Vocoder for a mobile terminal using discontinuous transmission |
FR2833103B1 (en) * | 2001-12-05 | 2004-07-09 | France Telecom | NOISE SPEECH DETECTION SYSTEM |
US20040002856A1 (en) * | 2002-03-08 | 2004-01-01 | Udaya Bhaskar | Multi-rate frequency domain interpolative speech CODEC system |
US7024353B2 (en) * | 2002-08-09 | 2006-04-04 | Motorola, Inc. | Distributed speech recognition with back-end voice activity detection apparatus and method |
SG119199A1 (en) * | 2003-09-30 | 2006-02-28 | Stmicroelectronics Asia Pacfic | Voice activity detector |
JP4601970B2 (en) * | 2004-01-28 | 2010-12-22 | 株式会社エヌ・ティ・ティ・ドコモ | Sound / silence determination device and sound / silence determination method |
CA2454296A1 (en) * | 2003-12-29 | 2005-06-29 | Nokia Corporation | Method and device for speech enhancement in the presence of background noise |
KR100677396B1 (en) * | 2004-11-20 | 2007-02-02 | 엘지전자 주식회사 | A method and a apparatus of detecting voice area on voice recognition device |
US7610197B2 (en) * | 2005-08-31 | 2009-10-27 | Motorola, Inc. | Method and apparatus for comfort noise generation in speech communication systems |
US8990073B2 (en) | 2007-06-22 | 2015-03-24 | Voiceage Corporation | Method and device for sound activity detection and sound signal classification |
JP5229234B2 (en) | 2007-12-18 | 2013-07-03 | 富士通株式会社 | Non-speech segment detection method and non-speech segment detection apparatus |
US8244528B2 (en) * | 2008-04-25 | 2012-08-14 | Nokia Corporation | Method and apparatus for voice activity determination |
-
2010
- 2010-10-18 PT PT108252859T patent/PT2491559E/en unknown
- 2010-10-18 EP EP10825285.9A patent/EP2491559B1/en active Active
- 2010-10-18 AU AU2010308597A patent/AU2010308597B2/en active Active
- 2010-10-18 JP JP2012535162A patent/JP5712220B2/en active Active
- 2010-10-18 CN CN2010800579955A patent/CN102667927B/en active Active
- 2010-10-18 CA CA2778342A patent/CA2778342C/en active Active
- 2010-10-18 EP EP20140181366 patent/EP2816560A1/en not_active Ceased
- 2010-10-18 US US13/502,962 patent/US9202476B2/en active Active
- 2010-10-18 WO PCT/SE2010/051116 patent/WO2011049514A1/en active Application Filing
-
2012
- 2012-04-13 IN IN3221DEN2012 patent/IN2012DN03221A/en unknown
-
2015
- 2015-11-19 US US14/945,495 patent/US9418681B2/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1239456A1 (en) * | 1991-06-11 | 2002-09-11 | QUALCOMM Incorporated | Variable rate vocoder |
WO2004012097A1 (en) * | 2002-07-26 | 2004-02-05 | Motorola, Inc. | Method for fast dynamic estimation of background noise |
WO2008143569A1 (en) * | 2007-05-22 | 2008-11-27 | Telefonaktiebolaget Lm Ericsson (Publ) | Improved voice activity detector |
Non-Patent Citations (2)
Title |
---|
"Digital cellular telecommunications system (Phase 2+); Universal Mobile Telecommunications System (UMTS); Mandatory speech codec speech processing functions; Adaptive Multi-Rate (AMR) speech codec; Voice Activity Detector (VAD)(3GPP TS 26.094 version 7.0.0 Release 7)", ETSI TS 126 094 ETSI STANDARDS, 1 June 2007 (2007-06-01), LIS, SOPHIA ANTIPOLIS CEDEX, FRANCE * |
See also references of EP2491559A4 * |
Cited By (44)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103578477B (en) * | 2012-07-30 | 2017-04-12 | 中兴通讯股份有限公司 | Denoising method and device based on noise estimation |
CN103578477A (en) * | 2012-07-30 | 2014-02-12 | 中兴通讯股份有限公司 | Denoising method and device based on noise estimation |
CN102820035A (en) * | 2012-08-23 | 2012-12-12 | 无锡思达物电子技术有限公司 | Self-adaptive judging method of long-term variable noise |
US9997174B2 (en) | 2012-08-31 | 2018-06-12 | Telefonaktiebolaget Lm Ericsson (Publ) | Method and device for voice activity detection |
JP2017151455A (en) * | 2012-08-31 | 2017-08-31 | テレフオンアクチーボラゲット エルエム エリクソン(パブル) | Method and device for voice activity detection |
US11900962B2 (en) | 2012-08-31 | 2024-02-13 | Telefonaktiebolaget Lm Ericsson (Publ) | Method and device for voice activity detection |
US11417354B2 (en) | 2012-08-31 | 2022-08-16 | Telefonaktiebolaget Lm Ericsson (Publ) | Method and device for voice activity detection |
CN107195313B (en) * | 2012-08-31 | 2021-02-09 | 瑞典爱立信有限公司 | Method and apparatus for voice activity detection |
US10607633B2 (en) | 2012-08-31 | 2020-03-31 | Telefonaktiebolaget Lm Ericsson (Publ) | Method and device for voice activity detection |
US9472208B2 (en) | 2012-08-31 | 2016-10-18 | Telefonaktiebolaget Lm Ericsson (Publ) | Method and device for voice activity detection |
EP3113184A1 (en) * | 2012-08-31 | 2017-01-04 | Telefonaktiebolaget LM Ericsson (publ) | Method and device for voice activity detection |
CN104603874A (en) * | 2012-08-31 | 2015-05-06 | 瑞典爱立信有限公司 | Method and device for voice activity detection |
WO2014035328A1 (en) * | 2012-08-31 | 2014-03-06 | Telefonaktiebolaget L M Ericsson (Publ) | Method and device for voice activity detection |
EP3301676A1 (en) * | 2012-08-31 | 2018-04-04 | Telefonaktiebolaget LM Ericsson (publ) | Method and device for voice activity detection |
CN107195313A (en) * | 2012-08-31 | 2017-09-22 | 瑞典爱立信有限公司 | Method and apparatus for Voice activity detector |
EP3007169A4 (en) * | 2013-05-30 | 2017-06-14 | Huawei Technologies Co., Ltd. | Media data transmission method, device and system |
EP4235661A3 (en) * | 2013-05-30 | 2023-11-15 | Huawei Technologies Co., Ltd. | Comfort noise generation method and device |
KR102099752B1 (en) | 2013-05-30 | 2020-04-10 | 후아웨이 테크놀러지 컴퍼니 리미티드 | Signal encoding method and apparatus |
KR20160003192A (en) * | 2013-05-30 | 2016-01-08 | 후아웨이 테크놀러지 컴퍼니 리미티드 | Media data transmission method, device and system |
US9886960B2 (en) | 2013-05-30 | 2018-02-06 | Huawei Technologies Co., Ltd. | Voice signal processing method and device |
US10692509B2 (en) | 2013-05-30 | 2020-06-23 | Huawei Technologies Co., Ltd. | Signal encoding of comfort noise according to deviation degree of silence signal |
EP3438979A1 (en) | 2013-12-19 | 2019-02-06 | Telefonaktiebolaget LM Ericsson (publ) | Estimation of background noise in audio signals |
US9626986B2 (en) | 2013-12-19 | 2017-04-18 | Telefonaktiebolaget Lm Ericsson (Publ) | Estimation of background noise in audio signals |
WO2015094083A1 (en) * | 2013-12-19 | 2015-06-25 | Telefonaktiebolaget L M Ericsson (Publ) | Estimation of background noise in audio signals |
RU2720357C2 (en) * | 2013-12-19 | 2020-04-29 | Телефонактиеболагет Л М Эрикссон (Пабл) | Method for estimating background noise, a unit for estimating background noise and a computer-readable medium |
EP3084763A4 (en) * | 2013-12-19 | 2016-12-14 | ERICSSON TELEFON AB L M (publ) | Estimation of background noise in audio signals |
US11164590B2 (en) | 2013-12-19 | 2021-11-02 | Telefonaktiebolaget Lm Ericsson (Publ) | Estimation of background noise in audio signals |
CN105830154A (en) * | 2013-12-19 | 2016-08-03 | 瑞典爱立信有限公司 | Estimation of background noise in audio signals |
EP3719801A1 (en) | 2013-12-19 | 2020-10-07 | Telefonaktiebolaget LM Ericsson (publ) | Estimation of background noise in audio signals |
RU2618940C1 (en) * | 2013-12-19 | 2017-05-11 | Телефонактиеболагет Л М Эрикссон (Пабл) | Estimation of background noise in audio signals |
US10573332B2 (en) | 2013-12-19 | 2020-02-25 | Telefonaktiebolaget Lm Ericsson (Publ) | Estimation of background noise in audio signals |
US10311890B2 (en) | 2013-12-19 | 2019-06-04 | Telefonaktiebolaget Lm Ericsson (Publ) | Estimation of background noise in audio signals |
US9818434B2 (en) | 2013-12-19 | 2017-11-14 | Telefonaktiebolaget Lm Ericsson (Publ) | Estimation of background noise in audio signals |
WO2016018186A1 (en) | 2014-07-29 | 2016-02-04 | Telefonaktiebolaget L M Ericsson (Publ) | Estimation of background noise in audio signals |
RU2713852C2 (en) * | 2014-07-29 | 2020-02-07 | Телефонактиеболагет Лм Эрикссон (Пабл) | Estimating background noise in audio signals |
EP3582221A1 (en) | 2014-07-29 | 2019-12-18 | Telefonaktiebolaget LM Ericsson (publ) | Esimation of background noise in audio signals |
US11114105B2 (en) | 2014-07-29 | 2021-09-07 | Telefonaktiebolaget Lm Ericsson (Publ) | Estimation of background noise in audio signals |
US10347265B2 (en) | 2014-07-29 | 2019-07-09 | Telefonaktiebolaget Lm Ericsson (Publ) | Estimation of background noise in audio signals |
RU2665916C2 (en) * | 2014-07-29 | 2018-09-04 | Телефонактиеболагет Лм Эрикссон (Пабл) | Estimation of background noise in audio signals |
US11636865B2 (en) | 2014-07-29 | 2023-04-25 | Telefonaktiebolaget Lm Ericsson (Publ) | Estimation of background noise in audio signals |
EP3309784A1 (en) | 2014-07-29 | 2018-04-18 | Telefonaktiebolaget LM Ericsson (publ) | Esimation of background noise in audio signals |
US9870780B2 (en) | 2014-07-29 | 2018-01-16 | Telefonaktiebolaget Lm Ericsson (Publ) | Estimation of background noise in audio signals |
CN110114827A (en) * | 2016-11-17 | 2019-08-09 | 弗劳恩霍夫应用研究促进协会 | The device and method of audio signal are decomposed using variable thresholding |
CN110114827B (en) * | 2016-11-17 | 2023-09-29 | 弗劳恩霍夫应用研究促进协会 | Apparatus and method for decomposing an audio signal using a variable threshold |
Also Published As
Publication number | Publication date |
---|---|
JP5712220B2 (en) | 2015-05-07 |
CN102667927B (en) | 2013-05-08 |
EP2491559B1 (en) | 2014-12-10 |
US20120209604A1 (en) | 2012-08-16 |
EP2816560A1 (en) | 2014-12-24 |
CA2778342C (en) | 2017-08-22 |
AU2010308597A1 (en) | 2012-05-17 |
US20160078884A1 (en) | 2016-03-17 |
IN2012DN03221A (en) | 2015-10-23 |
US9418681B2 (en) | 2016-08-16 |
CA2778342A1 (en) | 2011-04-28 |
CN102667927A (en) | 2012-09-12 |
EP2491559A1 (en) | 2012-08-29 |
PT2491559E (en) | 2015-05-07 |
EP2491559A4 (en) | 2013-11-06 |
AU2010308597B2 (en) | 2015-10-01 |
JP2013508772A (en) | 2013-03-07 |
US9202476B2 (en) | 2015-12-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9418681B2 (en) | Method and background estimator for voice activity detection | |
US9990938B2 (en) | Detector and method for voice activity detection | |
US9401160B2 (en) | Methods and voice activity detectors for speech encoders | |
US11900962B2 (en) | Method and device for voice activity detection | |
KR102000227B1 (en) | Discrimination and attenuation of pre-echoes in a digital audio signal |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201080057995.5 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 10825285 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 3221/DELNP/2012 Country of ref document: IN |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2012535162 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2010308597 Country of ref document: AU Ref document number: 2778342 Country of ref document: CA Ref document number: 13502962 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2010825285 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 2010308597 Country of ref document: AU Date of ref document: 20101018 Kind code of ref document: A |