EP1269462B1 - Appareil et procede de detection d'activite vocale - Google Patents
Appareil et procede de detection d'activite vocale Download PDFInfo
- Publication number
- EP1269462B1 EP1269462B1 EP01958309A EP01958309A EP1269462B1 EP 1269462 B1 EP1269462 B1 EP 1269462B1 EP 01958309 A EP01958309 A EP 01958309A EP 01958309 A EP01958309 A EP 01958309A EP 1269462 B1 EP1269462 B1 EP 1269462B1
- Authority
- EP
- European Patent Office
- Prior art keywords
- audio parameter
- unit
- delay
- audio
- averaging
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Lifetime
Links
- 230000000694 effects Effects 0.000 title claims abstract description 33
- 238000001514 detection method Methods 0.000 title claims abstract description 18
- 238000000034 method Methods 0.000 title claims description 31
- 238000012935 Averaging Methods 0.000 claims abstract description 37
- 230000003111 delayed effect Effects 0.000 claims description 8
- 230000003595 spectral effect Effects 0.000 claims description 6
- 230000008569 process Effects 0.000 claims description 5
- 238000012545 processing Methods 0.000 claims description 2
- 230000001934 delay Effects 0.000 claims 1
- 239000013598 vector Substances 0.000 description 19
- 230000007704 transition Effects 0.000 description 6
- 230000003044 adaptive effect Effects 0.000 description 4
- 230000005540 biological transmission Effects 0.000 description 3
- 238000001228 spectrum Methods 0.000 description 3
- 101100126625 Caenorhabditis elegans itr-1 gene Proteins 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 2
- 238000003491 array Methods 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 230000006837 decompression Effects 0.000 description 2
- 206010019133 Hangover Diseases 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 238000007796 conventional method Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 230000005236 sound signal Effects 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/78—Detection of presence or absence of voice signals
Definitions
- the present invention relates to voice processing systems in general, and to methods and apparatus for detecting voice activity in a low resource environment, in particular.
- a voice activity detector operates under the assumption that speech is present only in part of the audio signals while there are many intervals, which exhibit only silence or background noise.
- a voice activity detector can be used for many purposes such as suppressing overall transmission activity in a transmission system, when there is no speech, thus potentially saving power and channel bandwidth.
- the VAD detects that speech activity has resumed, then it can reinitiate transmission activity.
- a voice activity detector can also be used in conjunction with speech storage devices, by differentiating audio portions which include speech from those that are "speechless". The portions including speech are then stored in the storage device and the "speechless" portions are not stored.
- Conventional methods for detecting voice are based at least in part on methods for detecting and assessing the power of a speech signal.
- the estimated power is compared to either constant or adaptive threshold, for determining a decision.
- the main advantage of these methods is their low complexity, which makes them suitable for low resources implementations.
- the main disadvantage of such methods is that background noise can result in "speech" being detected when none is present or "speech" which is present not being detected because it is obscured and difficult to detect.
- Some methods for detecting speech activity are directed at noisy mobile environments and are based on adaptive filtering of the speech signal. This reduces the noise content from the signal, prior to the final decision.
- the frequency spectrum and noise level may vary because the method will be used for different speakers and in different environments.
- the input filter and thresholds are often adaptive so as to track these variations. Examples of these methods are provided in GSM specification 06.42 Voice Activity Detector (VAD) for half rate, full rate, and enhanced full rate speech traffic channels respectively.
- VAD Voice Activity Detector
- Another such method is the "Multi-Boundary Voice Activity Detection Algorithm" as proposed in ITU G.729 annex B. These methods are more accurate in noisy environment but are significantly complex to implement.
- European Patent application No. 0785419A2 Benyassine et al. is directed to a method for voice activity detection which includes the following steps: extracting a predetermined set of parameters from the incoming speech signal for each frame and making a frame voicing decision of the incoming speech signal for each frame according to a set of difference measures extracted from the predetermined set of parameters.
- WO-A-0017856 relates to a method and apparatus for detecting voice activity in a speech signal.
- WO-A-0017856 has a filing date of 27 August 1999, a publication date of 30 March 2000, and claims a priority date of 18 September 1998.
- the present invention alleviates the disadvantages of the prior art by providing a method which utilizes conventional vocoder output data of a voice related stream for detecting voice activity therein.
- the method for voice activity detection is based on the analysis of audio parameters such as Line Spectral Frequencies (LSF) parameters.
- LSF Line Spectral Frequencies
- the detection is based on a stationarity estimate of spectral characteristics of the incoming speech frames, which is represented by LSF parameters.
- Apparatus 100 includes two delay arrays 102 and 110, a plurality of distance measure units 106A, 106B, 106C and 106D, an averaging unit 108, a subtraction unit 114 and decision logic unit (DLU) 116.
- Delay array 102 includes a plurality of delay units 104A, 104B, 104C, 104D, 104E, 104F and 104G, all connected in series, so that each adds a further delay to the previous one.
- Delay array 110 includes a plurality of delay units, 112A, 112B, 112C and 112D, all connected in series, so that each adds a further delay to the previous one.
- Apparatus 100 is further connected to a Line Spectral Frequencies (LSF) generation unit 120, which can be a part of the voice encoder (vocoder) apparatus of an audio system.
- LSF unit 120 produces LSF values for each received audio frame. It is noted that LSF unit 120 is only one example for an audio parameter generation unit.
- the output of the LSF unit 120 is coupled to the input of delay unit 104A.
- the input of each of delay units 104A, 104B, 104C and 104D is connected to a respective one of distance measure units 106A, 106B, 106C and 106D.
- the input of delay unit 104A is connected to distance measure unit 106A.
- Delay unit 104A has its output connected to distance measure unit 106B.
- Unit 104B has its output connected to unit 106C.
- Unit 104C has its output connected 106D.
- the output of delay units 104D, 104E, 104F and 104G is connected to a respective one of distance measure units 106A, 106B, 106C and 106D.
- the output of delay unit 104D is connected to distance measure unit 106A.
- the LSF value L (n) at the input of delay unit 104A is associated with the value L (n-4) at the output of delay unit 104D.
- each of the LSF values L (n-1), L (n-2) and L (n-3) is associated with a respective one of LSF values L (n-5), L (n-6) and L (n-7), at a respective one of distance measure units 106B, 106C and 106D.
- the system includes a different number of delay units and can combine more than two LSF values, which are at different distances from each other, such as L ⁇ n + L ⁇ ⁇ n - 4 + L ⁇ ⁇ n - 6 .
- the distance measure units 106A, 106B, 106C and 106D are all connected to the averaging unit 108.
- Averaging unit 108 is further connected to delay unit 112A, subtraction unit 114 and to DLU 116.
- the output of each of delay units 112A, 112B, 112C and 112D is connected to DLU 116.
- the output of delay unit 112A is further connected to the subtraction unit 114.
- Figure 2 is an illustration of a method for operating the apparatus 100 of Figure 1 , operative in accordance with another preferred embodiment of the present invention.
- a plurality of audio parameters are received.
- Each of the audio parameters is related to a predetermined audio frame.
- the audio parameters include LSF values, which represent the short-time frequency spectrum characteristics of the signal envelope for each audio frame.
- LSF parameters are derived from the Linear Prediction Coefficients (LPC's), which are widely used by many modern speech compression and analysis schemes and are discussed in detail in A. M. Kondoz, Digital Speech: Coding for Low Bit Rate Communications Systems, New York: John Wiley & Sons, 1994 .
- LPC's Linear Prediction Coefficients
- step 152 the audio parameters are grouped according to a predetermined pattern of audio frames.
- each audio frame is associated with a voice frame, which is four places ahead of it.
- the audio parameters of audio frame n are grouped with the audio parameters of audio frame n-4.
- any other number can be used for the distance between the frames.
- further combinations can also be used such as combination (n,n-2,n-7) and the like.
- distance measure unit 106A groups vector L ( n ) of frame n with vector L ( n -4) of frame n-4.
- Distance measure unit 106B groups vector L ( n -1) of frame n-1 with vector L ( n -5) of frame n-5.
- Distance measure unit 106C groups vector L ( n - 2) of frame n-2 with vector L ( n -6) of frame n-6.
- Distance measure unit 106D groups vector L ( n -3) of frame n-3 with vector L ( n -7) of frame n-7.
- each distance measure unit 106A, 106B, 106C and 106D performs a two-stage operation.
- the distance measure units 106A, 106B, 106C and 106D provide the D vectors D ⁇ n D ⁇ ⁇ n - 1 D ⁇ ⁇ n - 2 ... D ⁇ ⁇ n - M 1 - 1 2 to averaging unit 108.
- step 156 an average value is determined for all of the present characteristic values.
- the measure a(n) is applied to a second M 2 -stage delay line.
- the delay line includes four delay units 112A, 112B, 112C and 112D.
- Averaging unit 108 further provides the latest average value a ( n ) to DLU 116 and to subtraction unit 114.
- Delay unit 112A provides the previous average value a ( n-1 ) to the subtraction unit 114.
- step 160 a decision is produced according to the values, which are present. Reference is now made to Figures 3 and 4 .
- the implementation of each of the decision functions can be according to a Boolean expression, which compares the value e(n) and components of the averaging vector A ( n ) with predetermined or variable threshold values.
- decision logic can vary according to specific performance requirements to a trade-off between "false alarm”, "miss detect” statistics, and the like.
- the logic can be either constant or adapted to other components such as background noise characteristic estimator, voicing mode if available, periodicity check, and the like.
- the instantaneous decision result can further be applied to an additional hangover function.
- step 180 represents the initial stage of the decision phase, wherein the current state of the VAD (speech-on or speech off) is detected. If the current state of the VAD is speech on, then the system 100 proceeds to step 182. Otherwise, the system 100 proceeds to step 186.
- VAD speech-on or speech off
- step 182 compliance with a speech-on-to-off transition condition is detected.
- a condition includes a predetermined combination of a ( n ) and e ( n ) with respect to predetermined values (Note that the threshold can be adaptive in general case).
- the system proceeds to step 184, which performs a transition in the VAD state to speech-off. Otherwise, this step is repeated until such compliance is detected.
- DLU 116 detects if the received values comply with the predetermined condition.
- step 186 compliance with a speech-off-to-on transition condition is detected.
- a condition includes another predetermined combination of a ( n ) and e(n) with respect to predetermined values.
- step 188 which performs a transition in the VAD state to speech-on. Otherwise, this step is repeated until such compliance is detected.
- step 184 After performing a VAD mode transition (either of step 184 and 188), the system proceeds back to step 180.
- Apparatus 200 includes a multi stage delay unit 202, two delay arrays 204 and 210, a distance measure unit 206, an averaging unit 208, a subtraction unit 214 and decision logic unit (DLU) 216.
- Delay array 204 includes a plurality of delay units 218A, 218B and 218M 2 , all connected in series, so that each adds a further delay to the previous one.
- delay array 210 includes a plurality of delay units, 212A, 212B, 212C and 212D, all connected in series, so that each adds a further delay stage to the previous one.
- System 200 is further connected to a Line Spectral Frequencies (LSF) generation unit 220, which can be a part of the voice encoder (vocoder) apparatus of an audio system.
- LSF unit 220 produces LSF values for each received audio frame.
- the input of multi stage delay unit 202 is connected to LSF unit 220.
- the output of multi stage delay unit 202 is connected distance measure unit 206.
- the LSF value L(n) at the input of delay unit 218A is associated with an M 1 stage delayed value L(n-M 1 ) at the output of delay unit 202.
- the output of distance measure unit 206 is connected to averaging unit 208 and to delay array 204.
- the output of each of the delay units 218A, 218B and 218M 2 is connected to the averaging unit 208 so that each provides a previously delayed distance measure output value to the averaging unit 208.
- delay unit 218A provides a distance measure value, which is respective of the pair, L(n-1) and L(n-M 1 -1). Accordingly, only the first distance measure value has to be calculated and the rest are stored, delayed and provided to the averaging unit 208 at the appropriate timing.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Acoustics & Sound (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Signal Processing (AREA)
- Computational Linguistics (AREA)
- Geophysics And Detection Of Objects (AREA)
- Measuring Pulse, Heart Rate, Blood Pressure Or Blood Flow (AREA)
- Mobile Radio Communication Systems (AREA)
- Time-Division Multiplex Systems (AREA)
- Measurement Of Velocity Or Position Using Acoustic Or Ultrasonic Waves (AREA)
- Measuring Or Testing Involving Enzymes Or Micro-Organisms (AREA)
- Transmission Systems Not Characterized By The Medium Used For Transmission (AREA)
Claims (14)
- Appareil de détection d'activité vocale (100) comprenant :un générateur de paramètres audio (120) pour produire des paramètres audio à partir de trames audio reçues ;une unité de retard multi-étages (102) connectée au générateur de paramètres audio pour produire une série des paramètres audio retardés de quantités différentes ;des moyens de mesure de distance (1 06A, 106B) pour mesurer la distance entre des groupes prédéterminés de paramètres audio ayant des retards différents ;une pluralité d'unité de détermination (106A, 106B) pour déterminer une valeur caractéristique pour chaque groupe des paramètres audio dont les distances ont été mesurées ;et, connectée aux unités de détermination, une unité de calcul de moyenne (108) pour déterminer une valeur moyenne de toutes les valeurs caractéristiques.
- Appareil de détection d'activité vocale (100) selon la revendication 1, dans lequel l'unité de retard multi-étages comprend une pluralité d'unités de retard de paramètres audio (104A, 104B) pour retarder les paramètres audio, les unités de retard de paramètres audio étant connectées en série les unes aux autres, de telle sorte que chaque unité de retard de paramètres audio qui suit une unité de retard de paramètres audio précédente ajoute un retard supplémentaire au retard ajouté par la précédente, la première desdites unités de retard de paramètres audio étant connectée au générateur de paramètres audio (120) ;
dans lequel les moyens de mesure de distance comprennent une pluralité d'unités de mesure de distance (106A, 106B) qui sont chacune connectées à au moins deux desdites unités de retard de paramètres audio, pour regrouper les valeurs audio retardées produites par les unités de retard de paramètres audio en fonction du modèle de retard prédéterminé, et les unités de mesure de distance sont également les unités de détermination. - Appareil de détection d'activité vocale selon la revendication 2, dans lequel ladite première unité de retard de paramètres audio (104A) fonctionne de façon à recevoir une pluralité de valeurs de paramètres audio, concernant une période de parole prédéterminée, en provenance dudit générateur de paramètres audio, chacune desdites unités de retard de paramètres audio restantes (104A, 104B, 104C) peut fonctionner de façon à recevoir lesdites valeurs de paramètres audio en provenance d'une précédente desdites unités de retard de paramètres audio (104A, 104B, 104C), chacune desdites unités de mesure de distance peut fonctionner de façon à traiter les valeurs de paramètres audio reçues en provenance de certaines sélectionnées desdites unités de retard de paramètres audio qui lui sont connectées, pour ainsi produire des valeurs différentielles, l'unité de calcul de moyenne pouvant fonctionner de façon à produire une valeur moyenne à partir desdites valeurs différentielles.
- Appareil de détection d'activité vocale selon la revendication 1, la revendication 2 ou la revendication 3, et comprenant, en outre :une pluralité d'unités de retard de valeurs moyennes (112A, 112B) connectées en série les unes aux autres, la première desdites unités de retard de valeurs moyennes (112A) étant, en outre, connectée à la sortie de ladite unité de calcul de moyenne (108) ; etune unité logique numérique (116) connectée auxdites unités de retard de valeurs moyennes.
- Appareil de détection d'activité vocale selon la revendication 4, dans lequel ladite première unité de retard de valeurs moyennes (112A) peut fonctionner de façon à recevoir une pluralité de valeurs moyennes de paramètres audio traités en provenance de ladite unité de calcul de moyenne (108), chacune desdites unités de retard pouvant fonctionner de façon à retarder chacune desdites valeurs moyennes de paramètres audio traités, ladite unité logique numérique pouvant fonctionner de façon à recevoir une pluralité de valeurs moyennes de paramètres audio traités successives, la dernière desdites valeurs moyennes de paramètres audio traités successives étant reçue en provenance de ladite unité de calcul de moyenne et le restant desdites valeurs moyennes de paramètres audio traités successives étant reçu en provenance de ladite unité de retard de valeurs moyennes, ladite unité logique numérique pouvant fonctionner de façon à traiter lesdites valeurs moyennes de paramètres audio traités successives de façon à produire une indication de présence de parole.
- Appareil de détection d'activité vocale selon la revendication 4 ou la revendication 5, dans lequel ladite première unité de retard de paramètres audio (104A) peut fonctionner de façon à recevoir une pluralité de valeurs de paramètres audio en provenance dudit générateur de paramètres audio (120), chacune desdites unités de retard de paramètres audio restantes (104B, 104C, 104D) peut fonctionner de façon à recevoir lesdites valeurs de paramètres audio en provenance d'une précédente desdites unités de retard de paramètres audio (104A, 104B, 104C), et chacune desdites unités de mesure de distance (106A, 106B) peut fonctionner de façon à traiter ensemble les valeurs de paramètres audio reçues en provenance de certaines sélectionnées desdites unités de retard de paramètres audio, qui lui sont connectées, pour ainsi produire des valeurs différentielles, ladite unité de calcul de moyenne pouvant fonctionner de façon à produire une valeur moyenne de paramètres audio traités à partir de chaque ensemble desdites valeurs différentielles, et
dans lequel ladite première unité de retard de valeurs moyennes peut fonctionner de façon à recevoir lesdites valeurs moyennes de paramètres audio traités en provenance de ladite unité de calcul de moyenne, chacun des unités de retard pouvant fonctionner de façon à retarder chacune des valeurs moyennes de paramètres audio traités, ladite unité logique numérique reçoit une pluralité de valeurs moyennes de paramètres audio traités successives, l'unité logique numérique pouvant fonctionner de façon à recevoir une pluralité de valeurs moyennes de paramètres audio traités successives, la dernière desdites valeurs moyennes de paramètres audio traités successives étant reçue en provenance de ladite unité de calcul de moyenne et le restant desdites valeurs moyennes de paramètres audio traités successives étant reçu en provenance de ladite unité de retard de valeurs moyennes, ladite unité logique numérique pouvant fonctionner de façon à traiter lesdites valeurs moyennes de paramètres audio traités successives de façon à produire une indication de présence de parole. - Appareil de détection d'activité vocale selon la revendication 1, la revendication 4 ou la revendication 5, dans lequel chacune des unités de détermination (218A, 218B) peut fonctionner de façon à envoyer une sortie de mesure de distance préalablement retardée à l'unité de calcul de moyenne (208), la première (218A) desdites unités de détermination étant connectée au générateur de paramètres audio (120) par l'intermédiaire d'une unité de mesure de distance (206) qui peut fonctionner de façon à mesurer une distance entre chacun d'une série de paramètres audio différemment retardés produite par l'unité de retard multi-étages (206) et une sortie non retardée du générateur de paramètres audio (120).
- Appareil de détection d'activité vocale selon l'une quelconque des revendications précédentes, dans lequel ledit paramètre audio comprend des fréquences de raies spectrales.
- Appareil de détection d'activité vocale selon la revendication 8, dans lequel ledit générateur de paramètres audio comprend un générateur de fréquences de raies spectrales.
- Appareil de détection d'activité vocale selon l'une quelconque des revendications 4 à 9, et comprenant, en outre, une unité de soustraction (114) connectée entre l'entrée et la sortie de ladite première unité de retard de valeurs moyennes et en plus de ladite unité logique numérique,
dans lequel ladite unité de soustraction peut fonctionner de façon à produire des valeurs de différence à partir des valeurs moyennes de paramètres audio traités reçues en provenance de ladite unité de calcul de moyenne et à partir des valeurs moyennes de paramètres audio traités retardées par la première unité de retard de valeurs moyennes, et
dans lequel ladite unité logique numérique peut fonctionner de façon à traiter lesdites valeurs de différence en même temps que lesdites valeurs moyennes de paramètres audio traités successives, pour produire une indication de présence de parole. - Procédé d'utilisation de l'appareil selon l'une quelconque des revendications précédentes pour la détection d'une activité de parole, comprenant les étapes consistant à :regrouper des paramètres audio, qui sont associés à une combinaison prédéterminée de trames audio, pour ainsi produire une pluralité de groupes ;déterminer une valeur caractéristique pour chacun desdits groupes ;déterminer une valeur moyenne pour chacune d'une pluralité de sélections d'une pluralité desdites valeurs caractéristiques ; etdéterminer la présence d'une activité de parole à partir de certaines sélectionnées desdites valeurs moyennes.
- Procédé selon la revendication 11, comprenant, en outre, l'étape consistant à détecter l'énergie d'échantillons audio associés auxdits paramètres audio, avant ladite étape de détermination de la présence d'une activité de parole.
- Procédé selon la revendication 11 ou la revendication 12, et comprenant, en outre, l'étape préliminaire consistant à recevoir lesdits paramètres audio en provenance d'un générateur audio.
- Procédé selon la revendication 11, la revendication 12 ou la revendication 13, et comprenant, en outre, l'étape préliminaire consistant à produire lesdits paramètres audio à partir d'une pluralité d'échantillons audio.
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GB0006312 | 2000-03-15 | ||
GB0006312A GB2360428B (en) | 2000-03-15 | 2000-03-15 | Voice activity detection apparatus and method |
PCT/IB2001/001603 WO2001080220A2 (fr) | 2000-03-15 | 2001-03-14 | Appareil et procede de detection d'activite vocale |
Publications (2)
Publication Number | Publication Date |
---|---|
EP1269462A2 EP1269462A2 (fr) | 2003-01-02 |
EP1269462B1 true EP1269462B1 (fr) | 2008-05-14 |
Family
ID=9887716
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP01958309A Expired - Lifetime EP1269462B1 (fr) | 2000-03-15 | 2001-03-14 | Appareil et procede de detection d'activite vocale |
Country Status (6)
Country | Link |
---|---|
EP (1) | EP1269462B1 (fr) |
AT (1) | ATE395683T1 (fr) |
AU (1) | AU2001280027A1 (fr) |
DE (1) | DE60133998D1 (fr) |
GB (1) | GB2360428B (fr) |
WO (1) | WO2001080220A2 (fr) |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2317084B (en) * | 1995-04-28 | 2000-01-19 | Northern Telecom Ltd | Methods and apparatus for distinguishing speech intervals from noise intervals in audio signals |
US5774849A (en) * | 1996-01-22 | 1998-06-30 | Rockwell International Corporation | Method and apparatus for generating frame voicing decisions of an incoming speech signal |
US6385548B2 (en) * | 1997-12-12 | 2002-05-07 | Motorola, Inc. | Apparatus and method for detecting and characterizing signals in a communication system |
US6188981B1 (en) * | 1998-09-18 | 2001-02-13 | Conexant Systems, Inc. | Method and apparatus for detecting voice activity in a speech signal |
-
2000
- 2000-03-15 GB GB0006312A patent/GB2360428B/en not_active Expired - Fee Related
-
2001
- 2001-03-14 EP EP01958309A patent/EP1269462B1/fr not_active Expired - Lifetime
- 2001-03-14 WO PCT/IB2001/001603 patent/WO2001080220A2/fr active IP Right Grant
- 2001-03-14 AU AU2001280027A patent/AU2001280027A1/en not_active Abandoned
- 2001-03-14 AT AT01958309T patent/ATE395683T1/de not_active IP Right Cessation
- 2001-03-14 DE DE60133998T patent/DE60133998D1/de not_active Expired - Lifetime
Also Published As
Publication number | Publication date |
---|---|
GB0006312D0 (en) | 2000-05-03 |
WO2001080220A3 (fr) | 2002-05-23 |
ATE395683T1 (de) | 2008-05-15 |
GB2360428B (en) | 2002-09-18 |
AU2001280027A1 (en) | 2001-10-30 |
WO2001080220A2 (fr) | 2001-10-25 |
DE60133998D1 (de) | 2008-06-26 |
GB2360428A (en) | 2001-09-19 |
EP1269462A2 (fr) | 2003-01-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR100883712B1 (ko) | 음원 방향 추정 방법, 및 음원 방향 추정 장치 | |
US6427134B1 (en) | Voice activity detector for calculating spectral irregularity measure on the basis of spectral difference measurements | |
EP3493205B1 (fr) | Procédé et appareil permettant de détecter de façon adaptative une activité vocale dans un signal audio d'entrée | |
EP0551803B1 (fr) | Méthode de synchronisation et d'estimation de canal dans un système de radiocommunication à accès multiple et à division de temps | |
JP5006279B2 (ja) | 音声活性検出装置及び移動局並びに音声活性検出方法 | |
KR100770839B1 (ko) | 음성 신호의 하모닉 정보 및 스펙트럼 포락선 정보,유성음화 비율 추정 방법 및 장치 | |
JP3878482B2 (ja) | 音声検出装置および音声検出方法 | |
US8818811B2 (en) | Method and apparatus for performing voice activity detection | |
US20030097257A1 (en) | Sound signal process method, sound signal processing apparatus and speech recognizer | |
KR100735343B1 (ko) | 음성신호의 피치 정보 추출장치 및 방법 | |
US6876965B2 (en) | Reduced complexity voice activity detector | |
EP1548703B1 (fr) | Dispositif et procédé de détection de l'activité vocale | |
JP3418005B2 (ja) | 音声ピッチ検出装置 | |
EP1269462B1 (fr) | Appareil et procede de detection d'activite vocale | |
KR20090080777A (ko) | 신호 검출 방법 및 장치 | |
JPH08221097A (ja) | 音声成分の検出法 | |
Beritelli et al. | A low‐complexity speech‐pause detection algorithm for communication in noisy environments | |
CA2279264C (fr) | Amelioration de l'insensibilite aux signaux vocaux dans un detecteur dtmf fonde sur la prediction lineaire | |
US5734679A (en) | Voice signal transmission system using spectral parameter and voice parameter encoding apparatus and decoding apparatus used for the voice signal transmission system | |
JPS61184912A (ja) | 定数可変型聴感的重み付けフイルタ | |
US6993478B2 (en) | Vector estimation system, method and associated encoder | |
JPH0832527A (ja) | フェージングピッチ推定装置 | |
GB2437868A (en) | Estimating noise power spectrum, sorting time frames, calculating the quantile and interpolating values over all remaining frequencies | |
JPH10304023A (ja) | 電話機 | |
JPH0311479B2 (fr) |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
AK | Designated contracting states |
Kind code of ref document: A2 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE TR |
|
AX | Request for extension of the european patent |
Free format text: AL;LT;LV;MK;RO;SI |
|
17P | Request for examination filed |
Effective date: 20021125 |
|
17Q | First examination report despatched |
Effective date: 20061201 |
|
GRAP | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOSNIGR1 |
|
GRAS | Grant fee paid |
Free format text: ORIGINAL CODE: EPIDOSNIGR3 |
|
GRAA | (expected) grant |
Free format text: ORIGINAL CODE: 0009210 |
|
AK | Designated contracting states |
Kind code of ref document: B1 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE TR |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: FG4D |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: EP |
|
REG | Reference to a national code |
Ref country code: IE Ref legal event code: FG4D Free format text: LANGUAGE OF EP DOCUMENT: FRENCH |
|
REF | Corresponds to: |
Ref document number: 60133998 Country of ref document: DE Date of ref document: 20080626 Kind code of ref document: P |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: FI Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20080514 Ref country code: ES Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20080825 |
|
NLV1 | Nl: lapsed or annulled due to failure to fulfill the requirements of art. 29p and 29m of the patents act | ||
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: AT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20080514 Ref country code: NL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20080514 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: PT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20081014 Ref country code: DK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20080514 Ref country code: SE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20080814 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: BE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20080514 |
|
PLBE | No opposition filed within time limit |
Free format text: ORIGINAL CODE: 0009261 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT |
|
26N | No opposition filed |
Effective date: 20090217 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: IT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20080514 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: MC Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20090331 |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: PL |
|
REG | Reference to a national code |
Ref country code: FR Ref legal event code: ST Effective date: 20091130 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: IE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20090314 Ref country code: LI Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20090331 Ref country code: CH Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20090331 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: FR Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20091123 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: GR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20080815 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LU Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20090314 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: TR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20080514 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: CY Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20080514 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: GB Payment date: 20200327 Year of fee payment: 20 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: DE Payment date: 20200528 Year of fee payment: 20 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R071 Ref document number: 60133998 Country of ref document: DE |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: PE20 Expiry date: 20210313 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: GB Free format text: LAPSE BECAUSE OF EXPIRATION OF PROTECTION Effective date: 20210313 |